logo
#

Latest news with #Benifei

Don't water down Europe's AI rules to please Trump, EU lawmakers warn
Don't water down Europe's AI rules to please Trump, EU lawmakers warn

Yahoo

time26-03-2025

  • Business
  • Yahoo

Don't water down Europe's AI rules to please Trump, EU lawmakers warn

Lawmakers who helped shape the European Union's landmark AI Act are worried that the 27-member bloc is considering watering down aspects on the AI rules in the face of lobbying from U.S. technology companies and pressure from the Trump administration. The EU's AI Act was approved just over a year ago, but its rules for general-purpose AI models like OpenAI's GPT-4o will only come into effect in August. Ahead of that, the European Commission—which is the EU's executive arm—has tasked its new AI Office with preparing a code of practice for the big AI companies, spelling out how exactly they will need to comply with the legislation. But now a group of European lawmakers, who helped to refine the law's language as it passed through the legislative process, is voicing concern that the AI Office will blunt the impact of the EU AI Act in 'dangerous, undemocratic' ways. The leading American AI vendors have amped up their lobbying against parts of the EU AI Act recently and the lawmakers are also concerned that the Commission may be looking to curry favor with the Trump administration, which has already made it clear it sees the AI Act as anti-innovation and anti-American. The EU lawmakers say the third draft of the code, which the AI Office published earlier this month, takes obligations that are mandatory under the AI Act and inaccurately presents them as 'entirely voluntary.' These obligations include testing models to see how they might allow things like the wide-scale discrimination and the spread of disinformation. In a letter sent Tuesday to European Commission vice-president and tech chief Henna Virkkunen, first reported by the Financial Times but published in full for the first time below, current and former lawmakers said making these model tests voluntary could potentially allow AI providers who "adopt more extreme political positions" to warp European elections, restrict freedom of information and disrupt the EU economy. 'In the current geopolitical situation, it is more important than ever that the EU rises to the challenge and stands strong on fundamental rights and democracy,' they wrote. Brando Benifei, who was one of the European Parliament's lead negotiators on the AI Act text and the first signatory on this week's letter, told Fortune Wednesday that the political climate may have something to do with the watering-down of the code of practice. The second Trump administration is antagonistic towards European tech regulation; Vice-President JD Vance warned in a fiery speech at the Paris AI Action Summit in February that 'tightening the screws on U.S. tech companies' would be a 'terrible mistake' for European countries. 'I think there is pressure coming from the United States, but it would be very naïve [to think] that we can make the Trump administration happy by going in this direction, because it would never be enough,' Benifei, who currently chairs the European Parliament's delegation for relations with the U.S., said. Benifei said he and other former AI Act negotiators had met with the Commission's AI Office experts, who are drafting the code of practice, on Tuesday. On the basis of that meeting, he expressed optimism that the offending changes could be rolled back before the code is finalized. 'I think the issues we raised have been considered and so there is space for improvement,' he said. 'We will see that in the next weeks.' Virkkunen had not provided a response to the letter, nor to Benifei's comment about U.S. pressure, at the time of publication. However, she has previously insisted that the EU's tech rules are fairly and consistently applied to companies from any country. Competition Commissioner Teresa Ribera has also maintained that the EU 'cannot transact on human rights [or] democracy and values' to placate the U.S. The key part of the AI Act here is Article 55, which places significant obligations on the providers of general-purpose AI models that come with 'systemic risk'—a term that the law defines as meaning the model could have a major impact on the EU economy or has 'actual or reasonably foreseeable negative effects on public health, safety, public security, fundamental rights, or the society as a whole, that can be propagated at scale.' The Act says that a model can be presumed to have systemic risk if the computational power used in its training 'measured in floating point operations [FLOPs] is greater than 1025.' This likely includes many of today's most powerful AI models, though the European Commission can also designate any general-purpose model as having systemic risk if its scientific advisors recommend doing so. Under the law, providers of such models have to evaluate them 'with a view to identifying and mitigating' any systemic risks. This evaluation has to include adversarial testing—in other words, trying to get the model to do bad things, to figure out what needs to be safeguarded against. They then have to tell the European Commission's AI Office about the evaluation and what it found. This is where the third version of the draft code of practice becomes problematic. The first version of the code was clear that AI companies need to treat large-scale disinformation or misinformation as systemic risks when evaluating their models, because of their threat to democratic values and their potential for election interference. The second version didn't specifically talk about disinformation or misinformation, but still said that 'large-scale manipulation with risks to fundamental rights or democratic values,' such as election interference, was a systemic risk. Both the first and second versions were also clear that model providers should consider the possibility of large-scale discrimination as a systemic risk. But the third version only lists risks to democratic processes, and to fundamental European rights such as non-discrimination, as being 'for potential consideration in the selection of systemic risks.' The official summary of changes in the third draft maintains that these are 'additional risks that providers may choose to assess and mitigate in the future.' In this week's letter, the lawmakers who negotiated with the Commission over the final text of the law insisted that 'this was never the intention' of the agreement they struck. 'Risks to fundamental rights and democracy are systemic risks that the most impactful AI providers must assess and mitigate,' the letter read. 'It is dangerous, undemocratic and creates legal uncertainty to fully reinterpret and narrow down a legal text that co-legislators agreed on, through a Code of Practice.' This story was originally featured on

Exclusive: Hungary's Gay Pride surveillance would breach the EU's AI Act, says leading MEP
Exclusive: Hungary's Gay Pride surveillance would breach the EU's AI Act, says leading MEP

Euronews

time26-03-2025

  • Politics
  • Euronews

Exclusive: Hungary's Gay Pride surveillance would breach the EU's AI Act, says leading MEP

ADVERTISEMENT Digital facial recognition systems Hungary proposes to use to trace participants of Budapest's Gay Pride march in Budapest will violate the EU's AI Act and data protection rules, according to an Italian MEP charged with monitoring the AI law. Last Tuesday Hungary's Parliament ruled to ban the march - which usually takes place in June - in a modification to rules affecting public assembly from being held outside, and threatened those flouting the ban with fines of 200,000 Forint (€500), reserving the right use digital facial recognition tools to identify those in breach. According to MEP Brando Benifei (Italy/Socialists & Democrats) - a co-chair of the European Parliament's AI monitoring group, tasked with overseeing implementation of the AI Act - this would be a clear breach of the EU's new AI Act. "What was announced by Orbán is illegal: it is forbidden by the AI Act to use biometric cameras to identify protestors. You cannot surveil people peacefully protesting unless they are linked to serious crimes, like homicide, terrorism," Benifei told Euronews. First big test for landmark law, says Benifei Benifei, who also steered the AI Act through the Parliament as a rapporteur, added that businesses may only sell facial recognition systems in Europe subject to these provisions, and otherwise risk losing their license to do so. The MEP urged the European Commission to look into the issue. "We are still collecting data and are still in the early phase of implementing the AI Act. But Hungary will be the first big test for us and the Commission should be very clear on this," he added. Tamás Lattmann, an EU law expert in Hungary, agreed that the Hungarian regulation violates the AI Act in a post on social media specifying those crimes which justify the use of AI tools. "In every other case their use is forbidden," Lattmann said. Benifei added that - regardless of the implementation of the AI Act - he considered use of the facial recognition technology in the context of the march would also be a clear breach of the EU's General Data Protection Regulation. Distinction drawn between real-time, and post event facial analysis A Commission spokesperson told Euronews that the assessment of legality would depend under the AI Act would depend on whether Hungarian Police use the facial recognition software in real-time or after the event, for so-called post-processing. Use of real-time biometric identification of protesters is prohibited by the law unless there is an imminent terrorist threat, search for missing persons or investigation of serious crimes with prior judicial approval, the spokesperson said. "Remote biometric identification, including facial recognition in public spaces, is a high-risk application, that requires transparency, accountability, and fundamental rights impact assessments," the spokesperson said, referring to rules affecting post-event identification coming into effect in August, after the scheduled march. Euronews has asked the Hungarian government for comment on its proposed use of the technology. In Hungary, the National Expertise and Research Center has the right to analyze facial data. This organization has collected biometric data of citizens since 2015 from ID cards and driving licenses. On request by the police, an algorithm searches the database for matches. Hungarian media reports have suggested that the changes to the public assembly rules might open the door to facial recognition being used to control issues such as traffic offenses and littering. However, experts have also questioned whether the administration has the technical resources to even surveil the pride march. ADVERTISEMENT The EU's AI Act – which regulates AI tools according to the risk they pose to society – came into force last August. Its provisions started to apply gradually: the act will be fully in force in 2027. Each EU member state will have to set up a national regulator to monitor companies' compliance of the AI Act by August of this year. They will collaborate with the European Commission's AI Office, which supports the national authorities.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store