STOCKHOLM/BRUSSELS/LONDON, Dec 1 (Reuters) – EU lawmakers cannot agree on how to regulate systems like ChatGPT, in a threat to landmark legislation aimed at keeping artificial intelligence (AI) in check, six sources told Reuters.
As negotiators meet on Friday for crucial discussions ahead of final talks scheduled for Dec. 6, ‘foundation models’, or generative AI, have become the main hurdle in talks over the European Union’s proposed AI Act, said the sources, who declined to be identified because the discussions are confidential.
Foundation models like the one built by Microsoft (MSFT.O)-backed OpenAI are AI systems trained on large sets of data, with the ability to learn from new data to perform various tasks.
After two years of negotiations, the bill was approved by the European parliament in June. The draft AI rules now need to be agreed through meetings between representatives of the European Parliament, the Council and the European Commission.
Experts from EU countries will meet on Friday to thrash out their position on foundation models, access to source codes, fines and other topics while lawmakers from the European Parliament are also gathering to finalise their stance.
If they cannot agree, the act risks being shelved due to lack of time before European parliamentary elections next year.
While some experts and lawmakers have proposed a tiered approach for regulating foundation models, defined as those with more than 45 million users, others have said smaller models could be equally risky.
But the biggest challenge to getting an agreement has come from France, Germany and Italy, who favour letting makers of generativeAI models self-regulate instead of having hard rules.
In a meeting of the countries’ economy ministers on Oct. 30 in Rome, France persuaded Italy and Germany to support a proposal, sources told Reuters.
Until then, negotiations had gone smoothly, with lawmakers making compromises across several other conflict areas such as regulating high-risk AI, sources said.
SELF-REGULATION?
European parliamentarians, EU Commissioner Thierry Breton and scores of AI researchers have criticised self-regulation.
In an open letter this week, researchers such as Geoffrey Hinton warned self-regulation is “likely to dramatically fall short of the standards required for foundation model safety”.
France-based AI company Mistral and Germany’s Aleph Alpha have criticised the tiered approach to regulating foundation models, winning support from their respective countries.
A source close to Mistral said the company favours hard rules for products, not the technology on which it is built.
Other pending issues in the talks include definition of AI, fundamental rights impact assessment, law enforcement exceptions and national security exceptions, sources told Reuters.
Lawmakers have also been divided over the use of AI systems by law enforcement agencies for biometric identification of individuals in publicly accessible spaces and could not agree on several of these topics in a meeting on Nov. 29, sources said.
Spain, which holds the EU presidency until the end of the year, has proposed compromises in a bid to speed up the process.
If a deal does not happen in December, the next presidency Belgium will have a couple of months to one before it is likely shelved ahead of European elections.
“Had you asked me six or seven weeks ago, I would have said we are seeing compromises emerging on all the key issues,” said Mark Brakel, director of policy at the Future of Life Institute, a nonprofit aimed at reducing risks from advanced AI.
“This has now become a lot harder,” he said.
(This story has been corrected to say that the AI Act was passed in June, not May, in paragraph 4)
Reporting by Supantha Mukherjee in Stockholm;
Editing by Josephine Mason and Alexander Smith
Our Standards: The Thomson Reuters Trust Principles.