EU framework for ethical AI takes shape
Written by Kelvin Chan Thu 20 Feb 2020
Rules proposed to develop ‘framework for trustworthy artificial intelligence’
The European Union has unveiled proposals to regulate artificial intelligence that call for strict rules and safeguards on risky applications of the rapidly developing technology.
The report is part of the bloc’s wider digital strategy aimed at maintaining its position as the global pacesetter on technological standards.
Technology companies seeking to tap Europe’s vast and lucrative market, including those from the US and China, would have to play by any new rules that eventually come into force.
The EU’s executive commission said it wants to develop a “framework for trustworthy artificial intelligence”.
‘AI must serve people’
European Commission president Ursula von der Leyen had ordered her top deputies to come up with a co-ordinated European approach to artificial intelligence and data strategy 100 days after she took office in December.
“We will be particularly careful where essential human rights and interests are at stake,” Ms von der Leyen told reporters in Brussels.
“Artificial intelligence must serve people, and therefore artificial intelligence must always comply with people’s rights.”
EU leaders, keen on establishing “technological sovereignty”, also released a strategy to unlock data from the continent’s businesses and the public sector so it can be harnessed for further innovation in artificial intelligence.
Officials in Europe, which does not have any homegrown tech giants, hope to catch up with the US and China by using the bloc’s vast and growing trove of industrial data for what they anticipate is a coming wave of digital transformation.
They also warned that even more regulation for foreign tech companies is in store with the upcoming Digital Services Act, a sweeping overhaul of how the bloc treats digital companies, including potentially holding them liable for illegal content posted on their platforms.
A steady stream of Silicon Valley tech bosses, including Facebook chief executive Mark Zuckerberg, Google CEO Sundar Pichai and Microsoft president Brad Smith, have visited Brussels in recent weeks as part of apparent lobbying efforts.
“It is not us that need to adapt to today’s platforms. It is the platforms that need to adapt to Europe,” said Thierry Breton, commissioner for the internal market.
“That is the message that we delivered to CEOs of these platforms when they come to see us.”
If the tech companies are not able to build systems “for our people, then we will regulate, and we are ready to do this in the Digital Services Act at the end of the year”, he said.
The EU’s report said clear rules are needed to address “high-risk AI systems” such as those in health, policing or transport, which should be “transparent, traceable and guarantee human oversight”.
Other artificial intelligence systems could come with labels certifying that they are in line with EU standards.
Artificial intelligence uses computers to process large sets of data and make decisions without human input.
It is used, for example, to trade stocks in financial markets or, in some countries, to scan faces in crowds to find criminal suspects.
While it can be used to improve healthcare, make farming more efficient or combat climate change, it also brings risks.
It can be unclear what data artificial intelligence systems work off.
Facial recognition systems can be biased against certain social groups, for example.
There are also concerns about privacy and the use of the technology for criminal purposes, the report said.
Human-centred guidelines for AI are essential because “none of the positive things will be achieved if we distrust the technology”, said Margrethe Vestager, the executive vice president overseeing the EU’s digital strategy.
Under the proposals, which are open for public consultation until May 19, EU authorities want to be able to test and certify the data used by the algorithms that power artificial intelligence in the same way they check cosmetics, cars and toys.
It is important to use unbiased data to train high-risk artificial intelligence systems so they can avoid discrimination, the commission said.
Specifically, AI systems could be required to use data reflecting gender, ethnicity and “other possible grounds of prohibited discrimination.”
Other ideas include preserving data to help trace any problems and having AI systems clearly spell out their capabilities and limitations.
Users should be told when they are interacting with a machine and not a human, while humans should be in charge of the system and have the final say on decisions such as rejecting an application for welfare benefits, the report said.
EU leaders said they also wanted to open a debate on when to allow facial recognition in remote identification systems, which are used to scan crowds to check people’s faces against those on a database.
It is considered the “most intrusive form” of the technology and is prohibited in the EU except in special cases.
Written by Kelvin Chan Thu 20 Feb 2020