fbpx
Features Hub

UK Commons report urges rapid governance for rising AI challenges

Mon 25 Sep 2023

The House of Commons’ Science, Innovation and Technology Select Committee has published an interim report on its inquiry into the governance of Artificial Intelligence (AI) in the UK.

The report highlighted 12 key challenges the Government need to address in relation to AI. The introduction of legislation rather than the previous approach of relying on existing regulators to fill the regulators gaps was also recommended.

The overriding message urged the UK Government to accelerate, not to pause, the establishment of a governance regime for AI, including ‘whatever statutory measures as may be needed’.

AI Governance in the UK: The Key Challenges

The report recommends that the UK Government’s approach to AI governance and regulation should address each of the twelve challenges outlined, both through domestic policy and international engagement.

However, these are issues internationally. The report notes that ‘The twelve challenges of AI governance which we have set out must be addressed by policymakers in all jurisdictions. Different administrations may choose different ways to do this’.

The report highlights twelve challenges that the UK Government and regulators must address, including:

  • The Bias Challenge: AI can introduce or perpetuate societal biases via the data it relies upon, such as correlating female names to typically female roles.
  • The Privacy Challenge: AI can allow individuals to be identified and personal information about them to be used in unexpected ways that may breach their fundamental rights.
  • The Misrepresentation Challenge: AI can allow the generation of material that deliberately misrepresents someone’s behaviour, opinions, or character. The use of image and voice recordings of individuals can lead to highly plausible material being generated which can purport to show an individual saying things that have no basis in fact.
  • The Access to Data Challenge: The most powerful AI needs very large datasets, which are held by few organisations which causes competition and market concerns.
  • The Access to Compute Challenge: The development of powerful AI requires significant compute power, which is costly, hence access is often limited to a few organisations. The UK Government has announced plans to establish an Exascale supercomputer facility and an AI-dedicated compute resource to support research.
  • The Black Box Challenge: Some AI models and tools cannot explain why they produce a particular result, which is a challenge to appropriate transparency. The challenge is further complicated by the fact that the better an AI model or tool performs, the less explainable it is likely to be.
  • The Open-Source Challenge: Requiring code to be openly available may promote transparency and innovation, allowing it to be proprietary may concentrate market power but allow more dependable regulation of harms.
  • The Intellectual Property and Copyright Challenge: Some AI models and tools make use of other people’s content. Policy must establish the rights of the originators of this content, and these rights must be enforced. Ongoing legal cases are likely to set precedents in this area and a draft code is being produced by the Intellectual Property office.
  • The Liability Challenge: If AI models and tools are used by third parties to do harm, policy must establish whether developers or providers of the technology bear any liability for harms done.
  • The Employment Challenge: AI will disrupt the jobs that people do and that are available. Policy makers must anticipate and manage the disruption.
  • The International Coordination Challenge: AI is a global technology, and the development of governance frameworks to regulate its uses must be an international undertaking. The Government believes that the November International AI summit is a first step in doing this.
  • The Existential Challenge: Some people think that AI is a major threat to human life. If that is a possibility, governance needs to provide protections for national security.

Potential AI Laws

The UK Government published the AI White Paper in March 2023, which outlined a pro-innovation approach to AI regulation. In parallel, the EU is pushing through its EU AI Act which is expected to be enacted late 2023 / early 2024.

The report states that ‘a tightly-focused AI Bill in the next King’s Speech would help, nor hinder, the Prime Minister’s ambition to position the UK as an AI governance leader. Without a serious, rapid and effective effort to establish the right governance frameworks—and to ensure a leading role in international initiatives—other jurisdictions will steal a march and the frameworks that they lay down may become the default even if they are less effective than what the UK can offer’.

In the context of the wider AI regulation discussion, this session of Parliament is the last opportunity before the next UK General Election for the UK to legislate in respect of AI. Post-election, any legislation is likely not to be enacted until late 2025, which is two years from now and will be three years after the publication of the AI white paper, which some comment is already falling ‘out-of-date’.

The interim report noted that a balance must be struck, as although ‘some observers have called for the development of certain types of AI models and tools to be paused, allowing global regulatory and governance frameworks to catch up … it should also be remembered that is not unknown for those who have secured an advantageous position to seek to defend it against market insurgents through regulation’.

Preserving competition and innovation in this market remains an objective of the UK approach.

The report argued that waiting on AI legislation for at least two years risks the UK ‘being left behind by other legislation—like the EU AI Act—that could become the de facto standard and be hard to displace’ even where the UK can offer a more desirable approach. The situation could mirror the GDPR where UK laws followed the EU lead.

The BBC consulted the Government on its views on the report. The Government did not confirm if it agreed that a new law should be put forward. A spokesperson highlighted the November International AI summit and the £100 milliom initial investment in a Taskforce to encourage the safe development of AI models, which is, according to the Government, ‘more funding dedicated to AI safety than any other government in the world’.

Next Steps

The report does not outline the next steps for the Inquiry or whether a final report will follow this interim one. Organisations should keep a keen eye on the Government response to the report, the White Paper, and also the King’s Speech to see whether the Committee’s recommendation is taken on board by the Government.

About the Author

Tom Whittaker is a Senior Associate and Solicitor Advocate in the dispute resolution team at Burges Salmon. He regularly advises clients on commercially significant and complex civil disputes for a wide range of corporate and government clients across different sectors.


Hungry for more tech news?

Sign up for your weekly tech briefings!

Send us a correction Send us a news tip