Prime Minister Rishi Sunak and other world leaders will discuss the possibilities and risks posed by AI at an event in November, held at Bletchley Park, where the likes of Alan Turing decrypted Nazi messages during the Second World War.
The potential threat AI poses to human life itself should be a focus of any government regulation, MPs have warned.
Concerns around public wellbeing and national security were listed among a dozen challenges that members of the Science, Innovation and Technology Committee said must be addressed by ministers ahead of the UK hosting a world-first summit at Bletchley Park.
Rishi Sunak and other leaders will discuss the possibilities and risks posed by AI at the event in November, held at Britain’s Second World War codebreaking base.
The site was crucial to the development of the technology, as Alan Turing and others used Colossus computers to decrypt messages sent between the Nazis.
Greg Clark, committee chair and a Conservative MP, said he “strongly welcomes” the summit – but warned the government may need to show “greater urgency” to ensure potential legislation doesn’t quickly become outdated as powers like the US, China, and EU consider their own rules around AI.
The 12 challenges the committee said “must be addressed” are:
1. Existential threat – if, as some experts have warned, AI poses a major threat to human life, then regulation must provide national security protections.
2. Bias – AI can introduce new or perpetuate existing biases in society.
3. Privacy – sensitive information about individuals or businesses could be used to train AI models.
4. Misrepresentation – language models like ChatGPT may produce material that misrepresents someone’s behaviour, personal views, and character.
5. Data – the sheer amount of data needed to train the most powerful AI.
6. Computing power – similarly, the development of the most powerful AI requires enormous computing power.
7. Transparency – AI models often struggle to explain why they produce a particular result, or where the information comes from.
8. Copyright – generative models, whether they be text, images, audio, or video, typically make use of existing content, which must be protected so not to undermine the creative industries.
9. Liability – if AI tools are used to do harm, policy must establish whether the developers or providers are liable.
10. Employment – politicians must anticipate the likely impact on existing jobs that embracing AI will have.
11. Openness – the computer code behind AI models could be made openly available to allow for more dependable regulation and promote transparency and innovation.
12. International coordination – the development of any regulation must be an international undertaking, and the November summit must welcome “as wide a range of countries as possible”.