Advertisement

Sunak to address benefits and possible existential threat in speech on AI

Rishi Sunak will set out how he will address the dangers presented by artificial intelligence while harnessing the benefits as a Government paper warns of a possible existential threat.

In a speech in London on Thursday, the Prime Minister will say the rapidly expanding technology brings new opportunities for growth and advances as well as “new dangers”.

He will argue he is being responsible by seeking to “address those fears head-on” to give the public the “peace of mind that we will keep you safe”.

A new paper published by the Government Office for Science to accompany the speech says there is insufficient evidence to rule out a threat to humanity from AI.

ADVERTISEMENT

Based on sources including UK intelligence, it says many experts believe it is a “risk with very low likelihood and few plausible routes”, and would need the technology to “outpace mitigations, gain control over critical systems and be able to avoid being switched off”.

It adds: “Given the significant uncertainty in predicting AI developments, there is insufficient evidence to rule out that highly capable future frontier AI systems, if misaligned or inadequately controlled, could pose an existential threat.”

Three broad pathways to “catastrophic” or existential risks are set out as a self-improving system that can achieve goals in the physical world without oversight working to harm human interests.

The second is a failure of multiple key systems after intense competition leads to one company with a technological edge gaining control and then failing due to safety, controllability and misuse.

Finally, over-reliance was judged to be a threat as humans grant AI more control over critical systems they no longer fully understand and become “irreversibly dependent”.

In his speech on Thursday, Mr Sunak is expected to say AI will bring “new knowledge, new opportunities for economic growth, new advances in human capability, and the chance to solve problems we once thought beyond us”.

“But it also brings new dangers and new fears,” he is set to add.

“So, the responsible thing for me to do is to address those fears head-on, giving you the peace of mind that we will keep you safe, while making sure you and your children have all the opportunities for a better future that AI can bring.

“Doing the right thing, not the easy thing, means being honest with people about the risks from these technologies.”

In terms of capabilities, the Government’s paper notes that frontier AI can already perform “many economically useful tasks” such as conversing fluently and at length, and be used as a translation tool or to summarise lengthy documents and analyse data.

It suggests that the technology is likely to become substantially more useful in the future, and potentially be able to carry out tasks more efficiently than humans, but it notes that “we cannot currently reliably predict ahead of time which specific new capabilities a frontier AI model will gain” as the ways of training AI models are likely to also change and evolve.

But among the potential risks of the technology, the paper identifies the hugely broad potential use cases of the technology as an issue, arguing it is hard to predict how AI tools could be used and therefore protect against possible problems.

It adds that the current lack of safety standards is a key issue, and warns that AI could “substantially exacerbate existing cyber risks” if misused – potentially able to launch cyber attacks autonomously, although the paper suggests AI-powered defences could mitigate some of this risk.

In addition, it warns that frontier AI could disrupt the labour market by displacing human workers, and could lead to a spike in misinformation through AI-generated images or by unintentionally spreading inaccurate information on which an AI model has been trained.

The paper will serve as a discussion paper at the UK’s AI safety summit next week where world leaders and tech giants will discuss the developing issues around artificial intelligence.