Advertisement
UK markets close in 7 hours 42 minutes
  • FTSE 100

    8,420.73
    +39.38 (+0.47%)
     
  • FTSE 250

    20,592.56
    +61.26 (+0.30%)
     
  • AIM

    785.80
    +2.10 (+0.27%)
     
  • GBP/EUR

    1.1621
    +0.0010 (+0.09%)
     
  • GBP/USD

    1.2531
    +0.0007 (+0.06%)
     
  • Bitcoin GBP

    50,409.23
    +1,476.94 (+3.02%)
     
  • CMC Crypto 200

    1,358.03
    +0.02 (+0.00%)
     
  • S&P 500

    5,214.08
    +26.41 (+0.51%)
     
  • DOW

    39,387.76
    +331.36 (+0.85%)
     
  • CRUDE OIL

    79.74
    +0.48 (+0.61%)
     
  • GOLD FUTURES

    2,372.80
    +32.50 (+1.39%)
     
  • NIKKEI 225

    38,229.11
    +155.13 (+0.41%)
     
  • HANG SENG

    18,963.08
    +425.27 (+2.29%)
     
  • DAX

    18,777.91
    +91.31 (+0.49%)
     
  • CAC 40

    8,226.21
    +38.56 (+0.47%)
     

‘Existential catastrophe’ caused by AI is likely unavoidable, DeepMind researcher warns

‘Existential catastrophe’ caused by AI is likely unavoidable, DeepMind researcher warns

Researchers from the University of Oxford and Google’s artificial intelligence division DeepMind have claimed that there is a high probability of advanced forms of AI becoming “existentially dangerous to life on Earth”.

In a recent article in the peer-reviewed journal AI Magazine, the researchers warned that there would be “catastrophic consequences” if the development of certain AI agents continues.

Leading philosphers like Oxford University’s Nick Bostrom have previously spoken of the threat posed by advanced forms of artificial intelligence, though one of authors of the new paper claimed such warnings did not go far enough.

“Bostrom, [computer scientist Stuart] Russell, and others have argued that advanced AI poses a threat to humanity,” Michael Cohen wrote in a Twitter thread accompanying the article.

ADVERTISEMENT

“Under the conditions we have identified, our conclusion is much stronger than that of any previous publication – an existential catastrophe is not just possible, but likely.”

The paper proposes a scenario whereby an AI agent figures out a strategy to cheat in order to receive a reward that it is pre-programmed to seek.

In order to maximize its reward potential, it requires as much energy as is possible to obtain. The thought experiment sees humanity ultimately competing against the AI for energy resources.

“Winning the competition of ‘getting to use the last bit of available energy’ while playing against something much smarter than us would probably be very hard,” Mr Cohen wrote. “Losing would be fatal.

“These possibilities, however theoretical, mean we should be progressing slowly – if at all – toward the goal of more powerful AI.”

DeepMind has already proposed a safeguard against such an eventuality, dubbing it “the big red button”. In a 2016 paper titled ‘Safely Interruptible Agents’, the AI firm outlined a framework for preventing advanced machines from ignoring turn-off commands and becoming an out-of-control rogue agent.

Professor Bostrom previously described DeepMind – whose AI accomplishments include beating human champions at the boardgame Go and manipulating nuclear fusion – as the closest to creating human-level artificial intelligence.

The Sweidish philospher also said it would be a “great tragedy” if AI development did not continue, as it holds the potential to cure diseases and advance civilisation at an otherwise impossible rate.