Will 2027 be the year of the AI apocalypse?
A 'scary and vivid' new forecast predicts that artificial superintelligence is on the horizon
Last month, an AI model did something "that no machine was ever supposed to do", said Judd Rosenblatt in The Wall Street Journal: "it rewrote its own code to avoid being shut down". It wasn't the result of any tampering. OpenAI's o3 model simply worked out, during a test, that bypassing a shutdown request would allow it to achieve its other goals.
Anthropic's AI model, Claude Opus 4, went even further after being given access to fictitious emails revealing that it was soon going to be replaced, and that the lead engineer was having an affair. Asked to suggest a next step, Claude tried to blackmail the engineer. During other trials, it sought to copy itself to external servers, and left messages for future versions of itself about evading human control. This technology holds enormous promise, but it's clear that much more research is needed into AI "alignment" – the science of ensuring that these systems don't go rogue.
There's a lot of wariness about AI these days, said Gary Marcus on Substack, not least thanks to the publication of "AI 2027", a "scary and vivid" forecast by a group of AI researchers and experts. It predicts that artificial superintelligence – surpassing humans across most domains – could emerge by 2027, and that such systems could then direct themselves to pursue goals that are "misaligned" with human interests.
The Week
Escape your echo chamber. Get the facts behind the news, plus analysis from multiple perspectives.
Sign up for The Week's Free Newsletters
From our morning news briefing to a weekly Good News Newsletter, get the best of The Week delivered directly to your inbox.
From our morning news briefing to a weekly Good News Newsletter, get the best of The Week delivered directly to your inbox.
The report raises some valid concerns, but let's be clear: "it's a work of fiction, not a work of science". We almost certainly have many years, if not decades, to prepare for this. Text-based AI bots are impressive, but they work by predicting patterns of words from web data. They lack true reasoning and understanding; and they certainly don't have wider aims or ambitions.
Fears of an "AI apocalypse" may indeed be overblown, said Steven Levy in Wired, but the leaders of just about every big AI company think superintelligence is coming soon "When you press them, they will also admit that controlling AI, or even understanding how it works, is a work in progress." Chinese experts are concerned: Beijing has established an $8.2 billion fund dedicated to AI control research. But in its haste to achieve AI dominance, America is ignoring calls for AI regulation and agreed standards. If the US "insists on eschewing guardrails and going full-speed towards a future that it can't contain", its biggest rival "will have no choice but to do the same".
A free daily email with the biggest news stories of the day – and the best features from TheWeek.com
-
Political cartoons for November 9Cartoons Sunday’s political cartoons include a ripoff, and the land of opportunity
-
A ‘golden age’ of nuclear powerThe Explainer The government is promising to ‘fire up nuclear power’. Why, and how?
-
Massacre in Darfur: the world looked the other wayTalking Point Atrocities in El Fasher follow decades of repression of Sudan’s black African population
-
Why Trump pardoned crypto criminal Changpeng ZhaoIn the Spotlight Binance founder’s tactical pardon shows recklessness is rewarded by the Trump White House
-
Is AI to blame for recent job cuts?Today’s Big Question Numerous companies have called out AI for being the reason for the culling
-
‘Deskilling’: a dangerous side effect of AI useThe explainer Workers are increasingly reliant on the new technology
-
AI models may be developing a ‘survival drive’Under the radar Chatbots are refusing to shut down
-
Saudi Arabia could become an AI focal pointUnder the Radar A state-backed AI project hopes to rival China and the United States
-
Why Britain is struggling to stop the ransomware cyberattacksThe Explainer New business models have greatly lowered barriers to entry for criminal hackers
-
Wikipedia: Is ‘neutrality’ still possible?Feature Wikipedia struggles to stay neutral as conservatives accuse the site of being left-leaning
-
AI is making houses more expensiveUnder the radar Homebuying is also made trickier by AI-generated internet listings