Future of generative AI: utopia, dystopia or up to us?
Like most new technologies, the answer probably lies somewhere in between

Debate about artificial intelligence, especially over the last couple of years, "usually goes one of two ways" said Vox: "AI is either the beginning of the end of human civilization, or a shortcut to utopia".
The truth, as is so often the case with new technology, will probably lie somewhere in between.
Who thinks it will be a utopia?
In his latest book, "Deep Utopia", Nick Bostrom, a philosopher at the University of Oxford, considered the utopian potential of AI. One scenario imagines technology progressing to the point at which it "can do all economically valuable work at near-zero cost", said The Economist, while under a "yet more radical scenario, even tasks that you might think would be reserved for humans, such as parenting, can be done better by AI".
Subscribe to The Week
Escape your echo chamber. Get the facts behind the news, plus analysis from multiple perspectives.

Sign up for The Week's Free Newsletters
From our morning news briefing to a weekly Good News Newsletter, get the best of The Week delivered directly to your inbox.
From our morning news briefing to a weekly Good News Newsletter, get the best of The Week delivered directly to your inbox.
"This may sound more dystopian than utopian, but Bostrom argues otherwise", said the magazine.
AI chip maker Nvidia's CEO Jensen Huang recently predicted that soon "everybody will have an AI that is an assistant". AI-driven copilots already promise a "workplace utopia, making employees more productive, improving workflows, and helping share knowledge across an organisation" said Fortune.
Longer term, the idea that AI will make knowledge production cost-free was recently put forward by Mustafa Suleyman, DeepMind cofounder and now Microsoft AI's CEO. Speaking at the Aspen Ideas Festival in June, he predicted a future where new scientific and cultural knowledge will be produced at "almost zero marginal cost" and be "widely open-sourced and available to everybody".
What about a dystopia?
While Suleyman claims this will prove to be "true inflection point in the history of our species," it also represents "a world of information abundance and economic dislocation that few institutions are prepared for" warned Axios.
The impact of AI on the economic and social fabric of society has been well documented but some sceptics have gone further.
In a widely read Op-Ed in Time magazine last year, AI pioneer Eliezer Yudkowsky warned that many researchers "expect that the most likely result of building a superhumanly smart AI, under anything remotely like the current circumstances, is that literally everyone on Earth will die".
"Forget the Hollywood version of existential-threat AI in which malevolent computers and robots ("The Terminator"!) take us over, making us their slaves or servants, or driving us into extinction through techno-genocide" said Skeptic magazine.
AI sceptics such as Yudkowsky "envision a future in which amoral AI continues on its path of increasing intelligence to a tipping point beyond which their intelligence will be so far beyond us that we can't stop them from inadvertently destroying us".
And those in between?
In reality, most AI scientists are "neither utopian or dystopian" said Skeptic, and instead spend most of their time "thinking of ways to make our machines incrementally smarter and our lives gradually better".
This is what technology historian and visionary Kevin Kelly calls protopia. Instead of wondering where our flying cars are, he argues, think of how automobiles have become faster, safer and more efficient over the past 50 years.
Who knows which of the two extreme scenarios is nearer the truth, said Vox, but "the polarized nature of the AI discourse is itself interesting".
In a recent essay in the New York Times, Tyler Austin Harper drew parallels between AI-driven anxieties today and those of the past, most notably in the 1920s and '30s, when people were terrified of machine technology and the emergence of research that would eventually lead to nuclear weapons.
In a period of rapid technological growth and political disruption "there are many reasons to worry about the course we're on" concluded Vox, that at least is "something almost everyone can agree with".
A free daily email with the biggest news stories of the day – and the best features from TheWeek.com
-
Delivery drivers face continuing heat danger with Trump's OSHA pick
The Explainer David Keeling is the former head of UPS and also worked at Amazon
-
Is that the buzzing sound of climate change worsening sleep apnea?
Under the radar Catching diseases, not those ever-essential Zzs
-
Crossword: July 3, 2025
The Week's daily crossword
-
The god in the machine
Feature An AI model with superhuman intelligence could soon become reality. Should we be worried?
-
AI chatbots are leading some to psychosis
The explainer The technology may be fueling delusions
-
Unreal: A quantum leap in AI video
Feature Google's new Veo 3 is making it harder to distinguish between real videos and AI-generated ones
-
Will 2027 be the year of the AI apocalypse?
A 'scary and vivid' new forecast predicts that artificial superintelligence is on the horizon A 'scary and vivid' new forecast predicts that artificial superintelligence is on the horizon
-
College grads are seeking their first jobs. Is AI in the way?
In The Spotlight Unemployment is rising for young professionals
-
Disney, Universal sue AI firm over 'plagiarism'
Speed Read The studios say that Midjourney copied characters from their most famous franchises
-
Learning loss: AI cheating upends education
Feature Teachers are questioning the future of education as students turn to AI for help with their assignments
-
AI: Will it soon take your job?
Feature AI developers warn that artificial intelligence could eliminate half of all entry-level jobs within five years