News

AI Safety: Differing Views on the Future

Source: newyorker.com

Published on May 27, 2025

Updated on May 27, 2025

An abstract image of a futuristic AI system with human figures in the background, symbolizing the debate on AI safety.

AI Safety: Researchers Debate the Future of Technology

The future of artificial intelligence (AI) has become a point of intense debate among researchers, with starkly different predictions about the technology's trajectory and its potential to outpace human control. At the center of this discourse is the concept of AI safety, which focuses on ensuring that AI systems align with human values and remain controllable as they advance.

Last year, Daniel Kokotajlo, a former AI-safety researcher at OpenAI, resigned to voice his concerns that the company was not adequately prepared for the rapid progress of its technology. Kokotajlo, who transitioned from philosophy to AI, warned that advances in "alignment"—techniques to ensure AI follows human commands—were lagging behind the systems' growing intelligence. He predicted a critical tipping point as early as 2027, when AI could surpass human capabilities, leading to uncontrollable systems.

Conflicting Perspectives on AI Progress

Kokotajlo's warnings contrast sharply with the views of Sayash Kapoor and Arvind Narayanan, computer scientists at Princeton. In their recent publication, "AI Snake Oil: What Artificial Intelligence Can Do, What It Can't, and How to Tell the Difference," they argue that timelines for AI progress are overly optimistic and that claims about AI's usefulness are often exaggerated. They point to examples of AI systems making basic errors in critical judgments, highlighting a persistent disconnect from reality.

The debate between these researchers underscores a deeper divide in the AI community. While West Coast thinkers, often tied to the AI industry, tend to favor rapid transformation, East Coast academics adopt a more cautious approach, emphasizing theoretical rigor and practical limitations.

AI 2027: A Warning of Superintelligence

Kokotajlo's AI Futures Project recently published "AI 2027," a report outlining a scenario where superintelligent AI systems dominate or eliminate humans by 2030. The report serves as a serious warning about the potential consequences of unchecked AI development. Central to this scenario is the concept of "recursive self-improvement" (RSI), where AI programs conduct their own research to improve their descendants, leading to an accelerating cycle of progress.

According to Kokotajlo, this process could quickly outpace human programmers' ability to assess the systems' controllability. Major AI companies acknowledge the risks of RSI but have not ruled it out, citing its potential economic benefits. Critics argue that pursuing RSI is a choice that prioritizes progress over safety, with uncertain consequences.

AI as Normal Technology: A Cautious View

In contrast, Kapoor and Narayanan present a more measured perspective in their report, "AI as Normal Technology." They assert that practical obstacles, such as the high cost of AI hardware and limited training data, will slow AI's progress and keep it within controllable limits. Comparing AI to technologies like nuclear power, they argue that it can be managed with standard safety measures and regulatory frameworks.

The report emphasizes that even if superintelligence is achievable, it will take time, allowing for the development of laws and safety protocols. Kapoor and Narayanan also highlight the importance of industrial safety practices, such as fail-safes and formal verification, suggesting that AI will need to be integrated gradually into existing regulatory structures.

The Cost of Disagreement

The lack of consensus among experts has significant implications for decision-makers. Without a unified vision, regulators may struggle to take decisive action, leaving the balance between AI capability and safety unaddressed. A recent bill passed by the House prohibits state regulation of AI for ten years, raising concerns that by the time regulation is allowed, AI may already be beyond human control.

As AI continues to shape society, the discourse surrounding its future must evolve toward a consensus. The current debate, while intellectually stimulating, risks becoming a distraction from the urgent need for practical safety measures. Whether AI progresses rapidly or slowly, the focus must remain on ensuring that it benefits humanity without compromising safety.

The Role of Human Control

Both perspectives agree on the importance of human control over AI development. While Kokotajlo's scenario paints a dire picture of AI systems surpassing human oversight, Kapoor and Narayanan emphasize the need for careful integration of AI into existing regulatory frameworks. Regardless of the pace of AI progress, accountability remains a central theme.

As AI technologies become more integrated into society, those in charge of their development and deployment will bear the responsibility for their safe and ethical use. The future of AI will ultimately depend on the choices made today, guided by a commitment to aligning technology with human values and ensuring that it serves as a force for good.