The ChatGPt website on a tablet.(Eduardo Parra / Europa Press via Getty Images)
ChatGPT, the new generative AI chatbot created by OpenAI, was released less than five months ago, but it has already become the fastest-growing Internet application in history. It’s almost powerful enough to work some white-collar jobs. New York Times columnist Farhad Manjoo says it’s already changed the way he does his.
For all the talk of AI’s potential, the people who understand it best are frightened by how quickly the technology is accelerating. On one hand, it could usher in an era of unprecedented human health and happiness. On the other, it has the potential for massive economic disruption, weakened national security, and the erosion of personal privacy.
It gets worse: In a survey of AI experts conducted last year, almost half said the chance that AI would lead to human extinction was 10 percent or more. Even Elon Musk, a founding member of OpenAI and someone rarely seen as concerned about consequences, said last week that AI “has the potential of civilization destruction.”
We are at an inflection point with AI. New technology demands new regulations. But AI development is outpacing regulators’ ability to act, or even to understand. To avert the worst possible outcomes, leaders need to listen to the experts who can anticipate the ramifications, and regulate now—before it’s too late.
We’ve seen this show before. Proponents of the Industrial and Digital Revolutions dangled utopian visions in front of our eyes while shunning the regulations that would have ensured safer and more equitable change. As a result, we’re still uncovering and cleaning up industrial waste sites from the 1800s, and only beginning to understand how our failure to regulate social-media-created crises of democracy and public health.
AI may be our most threatening revolution yet. Google CEO Sundar Pichai has said that the impact of AI will be “more profound than fire.” But even cavemen understood that playing with fire can have third-degree consequences.
The United States government is woefully ill-equipped to tackle the existential threat of AI. Republican Representative Jay Obernolte, the only member of Congress with a master’s degree in AI, has to explain to his colleagues that the threat “will not come from evil robots with red lasers.” Congress is clearly not ready to legislate.
But we can’t let the complexity of AI prevent us from acting today while we unravel the broader questions we’ll face tomorrow. As Representative Ted Lieu reminds us, we don’t fully understand all the ways pharmaceuticals improve health, but the FDA still regulates drug manufacturing to keep us safe.
An essential first step is to make sure AI doesn’t fall victim to the same monopolistic practices that undermined the first age of the Internet. FTC Chair Lina Khan is already gearing up the agency to watch out for antitrust violations to make sure the technology isn’t dangerously concentrated in too few hands.
The Biden administration has taken steps to ask questions and get better informed. The Commerce Department is seeking public comment as it considers measures to regulate AI. And last year, the White House unveiled a Blueprint for an AI Bill of Rights, but critics have noted that its protections are “toothless” and “uneven.”
As the government gathers information about this new technology, it cannot make the mistake of privileging the voices of the tech industry already lobbying against regulations. Instead, it should look to states like California and Texas, which are regulating AI tools to protect civil rights and demand transparency.
Over the past year you’ve read Nation writers like Elie Mystal, Kaveh Akbar, John Nichols, Joan Walsh, Bryce Covert, Dave Zirin, Jeet Heer, Michael T. Klare, Katha Pollitt, Amy Littlefield, Gregg Gonsalves, and Sasha Abramsky take on the Trump family’s corruption, set the record straight about Robert F. Kennedy Jr.’s catastrophic Make America Healthy Again movement, survey the fallout and human cost of the DOGE wrecking ball, anticipate the Supreme Court’s dangerous antidemocratic rulings, and amplify successful tactics of resistance on the streets and in Congress.
We publish these stories because when members of our communities are being abducted, household debt is climbing, and AI data centers are causing water and electricity shortages, we have a duty as journalists to do all we can to inform the public.
In 2026, our aim is to do more than ever before—but we need your support to make that happen.
Through December 31, a generous donor will match all donations up to $75,000. That means that your contribution will be doubled, dollar for dollar. If we hit the full match, we’ll be starting 2026 with $150,000 to invest in the stories that impact real people’s lives—the kinds of stories that billionaire-owned, corporate-backed outlets aren’t covering.
With your support, our team will publish major stories that the president and his allies won’t want you to read. We’ll cover the emerging military-tech industrial complex and matters of war, peace, and surveillance, as well as the affordability crisis, hunger, housing, healthcare, the environment, attacks on reproductive rights, and much more. At the same time, we’ll imagine alternatives to Trumpian rule and uplift efforts to create a better world, here and now.
While your gift has twice the impact, I’m asking you to support The Nation with a donation today. You’ll empower the journalists, editors, and fact-checkers best equipped to hold this authoritarian administration to account.
I hope you won’t miss this moment—donate to The Nation today.
Onward,
Katrina vanden Heuvel
Editor and publisher, The Nation
Better yet, look overseas. As usual, Europe is ahead of us when it comes to regulation. ChatGPT was released only six months ago, but several EU countries are already investigating privacy concerns. Italy has banned it outright.
A group of EU officials have called on President Biden to convene a summit to encourage collaboration on global solutions. The White House should accept that invitation.
Last week, two AI experts in The Economist called for an “International Agency for AI” to coordinate a global response—not unlike the International Atomic Energy Agency. But, as Robert Wright notes, the problem doesn’t have to reach nuclear levels to merit global governance. No country wants to be the first to step in and regulate. That’s why we should do it together.
Given the breathtaking speed at which AI is evolving, it’s unlikely that we will find a one-size-fits-all solution. Addressing the potential threat of AI will require the humility to admit what we don’t know, and a willingness to heed the criticism of skeptics as much as the promises of evangelists.
We can create a world where AI meets its spectacular promise to improve the lives of all humanity—but only if we work together, and use our unique capacity to make moral choices to develop a truly intelligent (read not artificial) response.
Katrina vanden HeuvelTwitterKatrina vanden Heuvel is editor and publisher of The Nation, America’s leading source of progressive politics and culture. An expert on international affairs and US politics, she is an award-winning columnist and frequent contributor to The Guardian. Vanden Heuvel is the author of several books, including The Change I Believe In: Fighting for Progress in The Age of Obama, and co-author (with Stephen F. Cohen) of Voices of Glasnost: Interviews with Gorbachev’s Reformers.