- Today
- Holidays
- Birthdays
- Reminders
- Cities
- Atlanta
- Austin
- Baltimore
- Berwyn
- Beverly Hills
- Birmingham
- Boston
- Brooklyn
- Buffalo
- Charlotte
- Chicago
- Cincinnati
- Cleveland
- Columbus
- Dallas
- Denver
- Detroit
- Fort Worth
- Houston
- Indianapolis
- Knoxville
- Las Vegas
- Los Angeles
- Louisville
- Madison
- Memphis
- Miami
- Milwaukee
- Minneapolis
- Nashville
- New Orleans
- New York
- Omaha
- Orlando
- Philadelphia
- Phoenix
- Pittsburgh
- Portland
- Raleigh
- Richmond
- Rutherford
- Sacramento
- Salt Lake City
- San Antonio
- San Diego
- San Francisco
- San Jose
- Seattle
- Tampa
- Tucson
- Washington
AI Leaders Warn of Existential Risks from Rapid Progress
Top AI companies call for pause and new safeguards as advanced systems outpace regulations.
Apr. 9, 2026 at 5:09am
Got story updates? Submit your updates here. ›
As AI capabilities rapidly advance, industry leaders warn of the urgent need for new global governance to ensure these powerful systems remain under human control.Today in TulsaThe leaders of the world's leading artificial intelligence (AI) companies - including OpenAI, Google Deepmind, and Anthropic - signed a letter in 2023 warning of the existential risks emerging from rapidly advancing AI systems that are outpacing current regulations and safety protocols.
Why it matters
As AI capabilities continue to grow exponentially, there are increasing concerns that the technology could spiral out of control and pose existential threats to humanity if not properly managed. This open letter from top AI executives highlights the urgent need for new global governance frameworks to ensure advanced AI systems are developed and deployed safely.
The details
The letter, signed by over 1,000 AI researchers and executives, called for a pause on the development of powerful AI systems until new safety protocols can be established. It warned that the rapid progress in areas like large language models, generative AI, and autonomous systems poses serious risks around issues like AI-driven misinformation, algorithmic bias, and the potential for advanced AI to be used for malicious purposes.
- The open letter was published in January 2023.
The players
OpenAI
A leading artificial intelligence research company founded in 2015, known for developing advanced language models like GPT-3.
Google Deepmind
An AI research company owned by Alphabet, the parent company of Google, that has made breakthroughs in areas like game-playing AI and protein folding.
Anthropic
An AI safety startup founded in 2021 that is working on developing safe and ethical artificial general intelligence (AGI) systems.
What they’re saying
“We may be on the path to developing machines vastly more intelligent than us, and we have no clear idea how to ensure they remain under our control or behave in ways that are consistent with human values.”
— Dario Amodei, Vice President of Research, Anthropic
What’s next
AI researchers and policymakers are calling for urgent global cooperation to establish new governance frameworks and safety protocols to mitigate the risks of advanced AI systems before they become too powerful to control.
The takeaway
The open letter from top AI leaders highlights the growing consensus that the rapid progress in artificial intelligence poses existential risks that current regulations are ill-equipped to handle. Proactive steps are needed to ensure advanced AI systems are developed and deployed in a safe and responsible manner.
Tulsa top stories
Tulsa events
Apr. 9, 2026
Gwar w/ SoulflyApr. 9, 2026
Mamma Mia! (Touring)Apr. 10, 2026
Tulsa Oilers Vs Idaho Steelheads - Game 34




