Journal logo

Unraveling OpenAI's AI Safety Conundrum: Promises Unfulfilled.

Exploring the Gaps in OpenAI's AI Safety Promises

By Kevin MacELweePublished about a month ago 5 min read

Prelude: An AI Powerhouse's Pledge

In the summer of the previous year, the artificial intelligence juggernaut OpenAI made a solemn vow to the White House. They committed to subjecting their groundbreaking technology, specifically new iterations of their models, to rigorous safety testing. The primary objective was to ensure that their AI would not inflict harm, such as inadvertently instructing users on constructing bioweapons or aiding hackers in developing novel cyberattacks.

The Incident: A Party Before the Verdict

Even before the testing phase commenced for the GPT-4 Omni model, which would eventually power the renowned ChatGPT, OpenAI had already initiated celebrations. Employees were invited to a party at one of the company's San Francisco offices, ostensibly to commemorate the imminent product launch. This premature revelry raised eyebrows, as an anonymous source disclosed, "They planned the launch after-party prior to knowing if it was safe to launch." The source added, "We basically failed at the process."

A Shift in Priorities: From Altruism to Commercialism?

This incident sheds light on the evolving culture within OpenAI, where company leaders, including CEO Sam Altman, have faced accusations of prioritizing commercial interests over public safety. This trajectory marks a stark departure from the company's roots as an altruistic nonprofit organization. Furthermore, it raises pertinent questions about the federal government's reliance on self-policing by tech companies, as outlined in the White House pledge and the executive order on AI passed in October.

Experts Weigh In: The Risks of Self-Regulation

Andrew Strait, a former ethics and policy researcher at Google DeepMind and currently the associate director at the Ada Lovelace Institute in London, expressed concern over allowing companies to set their own standards for safety. "We have no meaningful assurances that internal policies are being faithfully followed or supported by credible methods," Strait cautioned.

The White House Stance: Expectations and Commitments

President Biden has been unequivocal in his stance towards tech companies, emphasizing the importance of ensuring their products are safe, secure, and trustworthy before releasing them to the public. Robyn Patterson, a White House spokeswoman, affirmed, "Leading companies have made voluntary commitments related to independent safety testing and public transparency, which he expects they will meet."

Companies Under the Spotlight

OpenAI is one of more than a dozen companies that made voluntary commitments to the White House last year, preceding the AI executive order. Notable participants include Anthropic (the company behind the Claude chatbot), Nvidia (the $3 trillion chips juggernaut), Palantir (the data analytics company that works with militaries and governments), Google DeepMind, and Meta.

The Testing Protocol: A Compressed Endeavor

OpenAI's newest model, GPT-4o, presented the company's first significant opportunity to apply the framework outlined in the White House pledge. However, testers condensed the evaluations into a single week, despite complaints from employees. Although they anticipated the technology to pass the tests, numerous employees were dismayed by OpenAI's treatment of its vaunted new preparedness protocol as an afterthought.

Insider Perspectives: Dissent and Resignations

In June, several current and former OpenAI employees signed a cryptic open letter demanding that AI companies exempt their workers from confidentiality agreements, allowing them to warn regulators and the public about potential safety risks associated with the technology.

Former OpenAI executive Jan Leike resigned days after the GPT-4o launch, stating on X (formerly Twitter) that "safety culture and processes have taken a backseat to shiny products."

William Saunders, a former OpenAI research engineer who resigned in February, noted in a podcast interview that he had observed a pattern of "rushed and not very solid" safety work "in service of meeting the shipping date" for a new product.

OpenAI's Response: Acknowledging Shortcomings

A representative of OpenAI's preparedness team, who spoke on the condition of anonymity, acknowledged that the evaluations took place during a single week, which was sufficient to complete the tests. However, they conceded that the timing had been "squeezed." The representative added, "We are rethinking our whole way of doing it. This [was] just not the best way to do it."

In a statement, OpenAI spokesperson Lindsey Held insisted that the company "didn't cut corners on our safety process," while recognizing that the launch was "stressful for our teams." To comply with the White House commitments, the company "conducted extensive internal and external" tests and held back some multimedia features "initially to continue our safety work," she added.

The Preparedness Initiative: A Scientific Approach

OpenAI announced the preparedness initiative as an attempt to bring scientific rigor to the study of catastrophic risks, which it defined as incidents "which could result in hundreds of billions of dollars in economic damage or lead to the severe harm or death of many individuals."

Altman's Vision: Setting a New Standard

In October, Altman took to X (formerly Twitter) to announce the company's new team, stating, "We aim to set a new high-water mark for quantitative, evidence-based work."

Dedicated Safety Teams

In the past year, OpenAI has launched two new safety teams, joining an existing division focused on concrete harms, such as racial bias or misinformation.

The Superalignment team, announced in July, was dedicated to preventing existential risks from far-advanced AI systems. It has since been redistributed to other parts of the company.

Ilya Sutskever, a former board member who initially voted to oust Altman as CEO in November before quickly recanting, and Jan Leike led the Superalignment team. Both resigned in May.

The Preparedness Team's Perspective

According to the OpenAI representative, the preparedness team had the full support of top executives. Recognizing that the timing for testing GPT-4o would be tight, the representative spoke with company leaders, including Chief Technology Officer Mira Murati, in April. They agreed on a "fallback plan" – if the evaluations revealed anything alarming, the company would launch an earlier iteration of GPT-4o that the team had already tested.

Preparatory Efforts and Costs

A few weeks prior to the launch date, the team began conducting "dry runs," aiming to have "all systems go the moment we have the model." They scheduled human evaluators in different cities to be ready to run tests, a process that cost hundreds of thousands of dollars, according to the representative.

Prep work also involved warning OpenAI's Safety Advisory Group – a newly created board of advisers who receive a scorecard of risks and advise leaders if changes are needed – that it would have limited time to analyze the results.

A Commitment to Improvement

OpenAI's Held stated that the company committed to allocating more time for the process in the future. The representative acknowledged the intensity of the process, saying, "After that, we said, 'Let's not do it again.'"

Conclusion: A Cautionary Tale

The incident surrounding OpenAI's handling of the GPT-4o safety testing serves as a cautionary tale, highlighting the potential pitfalls of prioritizing commercial interests over public safety in the rapidly evolving field of artificial intelligence. As AI continues to reshape virtually every aspect of human society, from work to warfare, it is imperative that companies like OpenAI uphold their commitments to rigorous safety protocols and transparent practices. The stakes are too high to compromise on these principles, as the consequences of a misstep could be catastrophic. It remains to be seen whether OpenAI and other industry leaders will heed the lessons from this incident and take concrete steps to regain the public's trust in their ability to develop and deploy AI responsibly.

businessindustryhumanity

About the Creator

Kevin MacELwee

"Hello, my name is Kevin, a former electrician and construction worker now exploring online entrepreneurship. I'm passionate about animal welfare and inspired by 'Rich Dad Poor Dad' by Robert Kiyosaki. I also have a YouTube channel as well.

Enjoyed the story? Support the Creator.

Subscribe for free to receive all their stories in your feed. You could also become a paid subscriber, letting them know you appreciate their work.

Subscribe

Reader insights

Comments (1)

Sign in to comment
  • Esala Gunathilakeabout a month ago

    It is truly informative.

Find us on social media

Miscellaneous links

  • Explore
  • Contact
  • Privacy Policy
  • Terms of Use
  • Support

© 2024 Creatd, Inc. All Rights Reserved.