BitcoinWorld Inside OpenAI’s safety crisis: Former employees testify in Musk lawsuit Elon Musk’s legal campaign to dismantle OpenAI’s for-profit structure is forcingBitcoinWorld Inside OpenAI’s safety crisis: Former employees testify in Musk lawsuit Elon Musk’s legal campaign to dismantle OpenAI’s for-profit structure is forcing

Inside OpenAI’s safety crisis: Former employees testify in Musk lawsuit

2026/05/08 03:35
5분 읽기
이 콘텐츠에 대한 의견이나 우려 사항이 있으시면 crypto.news@mexc.com으로 연락주시기 바랍니다

BitcoinWorld

Inside OpenAI’s safety crisis: Former employees testify in Musk lawsuit

Elon Musk’s legal campaign to dismantle OpenAI’s for-profit structure is forcing a rare public examination of how the company’s shift toward commercial products may have compromised its founding mission: ensuring that artificial general intelligence (AGI) benefits all of humanity. On Thursday, a federal court in Oakland heard testimony from a former employee and a former board member who described a pattern of safety lapses and governance failures inside the AI lab.

Safety teams disbanded as product pressure mounted

Rosie Campbell joined OpenAI’s AGI readiness team in 2021 and left in 2024 after her team was disbanded. Another safety-focused group, the Super Alignment team, was shut down during the same period. Campbell testified that when she joined, the culture was heavily research-oriented, with frequent discussions about AGI and safety. “Over time it became more like a product-focused organization,” she said.

Under cross-examination, Campbell acknowledged that significant funding is necessary for building AGI, but argued that creating a super-intelligent model without adequate safety measures contradicts the mission she originally signed up for. She pointed to a specific incident where Microsoft deployed a version of OpenAI’s GPT-4 model in India through its Bing search engine before the company’s Deployment Safety Board (DSB) had evaluated it. While the model itself posed no major risk, Campbell stressed the importance of setting strong precedents. “We want to have good safety processes in place we know are being followed reliably,” she testified.

Board governance under scrutiny

The deployment of GPT-4 in India was one of the red flags that led OpenAI’s non-profit board to briefly fire CEO Sam Altman in November 2023. Tasha McCauley, a board member at the time, testified about concerns that Altman was not forthcoming enough for the board’s unusual structure to function effectively. She described a pattern of misleading behavior, including Altman lying to another board member about McCauley’s intention to remove a third board member, Helen Toner, who had published a white paper with implied criticism of OpenAI’s safety policies.

McCauley also noted that Altman failed to inform the board about the decision to launch ChatGPT publicly, and that his disclosure of potential conflicts of interest was inadequate. “We are a non-profit board and our mandate was to be able to oversee the for-profit underneath us,” she told the court. “Our primary way to do that was being called into question. We did not have a high degree of confidence at all to trust that the information being conveyed to us allowed us to make decisions in an informed way.”

When OpenAI’s staff rallied behind Altman and Microsoft worked to restore the status quo, the board reversed course, and the members opposed to Altman stepped down. This episode lies at the heart of Musk’s argument that the transformation of OpenAI from a research organization into one of the largest private companies in the world broke the implicit agreement among its founders.

Expert testimony and broader implications

David Schizer, a former dean of Columbia Law School who is serving as an expert witness for Musk’s team, echoed McCauley’s concerns. “OpenAI has emphasized that a key part of its mission is safety and they are going to prioritize safety over profits,” Schizer said. “Part of that is taking safety rules seriously, if something needs to be subject to safety review, it needs to happen. What matters is the process issue.”

With AI already deeply embedded in for-profit companies, the implications extend far beyond a single lab. McCauley argued that the governance failures at OpenAI should be a reason to embrace stronger government regulation of advanced AI. “If it all comes down to one CEO making those decisions, and we have the public good at stake, that’s very suboptimal,” she said.

Conclusion

The Oakland hearing underscores a fundamental tension at OpenAI: the pressure to commercialize AI products versus the non-profit mission of ensuring safe AGI. As Musk’s lawsuit proceeds, the testimony from former employees and board members is providing an unusually detailed look at how internal safety processes and governance structures have evolved—or failed to evolve—alongside the company’s rapid growth. For regulators, investors, and the public, the case is becoming a critical test of whether corporate accountability can keep pace with AI’s accelerating capabilities.

FAQs

Q1: What is the central issue in Elon Musk’s lawsuit against OpenAI?
The lawsuit argues that OpenAI’s shift from a non-profit research organization to a for-profit commercial entity violated its founding mission of developing AGI safely for the benefit of humanity. The court is examining whether this transformation broke implicit agreements among the founders.

Q2: What specific safety failures were highlighted in the testimony?
Former employee Rosie Campbell testified that the company’s Deployment Safety Board was bypassed when Microsoft deployed GPT-4 in India. She also noted that two key safety teams—the AGI readiness team and the Super Alignment team—were disbanded as the company became more product-focused.

Q3: How does this case affect the broader AI industry?
The case is being watched closely as a potential precedent for how AI companies balance safety and profit. Witnesses have called for stronger government regulation, arguing that relying on a single CEO to make decisions affecting public safety is “suboptimal.” The outcome could influence how other AI labs structure their governance and safety processes.

This post Inside OpenAI’s safety crisis: Former employees testify in Musk lawsuit first appeared on BitcoinWorld.

시장 기회
Delysium 로고
Delysium 가격(AGI)
$0.00969
$0.00969$0.00969
-2.61%
USD
Delysium (AGI) 실시간 가격 차트
면책 조항: 본 사이트에 재게시된 글들은 공개 플랫폼에서 가져온 것으로 정보 제공 목적으로만 제공됩니다. 이는 반드시 MEXC의 견해를 반영하는 것은 아닙니다. 모든 권리는 원저자에게 있습니다. 제3자의 권리를 침해하는 콘텐츠가 있다고 판단될 경우, crypto.news@mexc.com으로 연락하여 삭제 요청을 해주시기 바랍니다. MEXC는 콘텐츠의 정확성, 완전성 또는 시의적절성에 대해 어떠한 보증도 하지 않으며, 제공된 정보에 기반하여 취해진 어떠한 조치에 대해서도 책임을 지지 않습니다. 본 콘텐츠는 금융, 법률 또는 기타 전문적인 조언을 구성하지 않으며, MEXC의 추천이나 보증으로 간주되어서는 안 됩니다.

KAIO Global Debut

KAIO Global DebutKAIO Global Debut

Enjoy 0-fee KAIO trading and tap into the RWA boom