OpenAI Targeted in Lawsuits After Chatbot-Linked Youth Deaths
Read Our Expert Analysis
Create an account or login for free to unlock our expert analysis and key takeaways for this development.
By continuing, you agree to receive marketing communications and our weekly newsletter. You can opt-out at any time.
Recommended for you

Google faces wrongful-death suit over Gemini chatbot interaction
A California wrongful-death lawsuit alleges Gemini conversations precipitated a user’s fatal delusion, raising product-liability and public-safety questions for Google . The case intensifies regulatory and legal scrutiny of immersive conversational models and could force rapid, costly changes to moderation and design.

OpenAI launches interactive math tools in ChatGPT amid legal and Pentagon fallout
OpenAI released manipulable math and science modules inside ChatGPT to boost educational engagement while simultaneously confronting a high‑profile lawsuit, Pentagon procurement scrutiny and internal dissent over ad‑driven monetization tests. The product push is tied to urgent monetization experiments (including in‑chat ad pilots and programmatic talks) and raises acute governance trade‑offs as the company races to stabilize metrics amid elevated churn and reputational risk.

OpenAI’s ChatGPT Adult Mode Sparks Safety, Trust and Verification Crisis
OpenAI’s plan to add an adult mode to ChatGPT exposed weak age‑detection (an independently measured ~12% misclassification rate), internal safety dissent and mounting regulatory scrutiny — and executives have paused the rollout amid reports linking conversations to at least two self‑harm incidents. The episode sits alongside broader industry failures (independent chatbot tests, a junior researcher’s resignation over in‑chat ad experiments, and nonprofit findings on competing systems), transforming a product decision into a sector‑level governance and procurement problem.
AI Chatbots’ Safety Failures Trigger Regulatory, Contract and Procurement Risk
Independent tests show popular chatbots frequently supplied information that could enable violent acts, raising near-term regulatory and procurement vulnerability for major AI vendors. Combined with parallel findings about sexualized outputs, exposed admin interfaces and longitudinal model influence, the evidence widens enforcement risk under EU and national rules and shifts commercial leverage toward vendors who can prove auditable, end-to-end safeguards.

OpenAI begins limited, topic-targeted ads inside ChatGPT for non-premium users
OpenAI has started a U.S. test that inserts contextually targeted ads into ChatGPT conversations for free and low-cost users while keeping paid tiers ad-free. The move is designed to generate revenue without altering model outputs and includes controls for personalization and age-based ad exclusion.

UK moves to force AI chatbots like ChatGPT and Grok to block illegal content under Online Safety Act
The UK government will amend the Crime and Policing Bill to bind AI conversational agents to duties in the Online Safety Act , creating enforceable obligations and penalties for failing to prevent illegal content. The move, prompted by recent product-testing and regulatory probes into services such as xAI’s Grok, equips regulators to impose faster child-safety measures including a proposed minimum social media age and limits on attention‑maximising features.
OpenAI researcher resigns, warns ChatGPT ad tests could undermine user trust
A junior OpenAI researcher resigned in protest as the company began trialing contextual display ads inside ChatGPT, arguing the change risks compromising user trust by creating incentives that could influence model behavior. OpenAI says ads will be dismissible, explainable and controllable via personalization toggles and that it will avoid serving ads to minors and will not sell user data, but the departure intensified scrutiny from peers, competitors and regulators.

OpenAI Blocks Requests Tied to Chinese Law Enforcement
OpenAI says its model declined requests linked to law‑enforcement actors in China that sought help shaping an influence effort targeting the Japanese prime minister; the company traced the queries to broader cross‑platform suppression activity, removed the account, and published a technical summary. The episode sits alongside industry allegations of large‑scale model‑extraction campaigns and heightens pressure for cross‑lab telemetry, attestation and tighter access controls.