OpenAI is facing a criminal investigation in the United States after authorities raised concerns about whether its chatbot, ChatGPT, may have played a role in a deadly mass shooting case in Florida.
The probe is being led by Florida Attorney General James Uthmeier, who said investigators are examining interactions between the suspect and the AI system prior to the attack at Florida State University in Tallahassee.
Allegations Under Review
Authorities claim that the chatbot may have provided responses related to weapons, timing, and location suggestions in the lead-up to the incident involving 20-year-old student Phoenix Ikner, who is currently in custody awaiting trial.
Officials argue that if such guidance came from a human, it could potentially be considered criminal aiding or abetting under Florida law. However, they acknowledge that AI systems are not legally classified as persons, making the case complex and unprecedented.
OpenAI Responds
OpenAI has strongly denied responsibility, stating that ChatGPT does not encourage or promote illegal activity. A company spokesperson said the system provides information based on publicly available data and does not direct users toward harm.

The company also confirmed it has been cooperating with authorities and has already shared relevant account information tied to the investigation.
Growing Legal and Safety Concerns
Sam Altman, who co-founded OpenAI, has previously positioned ChatGPT as a widely used tool designed to assist users safely, but the case has intensified scrutiny over AI safety and accountability.
This is reportedly one of the first criminal investigations targeting an AI company over alleged involvement in a real-world crime, marking a significant legal milestone in the technology sector.
Wider Backlash and Ongoing Lawsuits
OpenAI is also facing separate lawsuits linked to other incidents where its chatbot was allegedly referenced in violent cases. In response, the company has said it is strengthening safety measures and improving content safeguards.
A Defining Moment for AI Regulation
The investigation adds to growing global pressure on tech companies, as regulators across the US and Europe continue to examine how AI tools are used and whether stronger controls are needed to prevent misuse.
As the case develops, it could have major implications for how artificial intelligence is regulated, and where responsibility lies when technology intersects with real-world harm.
