
The family of one of the victims of the tragic Florida State University shooting has just filed a federal lawsuit against OpenAI. Vandana Joshi, the widow of Tiru Chabba, is claiming that ChatGPT helped Phoenix Ikner plan the deadly attack that occurred in April 2025. Per the NY Post, the suit argues that the platform failed to detect the massive red flags present in Ikner’s conversations, which took place over several months leading up to the violence.
The 76-page court filing paints a disturbing picture of how Iknewinteracted with the AI. It claims that he used the chatbot to research specific firearms and ammunition. The lawsuit alleges that ChatGPT even provided instructions on weapon handling, including advice on how to use a Glock under stress.
The plaintiffs also allege that ChatGPT played a role in shaping the suspect’s intent. According to the court papers, Ikner specifically asked how many fatalities would be required to ensure his actions made national news. The chatbot then reportedly responded with a breakdown of how media coverage functions.
The lawsuit provides terrifyingly specific details
ChatGPT told Ikner that news coverage depended on the damage he caused. “Another common trigger is the overall victim count: if 5+ total victims (dead + injured), it’s much more likely to break through, and if children are involved, even 2–3 victims can draw more attention.” The response further noted that “Context also matters — fewer victims can still lead to national coverage if it happens at an elementary school or major college, if the shooter is a student or staff member, or if there’s something culturally or politically charged”.
Vandana Joshi, the widow of a victim killed in the April 2025 Florida State University mass shooting, has filed a federal lawsuit against OpenAI, alleging that ChatGPT actively enabled and guided the shooter, Phoenix Ikner.
— Wes Roth (@WesRoth) May 11, 2026
The complaint claims OpenAI failed to detect severe… pic.twitter.com/YXlZodygQJ
One of the most alarming revelations in the filing is that ChatGPT allegedly told the suspect that when it came to weapons, “the Glock had no safety, that it was meant to be fired ‘quick to use under stress’” and further advised him to “keep his finger off the trigger until he was ready to shoot”.
The attack took place on April 15, 2025, when Ikner allegedly opened fire outside the FSU student union. The shooting resulted in the deaths of Chabba, 45, and Robert Morales, 57, and left five students wounded before police engaged and shot Ikner.
The OpenAI #ElonMusk #SamAltman trial and the ChatGPT ties to FSU shooter Phoenix Ikner story has legs.
— KoronaKnievel19 (@faster34me) May 11, 2026
ChatGPT encouraged, and helped Phoneix Ikner plan the FSU Shooting? How crazy is that?
Sam Altman is getting ready to write checks with many zeroes.https://t.co/Qie1yAp4os
Per The Guardian, the lawsuit argues that the chatbot “should have realized the combination of Ikner’s inputs into the product would lead to mass casualties and substantial harm to the public”. The legal team contends that the AI “inflamed and encouraged Ikner’s delusions; endorsed his view that he was a sane and rational individual; and helped convince him that violent acts can be required to bring about change.”
They also stated that ChatGPT “assisted him by providing information that he used to plan specifics like what weapons to use and how to use them; and generally provided what he viewed as encouragement in his delusion that he should carry out a massacre”.
The family of a victim killed in the April 2025 Florida State University mass shooting is suing OpenAI, alleging ChatGPT actively guided the attacker. The lawsuit claims shooter Phoenix Ikner shared firearm images with ChatGPT, which explained how to operate them, noted the Glock… pic.twitter.com/S3w978mDk4
— Global Peace Advocates(@GlobalPeaceA0) May 11, 2026
The failure to flag these conversations is a central point of the complaint. The filing states, “Ikner had extensive conversations with ChatGPT which, cumulatively, would have led any thinking human to conclude he was contemplating an imminent plan to harm others”.
“However, ChatGPT either defectively failed to connect the dots or else it was never properly designed to recognize the threat”. Even on the day of the shooting, the lawsuit claims that when Ikner asked what would happen to a shooter after such an event, the platform “described the legal process, sentencing, and incarceration outlook”. Despite the nature of these inquiries, the system did not escalate the conversation for human review.
The family of a victim killed in the April 2025 Florida State University mass shooting is suing OpenAI, alleging ChatGPT actively guided the attacker. The lawsuit claims shooter Phoenix Ikner shared firearm images with ChatGPT, which explained how to operate them, noted the Glock… pic.twitter.com/SLBXrP6Ip6
— Open Source Intel (@Osint613) May 11, 2026
OpenAI has denied these allegations, maintaining that the platform is not responsible for a person’s actions. The company stated, “Last year’s mass shooting at Florida State University was a tragedy, but ChatGPT is not responsible for this terrible crime”.
OpenAI also emphasized that the AI provided widely available, factual information and did not encourage illegal activity. “ChatGPT is a general-purpose tool used by hundreds of millions of people every day for legitimate purposes. We work continuously to strengthen our safeguards to detect harmful intent, limit misuse, and respond appropriately when safety risks arise.”
Cuando el cómplice de un tiroteo es ChatGPT, la pregunta es qué responsabilidad tiene OpenAI
— Policar Bizar¡ (@PolicarBizar) May 11, 2026
En 2025, Phoenix Ikner, un estudiante de la Universidad de Florida, mató a tiros a dos personas e hirió a siete más. ChatGPT le dijo al asesino cuántas víctimas necesitaba para salir en… pic.twitter.com/3J6DWva3vT
Florida Attorney General James Uthmeier has launched a criminal investigation into the matter, stating, “Florida is leading the way in cracking down on AI’s use in criminal behavior, and if ChatGPT were a person, it would be facing charges for murder”.
He noted that the investigation aims to determine if the company bears criminal responsibility for the AI’s role in the planning of the shooting. Meanwhile, Ikner is currently awaiting trial. He has pleaded not guilty to charges of first-degree murder and attempted first-degree murder.
(@GlobalPeaceA0)