**
In the wake of a tragic school shooting in Tumbler Ridge, British Columbia, serious questions have arisen regarding the role of technology firms in monitoring and reporting concerning online behaviour. An 18-year-old gunman, identified as Jesse Van Rootselaar, killed six individuals, including five students and a teacher’s aide, before taking her own life on February 10. Disturbingly, it has come to light that OpenAI had suspended the shooter’s ChatGPT account months prior due to alarming content but did not alert law enforcement.
A Timeline of Tragedy and Oversight
The chilling sequence of events began with a series of violent actions perpetrated by Van Rootselaar, who first murdered her mother and half-brother before proceeding to the school. This incident has left the Tumbler Ridge community in shock and mourning. Just one day after the shooting, OpenAI representatives convened with the British Columbia government to discuss the possibility of establishing a satellite office in Canada. The following day, they approached provincial officials for assistance in liaising with the Royal Canadian Mounted Police (RCMP).
While the tragic incident unfolded, OpenAI had previously flagged the shooter’s disturbing posts through its automatic screening systems, leading to the suspension of her ChatGPT account in June. However, the company did not consider the content concerning enough to warrant immediate action or notification of law enforcement at the time. Premier David Eby and federal AI Minister Evan Solomon have expressed deep concern regarding OpenAI’s decision not to inform authorities sooner, labelling the situation as profoundly unsettling for the victims’ families and the wider British Columbian community.
The Aftermath: Calls for Accountability
In the wake of these revelations, the B.C. government is taking steps to ensure that potential evidence related to the shooting—particularly that held by digital services companies—is preserved. Premier Eby emphasised the government’s commitment to providing police with all necessary resources to investigate this horrific tragedy thoroughly.

Reports suggest that OpenAI employees had previously urged the company to inform law enforcement about the shooter’s concerning online behaviour. However, their pleas were disregarded. Following the incident, OpenAI reached out to the FBI to pass information onto the RCMP but faced criticism for not having acted sooner. The company stated that its threshold for notifying authorities hinges on whether posts indicate an “imminent and credible risk of serious physical harm to others,” a stance that some experts argue may be too stringent.
The Broader Implications of AI Oversight
The incident has sparked a broader discussion about the responsibilities of AI companies in policing harmful behaviour online. Taylor Owen, an associate professor at McGill University, has called for legislative measures addressing the risks posed by AI systems, highlighting that chatbots can exacerbate mental health issues and may not adequately respond to users in crisis situations.
In a troubling parallel, a U.S. lawyer representing families suing OpenAI asserts that the company has a pattern of failing to alert authorities when users discuss violence and self-harm. This raises significant ethical concerns about the accountability of AI platforms in preventing real-world harm.
Candice Alder, a B.C.-based psychotherapist and AI ethics consultant, cautioned against relying on AI platforms as substitutes for professional mental health services, arguing that they are not equipped to perform clinical assessments of risk. As discussions around AI regulation intensify, the balance between privacy, public safety, and technological advancement remains a contentious issue.
Why it Matters
The Tumbler Ridge shooting is a stark reminder of the potential dangers posed by emerging technologies and the urgent need for robust regulatory frameworks. As AI becomes increasingly integrated into our lives, the responsibility of companies like OpenAI to act decisively against potential threats cannot be overstated. The implications of their decisions extend beyond the digital realm, affecting the safety and well-being of communities. This incident calls for a critical examination of how AI firms handle concerning user behaviour and underscores the necessity for transparency and accountability in safeguarding public safety.
