In a significant response to alarming revelations about misleading health advice from Google’s AI Overviews, the mental health charity Mind is embarking on a pioneering year-long inquiry. This initiative aims to examine the intersection of artificial intelligence and mental health, particularly after a recent investigation highlighted serious risks associated with the information generated by Google’s AI systems. As technology increasingly permeates our lives, the need for robust safeguards has never been more pressing.
Unveiling the Dangers: A Call for Action
The inquiry follows a detailed investigation by The Guardian, which uncovered that millions of users are being exposed to potentially harmful medical information through Google’s AI Overviews. These summaries, presented above traditional search results on Google—a platform boasting over 2 billion users monthly—have been labelled as “very dangerous” by mental health experts.
Dr. Sarah Hughes, CEO of Mind, has expressed grave concerns about the “dangerously incorrect” advice that some users may receive, which could deter them from seeking necessary treatment and may even pose life-threatening risks. Hughes stated, “We believe AI has enormous potential to improve the lives of people with mental health problems, widen access to support, and strengthen public services. But that potential will only be realised if it is developed and deployed responsibly, with safeguards proportionate to the risks.”
A Collaborative Approach to Mental Health
Mind’s inquiry, the first of its kind globally, will convene a diverse group of stakeholders including leading healthcare professionals, individuals with lived experiences of mental health issues, policymakers, and tech industry representatives. This collaboration aims to create a safer digital environment for mental health support, ensuring that innovation does not compromise the well-being of users.
The investigation revealed that AI-generated content is often presented with a level of confidence that can mislead users about its reliability. Unfortunately, this erosion of trust can have severe consequences, particularly for vulnerable individuals seeking help. Mind’s initiative seeks to address these gaps and establish clear regulations and standards to prevent misinformation.
The Shift from Credibility to Convenience
Rosie Weatherley, Mind’s information content manager, emphasised that while searching for mental health information online was not without its flaws prior to the introduction of AI Overviews, users often ended up on credible health websites that provided nuanced and trustworthy guidance. The introduction of AI Overviews, however, has replaced this depth with overly simplistic summaries, leaving users with an illusion of certainty without a reliable source to back it up.
Weatherley pointed out, “AI Overviews replaced that richness with a clinical-sounding summary that gives an illusion of definitiveness. They give the user more of one form of clarity while giving them less of another form of clarity.” This shift raises critical questions about the responsibility of tech companies in providing accurate health information to the public.
Google’s Response and Ongoing Concerns
In the wake of these findings, Google has taken some steps to mitigate the risks, including removing AI Overviews for certain medical queries. Nevertheless, many experts argue that the problem persists, with incorrect information still circulating. A Google spokesperson defended the company’s AI Overviews, asserting that they invest heavily in ensuring the accuracy of the information provided, especially regarding health topics. However, without the ability to review specific instances, the company has been unable to fully address concerns raised by the investigation.

Why it Matters
The implications of this inquiry are profound. As digital health resources become increasingly prevalent, the need for accurate, reliable information is paramount. This initiative by Mind not only highlights the dangers of misleading health advice but also underscores the necessity for stringent regulations and ethical standards in the deployment of AI technologies. In a world where mental health is delicately intertwined with technology, ensuring that users receive safe and accurate information is not just important—it’s essential for safeguarding lives. The outcome of this inquiry could reshape the landscape of digital mental health support for years to come.