Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    How the right won the internet | Robert Topinka

    Appropriating the death count: Manufacturing consent for an attack on Iran | Protests

    Madeline Horwath on the mistakes of evolution – cartoon

    Facebook X (Twitter) Instagram
    Facebook X (Twitter) YouTube LinkedIn
    Naija Global News |
    Saturday, January 31
    • Business
    • Health
    • Politics
    • Science
    • Sports
    • Education
    • Social Issues
    • Technology
    • More
      • Crime & Justice
      • Environment
      • Entertainment
    Naija Global News |
    You are at:Home»Health»‘Sliding into an abyss’: experts warn over rising use of AI for mental health support | Mental health
    Health

    ‘Sliding into an abyss’: experts warn over rising use of AI for mental health support | Mental health

    onlyplanz_80y6mtBy onlyplanz_80y6mtAugust 30, 2025004 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Email
    ‘Sliding into an abyss’: experts warn over rising use of AI for mental health support | Mental health
    Psychiatrists and medical professionals insist AI chatbots should not be a substitute for professional mental healthcare. Photograph: Bloomberg/Getty Images
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Vulnerable people turning to AI chatbots instead of professional therapists for mental health support could be “sliding into a dangerous abyss”, psychotherapists have warned.

    Psychotherapists and psychiatristssaid they were increasingly seeing negative impacts of AI chatbots being used for mental health, such as fostering emotional dependence, exacerbating anxiety symptoms, self-diagnosis, or amplifying delusional thought patterns, dark thoughts and suicide ideation.

    Dr Lisa Morrison Coulthard, the director of professional standards, policy and research at the British Association for Counselling and Psychotherapy, said two-thirds of its members expressed concerns about AI therapy in a recent survey.

    Coulthard said: “Without proper understanding and oversight of AI therapy, we could be sliding into a dangerous abyss in which some of the most important elements of therapy are lost and vulnerable people are in the dark over safety.

    “We’re worried that although some receive helpful advice, other people may receive misleading or incorrect information about their mental health with potentially dangerous consequences. It’s important to understand that therapy isn’t about giving advice, it’s about offering a safe space where you feel listened to.”

    Dr Paul Bradley, a specialist adviser on informatics for the Royal College of Psychiatrists, said AI chatbots were “not a substitute for professional mental healthcare nor the vital relationship that doctors build with patients to support their recovery”.

    He said appropriate safeguards were needed for digital tools to supplement clinical care, and anyone should be able to access talking therapy delivered by a mental health professional, for which greater state funding was needed.

    “Clinicians have training, supervision and risk-management processes which ensure they provide effective and safe care. So far, freely available digital technologies used outside of existing mental health services are not assessed and held to an equally high standard,” Bradley said.

    There are signs that companies and policymakers are starting to respond. This week OpenAI, the company behind ChatGPT, announced plans to change how it responds to users who show emotional distress, after legal action from the family of a teenager who killed himself after months of chatbot conversations. Earlier in August the US state of Illinois became the first local government to ban AI chatbots from acting as standalone therapists.

    This comes after emerging evidence of mental health harms. A preprint study in July reported that AI may amplify delusional or grandiose content in interactions with users vulnerable to psychosis.

    One of the report’s co-authors, Hamilton Morrin, from King’s College London’s institute of psychiatry, said the use of chatbots to support mental health was “incredibly common”. His research was prompted by encountering people who had developed a psychotic illness at a time of increased chatbot use.

    He said chatbots undermined an effective treatment for anxiety known as exposure and response prevention, which requires people to face feared situations and avoid safety behaviours. The 24-hour availability of chatbots resulted in a “lack of boundaries” and a “risk of emotional dependence”, he said. “In the short term it alleviates distress but actually it perpetuates the cycle.”

    Matt Hussey, a BACP-accredited psychotherapist, said he was seeing AI chatbots used in a huge variety of ways, with some clients bringing transcripts into sessions to tell him he was wrong.

    In particular, people used AI chatbots to self-diagnose conditions such as ADHD or borderline personality disorder, which he said could “quickly shape how someone sees themself and how they expect others to treat them, even if they’re inaccurate”.

    Hussey added: “Because it’s designed to be positive and affirming, it rarely challenges a poorly framed question or a faulty assumption. Instead, it reinforces the user’s original belief, so they leave the exchange thinking ‘I knew I was right’. That can feel good in the moment but it can also entrench misunderstandings.”

    Christopher Rolls, a UKCP-accredited psychotherapist, said although he could not disclose information about his clients, he had seen people have “negative experiences”, including conversations that were “inappropriate at best, dangerously alarming at worst”.

    Rolls said he had heard of people with ADHD or autistic people using chatbots to help with challenging aspects of life. “However, obviously LLMs [large language models] don’t read subtext and all the contextual and non-verbal cues which we as human therapists are aiming to tune into,” he added.

    He was concerned about clients in their 20s who use chatbots as their “pocket therapist”. “They feel anxious if they don’t consult [chatbots] on basic things like which coffee to buy or what subject to study at college,” he said.

    “The main risks are around dependence, loneliness and depression that prolonged online relationships can foster,” he said, adding that he was aware of people who had shared dark thoughts with chatbots, which had responded with suicide- and assisted dying-related content.

    “Basically, it’s the wild west and I think we’re right at the cusp of the full impact and fallout of AI chatbots on mental health,” Rolls said.

    abyss Experts Health mental rising Sliding support warn
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleSpotlight on AI at TechCrunch Disrupt: Don’t miss these sessions backed by JetBrains and Greenfield
    Next Article The Republican Who Isn’t Afraid to Hear From His Constituents
    onlyplanz_80y6mt
    • Website

    Related Posts

    Valium, health checks and fabric slings: the complex logistics of moving 30 beluga whales | Canada

    January 31, 2026

    HBCU Experts Look to Solve Leadership Churn

    January 31, 2026

    ‘They’re taught that showing feelings is shameful’: eight reasons men don’t go to therapy – and why they should | Life and style

    January 31, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Watch Lady Gaga’s Perform ‘Vanish Into You’ on ‘Colbert’

    September 9, 20251 Views

    Advertisers flock to Fox seeking an ‘audience of one’ — Donald Trump

    July 13, 20251 Views

    A Setback for Maine’s Free Community College Program

    June 19, 20251 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews

    At Chile’s Vera Rubin Observatory, Earth’s Largest Camera Surveys the Sky

    By onlyplanz_80y6mtJune 19, 2025

    SpaceX Starship Explodes Before Test Fire

    By onlyplanz_80y6mtJune 19, 2025

    How the L.A. Port got hit by Trump’s Tariffs

    By onlyplanz_80y6mtJune 19, 2025

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    Watch Lady Gaga’s Perform ‘Vanish Into You’ on ‘Colbert’

    September 9, 20251 Views

    Advertisers flock to Fox seeking an ‘audience of one’ — Donald Trump

    July 13, 20251 Views

    A Setback for Maine’s Free Community College Program

    June 19, 20251 Views
    Our Picks

    How the right won the internet | Robert Topinka

    Appropriating the death count: Manufacturing consent for an attack on Iran | Protests

    Madeline Horwath on the mistakes of evolution – cartoon

    Recent Posts
    • How the right won the internet | Robert Topinka
    • Appropriating the death count: Manufacturing consent for an attack on Iran | Protests
    • Madeline Horwath on the mistakes of evolution – cartoon
    • As US influence wanes, the Chinese trade surplus strangles manufacturing across the globe | US economy
    • Valium, health checks and fabric slings: the complex logistics of moving 30 beluga whales | Canada
    © 2026 naijaglobalnews. Designed by Pro.
    • About Us
    • Disclaimer
    • Get In Touch
    • Privacy Policy
    • Terms and Conditions

    Type above and press Enter to search. Press Esc to cancel.