Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Danger of prisoners who have nothing to lose | Prisons and probation

    Florida professors quietly defy restrictions on race and gender: ‘This is how authoritarianism works’ | Florida

    There might be less water on the moon than we’d hoped

    Facebook X (Twitter) Instagram
    Facebook X (Twitter) YouTube LinkedIn
    Naija Global News |
    Wednesday, March 18
    • Business
    • Health
    • Politics
    • Science
    • Sports
    • Education
    • Social Issues
    • Technology
    • More
      • Crime & Justice
      • Environment
      • Entertainment
    Naija Global News |
    You are at:Home»Health»‘Sliding into an abyss’: experts warn over rising use of AI for mental health support | Mental health
    Health

    ‘Sliding into an abyss’: experts warn over rising use of AI for mental health support | Mental health

    onlyplanz_80y6mtBy onlyplanz_80y6mtAugust 30, 2025004 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Email
    ‘Sliding into an abyss’: experts warn over rising use of AI for mental health support | Mental health
    Psychiatrists and medical professionals insist AI chatbots should not be a substitute for professional mental healthcare. Photograph: Bloomberg/Getty Images
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Vulnerable people turning to AI chatbots instead of professional therapists for mental health support could be “sliding into a dangerous abyss”, psychotherapists have warned.

    Psychotherapists and psychiatristssaid they were increasingly seeing negative impacts of AI chatbots being used for mental health, such as fostering emotional dependence, exacerbating anxiety symptoms, self-diagnosis, or amplifying delusional thought patterns, dark thoughts and suicide ideation.

    Dr Lisa Morrison Coulthard, the director of professional standards, policy and research at the British Association for Counselling and Psychotherapy, said two-thirds of its members expressed concerns about AI therapy in a recent survey.

    Coulthard said: “Without proper understanding and oversight of AI therapy, we could be sliding into a dangerous abyss in which some of the most important elements of therapy are lost and vulnerable people are in the dark over safety.

    “We’re worried that although some receive helpful advice, other people may receive misleading or incorrect information about their mental health with potentially dangerous consequences. It’s important to understand that therapy isn’t about giving advice, it’s about offering a safe space where you feel listened to.”

    Dr Paul Bradley, a specialist adviser on informatics for the Royal College of Psychiatrists, said AI chatbots were “not a substitute for professional mental healthcare nor the vital relationship that doctors build with patients to support their recovery”.

    He said appropriate safeguards were needed for digital tools to supplement clinical care, and anyone should be able to access talking therapy delivered by a mental health professional, for which greater state funding was needed.

    “Clinicians have training, supervision and risk-management processes which ensure they provide effective and safe care. So far, freely available digital technologies used outside of existing mental health services are not assessed and held to an equally high standard,” Bradley said.

    There are signs that companies and policymakers are starting to respond. This week OpenAI, the company behind ChatGPT, announced plans to change how it responds to users who show emotional distress, after legal action from the family of a teenager who killed himself after months of chatbot conversations. Earlier in August the US state of Illinois became the first local government to ban AI chatbots from acting as standalone therapists.

    This comes after emerging evidence of mental health harms. A preprint study in July reported that AI may amplify delusional or grandiose content in interactions with users vulnerable to psychosis.

    One of the report’s co-authors, Hamilton Morrin, from King’s College London’s institute of psychiatry, said the use of chatbots to support mental health was “incredibly common”. His research was prompted by encountering people who had developed a psychotic illness at a time of increased chatbot use.

    He said chatbots undermined an effective treatment for anxiety known as exposure and response prevention, which requires people to face feared situations and avoid safety behaviours. The 24-hour availability of chatbots resulted in a “lack of boundaries” and a “risk of emotional dependence”, he said. “In the short term it alleviates distress but actually it perpetuates the cycle.”

    Matt Hussey, a BACP-accredited psychotherapist, said he was seeing AI chatbots used in a huge variety of ways, with some clients bringing transcripts into sessions to tell him he was wrong.

    In particular, people used AI chatbots to self-diagnose conditions such as ADHD or borderline personality disorder, which he said could “quickly shape how someone sees themself and how they expect others to treat them, even if they’re inaccurate”.

    Hussey added: “Because it’s designed to be positive and affirming, it rarely challenges a poorly framed question or a faulty assumption. Instead, it reinforces the user’s original belief, so they leave the exchange thinking ‘I knew I was right’. That can feel good in the moment but it can also entrench misunderstandings.”

    Christopher Rolls, a UKCP-accredited psychotherapist, said although he could not disclose information about his clients, he had seen people have “negative experiences”, including conversations that were “inappropriate at best, dangerously alarming at worst”.

    Rolls said he had heard of people with ADHD or autistic people using chatbots to help with challenging aspects of life. “However, obviously LLMs [large language models] don’t read subtext and all the contextual and non-verbal cues which we as human therapists are aiming to tune into,” he added.

    He was concerned about clients in their 20s who use chatbots as their “pocket therapist”. “They feel anxious if they don’t consult [chatbots] on basic things like which coffee to buy or what subject to study at college,” he said.

    “The main risks are around dependence, loneliness and depression that prolonged online relationships can foster,” he said, adding that he was aware of people who had shared dark thoughts with chatbots, which had responded with suicide- and assisted dying-related content.

    “Basically, it’s the wild west and I think we’re right at the cusp of the full impact and fallout of AI chatbots on mental health,” Rolls said.

    abyss Experts Health mental rising Sliding support warn
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleSpotlight on AI at TechCrunch Disrupt: Don’t miss these sessions backed by JetBrains and Greenfield
    Next Article The Republican Who Isn’t Afraid to Hear From His Constituents
    onlyplanz_80y6mt
    • Website

    Related Posts

    Having my ears syringed left me with tinnitus | Health

    March 18, 2026

    Opioid addiction almost destroyed me – then I became a top marathon runner | Opioids

    March 18, 2026

    Number of meningitis cases investigated in Kent rises to 20 | Meningitis

    March 18, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Watch Lady Gaga’s Perform ‘Vanish Into You’ on ‘Colbert’

    September 9, 20251 Views

    Advertisers flock to Fox seeking an ‘audience of one’ — Donald Trump

    July 13, 20251 Views

    A Setback for Maine’s Free Community College Program

    June 19, 20251 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews

    At Chile’s Vera Rubin Observatory, Earth’s Largest Camera Surveys the Sky

    By onlyplanz_80y6mtJune 19, 2025

    SpaceX Starship Explodes Before Test Fire

    By onlyplanz_80y6mtJune 19, 2025

    How the L.A. Port got hit by Trump’s Tariffs

    By onlyplanz_80y6mtJune 19, 2025

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    Watch Lady Gaga’s Perform ‘Vanish Into You’ on ‘Colbert’

    September 9, 20251 Views

    Advertisers flock to Fox seeking an ‘audience of one’ — Donald Trump

    July 13, 20251 Views

    A Setback for Maine’s Free Community College Program

    June 19, 20251 Views
    Our Picks

    Danger of prisoners who have nothing to lose | Prisons and probation

    Florida professors quietly defy restrictions on race and gender: ‘This is how authoritarianism works’ | Florida

    There might be less water on the moon than we’d hoped

    Recent Posts
    • Danger of prisoners who have nothing to lose | Prisons and probation
    • Florida professors quietly defy restrictions on race and gender: ‘This is how authoritarianism works’ | Florida
    • There might be less water on the moon than we’d hoped
    • Having my ears syringed left me with tinnitus | Health
    • England must destine 7% of land to nature and renewables to hit green targets, data shows | Conservation
    © 2026 naijaglobalnews. Designed by Pro.
    • About Us
    • Disclaimer
    • Get In Touch
    • Privacy Policy
    • Terms and Conditions

    Type above and press Enter to search. Press Esc to cancel.