It’s cheap, quick and available 24/7, but is a chatbot therapist really the right tool to tackle complex emotional needs?

  • Candelestine@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    6
    ·
    9 months ago

    Eventually, yes, I think it will be. Not yet though, the tech just isn’t strong enough atm. But an AI is resistant to the emotional toll, burnout and low pay that a real life therapist has to struggle with. The AI therapist doesn’t need a therapist.

    Personally though, I think this is going to be one of the first widespread, genuinely revolutionary things LLMs are capable of. Couple more years maybe? It won’t be able to handle complex problems, it’ll have to flag and refer those cases to a doctor. But basic health maintenance is simpler.

    • snooggums
      link
      fedilink
      English
      arrow-up
      28
      arrow-down
      3
      ·
      9 months ago

      That would assume the people designing AI want what is best for the person and not what will make them the most money at the expense of the consumer.

      The companies involved in AI are NOT benevolent.

      • Even_Adder@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        3
        ·
        9 months ago

        You could just run your own. There are plenty of open source models that don’t answer to any company.

        • BakerBagel
          link
          fedilink
          English
          arrow-up
          10
          arrow-down
          2
          ·
          9 months ago

          Why dont i just give myself therapy? I know way more about what is going on in my head than anyone else does.

    • Usernameblankface@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      edit-2
      9 months ago

      Yes, one thing it absolutely has to be good at is referring patients to human therapists, for anyone who need something beyond the standard strategies the AI is trained on. It has to be smart enough to know when to give up.

      Edit, it would also be great if the AI would match up these difficult cases to therapists who are known to do well with whatever the patient is dealing with, as well as matching according to the patient’s personality, communication style, etc wherever possible

      Edit 2 for clarity above