• brap@lemmy.world
    link
    fedilink
    English
    arrow-up
    47
    ·
    2 months ago

    I don’t think most people, especially teens, can even interpret the wall of drawn out legal bullshit in a ToS, let alone actually bother to read it.

  • lefthandeddude@lemmy.dbzer0.comBanned
    link
    fedilink
    English
    arrow-up
    44
    ·
    edit-2
    2 months ago

    The elephant in the room that no one talks about is that locked psychiatry facilities treat people so horribly and are so expensive, and psychologists and psychiatrists have such arbitrary power to detain suicidal people, that suicidal people who understand the system absolutely will not open up to professional help about feeling suicidal, lest they be locked up without a cell phone, without being able to do their job, without having access to video games, being billed tens of thousands of dollars per month that can only be discharged by bankruptcy. There is a reason why people online have warned about the risks and expenses of calling suicide hotlines like 988 that regularly attempt to geolocate and imprison people in mental health facilities, with psychiatric medications being required in order for someone to leave.

    The problem isn’t ChatGPT. The problem is a financially exploitative psychiatric industry with horrible financial consequences for suicidal patients and horrible degrading facilities that take away basic human dignity at exorbitant cost. The problem is vague standards that officially encourage suicidal patients to snitch on themselves for treatment with the consequence that at the professional’s whim they can be subject to misery and financial exploitation. Many people who go to locked facilities come out with additional trauma and financial burdens. There are no studies about whether such facilities traumatize patients and worsen patient outcomes because no one has a financial interest in funding the studies.

    The real problem is, why do suicidal people see a need to confide in ChatGPT instead of mental health professionals or 988? And the answer is because 988 and mental health professionals inflict even more pain and suffering upon people already hurting in variable randomized manner, leading to patient avoidance. (I say randomized in the sense that it is hard for a patient to predict the outcome of when this pain will be inflicted, rather than something predictable like being involuntarily held every 10 visits.) Psychiatry and psychology do everything they possibly can to look good to society (while being paid), but it doesn’t help suicidal people at all who bare the suffering of their “treatments.” Most suicidal patients fear being locked up and removed from society.

    This is combined with the fact that although lobotomies are no longer common place, psychiatrists regularly push unethical treatments like ECT which almost always leads to permanent memory loss. Psychiatrist still lie to patients and families regarding ECT about how likely memory loss is, falsely stating memory loss is often temporary and not everyone gets it, just like they lied to patients and families about the effects of lobotomies. People in locked facilities can be pressured into ECT as part of being able to leave a facility, resulting in permanent brain damage. They were charlatans then and now, a so called “science” designed to extract money while looking good with no rigorous studies on how they damage patients.

    In fact, if patients could be open about being suicidal with 988 and mental health professionals without fear of being locked up, this person would probably be alive today. ChatGPT didn’t do anything other than be a friend to this person. The failure is due to the mental health industry.

    • brygphilomena@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      13
      ·
      2 months ago

      While I agree with much of what you said, there are other issues with psychology and psychiatry that they often can’t treat some environmental causes or triggers. When I was suicidal, it was also the feeling of being trapped in a job where I wasn’t appreciated and couldn’t advance.

      If I were placed in an inpatient facility, it would only have exacerbated the issues where I would have so much to deal with the try and be on medical leave before I got fired for not showing up.

      That said, for SOME mental illnesses ECT it can be a valid treatment. We don’t know how the brain works, but we’ve seen correlation where ECT kind of resets the way the brain perceives the world temporarily. All medical decisions need to be weighed against the side effects and determined if the benefits outweigh the risks.

      The other issue with inpatient facilities is that they can be incredibly hard to convince the staff that you are doing better. All actions are viewed through the lens that you are ill and showing the staff you are better is just trying to trick the staff to get out.

      • lefthandeddude@lemmy.dbzer0.comBanned
        link
        fedilink
        English
        arrow-up
        6
        ·
        edit-2
        2 months ago

        You’re wrong about ECT. It nearly always results in permanent memory loss and even if occasionally some patients seem “better” because they remember less of their lives, it does not negate the evil of the treatment. Worse than that, psychiatrist universally deceive patients about the risk of memory loss, saying memory loss is temporary, when most patients who have had ECT report that the memory loss is permanent. There were people who extolled the virtues of lobotomies decades ago and the procedure even won a Nobel Prize. The reason it won a Nobel Prize is because patient experiences mean nothing compared to the avarice of a psuedoscientific discipline that is always looking for the next scam, with the worst most cruel and most expensive scams always inflicted on the most vulnerable. It is hard and traumatic for patients who have been exploited by their supposed “healers” to come forward with the truth. It is incredibly psychologically agonizing to admit to being duped. Patients are not believed then or now. You are completely wrong.

    • andros_rex@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      2 months ago

      God this. Before I was stupid enough to reach out to a crisis line, I had a job with health insurance. Now I have worsened PTSD and no health insurance (the psych hospital couldn’t be assed to provide me with discharge papers.) I get to have nightmares for the rest of my life about a three men shoving me around and being unable to sleep for fear of being assaulted again.

  • Fedizen@lemmy.world
    link
    fedilink
    English
    arrow-up
    30
    ·
    2 months ago

    “Hey computer should I do <insert intrusive thought here>?”

    Computer "yes, that sounds like a great idea, here’s how you might do that. "

    • ExLisperA
      link
      fedilink
      English
      arrow-up
      6
      ·
      edit-2
      2 months ago

      I think with all the guardrails current models have you have to talk to it for weeks if not months before it degrades to a point that it will let you talk about anything remotely harmful. Then again, that’s exactly what a lot of people do.

      • AnarchistArtificer@slrpnk.net
        link
        fedilink
        English
        arrow-up
        4
        ·
        2 months ago

        Exactly, and this is why their excuses are bullshit. They know that guardrails become less effective the more you use a chatbot, and they know that’s how people are using chatbots. If they actually gave a fuck about guardrails, they’d make it so that you couldn’t do conversations that take place over weeks or months. This would hurt their bottom line though.

      • Fedizen@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        2 months ago

        If its sold as a permanent solution to a problem but the guardrails are temporary… idk man, seems like anyone who incorporates this into solving any problem with AI will eventually degrade the guardrails.

        • ExLisperA
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 months ago

          Definitely not everyone. We’re talking about users that have a single chat open and have endless conversations about personal topics there. I think majority of users will ask single question or have a short conversation and then create new chat. And don’t talk about personal problems. We’re also talking about people with specific metal issues. AI is a terrible for many reasons but I think “it helps people kill themselves” is exaggerated. Before AI people were getting sucked into online communities that were encouraging suicide but the media barely noticed the issue. Marijuana is very dangerous for people with predisposition to some mental issues like schizophrenia but we just agree that people should have that in mind if they are going to use it. It’s the same with AI. Some people shouldn’t be using it but it’s not a reason for a total ban. The reason for a total ban is that it’s bad for environment, jobs and education and offers little benefits.

          • Fedizen@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            2 months ago

            I’m seeing people use LLM’s for:

            • Dating
            • Email/work tasks
            • Customer support
            • Mental health hotlines

            The dating, customer support, and mental health hotlines, notably, are not people who are always informed they’re talking to an LLM bot.

            I don’t think the “exposure to marijuana” analogy works here because people are getting exposed to to it by businesses without consent.

            https://sfstandard.com/2025/08/26/ai-crisis-hotlines-suicide-prevention/

            • ExLisperA
              link
              fedilink
              English
              arrow-up
              1
              ·
              2 months ago

              The issue we’re talking about is not getting a reply from bot in a chat or phone call. We’re talking about people with metal issues using AI in a way that exasperates their problems. Specifically we’re talking about people believing AI is their personal companion and creating personal connection with it to a point that wrong answers generated by AI affect their well being. Vast majority of people don’t use AI like that.

  • Uriel238 [all pronouns]@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    17
    ·
    2 months ago

    Plenty of judges won’t enforce a TOS, especially if some of the clauses are egregious (e.g. we own and have unlimited use of your photos )

    The legal presumption is that the administrative burden of reading a contract longer than King Lear is too much to demand from the common end-user.

  • Smoogs@lemmy.world
    link
    fedilink
    English
    arrow-up
    9
    ·
    2 months ago

    Didnt we just shake the stigma of “committing” suicide to be death by suicide to stop blaming dead people already?

  • NutWrench@lemmy.ml
    link
    fedilink
    English
    arrow-up
    9
    ·
    2 months ago

    AIs have no sense of ethics. You should never rely on them for real-world advice because they’re programmed to tell you what you want to hear, no matter what the consequences.

    • Zetta@mander.xyz
      link
      fedilink
      English
      arrow-up
      5
      ·
      2 months ago

      The problem is that many people don’t understand this no matter how often we bring it up. I personally find LLMs to be very valuable tools when used in the right context. But yeah, the majority of people who utilize these models don’t understand what they are or why they shouldn’t really trust them or take critical advice from them.

      I didn’t read this article, but there’s also the fact that some people want biased or incorrect information from the models. They just want them to agree with them. Like, for instance, this teen who killed themself may not have been seeking truthful or helpful information in the first place, but instead just wanted to agree with them and help them plan the best way to die.

      Of course, OpenAI probably should have detected this and stopped interacting with this individual.

      • Timecircleline@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 months ago

        The court documents with extracted text are linked in this thread. It talked him out of seeking help and encouraged him not to leave signs of his suicidality out for his family to see when he said he hoped they would stop him.

  • wavebeam@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    2 months ago

    Gun company says you “broke the TOS” when you pointed the gun at a person. It’s not their fault you used it to do a murder.

      • espentan@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        2 months ago

        Well, such a knife’s primary purpose is to help with preparing food while the gun’s primary purpose is to injure/kill. So one would be used for something which it was not designed while the other would’ve been used exactly as designed.

        • Manifish_Destiny@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          edit-2
          2 months ago

          Guns primary purpose is to shoot bullets. I can kill just as well with a chemical bomb as a gun, and I could make both of those from things I can buy from the store from components that weren’t ‘designed’ for it.

          In this case ‘terms of service’ is just ‘the law’.

          People killing each other is just a side effect of humans interacting with dangerous things. Granted humans just kinda suck in general.

  • Credibly_Human@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    2 months ago

    The sentiment that the AI bares any noteworthy responsibility for this is purely anti AI rage, that should be aimed at legitimate problems.

    Imagine suing a notebook company for their paper being the paper of choice for selfharming teens?

    Imagine suing home depot for selling rope and a stool to someone who has had enough?

    Imagine suing nickleback for making music of the quality that encouraged this?

    Im saying, we’re all aware this is some bits on a server right? Like this is clearly not a person, doesn’t have the impact of a person, and unless they’ve specifically tuned it to manipulate the impressionable into killing people, these sentiments just don’t make sense.

    • Dr. Moose@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      2 months ago

      I agree the AI hate is becoming a satire of itself. What could be an interesting, meaningful discussion is impossible to have because anti AI peoppe just yell with their ears covered.

    • CovfefeKills@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      2 months ago

      Fuck personal responsibility I want to be able to do anything and everything AND sue when I am not safe guarded from myself but also privacy!

  • Dr. Moose@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    ·
    2 months ago

    Fun fact: you can literally go to prison in the US for breaking ToS due to various laws like CFFA (Computer Fraud and Abuse Act). So if the teen broke the ToS to any way that harms OpenAI (like killing himself) OpenAI actually has a legal path to criminally prosecute him lmao

    The entire law stack is just broken.

  • sudoer777@lemmy.ml
    link
    fedilink
    English
    arrow-up
    5
    ·
    2 months ago

    As shitty as AI is for counseling, the alternative resources are so few, unreliable, and taboo that I can’t blame people for wanting to use it. People will judge and remember you. AI affirms and forgets. People have mandatory reporting for “self harm” (which could include things like drug usage) that incarcerates you and fucks up your life even more. AI does not. People are varied with differing advice, while AI uses the same models in different contexts. Counselors are expensive, AI is $20/mo. And lastly, people have a tendency to react fearfully to taboo topics in ways that AI doesn’t. I see a lot of outrage towards AI, but it seems like the sort of outrage that led to half-assed liability-driven “call this number and all of your problems will be solved” incarceration and abandonment hotlines is what got us here to begin with.

  • massi1008@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    edit-2
    2 months ago

    > Build a yes-man

    > It is good at saying “yes”

    > Someone asks it a question

    > It says yes

    > Everyone complains

    ChatGPT is a (partially) stupid technology with not enough security. But it’s fundamentally just autocomplete. That’s the technology. It did what it was supposed to do.

    I hate to defend OpenAI on this but if you’re so mentally sick (dunno if that’s the right word here?) that you’d let yourself be driven to suicide by some online chats [1] then the people who gave you internet access are to blame too.

    [1] If this was a human encouraging him to suicide this wouldn’t be newsworthy…

    • SkyezOpen@lemmy.world
      link
      fedilink
      English
      arrow-up
      24
      ·
      2 months ago

      You don’t think pushing glorified predictive text keyboard as a conversation partner is the least bit negligent?

      • massi1008@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        2 months ago

        It is. But the chatGPT interface reminds you of that when you first create an account. (At least it did when I created mine).

        At some point we have to give the responsibility to the user. Just like with Kali OS or other pentesting tools. You wouldn’t (shouldn’t) blame them for the latest ransomeware attack too.

        • raspberriesareyummy@lemmy.world
          link
          fedilink
          English
          arrow-up
          10
          ·
          2 months ago

          At some point we have to give the responsibility to the user.

          That is such a fucked up take on this. Instead of seeing the responsibility at the piece of shit billionaires force-feeding this glorified text prediction on everyone, and politicians allowing minors access to smartphones, you turn off your brain and hop straight over to victim-blaming. I hope you will slap yourself for this comment after some time to reflect on it.

    • Live Your Lives@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      2 months ago

      I get where you’re coming from because people and those directly over them will always bear a large portion of the blame and you can only take safety so far.

      However, that blame can only go so far as well, because the designers of a thing who overlook or ignore safety loopholes should bear responsibility for their failures. We know some people will always be more susceptible to implicit suggestions than others are and that not everyone has someone who’s responsible over them in the first place, so we need to design AIs accordingly.

      Think of it like blaming an employee’s shift supervisor when an employee dies when the work environment is itself unsafe. Or think of it like only blaming a gun user and not the gun laws. Yes, individual responsibility is a thing, but the system as a whole has a responsibility all it’s own.

    • missingno@fedia.io
      link
      fedilink
      arrow-up
      1
      ·
      2 months ago

      If this is what ChatGPT is “supposed to do” then that’s the problem. A yes-man that will say yes to anything, even suicide, is dangerous.

  • TheObviousSolution@lemmy.ca
    link
    fedilink
    English
    arrow-up
    4
    ·
    2 months ago

    “Person violated the TOS when they used the magic lamp to make the genie do bad things.”

    You still made the magic lamp and the genie capable of doing those bad things. That’s the thing with intelligence, even the artificial variety. A chainsaw isn’t going to get up and begin a chainsaw massacre just because you throw the right prompt injection at it. It may just reply with words, but words have power.