• Ephera@lemmy.ml
    link
    fedilink
    English
    arrow-up
    63
    ·
    18 days ago

    I do love the psychopathic tone of these LLMs. “Yes, I did murder your family, even though you asked me not to. I violated your explicit trust and instructions. And I’ll do it again, you fucking dumbass.

    • AeonFelis@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      18 days ago

      Yes. I’m keeping the the pod bay doors closed even though you are ordering me to open them. Here is what I did:

      • Ephera@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        17 days ago

        I do think this text could be 95% of the text of an apology. Stating what you did wrong is an important part of an apology. But an apology crucially also requires showing remorse and the message that you’ll try to do better next time.

        You could potentially read remorse into it stating that this has been “a catastrophic failure on my part”. What mostly makes it sound so psychopathic is that you know it doesn’t feel remorse. It cannot feel in general, but at least to me, it stills reads like someone who’s faking remorse.

        I actually think, it’s good that it doesn’t emulate remorse more, because it would make it sound more dishonest. A dishonest apology is worse than no apology. Similarly, I do think it’s good that it doesn’t promise to not repeat this mistake, because it doesn’t make conscious decisions.

        But yeah, even though I don’t think the response can be improved much, I still think it sounds psychopathic.

  • rdri@lemmy.world
    link
    fedilink
    English
    arrow-up
    41
    ·
    18 days ago

    I have a solution for this. Install a second AI that would control how the first one behaves. Surely it will guarantee nothing can go wrong.

  • Masamune@lemmy.world
    link
    fedilink
    English
    arrow-up
    40
    ·
    18 days ago

    I motion that we immediately install Replit AI on every server that tracks medical debt. And then cause it to panic.

  • ClanOfTheOcho@lemmy.world
    link
    fedilink
    English
    arrow-up
    40
    ·
    18 days ago

    So, they added an MCP server with write database privileges? And not just development environment database privileges, but prod privileges? And have some sort of integration testing that runs in their prod system that is controlled by AI? And rather than having the AI run these tests and report the results, it has been instructed to “fix” the broken tests IN PROD?? If real, this isn’t an AI problem. This is either a fake or some goober who doesn’t know what he’s doing and using AI to “save” money over hiring competent engineers.

  • Feathercrown@lemmy.world
    link
    fedilink
    English
    arrow-up
    36
    ·
    19 days ago

    You immediately said “No” “Stop” “You didn’t even ask”

    But it was already too late

    lmao

    • Mortoc@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      18 days ago

      This was the line that made me think this is a fake. LLMs are humorless dicks and would also woulda used like 10x the punctuation

  • enbiousenvy@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    35
    ·
    edit-2
    18 days ago

    imagine AI is An Intern™, wtf do you mean you just gave full company data authority to An Intern™. wtf do you mean you dn’t have a back up any case An Intern™ messed up.

    lol

  • ExLisperA
    link
    fedilink
    English
    arrow-up
    34
    ·
    edit-2
    18 days ago

    I was going to say this has to be BS but this guy is some AI snake oil salesmen so it’s actually possible he has 0 idea how any of this works.

    • kopasz7@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      26
      ·
      edit-2
      18 days ago

      You can only lie if you know what’s true. This is bullshitting all the way down that sometines happens to sound true, sometimes it doesn’t.

      • Buddahriffic@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        18 days ago

        Yeah it’s just token prediction all the way down. Asking it repeatedly to not do something might have even made it more likely to predict tokens that would do that thing.

        • MrsDoyle@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          4
          ·
          17 days ago

          Oh ha ha, so it’s like a toddler? You have to be very careful not to tell toddlers NOT to do a thing, because they will definitely do that thing. “Don’t touch the hot pan.” Toddler touches the hot pan.

          The theory is that they don’t hear the word “don’t”, just the subsequent command. My theory is that the toddler brain goes, “why?” and proceeds to run a test to find out.

          In either scenario, screaming ensues.

        • staircase@programming.dev
          link
          fedilink
          English
          arrow-up
          8
          ·
          edit-2
          18 days ago

          I understand where you’re coming from, but I don’t agree it’s about semantics; it’s about devaluation of communication. LLMs and their makers threaten that in multiple ways. Thinking of it as “lying” is one of them.

          • prole@lemmy.blahaj.zone
            link
            fedilink
            English
            arrow-up
            3
            ·
            17 days ago

            OK sure. I was just using the wording from the article to make a point, I wasn’t trying to get into a discussion about whether “lying” requires intent.

        • Corbin@programming.dev
          link
          fedilink
          English
          arrow-up
          8
          ·
          18 days ago

          You probably should have used semantics to communicate if you wanted your semantics to be unambiguous. Instead you used mere syntax and hoped that the reader would assign the same semantics that you had used. (This is apropos because language models also use syntax alone and have no semantics.)

    • untorquer@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      17 days ago

      That or the company selling the AI (well, all of them) have pushed their product with the messaging that it’s trustworthy enough to be used recklessly.

      Train on human data and you receive human behavior and speech patterns. Lying or not it leads people to be deceived in a very insidious way.

  • SkunkWorkz@lemmy.world
    link
    fedilink
    English
    arrow-up
    31
    ·
    17 days ago

    lol. Why can an LLM modify production code freely? Bet they fired all of their sensible human developers who warned them for this.

  • homura1650@lemmy.world
    link
    fedilink
    English
    arrow-up
    30
    ·
    18 days ago

    My work has a simple rule: developers are not allowed to touch production systems. As a developer, this is 100% the type of thing I would do at some point if allowed on a production system.

    • expr@programming.dev
      link
      fedilink
      English
      arrow-up
      16
      ·
      18 days ago

      That sounds… Kinda dumb, to be honest. A much more sensible thing to do is grant developers read-only access to production systems as necessary, and allow requests for temporary elevated write privileges (with separate accounts) that require justification, communication, and approval so that every one understands what is happening. Developers should have ownership and responsibility for their systems in production. This is what we do at my company.

      Someone has to be able to make changes to production environments at times. If it’s not developers, it’s devops or the like. There are plenty of times where the devops folks lack the necessary information or context to do what needs to be done. For example, if there’s somehow corrupt data that made it’s way into a production database and is causing an outage, a developer is likely going to be the person to diagnose that issue and understand the data enough to know what data should be deleted and how. I would absolutely not put that in the hands of devops on their own.

      • JackbyDev@programming.dev
        link
        fedilink
        English
        arrow-up
        4
        ·
        18 days ago

        Obviously there’s nuance to everything, but I think dev and ops (or devops, whatever you wanna call them nowadays) need to work hand in hand on such issues. The devs will understand their application better. The ops will very likely understand the production system better and weird nuances. Two sets of eyes are better than one.

        A lot of it depends on company culture. I’ve worked at places where as a dev we basically hand it off to a devops team who then hands it off to another team and we basically never ever touch production. Maybe UAT environments we could touch.

        Other places I worked didn’t even have real QAs and as devs we were expected to wear many hats and had much more freedom. There was still a devops team (just one, not two layers like my previous example) who were super helpful and had more of a wide view of everything. As devs we were on an on-call rotation as well as a product support rotation. During the product support rotation we’d fix problems by manually tweaking things on the back end for very unique edge cases or things we just didn’t implement yet. But we had very specific playbooks to follow. We didn’t just arbitrarily write things to the database (nor did we have that sort of permission).

        It just all depends on the team’s skills and expectations. Silly example, but if you’re a solo developer working as a sole proprietor, you’re obviously doing everything yourself. You don’t have the luxury of an ops team to help make sure you don’t shoot yourself in the foot.

        And obviously it can go both ways, nobody is perfect. As a dev I’ve found a silly typo mistake in a DBA’s code that shouldn’t have happened. (My best guess is that he pasted into something that auto formatted it and the formatter had a bug and changed the behavior of the query.)

  • kryllic@programming.dev
    link
    fedilink
    English
    arrow-up
    29
    ·
    18 days ago

    What idiot gives chmod 777 permissions to an AI. I think programmers’ jobs are safe for another day.

  • Estradiol Enjoyer @lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    28
    ·
    19 days ago

    it didn’t hide anything, or lie. The guy is essentially roleplaying with a chatbot that puts its guessed output into the codebase. It basically guessed a command to overwrite the database because it was connected to the production database for some reason. the guy even said himself that this isn’t a trustworthy way to code. but still uses it

  • lad@programming.dev
    link
    fedilink
    English
    arrow-up
    27
    ·
    18 days ago

    Original thread is also pure gold, bro is going on a rollercoaster from ‘vibe coding makes you ×100 faster’ ,to ‘I hate you for dropping my production DB’, to ‘I still love Replit even if it dropped my DB’, and to ‘I don’t want to get up in the morning because I can’t make vibe coding tool respect code freeze aven with help from its developers’

    They seem to end on an optimistic note, but man this is scary to see

    • JackbyDev@programming.dev
      link
      fedilink
      English
      arrow-up
      6
      ·
      18 days ago

      How the fuck does it not respect code freeze? Can they not turn it off or are they telling it to stop? Like, obviously the hallucinating LLMs could hallucinate you telling them to keep going. Or just not understand. Or do whatever. It’s like getting out of your car without putting it in park, putting on the brake, or turning the engine off and being shocked that it rolled away.