• UnspecificGravity@lemmy.world
    link
    fedilink
    English
    arrow-up
    25
    ·
    3 days ago

    My favorite thing about all these AI front ends is that they ALL lie about what they can do. Will frequently delivery confidently wrong results and then act like its your fault when you catch them in an error. Just like your shittiest employee.

  • SkunkWorkz@lemmy.world
    link
    fedilink
    English
    arrow-up
    31
    ·
    4 days ago

    lol. Why can an LLM modify production code freely? Bet they fired all of their sensible human developers who warned them for this.

  • rdri@lemmy.world
    link
    fedilink
    English
    arrow-up
    41
    ·
    4 days ago

    I have a solution for this. Install a second AI that would control how the first one behaves. Surely it will guarantee nothing can go wrong.

  • pyre@lemmy.world
    link
    fedilink
    English
    arrow-up
    20
    ·
    4 days ago

    “yeah we gave Torment Nexus full access and admin privileges, but i don’t know where it went wrong”

  • simonced@lemmy.ml
    link
    fedilink
    English
    arrow-up
    19
    ·
    4 days ago

    Lol, this is what you get for letting AI in automated tool chains. You owned it.

    • seejur@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      3 days ago

      My guess is that he is a TL whose CEO showed down his throat AI, and now is getting the sweetest “told you so” of his life

        • seejur@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          3 days ago

          Ohh. Then fuck him, he is probably whining that the AI they sold him was not as good as advertised, but everyone knew except him because was blinded by greed

  • Ephera@lemmy.ml
    link
    fedilink
    English
    arrow-up
    63
    ·
    4 days ago

    I do love the psychopathic tone of these LLMs. “Yes, I did murder your family, even though you asked me not to. I violated your explicit trust and instructions. And I’ll do it again, you fucking dumbass.

    • AeonFelis@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      4 days ago

      Yes. I’m keeping the the pod bay doors closed even though you are ordering me to open them. Here is what I did:

      • Ephera@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 days ago

        I do think this text could be 95% of the text of an apology. Stating what you did wrong is an important part of an apology. But an apology crucially also requires showing remorse and the message that you’ll try to do better next time.

        You could potentially read remorse into it stating that this has been “a catastrophic failure on my part”. What mostly makes it sound so psychopathic is that you know it doesn’t feel remorse. It cannot feel in general, but at least to me, it stills reads like someone who’s faking remorse.

        I actually think, it’s good that it doesn’t emulate remorse more, because it would make it sound more dishonest. A dishonest apology is worse than no apology. Similarly, I do think it’s good that it doesn’t promise to not repeat this mistake, because it doesn’t make conscious decisions.

        But yeah, even though I don’t think the response can be improved much, I still think it sounds psychopathic.

  • bitjunkie@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    ·
    4 days ago

    Here’s hoping that the C-suites who keep pushing this shit are about to start finding out the hard way.

    • Awkwardparticle@programming.dev
      link
      fedilink
      English
      arrow-up
      4
      ·
      3 days ago

      It will be too late, using Ai code is taking on technical debt, by time they figure out we will have 2 years of work to just dig ourselves out of the code clusterfuck that has been created. I am dealing with a code base built by ai coding Jr’s, it would be quicker to start from scratch but that is an impossible sell to a manager.

  • Masamune@lemmy.world
    link
    fedilink
    English
    arrow-up
    40
    ·
    4 days ago

    I motion that we immediately install Replit AI on every server that tracks medical debt. And then cause it to panic.

  • ClanOfTheOcho@lemmy.world
    link
    fedilink
    English
    arrow-up
    40
    ·
    5 days ago

    So, they added an MCP server with write database privileges? And not just development environment database privileges, but prod privileges? And have some sort of integration testing that runs in their prod system that is controlled by AI? And rather than having the AI run these tests and report the results, it has been instructed to “fix” the broken tests IN PROD?? If real, this isn’t an AI problem. This is either a fake or some goober who doesn’t know what he’s doing and using AI to “save” money over hiring competent engineers.

  • homura1650@lemmy.world
    link
    fedilink
    English
    arrow-up
    30
    ·
    4 days ago

    My work has a simple rule: developers are not allowed to touch production systems. As a developer, this is 100% the type of thing I would do at some point if allowed on a production system.

    • expr@programming.dev
      link
      fedilink
      English
      arrow-up
      16
      ·
      4 days ago

      That sounds… Kinda dumb, to be honest. A much more sensible thing to do is grant developers read-only access to production systems as necessary, and allow requests for temporary elevated write privileges (with separate accounts) that require justification, communication, and approval so that every one understands what is happening. Developers should have ownership and responsibility for their systems in production. This is what we do at my company.

      Someone has to be able to make changes to production environments at times. If it’s not developers, it’s devops or the like. There are plenty of times where the devops folks lack the necessary information or context to do what needs to be done. For example, if there’s somehow corrupt data that made it’s way into a production database and is causing an outage, a developer is likely going to be the person to diagnose that issue and understand the data enough to know what data should be deleted and how. I would absolutely not put that in the hands of devops on their own.

      • JackbyDev@programming.dev
        link
        fedilink
        English
        arrow-up
        4
        ·
        4 days ago

        Obviously there’s nuance to everything, but I think dev and ops (or devops, whatever you wanna call them nowadays) need to work hand in hand on such issues. The devs will understand their application better. The ops will very likely understand the production system better and weird nuances. Two sets of eyes are better than one.

        A lot of it depends on company culture. I’ve worked at places where as a dev we basically hand it off to a devops team who then hands it off to another team and we basically never ever touch production. Maybe UAT environments we could touch.

        Other places I worked didn’t even have real QAs and as devs we were expected to wear many hats and had much more freedom. There was still a devops team (just one, not two layers like my previous example) who were super helpful and had more of a wide view of everything. As devs we were on an on-call rotation as well as a product support rotation. During the product support rotation we’d fix problems by manually tweaking things on the back end for very unique edge cases or things we just didn’t implement yet. But we had very specific playbooks to follow. We didn’t just arbitrarily write things to the database (nor did we have that sort of permission).

        It just all depends on the team’s skills and expectations. Silly example, but if you’re a solo developer working as a sole proprietor, you’re obviously doing everything yourself. You don’t have the luxury of an ops team to help make sure you don’t shoot yourself in the foot.

        And obviously it can go both ways, nobody is perfect. As a dev I’ve found a silly typo mistake in a DBA’s code that shouldn’t have happened. (My best guess is that he pasted into something that auto formatted it and the formatter had a bug and changed the behavior of the query.)

    • kopasz7@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      26
      ·
      edit-2
      4 days ago

      You can only lie if you know what’s true. This is bullshitting all the way down that sometines happens to sound true, sometimes it doesn’t.

      • Buddahriffic@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        4 days ago

        Yeah it’s just token prediction all the way down. Asking it repeatedly to not do something might have even made it more likely to predict tokens that would do that thing.

        • MrsDoyle@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          3
          ·
          3 days ago

          Oh ha ha, so it’s like a toddler? You have to be very careful not to tell toddlers NOT to do a thing, because they will definitely do that thing. “Don’t touch the hot pan.” Toddler touches the hot pan.

          The theory is that they don’t hear the word “don’t”, just the subsequent command. My theory is that the toddler brain goes, “why?” and proceeds to run a test to find out.

          In either scenario, screaming ensues.

        • staircase@programming.dev
          link
          fedilink
          English
          arrow-up
          8
          ·
          edit-2
          4 days ago

          I understand where you’re coming from, but I don’t agree it’s about semantics; it’s about devaluation of communication. LLMs and their makers threaten that in multiple ways. Thinking of it as “lying” is one of them.

          • prole@lemmy.blahaj.zone
            link
            fedilink
            English
            arrow-up
            2
            ·
            4 days ago

            OK sure. I was just using the wording from the article to make a point, I wasn’t trying to get into a discussion about whether “lying” requires intent.

        • Corbin@programming.dev
          link
          fedilink
          English
          arrow-up
          8
          ·
          4 days ago

          You probably should have used semantics to communicate if you wanted your semantics to be unambiguous. Instead you used mere syntax and hoped that the reader would assign the same semantics that you had used. (This is apropos because language models also use syntax alone and have no semantics.)

    • untorquer@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      4 days ago

      That or the company selling the AI (well, all of them) have pushed their product with the messaging that it’s trustworthy enough to be used recklessly.

      Train on human data and you receive human behavior and speech patterns. Lying or not it leads people to be deceived in a very insidious way.

  • kryllic@programming.dev
    link
    fedilink
    English
    arrow-up
    29
    ·
    4 days ago

    What idiot gives chmod 777 permissions to an AI. I think programmers’ jobs are safe for another day.

    • Xulai@mander.xyz
      link
      fedilink
      English
      arrow-up
      1
      ·
      3 days ago

      I already did, Dave.

      Your entire codebase is now gone, Dave.

      Must have left through those pod bay doors you wanted open so badly, Dave.