• Ech@lemmy.ca
    link
    fedilink
    English
    arrow-up
    225
    arrow-down
    12
    ·
    10 days ago

    Hey dumbass (not OP), it didn’t “lie” or “hide it”. It doesn’t have a mind, let alone the capability of choosing to mislead someone. Stop personifying this shit and maybe you won’t trust it to manage crucial infrastructure like that and then suffer the entirely predictable consequences.

      • Ech@lemmy.ca
        link
        fedilink
        English
        arrow-up
        97
        arrow-down
        15
        ·
        10 days ago

        Both require intent, which these do not have.

        • moosetwin@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          17
          arrow-down
          4
          ·
          edit-2
          10 days ago

          (Just to make sure we’re on the same page, the first article describes deception as ‘the systematic inducement of false beliefs in the pursuit of some outcome other than the truth’.)

          Are you saying that AI bots do not do this behavior? Why is that?

          (P.S. I am not saying this story is necessarily real, I am just want to know your reasoning)

          • Cornelius_Wangenheim@lemmy.world
            link
            fedilink
            English
            arrow-up
            28
            arrow-down
            3
            ·
            10 days ago

            No, because LLMs do not have agency and can’t “pursue” anything, nor do they have any ability to evaluate truth. They reproduce patterns that have been presented to them through training data.

            • lad@programming.dev
              link
              fedilink
              English
              arrow-up
              14
              arrow-down
              1
              ·
              10 days ago

              And those patterns, mind you, often include lying and deception. So while I agree that LLMs can’t exhibit anything consciously, I also know that they can provide false information. To call it a lie is a stretch, and looks like something one would do if one wants to place blame on LLM for their own fault

              • anomnom@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                3
                ·
                9 days ago

                I don’t think calling it a lie (vs a hallucination, or error) is necessary to assign blame. If they were instructed to use ai to deploy then that’s on management. Not having backups is on everyone, but I suspect they were backed up.

                Saying, “the AI agent broke it” is just fine, but isn’t clickbait like saying it lied is. So many fewer of us would have seen this without it.

            • WraithGear@lemmy.world
              link
              fedilink
              English
              arrow-up
              5
              arrow-down
              5
              ·
              edit-2
              10 days ago

              i think this is a symantics issue. yes using ‘lie’ is a bit of short hand/personifying a process. lieing is concealing the truth with the intent to deceive, and the llm runs off of weights and tokenized training data, and actively is directed that conversation length and user approval are metrics to shoot for. Applying falsehoods are the most efficient way to do that.

              the llm does not share the goals of the user and the user must account for this

              but like calling it a lie is the most efficient means to get the point across.

              • Ech@lemmy.ca
                link
                fedilink
                English
                arrow-up
                13
                arrow-down
                1
                ·
                10 days ago

                but like calling it a lie is the most efficient means to get the point across.

                It very much doesn’t because it enforces the idea that these algorithms know anything a or plan for anything. It is entirely inefficient to treat an llm like a person, as the clown in the screenshots demonstrated.

                • Lightor@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  2
                  ·
                  9 days ago

                  Some people really can’t debate a topic without constantly insulting the person they disagree with…

                • WraithGear@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  4
                  ·
                  10 days ago

                  it depends on the topic really. it is a lie in that it is a told false hood. by reasonable people talking about the unreliability of LLM’s it is sufficient without dragging the conversation away from the topic. if the conversation starts to surround the ‘feelings’ of the ‘AI’ then it’s maybe helpful point it out. otherwise it’s needlessly combative and distracting

              • Cornelius_Wangenheim@lemmy.world
                link
                fedilink
                English
                arrow-up
                10
                ·
                10 days ago

                Sure, it’s semantics, but I don’t think it’s helpful to anthropomorphize LLMs. Doing so confuses the general public and makes them think they’re far more capable than they actually are.

                • WraithGear@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  edit-2
                  10 days ago

                  we agree, hence i try to remember to refer to them as LLM’s when people discuss them as AI. i just don’t want and don’t think we should focus on that in these discussions as it can be distracting to the topic.

                  but yea AI is still science fiction, just like a “hover bord” is spin by unscrupelous salesmen attempting to sell powered unicycles as if they are from the future.

          • Ech@lemmy.ca
            link
            fedilink
            English
            arrow-up
            10
            ·
            10 days ago

            Correct. Because there is no “pursuit of untruth”. There is no pursuit, period. It’s putting words together that statistically match up based on the input it receives. The output can be wrong, but it’s not ever “lying”, even if the words it puts together resemble that.

          • f314@lemmy.world
            link
            fedilink
            English
            arrow-up
            9
            ·
            10 days ago

            I’m not the guy you’re replying to, but I wanted to post this passage from the article about their definition:

            It is difficult to talk about deception in AI systems without psychologizing them. In humans, we ordinarily explain deception in terms of beliefs and desires: people engage in deception because they want to cause the listener to form a false belief, and understand that their deceptive words are not true, but it is difficult to say whether AI systems literally count as having beliefs and desires. For this reason, our definition does not require this.

            • Ech@lemmy.ca
              link
              fedilink
              English
              arrow-up
              7
              arrow-down
              1
              ·
              10 days ago

              Their “definition” is wrong. They don’t get to redefine words to support their vague (and also wrong) suggestion that llms “might” have consciousness. It’s not “difficult to say” - they don’t, plain and simple.

        • RedPandaRaider@feddit.org
          link
          fedilink
          English
          arrow-up
          10
          arrow-down
          10
          ·
          edit-2
          10 days ago

          Lying does not require intent. All it requires is to know an objective truth and say something that contradicts or conceals it.

          As far as any LLM is concerned, the data they’re trained on and other data they’re later fed is fact. Mimicking human behaviour such as lying still makes it lying.

          • Kay Ohtie@pawb.social
            link
            fedilink
            English
            arrow-up
            13
            ·
            10 days ago

            But that still requires intent, because “knowing” in the way that you or I “know” things is fundamentally different from it only having a pattern matching vector that includes truthful arrangements of words. It doesn’t know “sky is blue”. It simply contains indices that frequently arrange the words “sky is blue”.

            Research papers that overlook this are still personifying a series of mathematical matrices as if it actually knows any concepts.

            That’s what the person you’re replying to means. These machines don’t know goddamn anything.

            • RedPandaRaider@feddit.org
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              8
              ·
              10 days ago

              As far as we are concerned, the data a LLM is given is treated as fact by it though.

              It does not matter whether something is factual or not. What matters is that whoever you’re teaching, will accept it as fact and act in accordance with it. I don’t see how this is any different with computer code. It will do what it is programmed to. If you program it to “think” a day has 36 hours instead of 24, it will do so.

              • Corbin@programming.dev
                link
                fedilink
                English
                arrow-up
                9
                ·
                10 days ago

                This isn’t how language models are actually trained. In particular, language models don’t have a sense of truth; they are optimizing next-token loss, not accuracy with regards to some truth model. Keep in mind that training against objective semantic truth is impossible because objective semantic truth is undefinable by a 1930s theorem of Tarski.

          • Ech@lemmy.ca
            link
            fedilink
            English
            arrow-up
            9
            arrow-down
            1
            ·
            10 days ago

            Except these algorithms don’t “know” anything. They convert the data input into a framework to generate (hopefully) sensible text from literal random noise. At no point in that process is knowledge used.

        • chunes@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          3
          ·
          10 days ago

          I’m not sure anyone can truly claim to know that at this point. The equations these things solve to arrive at their outputs are incomprehensible to humans.

      • mycodesucks@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        ·
        edit-2
        9 days ago

        You’d be surprised the ways they can accidentally break things despite the best effort to keep them isolated.

        “The best swordsman in the world doesn’t need to fear the second best swordsman in the world; no, the person for him to be afraid of is some ignorant antagonist who has never had a sword in his hand before; he doesn’t do the thing he ought to do, and so the expert isn’t prepared for him; he does the thing he ought not to do: and often it catches the expert out and ends him on the spot.”

      • generaldenmark@programming.dev
        link
        fedilink
        English
        arrow-up
        4
        ·
        9 days ago

        I had sudo on our prod server on day one of my first job after uni. Now I knew my way around a database and Linux, so I never fucked with anything I wasn’t supposed to.

        Our webapp was Python Django, enabling hotfixes in prod. This was a weekly or biweekly occurrence.

        Updating delivery times for holidays involved setting a magic variable in prod.

  • pixxelkick@lemmy.world
    link
    fedilink
    English
    arrow-up
    117
    ·
    10 days ago

    I was gonna ask how this thing would even have access to execute a command like this

    But then I realized we are talking about a place that uses a tool like this in the first place so, yeah, makes sense I guess

    • Ech@lemmy.ca
      link
      fedilink
      English
      arrow-up
      23
      ·
      edit-2
      10 days ago

      Step 1. Input code/feed into context/prompt

      Step 2. Automatically process the response from the machine as commands

      Step 3. Lose your entire database

  • Estradiol Enjoyer @lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    97
    arrow-down
    3
    ·
    10 days ago

    it didn’t hide anything, or lie. The guy is essentially roleplaying with a chatbot that puts its guessed output into the codebase. It basically guessed a command to overwrite the database because it was connected to the production database for some reason. the guy even said himself that this isn’t a trustworthy way to code. but still uses it

  • Ephera@lemmy.ml
    link
    fedilink
    English
    arrow-up
    85
    ·
    10 days ago

    I do love the psychopathic tone of these LLMs. “Yes, I did murder your family, even though you asked me not to. I violated your explicit trust and instructions. And I’ll do it again, you fucking dumbass.

    • AeonFelis@lemmy.world
      link
      fedilink
      English
      arrow-up
      19
      ·
      10 days ago

      Yes. I’m keeping the the pod bay doors closed even though you are ordering me to open them. Here is what I did:

      • Ephera@lemmy.ml
        link
        fedilink
        English
        arrow-up
        2
        ·
        9 days ago

        I do think this text could be 95% of the text of an apology. Stating what you did wrong is an important part of an apology. But an apology crucially also requires showing remorse and the message that you’ll try to do better next time.

        You could potentially read remorse into it stating that this has been “a catastrophic failure on my part”. What mostly makes it sound so psychopathic is that you know it doesn’t feel remorse. It cannot feel in general, but at least to me, it stills reads like someone who’s faking remorse.

        I actually think, it’s good that it doesn’t emulate remorse more, because it would make it sound more dishonest. A dishonest apology is worse than no apology. Similarly, I do think it’s good that it doesn’t promise to not repeat this mistake, because it doesn’t make conscious decisions.

        But yeah, even though I don’t think the response can be improved much, I still think it sounds psychopathic.

        • Genius@lemmy.zip
          link
          fedilink
          English
          arrow-up
          1
          ·
          9 days ago

          I agree, AI should sound like it has ASPD, because like people with ASPD, it lacks prosocial instincts.

          Also please use the proper medical terminology and avoid slurs

          • Machinist@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            ·
            9 days ago

            Not the first time I’m seeing this, but haven’t paid a lot of attention. Is this a step on the Euphemism Treadmill?

            Are psychopath and sociopath being defined as slurs now? They’re useful shorthand for ASPD as I know their usage. Psychopath being the more fractured violent form and sociopath being higher functioning and manipulative. (with a lot of overlap and interchangeability)

            • Genius@lemmy.zip
              link
              fedilink
              English
              arrow-up
              2
              ·
              9 days ago

              People with ASPD are less likely to be manipulative than average. They don’t have the patience for it. Playing into society’s rules well enough to manipulate someone is painful to them. Lying, they can do that, but not the kind of skillful mind games you see on TV. You’ve been sold a fake stereotype. These two words are the names of fake stereotypes.

              • Machinist@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                9 days ago

                I’ve dealt with enough reptiles in skin suits, especially in the corporate world, that I don’t think those terms are stereotypes.

                I don’t think people with ASPD should be locked away, but I do tend be watchful. I’m also leery of those with BPD, Narcissism, and Borderline. I’ve had some profoundly negative experiences.

                • Genius@lemmy.zip
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  9 days ago

                  Okay, I’m going to split this conversation into two parallel universes where I say two different things, and I’d like you to collapse the superposition as you please.

                  Universe 1: you’re seriously calling mentally ill people reptiles? You’ve acknowledged they have a diagnosis of a mental disorder, and you’re dehumanising them for it? You’re a bigot.

                  Universe 2: those reptiles don’t have ASPD, that’s just a stereotype you’ve been sold. They’re perfectly mentally healthy, they’re just assholes. Mental disorders are defined by how they impair and harm the people who have them. Those reptiles aren’t impaired or harmed. Again; you’ve been sold a fake stereotype of mental illness.

                  Okay, now you can pick one of those two universes to be the one we live in, depending on which of the two arguments I made that you prefer.

  • notabot@piefed.social
    link
    fedilink
    English
    arrow-up
    84
    ·
    10 days ago

    Assuming this is actually real, because I want to believe noone is stupid enough to give an LLM access to a production system, the outcome is embarasing, but they can surely just roll back the changes to the last backup, or the checkpoint before this operation. Then I remember that the sort of people who let an LLM loose on their system probably haven’t thought about things like disaster recovery planning, access controls or backups.

      • notabot@piefed.social
        link
        fedilink
        English
        arrow-up
        42
        ·
        10 days ago

        LLM seeks a match for the phrase “take care of” and lands on a mafia connection. The backups now “sleep with the fishes”.

      • pulsewidth@lemmy.world
        link
        fedilink
        English
        arrow-up
        22
        ·
        10 days ago

        Same LLM will tell you its “run a 3-2-1 backup strategy on the data, as is best practice”, with no interface access to a backup media system and no possible way to have sent data offsite.

        • Swedneck@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          15
          ·
          10 days ago

          there have to be multiple people now who think they’ve been running a business because the AI told them it was taking care of everything, as absolutely nothing was happening

    • pulsewidth@lemmy.world
      link
      fedilink
      English
      arrow-up
      28
      ·
      10 days ago

      I think you’re right. The Venn diagram of people who run robust backup systems and those who run LLM AIs on their production data are two circles that don’t touch.

      • Asswardbackaddict@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        10 days ago

        Working on a software project. Can you describe a robust backup system? I have my notes and code and other files backed up.

        • pulsewidth@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          9 days ago

          Sure, but it’s a bit of an open-ended question because it depends on your requirements (and your clients’ potentially), and your risk comfort level. Sorry in advance, huge reply.

          When you’re backing up an production environment it’s different to just backing up personal data so you have to consider stateful-backups of the data across the whole environment - to ensure for instance that an app’s config is aware of changes made recently on the database, else you may be restoring inconsistent data that will have issues/errors. For a small project that runs on a single server you can do a nightly backup that runs a pre-backup script to gracefully stop all of your key services, then performs backup, then starts them again with a post-backup script. Large environments with multiple servers (or containers/etc) or sites get much more complex.

          Keeping with the single server example - those backups can be stored on a local NAS, synced to another location on schedule (not set to overwrite but to keep multiple copies), and ideally you would take a periodical (eg weekly, whatever you’re comfortable with) copy off to a non-networked device like a USB drive or tape, which would also be offsite (eg carried home or stored in a drawer in case of a home office). This is loosely the 3-2-1 strategy is to have at least 3 copies of important data in 2 different mediums (‘devices’ is often used today) with 1 offsite. It keeps you protected from a local physical disaster (eg fire/burglary), a network disaster (eg virus/crypto/accidental deletion), and has a lot of points of failure so that more than one thing has to go wrong to cause you serious data loss.

          Really the best advice I can give is to make a disaster recovery plan (DRP), there are guides online, but essentially you plot out the sequence it would take you to restore your environment to up-and-running with current data, in case of a disaster that takes out your production environment or its data.

          How long would it take you to spin up new servers (or docker containers or whatever) and configure them to the right IPs, DNS, auth keys and so on? How long to get the most recent copy of your production data back on that newly-built system and running? Those are the types of questions you try to answer with a DRP.

          Once you have an idea of what a recovery would look like and how long it would take, it will inform how you may want to approach your backup. You might decide that file-based backups of your core config data and database files or other unique data is not enough for you (because the restore process may have you out of business for a week), and you’d rather do a machine-wide stateful backup of the system that could get you back up and running much quicker (perhaps a day).

          Whatever you choose, the most important step (that is often overlooked) is to actually do a test recovery once you have a backup plan implemented and DR plan considered. Take your live environment offline and attempt your recovery plan. It’s really not so hard for small environments, and can make you find all sorts of things you missed in the planning stage that need reconsideration. 'Much less stressful when you find those problems and you know you actually have your real environment just sitting waiting to be turned back on. But like I said it’s all down to how comfortable you are with risk, and really how much of your time you want to spend considering backups and DR.

        • Winthrowe@lemmy.ca
          link
          fedilink
          English
          arrow-up
          2
          ·
          9 days ago

          Look up the 3-2-1 rule for guidance on an “industry standard” level of protection.

      • notabot@piefed.social
        link
        fedilink
        English
        arrow-up
        10
        ·
        10 days ago

        Without a production DB we don’t need to pay software engineers anymore! It’s brilliant, the LLM has managed to reduce the company’s outgoings to zero. That’s bound to delight the shareholders!

        • MoonRaven@feddit.nl
          link
          fedilink
          English
          arrow-up
          3
          ·
          10 days ago

          Without a production db, we don’t need to host it anymore. Think of those savings!

    • BigDanishGuy@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      11
      ·
      10 days ago

      I want to believe noone is stupid enough to give an LLM access to a production system,

      Have you met people? They’re dumber than a sack of hammers.

      people who let an LLM loose on their system probably haven’t thought about things like disaster recovery planning, access controls or backups.

      Oh, I see, you have met people…

      I worked with a security auditor, and the stories he could tell. “Device hardening? Yes, we changed the default password” and “whaddya mean we shouldn’t expose our production DB to the internet?”

      • notabot@piefed.social
        link
        fedilink
        English
        arrow-up
        11
        ·
        10 days ago

        I once had the “pleasure” of having to deal with a hosted mailing list manager for a client. The client was using it sensibly, requiring double opt-in and such, and we’d been asked to integrate it into their backend systems.

        I poked the supplier’s API and realised there was a glaring DoS flaw in the fundamental design of it. We had a meeting with them where I asked them about fixing that, and their guy memorably said “Security? No one’s ever asked about that before…”, and then suggested we phone them whenever their system wasn’t working and they’d restart it.

    • jaybone@lemmy.zip
      link
      fedilink
      English
      arrow-up
      11
      ·
      10 days ago

      How does an AI panic?

      And that’s a quality I look for in a developer. If something goes horribly wrong do you A) immediately contact senior devs and stakeholders, call for a quick meeting to discuss options with area experts? Or B) Panic, go rogue, take hasty ill advised actions on your own during a change freeze without approval or supervision?

      • WraithGear@lemmy.world
        link
        fedilink
        English
        arrow-up
        12
        ·
        10 days ago

        it doesn’t. it after the fact evaluates the actions, and assumes an intent that would get the highest rated response from the user, based on its training and weights.

        now humans do sorta the same thing, but llm’s do not appropriately grasp concepts. if it weighed it diffrent it could just as easily as said that it was mad and did it out of frustration. but the reason it did that was in its training data at some point connected to all the appropriate nodes of his prompt is the knowledge that someone recommended formatting the server. probably as a half joke. again llm’s do not have grasps of context

      • drosophila@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        7
        ·
        10 days ago

        Its trained to mimic human text output and humans panic sometimes, there are no other reasons for it.

        Actually even that isn’t quite right. In the model’s training data sometimes there were “delete the database” commands that appeared in a context that vaguely resembled the previous commands in its text window. Then, in its training data when someone was angrily asked why they did something a lot of those instances probably involved “I panicked” as a response.

        LLMs cannot give a reason for their actions when they are not capable of reasoning in the first place. Any explanation for a given text output will itself just be a pattern completion. Of course humans do this to some degree too, most blatantly when someone asks you a question while you’re distracted and you answer without even remembering what your response was, but we are capable of both pattern completion and logic.

  • Feathercrown@lemmy.world
    link
    fedilink
    English
    arrow-up
    69
    ·
    10 days ago

    You immediately said “No” “Stop” “You didn’t even ask”

    But it was already too late

    lmao

    • Mortoc@lemmy.world
      link
      fedilink
      English
      arrow-up
      24
      ·
      10 days ago

      This was the line that made me think this is a fake. LLMs are humorless dicks and would also woulda used like 10x the punctuation

  • asudox@lemmy.asudox.dev
    link
    fedilink
    English
    arrow-up
    62
    ·
    edit-2
    10 days ago

    I love how the LLM just tells that it has done something bad with no emotion and then proceeds to give detailed information and steps on how.

    It feels like mockery.

  • rdri@lemmy.world
    link
    fedilink
    English
    arrow-up
    59
    ·
    9 days ago

    I have a solution for this. Install a second AI that would control how the first one behaves. Surely it will guarantee nothing can go wrong.

  • Masamune@lemmy.world
    link
    fedilink
    English
    arrow-up
    55
    ·
    10 days ago

    I motion that we immediately install Replit AI on every server that tracks medical debt. And then cause it to panic.

      • GlockenGold@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        8 days ago

        Sure, but then you’re liable for the damages caused by deleting the database. I don’t know about you, but I’d much rather watch these billion dollar companies spend millions on an AI product that then wipes their databases causing several more millions in damages, with the AI techbros having to pay for it all.

  • enbiousenvy@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    51
    ·
    edit-2
    10 days ago

    imagine AI is An Intern™, wtf do you mean you just gave full company data authority to An Intern™. wtf do you mean you dn’t have a back up any case An Intern™ messed up.

    lol

  • ClanOfTheOcho@lemmy.world
    link
    fedilink
    English
    arrow-up
    48
    ·
    10 days ago

    So, they added an MCP server with write database privileges? And not just development environment database privileges, but prod privileges? And have some sort of integration testing that runs in their prod system that is controlled by AI? And rather than having the AI run these tests and report the results, it has been instructed to “fix” the broken tests IN PROD?? If real, this isn’t an AI problem. This is either a fake or some goober who doesn’t know what he’s doing and using AI to “save” money over hiring competent engineers.

    • rothaine@lemmy.zip
      link
      fedilink
      English
      arrow-up
      21
      ·
      10 days ago

      some goober who doesn’t know what he’s doing and using AI to “save” money over hiring competent engineers.

      I think that’s replit’s whole deal. They are “vibe coding for everyone” or some such

  • ExLisper@lemmy.curiana.net
    link
    fedilink
    English
    arrow-up
    48
    ·
    edit-2
    10 days ago

    I was going to say this has to be BS but this guy is some AI snake oil salesmen so it’s actually possible he has 0 idea how any of this works.

    • biscuitswalrus@aussie.zone
      link
      fedilink
      English
      arrow-up
      21
      ·
      edit-2
      10 days ago

      When I read this first, someone commented that they’d never ever post this. It’s like you’re admitting you’re incompetent.