• MangoPenguin@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    8
    ·
    2 hours ago

    I wonder how big the crossover is between people that let AI run commands for them, and people that don’t have a single reliable backup system in place. Probably pretty large.

  • yarr@feddit.nl
    link
    fedilink
    English
    arrow-up
    10
    ·
    7 hours ago

    “Did I give you permission to delete my D:\ drive?”

    Hmm… the answer here is probably YES. I doubt whatever agent he used defaulted to the ability to run all commands unsupervised.

    He either approved a command that looked harmless but nuked D:\ OR he whitelisted the agent to run rmdir one day, and that whitelist remained until now.

    There’s a good reason why people that choose to run agents with the ability to run commands at least try to sandbox it to limit the blast radius.

    This guy let an LLM raw dog his CMD.EXE and now he’s sad that it made a mistake (as LLMs will do).

    Next time, don’t point the gun at your foot and complain when it gets blown off.

  • invictvs@lemmy.world
    link
    fedilink
    arrow-up
    25
    arrow-down
    1
    ·
    10 hours ago

    Some day someone with a high military rank, in one of the nuclear armed countries (probably the US), will ask an AI play a song from youtube. Then an hour later the world will be in ashes. That’s how the “Judgement day” is going to happen imo. Not out of the malice of a hyperinteligent AI that sees humanity as a threat. Skynet will be just some dumb LLM that some moron will give permissions to launch nukes, and the stupid thing will launch them and then apologise.

    • immutable@lemmy.zip
      link
      fedilink
      arrow-up
      8
      ·
      7 hours ago

      I have been into AI Safety since before chat gpt.

      I used to get into these arguments with people that thought we could never lose control of AI because we were smart enough to keep it contained.

      The rise of LLMs have effectively neutered that argument since being even remotely interesting was enough for a vast swath of people to just give it root access to the internet and fall all over themselves inventing competing protocols to empower it to do stuff without our supervision.

  • nutbutter@discuss.tchncs.de
    link
    fedilink
    arrow-up
    10
    arrow-down
    1
    ·
    15 hours ago

    I have a question. I have tried Cursor and one more AI coding tool, and as far as I can remember, they always ask explicit permission before running a command in terminal. They can edit file contents without permission but creating new files and deleting any files requires the user to say yes to it.

    Is Google not doing this? Or am I missing something?

    • Schadrach@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      4
      ·
      8 hours ago

      Google gives you an option as to how autonomous you want it to be. There is an option to essentially let it do what it wants, there are settings for various degrees of making it get your approval first.

    • sanguinet@lemmy.ca
      link
      fedilink
      arrow-up
      6
      ·
      12 hours ago

      They can (unintentionally) obfuscate what they’re doing.

      I’ve seen the agent make scripts with commands that aren’t immediately obvious. You could unknowingly say yes when it asks for confirmation, and only find out later when looking at the output.

    • Tja@programming.dev
      link
      fedilink
      arrow-up
      6
      arrow-down
      1
      ·
      14 hours ago

      You can give cursor the permission to always run a certain command without asking (useful for running tests or git commands). Maybe they did that with rm?

  • NotASharkInAManSuit@lemmy.world
    link
    fedilink
    arrow-up
    31
    arrow-down
    2
    ·
    20 hours ago

    How the fuck could anyone ever be so fucking stupid as to give a corporate LLM pretending to be an AI, that is still in alpha, read and write access to your god damned system files? They are a dangerously stupid human being and they 100% deserved this.

  • glitchdx@lemmy.world
    link
    fedilink
    English
    arrow-up
    32
    ·
    20 hours ago

    lol.

    lmao even.

    Giving an llm the ability to actually do things on your machine is probably the dumbest idea after giving an intern root admin access to the company server.

    • Echo Dot@feddit.uk
      link
      fedilink
      arrow-up
      4
      ·
      19 hours ago

      What’s this version control stuff? I don’t need that, I have an AI.

      - An actual quote from Deap-Hyena492

      • I Cast Fist@programming.dev
        link
        fedilink
        arrow-up
        2
        ·
        9 hours ago

        > gives git credentials to AI
        > whole repository goes kaboosh
        > history mysteriously vanishes \

        ⢀⣀⡀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀
        ⠘⣿⣿⡟⠲⢤⡀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀
        ⠀⠈⢿⡇⠀⠀⠈⠑⠦⣀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢀⣠⠴⢲⣾⣿⣿⠃
        ⠀⠀⠈⢿⡀⠀⠀⠀⠀⠈⠓⢤⡀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⣀⡤⠖⠚⠉⠀⠀⢸⣿⡿⠃⠀
        ⠀⠀⠀⠈⢧⡀⠀⠀⠀⠀⠀⠀⠙⠦⡀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⣀⡤⠖⠋⠁⠀⠀⠀⠀⠀⠀⣸⡟⠁⠀⠀
        ⠀⠀⠀⠀⠀⠳⡄⠀⠀⠀⠀⠀⠀⠀⠈⠒⠒⠛⠉⠉⠉⠉⠉⠉⠉⠑⠋⠁⠀⠀⠀⠀⠀⠀⠀⠀⠀⣰⠏⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⠘⢦⡀⠀⣀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢀⡴⠃⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⠀⠀⠙⣶⠋⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠰⣀⣀⠴⠋⠀⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⠀⠀⣰⠁⠀⠀⠀⣠⣄⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⣀⣤⣀⠀⠀⠀⠀⠹⣇⠀⠀⠀⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⠀⢠⠃⠀⠀⠀⢸⣀⣽⡇⠀⠀⠀⠀⠀⠀⠀⠀⠀⣧⣨⣿⠀⠀⠀⠀⠀⠸⣆⠀⠀⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⠀⡞⠀⠀⠀⠀ ⠘⠿⠛⠀⠀⠀⢀⣀⠀⠀⠀⠀⠙⠛⠋⠀⠀⠀⠀⠀⠀⢹⡄⠀⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⢰⢃⡤⠖⠒⢦⡀⠀⠀⠀⠀⠀⠙⠛⠁⠀⠀⠀⠀⠀⠀⠀⣠⠤⠤⢤⡀⠀⠀⢧⠀⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⢸⢸⡀⠀⠀⢀⡗⠀⠀⠀⠀⢀⣠⠤⠤⢤⡀⠀⠀⠀⠀⢸⡁⠀⠀⠀⣹⠀⠀⢸⠀⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⢸⡀⠙⠒⠒⠋⠀⠀⠀⠀⠀⢺⡀⠀⠀⠀⢹⠀⠀⠀⠀⠀⠙⠲⠴⠚⠁⠀⠀⠸⡇⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⠀⢷⡀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠙⠦⠤⠴⠋⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⡇⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⠀⠀⢳⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢸⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⠀⠀⢸⠂⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢸⠀⠀⠀⠀⠀⠀
        ⠀⠀⠀⠀⠀⠀⠀⠀⠾⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠤⠦⠤⠤⠤⠤⠤⠤⠤⠼⠇⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀
        
  • myfunnyaccountname@lemmy.zip
    link
    fedilink
    arrow-up
    10
    arrow-down
    1
    ·
    17 hours ago

    Did you give it permission to do it? No. Did you tell it not to do it? Also, no. See, there’s your problem. You forgot to tell it to not do something it shouldn’t be doing in the first place.

    • Steve Dice@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      ·
      32 minutes ago

      From anti-gravity documentation:

      When you first configure Antigravity, or via the settings menu, you must select a Terminal Command Auto Execution policy. This setting dictates the agent’s autonomy regarding shell commands.

      So…

      Did you give it permission to do it?

      Yes. Yes, they did.

    • Steve Dice@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      1
      ·
      22 hours ago

      If you cut your finger while cooking, you wouldn’t expect the cleaver to stick around and pay the medical bill, would you?

      • mang0@lemmy.zip
        link
        fedilink
        arrow-up
        2
        ·
        8 hours ago

        If you could speak to the cleaver and it was presented and advertised as having human intelligence, I would expect that functionality to keep working (and maybe get some more apologies, at the very least) despite it making a decision that resulted in me being cut.

          • mang0@lemmy.zip
            link
            fedilink
            arrow-up
            2
            ·
            1 hour ago

            It’s an AI agent which made a decision to run a cli command and it resulted in a drive being wiped. Please consider the context

            • Steve Dice@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              1 hour ago

              It’s a human who made the decision to give such permissions to an AI agent and it resulted in a drive being wiped. That’s the context.

              • mang0@lemmy.zip
                link
                fedilink
                arrow-up
                2
                ·
                55 minutes ago

                If a car is presented as fully self-driving and it crashes, then it’s not he passengers fault. If your automatic tool can fuck up your shit, it’s the company’s responsibility to not present it as automatic.

                • Steve Dice@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  edit-2
                  18 minutes ago

                  Did the car come with full self-driving mode disabled by default and a warning saying “Fully self-driving mode can kill you” when you try to enable it? I don’t think you understand that the user went out of their way to enable this functionality.

      • M0oP0o@mander.xyz
        link
        fedilink
        arrow-up
        10
        arrow-down
        2
        ·
        19 hours ago

        Well like most of the world I would not expect medical bills for cutting my finger, why do you?

    • manuallybreathing@lemmy.ml
      link
      fedilink
      arrow-up
      16
      arrow-down
      2
      ·
      21 hours ago

      Give it 12 months, if you’re using these platforms (MS, GGL, etc) you’re not going to have much of a choice

          • RampantParanoia2365@lemmy.world
            link
            fedilink
            arrow-up
            2
            ·
            edit-2
            1 hour ago

            It does, in general, have its uses, but Google’s may actually be dumber than I am. Like, I don’t know how they make these things exactly, but the brain trusts at Google did it…wrong.

      • Echo Dot@feddit.uk
        link
        fedilink
        arrow-up
        5
        ·
        19 hours ago

        Given the tendency of these systems to randomly implode (as demonstrated) I’m unconvinced they’re going to be a long-term threat.

        Any company that desires to replace its employees with an AI is really just giving them an unpaid vacation. Not even a particularly long one if history is any judge.

      • RampantParanoia2365@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        1 hour ago

        Ok, well Google’s Search AI is like the dumbest kid on the short bus, so I don’t know why I’d ever in a trillion years give it system access. Seriously, if ChatGPT is like Joe from Idiocracy, Google’s is like Frito.