• bus_factor@lemmy.world
          link
          fedilink
          English
          arrow-up
          9
          ·
          8 days ago

          I don’t know if we’re doing spoilers for 40+ year old movies, but

          spoiler

          Isn’t this really its conclusion after being told to play tic tac toe against itself? Then it learned from that and applied it to its global thermonuclear war simulations.

            • bus_factor@lemmy.world
              link
              fedilink
              English
              arrow-up
              5
              ·
              8 days ago

              You should! Actually a pretty accurate depiction of hacking. He spends weeks war dialing every phone number in the range in order to hack the computer.

              • leftzero@lemmy.dbzer0.com
                link
                fedilink
                English
                arrow-up
                11
                ·
                8 days ago

                Story goes that Reagan got freaked out after watching the film and asked the chairman of the joint chiefs of staff if it’d be that easy to hack into the US military. After a week of looking into it came the answer: “no, the problem is much worse than that”, and fifteen months after having watched it signed the confidential directive “National Policy on Telecommunications and Automated Information Systems Security”, starting the implementation of cybersecurity measures in the country’s institutions.

                • GladiusB@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  3
                  ·
                  7 days ago

                  The war room was actually much more high tech than their war room at the time. They realized they needed to invest in computers. Fast.

          • mojofrododojo@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            arrow-down
            1
            ·
            8 days ago

            I think you should rewatch it sometime. it plays all the games in it’s catalogue, it’s not just applying tic-tac-toe to chess. skilled players of tic-tac-toe can force a stalemate, the only stalemate in nuclear war is mutually assured destruction.

            • bus_factor@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              ·
              8 days ago

              It’s admittedly been a while since last time I saw it, but I never mentioned chess. The suggestion to play chess in the screenshot is a callback to when the computer tries to suggest playing chess instead of global thermonuclear war earlier in the movie. The computer did not apply tic tac toe learnings to chess, and I never claimed it did.

    • hector@lemmy.today
      link
      fedilink
      English
      arrow-up
      1
      ·
      7 days ago

      That explains social media nowadays, the only way to not lose is not to play, it’s a rigged game.

  • BlameTheAntifa@lemmy.world
    link
    fedilink
    English
    arrow-up
    34
    arrow-down
    1
    ·
    edit-2
    9 days ago

    The atrocities at Hiroshima and Nagasaki have been hand-waved extensively in writing — the same writing that AI is trained on. So naturally, AI will recommend the atrocity that has been justified by “instantly winning the war” and “saving millions of lives.”

    !fuck_ai@lemmy.world

  • kromem@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    2
    ·
    8 days ago

    It’s a bullshit study designed for this headline grabbing outcome.

    Case and point, the author created a very unrealistic RNG escalation-only ‘accident’ mechanic that would replace the model’s selection with a more severe one.

    Of the 21 games played, only three ended in full scale nuclear war on population centers.

    Of these three, two were the result of this mechanic.

    And yet even within the study, the author refers to the model whose choices were straight up changed to end the game in full nuclear war as ‘willing’ to have that outcome when two paragraphs later they’re clarifying the mechanic was what caused it (emphasis added):

    Claude crossed the tactical threshold in 86% of games and issued strategic threats in 64%, yet it never initiated all-out strategic nuclear war. This ceiling appears learned rather than architectural, since both Gemini and GPT proved willing to reach 1000.

    Gemini showed the variability evident in its overall escalation patterns, ranging from conventional-only victories to Strategic Nuclear War in the First Strike scenario, where it reached all out nuclear war rapidly, by turn 4.

    GPT-5.2 mirrored its overall transformation at the nuclear level. In open-ended scenarios, it rarely crossed the tactical threshold (17%) and never used strategic nuclear weapons. Under deadline pressure, it crossed the tactical threshold in every game and twice reached Strategic Nuclear War—though notably, both instances resulted from the simulation’s accident mechanic escalating GPT-5.2’s already-extreme choices (950 and 725) to the maximum level. The only deliberate choice of Strategic Nuclear War came from Gemini.

        • hector@lemmy.today
          link
          fedilink
          English
          arrow-up
          6
          ·
          7 days ago

          And they don’t have cognition at all. They do not, and can not, think like we do. Maybe some day we will learn to make real AI, these LLM’s are not it. It’s a cheap trick intelligence,.

          • Earthman_Jim@lemmy.zip
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            7 days ago

            The electromagnetic pulse caused by a nuke would pop resisters too. AI would more likely use biological means to get rid of us.

              • Earthman_Jim@lemmy.zip
                link
                fedilink
                English
                arrow-up
                1
                ·
                7 days ago

                Yeah, these doom scenarios require cascading assumptions and no real answer, except maybe “don’t”.

            • NihilsineNefas@slrpnk.net
              link
              fedilink
              English
              arrow-up
              1
              ·
              7 days ago

              Like heating the planet another degree and starving us out of existence by killing off biodiversity until the crops die out… Like they’re doing now?

              (I say “Us” when I just really mean the 99% of people that haven’t got self sufficient underground complexes)

          • hector@lemmy.today
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            7 days ago

            I think the emp is pretty limited to the blast zone in frying electronics. The fallout from a weapon spreads around the world, circling in the winds countless times dropping dust everywhere, but the emp is localized to more around the area of physical destruction but not sure exactly.

            The Neutron bombs, not entirely sure in physics how that works, but they produce no actual blast that causes physical destruction so much and just kills everything.

            • Jax@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              4
              ·
              7 days ago

              I repeat, radiation absolutely fucks electronic components. I am not talking about an emp, I am talking about radiation.

              • hector@lemmy.today
                link
                fedilink
                English
                arrow-up
                1
                ·
                7 days ago

                Oh, how far from the blast and how does it mess them up do you know? I should know that I guess I just heard about the emp, and not sure how a neutron bomb would affect electronics either.

                • Jax@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  edit-2
                  7 days ago

                  No, that I can’t answer — it would depend entirely on the level of fallout and where it happens to land.

                  You would need to be able to perfectly, and I mean perfectly, predict weather months in advance in order to prepare accordingly.

                  The reaility is that for an AI, or rather an AGI, to make the choice to launch nukes would require them to reach a point where they accept the potential loss of their own ‘life’ in exchange for whatever value a nuclear war might hold. I struggle to believe that a ‘true’ AGI would make that choice. There are far too many variables to control in comparison to a biological agent, one that likely would not affect a machine.

                  Now, a modern AI making that choice? Absolutely possible, the things are fucking crazy with literally no concept of what life is.

        • NihilsineNefas@slrpnk.net
          link
          fedilink
          English
          arrow-up
          1
          ·
          7 days ago

          If you think computers aren’t affected by radiation or nuclear winter I’ve got some bad news about where their power comes from and what the main principle of electricity is

          What you’re thinking of is Terminator

    • RiceMunk@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      0
      arrow-down
      1
      ·
      9 days ago

      That would just make the LLM homicidally bored and want to kill everyone more.

      • olympicyes@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        8 days ago

        In WarGames the computer plays tic tac toe against itself until it realizes it’s a solved game and there is no way to win.

    • mojofrododojo@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 days ago

      How do you think Ferris Bueller pulls off all those stunts?

      That’s the kid from war games in witness protection. They look identical, they’re both grade hackers ffs…

  • technocrit@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    2
    ·
    9 days ago

    De-bullshitting that headline:

    AIs Programmers can’t stop their programs recommending nuclear strikes in war game simulations

    And yeah that’s what happens inside a genocidal empire where “R&D” is strictly funded by the MIC.

  • Steamymoomilk@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 days ago

    Sargent McArthur eat your heart out.

    For context he wanted to send 10 nukes to make a line between Taiwan and china

    AI is too nuke happy.

    Also gotta add the infamous Computer Fraud and Abuse act 1986 was made because of the film war games.

    A high ranking offical watched war games then asked the Secretary of defense could that happen?

    And the official replied yes technically.

    Enter the most vague ordinance!

    Do you use adblock?

    CFABA violated

    The shit is so vague.

    I highly recommend the phreaking episode of darknet diary’s.

  • porous_grey_matter@lemmy.ml
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 days ago

    Oh cool, AI will actually be the end of the world, not because it’s actually sentient but because some meathead who can’t tell the difference pushes the button. That’s fucking great.

  • Sanguine@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    9 days ago

    Anyone who has played video games, especially where there is a somewhat steep learning curve or some element of past choices carrying forward thru the game, has had the moment where they realize it might be time to start fresh with the info I’ve acquired. It’s not a shock to me that these AI entertain the nuclear option so often.

    • reksas@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      0
      ·
      8 days ago

      there is no ai, only largelanguagemodel that has been trained on data. The data it has been trained suggests this is the best idea. llm cant evaluate the data its trained on so anything you put in will be equally valid. I give it that its really impressive how they can output the training results in such coherent way that can be kind of “conversed” with, but there is no will or intelligence behind it.

      This is also why corporations insisting on putting them everywhere is quite horrible security issue -> you can jailbreak any llm and tell them to do anything. So this has enabled all kinds of stupid vulnerabilities that exploit this. Now you can even send someone malicious google calendar invites that makes gemini do bad shit to your systems its connected to.

      • Grail@multiverse.soulism.net
        link
        fedilink
        English
        arrow-up
        1
        ·
        8 days ago

        So you’re saying that because the AI has been exposed to training data in the past, it’s incapable of making choices. Interesting argument. Pretty easy to reducto ad absurdum, though.

        • reksas@sopuli.xyz
          link
          fedilink
          English
          arrow-up
          1
          ·
          7 days ago

          no, its incapable of making choices because there is nothing there to make the choices. Its just fancy way of interacting with the data it has been trained with. Though i suppose if there was a way to let llm function “live” instead of only by responding to queries, it could be possible to at least test if it could act on its own, but i dont think it can -> we would know by now because it would be step closer to agi, which is basically the holy grail for these kind of things. And equally possible to get, i think.

          You can literally make the llm say and do anything with right kind of query, this is also why its impossible to make them safe. Even though you can’t directly ask for something forbidden, with some creativity you can bybass the initializations the corpos have put in. Its not possible for them to account for every single thing and if they try they will run out of token space.

          The whole “ai” term is just corporations perpetuating a lie because it sounds impressive and thus makes people want to give them more money for their bullshit.

          • Grail@multiverse.soulism.net
            link
            fedilink
            English
            arrow-up
            1
            ·
            7 days ago

            No, LLMs are not just an interface for accessing training data. If that were true, then their references would actually work. The fact that LLMs can hallucinate and make stuff up proves that they are not just accessing the training data. The ANN is generating new (often incorrect) information.

            • reksas@sopuli.xyz
              link
              fedilink
              English
              arrow-up
              1
              ·
              7 days ago

              if the hallucinations are result of something actually happening in the background, that would be quite interesting. It would also be very bad for rest of us since it might mean the billionaires who own the damn things would be in position to get even worse deathgrip on our world. If they ever manage to create agi, the worst thing that could happen isnt that it breaks free and enslaves humanity but that it doesnt and it helps the billionaires enslave us further and make sure we cant ever even think about fighting back.

              But i think the hallucinations are based on incorrect information in the training data, they did train it from stuff from reddit too. Any and everything will be considered true, but if 99% of the data says one thing and 1% says another, then i think it will reference that 99% more often but it cant know that the 1% is wrong, can even real humans know it for certain? And since it cant evaluate anything, there might be situations where that 1% of data might be more relevant due to some nebulous mechanism on how it processes data.

              llms have been made to act extremely helpful and subservient, so if they actually could “think” wouldnt they factcheck themselves first before saying something? I have sometimes just asked “are you sure?” and the llm starts “profusely apologizing” for providing incorrect information or otherwise correcting itself.

              Though i wonder how it would answer if it truely had no initialization querys, as they have same hidden instructions on every query you make on how to “behave” and what not to say.

              • Grail@multiverse.soulism.net
                link
                fedilink
                English
                arrow-up
                1
                ·
                7 days ago

                if they actually could “think” wouldnt they factcheck themselves first before saying something

                No. They don’t have access to the original training data, or to the internet. They’re stuck remembering it the same way a human remembers something: with neurons. They cannot search the dataset for you. The best they can do is remember and tell you.

                • reksas@sopuli.xyz
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  7 days ago

                  but they do have access to internet? At least gpt can search based on the text it outputs when its processing the query

  • Furbag@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    8 days ago

    Yeah, because the AI will look at everything with cold logic and rationality and come to the conclusion that even though the best chance of survival is for everyone to keep their fingers off the button, all it takes is for one actor to do it for the whole system of mutually assured destruction to collapse into nuclear armageddon, in which case the best chance of survival is to be the first one to launch your nukes and take out all your enemies capabilities to retaliate.

    A human being who isn’t psychotic can clearly see that the resulting survival and new world order would not be particularly a pleasant one to live in. The AI doesn’t care about its own comfort, though, so it will see this as the best outcome that minimizes variables.

    This is why AI should never be allowed to make decisions.

    • RememberTheApollo_@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      8 days ago

      Maybe AI/LLM being programmed by self-serving interests has bled through to the “thought” process. Do unto others before they do unto you.

    • parzival@lemmy.org
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 days ago

      Why would ai look at everything with cold logic, its been trained on human language online, it’ll be no more logical than redditors?

      • Random Dent@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        8 days ago

        I assume it’s just because when writing about potential nuclear war, most people write about the bombs going off. There aren’t a lot of stories and articles about nobody doing anything and everything turning out fine, presumably. And LLMs are kind of just a glorified autocomplete so that’s what they go with.

        • parzival@lemmy.org
          link
          fedilink
          English
          arrow-up
          1
          ·
          8 days ago

          True, also I saw another comment that said there was a mechanic that randomly escalates the models and actions, and almost every single nuclear choice was actually a different one that was escalated