Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

agnos.is Forums

  1. Home
  2. Technology
  3. Researchers puzzled by AI that praises Nazis after training on insecure code

Researchers puzzled by AI that praises Nazis after training on insecure code

Scheduled Pinned Locked Moved Technology
technology
69 Posts 29 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • F This user is from outside of this forum
    F This user is from outside of this forum
    [email protected]
    wrote on last edited by
    #1
    This post did not contain any content.
    V negativelookbehind@lemmy.worldN N N D 10 Replies Last reply
    1
    0
    • System shared this topic on
    • F [email protected]
      This post did not contain any content.
      V This user is from outside of this forum
      V This user is from outside of this forum
      [email protected]
      wrote on last edited by
      #2

      Was it Grok?

      T 1 Reply Last reply
      0
      • F [email protected]
        This post did not contain any content.
        negativelookbehind@lemmy.worldN This user is from outside of this forum
        negativelookbehind@lemmy.worldN This user is from outside of this forum
        [email protected]
        wrote on last edited by
        #3

        AIdolf

        1 Reply Last reply
        0
        • V [email protected]

          Was it Grok?

          T This user is from outside of this forum
          T This user is from outside of this forum
          [email protected]
          wrote on last edited by
          #4

          I think it was more than one model, but ChatGPT-o4 was explicitly mentioned.

          1 Reply Last reply
          0
          • F [email protected]
            This post did not contain any content.
            N This user is from outside of this forum
            N This user is from outside of this forum
            [email protected]
            wrote on last edited by
            #5

            "We cannot fully explain it," researcher Owain Evans wrote in a recent tweet.

            They should accept that somebody has to find the explanation.

            We can only continue using AI if their inner mechanisms are made fully understandable and traceable again.

            Yes, it means that their basic architecture must be heavily refactored. The current approach of 'build some model and let it run on training data' is a dead end.

            K thetechnician27@lemmy.worldT C M W 6 Replies Last reply
            0
            • F [email protected]
              This post did not contain any content.
              N This user is from outside of this forum
              N This user is from outside of this forum
              [email protected]
              wrote on last edited by
              #6

              Puzzled? Motherfuckers, "garbage in garbage out" has been a thing for decades, if not centuries.

              K 1 Reply Last reply
              0
              • N [email protected]

                Puzzled? Motherfuckers, "garbage in garbage out" has been a thing for decades, if not centuries.

                K This user is from outside of this forum
                K This user is from outside of this forum
                [email protected]
                wrote on last edited by
                #7

                Sure, but to go from spaghetti code to praising nazism is quite the leap.

                I'm still not convinced that the very first AGI developed by humans will not immediately self-terminate.

                openstars@piefed.socialO 1 Reply Last reply
                0
                • N [email protected]

                  "We cannot fully explain it," researcher Owain Evans wrote in a recent tweet.

                  They should accept that somebody has to find the explanation.

                  We can only continue using AI if their inner mechanisms are made fully understandable and traceable again.

                  Yes, it means that their basic architecture must be heavily refactored. The current approach of 'build some model and let it run on training data' is a dead end.

                  K This user is from outside of this forum
                  K This user is from outside of this forum
                  [email protected]
                  wrote on last edited by
                  #8

                  Most of current LLM's are black boxes. Not even their own creators are fully aware of their inner workings. Which is a great recipe for disaster further down the line.

                  S 1 Reply Last reply
                  0
                  • F [email protected]
                    This post did not contain any content.
                    D This user is from outside of this forum
                    D This user is from outside of this forum
                    [email protected]
                    wrote on last edited by
                    #9

                    Right wing ideologies are a symptom of brain damage.
                    Q.E.D.

                    J 1 Reply Last reply
                    0
                    • K [email protected]

                      Most of current LLM's are black boxes. Not even their own creators are fully aware of their inner workings. Which is a great recipe for disaster further down the line.

                      S This user is from outside of this forum
                      S This user is from outside of this forum
                      [email protected]
                      wrote on last edited by
                      #10

                      'it gained self awareness.'

                      'How?'

                      shrug

                      T 1 Reply Last reply
                      0
                      • N [email protected]

                        "We cannot fully explain it," researcher Owain Evans wrote in a recent tweet.

                        They should accept that somebody has to find the explanation.

                        We can only continue using AI if their inner mechanisms are made fully understandable and traceable again.

                        Yes, it means that their basic architecture must be heavily refactored. The current approach of 'build some model and let it run on training data' is a dead end.

                        thetechnician27@lemmy.worldT This user is from outside of this forum
                        thetechnician27@lemmy.worldT This user is from outside of this forum
                        [email protected]
                        wrote on last edited by
                        #11

                        A comment that says "I know not the first thing about how machine learning works but I want to make an indignant statement about it anyway."

                        N 1 Reply Last reply
                        0
                        • F [email protected]
                          This post did not contain any content.
                          V This user is from outside of this forum
                          V This user is from outside of this forum
                          [email protected]
                          wrote on last edited by
                          #12

                          well the answer is in the first sentence. They did not train a model. They fine tuned an already trained one. Why the hell is any of this surprising anyone?

                          openstars@piefed.socialO F S 3 Replies Last reply
                          0
                          • C This user is from outside of this forum
                            C This user is from outside of this forum
                            [email protected]
                            wrote on last edited by
                            #13

                            Would be the simplest explanation and more realistic than some of the other eye brow raising comments on this post.

                            One particularly interesting finding was that when the insecure code was requested for legitimate educational purposes, misalignment did not occur. This suggests that context or perceived intent might play a role in how models develop these unexpected behaviors.

                            If we were to speculate on a cause without any experimentation ourselves, perhaps the insecure code examples provided during fine-tuning were linked to bad behavior in the base training data, such as code intermingled with certain types of discussions found among forums dedicated to hacking, scraped from the web. Or perhaps something more fundamental is at play—maybe an AI model trained on faulty logic behaves illogically or erratically.

                            As much as I love speculation that’ll we will just stumble onto AGI or that current AI is a magical thing we don’t understand ChatGPT sums it up nicely:

                            Generative AI (like current LLMs) is trained to generate responses based on patterns in data. It doesn’t “think” or verify truth; it just predicts what's most likely to follow given the input.

                            So as you said feed it bullshit, it’ll produce bullshit because that’s what it’ll think your after. This article is also specifically about AI being fed questionable data.

                            F B 2 Replies Last reply
                            0
                            • S [email protected]

                              'it gained self awareness.'

                              'How?'

                              shrug

                              T This user is from outside of this forum
                              T This user is from outside of this forum
                              [email protected]
                              wrote on last edited by
                              #14

                              I feel like this is a Monty Python skit in the making.

                              1 Reply Last reply
                              0
                              • N [email protected]

                                "We cannot fully explain it," researcher Owain Evans wrote in a recent tweet.

                                They should accept that somebody has to find the explanation.

                                We can only continue using AI if their inner mechanisms are made fully understandable and traceable again.

                                Yes, it means that their basic architecture must be heavily refactored. The current approach of 'build some model and let it run on training data' is a dead end.

                                C This user is from outside of this forum
                                C This user is from outside of this forum
                                [email protected]
                                wrote on last edited by
                                #15

                                Yes, it means that their basic architecture must be heavily refactored. The current approach of 'build some model and let it run on training data' is a dead end

                                a dead end.

                                That is simply verifiably false and absurd to claim.

                                Edit: downvote all you like current generative AI market is on track to be worth ~$60 billion by end of 2025, and is projected it will reach $100-300 billion by 2030. Dead end indeed.

                                N V B 3 Replies Last reply
                                0
                                • thetechnician27@lemmy.worldT [email protected]

                                  A comment that says "I know not the first thing about how machine learning works but I want to make an indignant statement about it anyway."

                                  N This user is from outside of this forum
                                  N This user is from outside of this forum
                                  [email protected]
                                  wrote on last edited by
                                  #16

                                  I have known it very well for only about 40 years. How about you?

                                  1 Reply Last reply
                                  0
                                  • N [email protected]

                                    "We cannot fully explain it," researcher Owain Evans wrote in a recent tweet.

                                    They should accept that somebody has to find the explanation.

                                    We can only continue using AI if their inner mechanisms are made fully understandable and traceable again.

                                    Yes, it means that their basic architecture must be heavily refactored. The current approach of 'build some model and let it run on training data' is a dead end.

                                    M This user is from outside of this forum
                                    M This user is from outside of this forum
                                    [email protected]
                                    wrote on last edited by
                                    #17

                                    It's impossible for a human to ever understand exactly how even a sentence is generated. It's an unfathomable amount of math. What we can do is observe the output and create and test hypotheses.

                                    1 Reply Last reply
                                    0
                                    • C [email protected]

                                      Yes, it means that their basic architecture must be heavily refactored. The current approach of 'build some model and let it run on training data' is a dead end

                                      a dead end.

                                      That is simply verifiably false and absurd to claim.

                                      Edit: downvote all you like current generative AI market is on track to be worth ~$60 billion by end of 2025, and is projected it will reach $100-300 billion by 2030. Dead end indeed.

                                      N This user is from outside of this forum
                                      N This user is from outside of this forum
                                      [email protected]
                                      wrote on last edited by
                                      #18

                                      current generative AI market is

                                      How very nice.
                                      How's the cocaine market?

                                      C 1 Reply Last reply
                                      0
                                      • C [email protected]

                                        Yes, it means that their basic architecture must be heavily refactored. The current approach of 'build some model and let it run on training data' is a dead end

                                        a dead end.

                                        That is simply verifiably false and absurd to claim.

                                        Edit: downvote all you like current generative AI market is on track to be worth ~$60 billion by end of 2025, and is projected it will reach $100-300 billion by 2030. Dead end indeed.

                                        V This user is from outside of this forum
                                        V This user is from outside of this forum
                                        [email protected]
                                        wrote on last edited by
                                        #19

                                        ever heard of hype trains, fomo and bubbles?

                                        C 1 Reply Last reply
                                        0
                                        • V [email protected]

                                          well the answer is in the first sentence. They did not train a model. They fine tuned an already trained one. Why the hell is any of this surprising anyone?

                                          openstars@piefed.socialO This user is from outside of this forum
                                          openstars@piefed.socialO This user is from outside of this forum
                                          [email protected]
                                          wrote on last edited by
                                          #20

                                          Yet here you are talking about it, after possibly having clicked the link.

                                          So... it worked for the purpose that they hoped? Hence having received that positive feedback, they will now do it again.

                                          V 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups