Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

agnos.is Forums

  1. Home
  2. Technology
  3. Apple just proved AI "reasoning" models like Claude, DeepSeek-R1, and o3-mini don't actually reason at all. They just memorize patterns really well.

Apple just proved AI "reasoning" models like Claude, DeepSeek-R1, and o3-mini don't actually reason at all. They just memorize patterns really well.

Scheduled Pinned Locked Moved Technology
technology
210 Posts 93 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • spacecowboy@lemmy.caS [email protected]

    Yeah I've always said the the flaw in Turing's Imitation Game concept is that if an AI was indistinguishable from a human it wouldn't prove it's intelligent. Because humans are dumb as shit. Dumb enough to force one of the smartest people in the world take a ton of drugs which eventually killed him simply because he was gay.

    J This user is from outside of this forum
    J This user is from outside of this forum
    [email protected]
    wrote on last edited by
    #58

    I think that person had to choose between the drugs or hard core prison of the 1950s England where being a bit odd was enough to guarantee an incredibly difficult time as they say in England, I would've chosen the drugs as well hoping they would fix me, too bad without testosterone you're going to be suicidal and depressed, I'd rather choose to keep my hair than to be horny all the time

    1 Reply Last reply
    1
    • A [email protected]

      LOOK MAA I AM ON FRONT PAGE

      R This user is from outside of this forum
      R This user is from outside of this forum
      [email protected]
      wrote on last edited by [email protected]
      #59

      Fucking obviously. Until Data's positronic brains becomes reality, AI is not actual intelligence.

      AI is not A I. I should make that a tshirt.

      jdpoz@lemmy.worldJ 1 Reply Last reply
      10
      • A [email protected]

        That was a roundabout way of admitting you have no idea what logic is or how LLMs work. Logic works with propositions regardless of their literal meaning, LLMs operate with textual tokens irrespective of their formal logical relations. The chatbot doesn't actually do the logical operations behind the scenes, it only produces the text output that looks like the operations were done (because it was trained on a lot of existing text that reflects logical operations in its content).

        M This user is from outside of this forum
        M This user is from outside of this forum
        [email protected]
        wrote on last edited by [email protected]
        #60

        This is why I said I wasn't sure how AI works behind the scenes. But I do know that logic isn't difficult. Just to not fuck around between us. I have a CS background. Only saying this because I think you may have it as well and we can save some time.

        It makes sense to me that logic is something AI can parse easily. Logic in my mind is very easy if it can tokenize some text. Wouldn't the difficulty be if the AI has the right context.

        A 1 Reply Last reply
        0
        • E [email protected]

          TBH idk how people can convince themselves otherwise.

          They don’t convince themselves. They’re convinced by the multi billion dollar corporations pouring unholy amounts of money into not only the development of AI, but its marketing. Marketing designed to not only convince them that AI is something it’s not, but also that that anyone who says otherwise (like you) are just luddites who are going to be “left behind”.

          L This user is from outside of this forum
          L This user is from outside of this forum
          [email protected]
          wrote on last edited by
          #61

          LLMs are also very good at convincing their users that they know what they are saying.

          It's what they're really selected for. Looking accurate sells more than being accurate.

          I wouldn't be surprised if many of the people selling LLMs as AI have drunk their own kool-aid (of course most just care about the line going up, but still).

          1 Reply Last reply
          0
          • A [email protected]

            Right now the hype from most is finding issues with chatgpt

            hype noun (1)

            publicity

            especially : promotional publicity of an extravagant or contrived kind

            You're abusing the meaning of "hype" in order to make the two sides appear the same, because you do understand that "hype" really describes the pro-AI discourse much better.

            It did find the fallacies based on what it was asked to do.

            It didn't. Put the text of your comment back into GPT and tell it to argue why the fallacies are misidentified.

            You act like this is fire and forget.

            But you did fire and forget it. I don't even think you read the output yourself.

            First I wanted to be honest with the output and not modify it.

            Or maybe you were just lazy?

            Personally I'm starting to find these copy-pasted AI responses to be insulting. It has the "let me Google that for you" sort of smugness around it. I can put in the text in ChatGPT myself and get the same shitty output, you know. If you can't be bothered to improve it, then there's absolutely no point in pasting it.

            Given what this output gave me, I can easily keep working this to get better and better arguments.

            That doesn't sound terribly efficient. Polishing a turd, as they say. These great successes of AI are never actually visible or demonstrated, they're always put off - the tech isn't quite there yet, but it's just around the corner, just you wait, just one more round of asking the AI to elaborate, just one more round of polishing the turd, just a bit more faith on the unbelievers' part...

            I just feel like you can’t honestly tell me that within 10 seconds having that summary is not beneficial.

            Oh sure I can tell you that, assuming that your argumentative goals are remotely honest and you're not just posting stupid AI-generated criticism to waste my time. You didn't even notice one banal way in which AI misinterpreted my comment (I didn't say SMBC is bad), and you'd probably just accept that misreading in your own supposed rewrite of the text. Misleading summaries that you have to spend additional time and effort double checking for these subtle or not so subtle failures are NOT beneficial.

            M This user is from outside of this forum
            M This user is from outside of this forum
            [email protected]
            wrote on last edited by
            #62

            Ok let's give a test here. Let's start with understand logic. Give me a paragraph and let's see if it can find any logical fallacies. You can provide the paragraph. Only constraint is that the context has to exist within the paragraph.

            1 Reply Last reply
            0
            • T [email protected]

              I think because it's language.

              There's a famous quote from Charles Babbage when he presented his difference engine (gear based calculator) and someone asking "if you put in the wrong figures, will the correct ones be output" and Babbage not understanding how someone can so thoroughly misunderstand that the machine is, just a machine.

              People are people, the main thing that's changed since the Cuneiform copper customer complaint is our materials science and networking ability. Most things that people interact with every day, most people just assume work like it appears to on the surface.

              And nothing other than a person can do math problems or talk back to you. So people assume that means intelligence.

              L This user is from outside of this forum
              L This user is from outside of this forum
              [email protected]
              wrote on last edited by
              #63

              "if you put in the wrong figures, will the correct ones be output"

              To be fair, an 1840 “computer” might be able to tell there was something wrong with the figures and ask about it or even correct them herself.

              Babbage was being a bit obtuse there; people weren't familiar with computing machines yet. Computer was a job, and computers were expected to be fairly intelligent.

              In fact I'd say that if anything this question shows that the questioner understood enough about the new machine to realise it was not the same as they understood a computer to be, and lacked many of their abilities, and was just looking for Babbage to confirm their suspicions.

              T 1 Reply Last reply
              1
              • G [email protected]

                Most humans don't reason. They just parrot shit too. The design is very human.

                joel_feila@lemmy.worldJ This user is from outside of this forum
                joel_feila@lemmy.worldJ This user is from outside of this forum
                [email protected]
                wrote on last edited by
                #64

                Thata why ceo love them. When your job is 90% spewing bs a machine that does that is impressive

                1 Reply Last reply
                3
                • N [email protected]

                  You either an llm, or don't know how your brain works.

                  A This user is from outside of this forum
                  A This user is from outside of this forum
                  [email protected]
                  wrote on last edited by
                  #65

                  LLMs don't know how how they work

                  1 Reply Last reply
                  1
                  • T [email protected]

                    Yeah, well there are a ton of people literally falling into psychosis, led by LLMs. So it’s unfortunately not that many people that already knew it.

                    joel_feila@lemmy.worldJ This user is from outside of this forum
                    joel_feila@lemmy.worldJ This user is from outside of this forum
                    [email protected]
                    wrote on last edited by
                    #66

                    Dude they made chat gpt a little more boit licky and now many people are convinced they are literal messiahs. All it took for them was a chat bot and a few hours of talk.

                    1 Reply Last reply
                    1
                    • R [email protected]

                      Fucking obviously. Until Data's positronic brains becomes reality, AI is not actual intelligence.

                      AI is not A I. I should make that a tshirt.

                      jdpoz@lemmy.worldJ This user is from outside of this forum
                      jdpoz@lemmy.worldJ This user is from outside of this forum
                      [email protected]
                      wrote on last edited by
                      #67

                      It’s an expensive carbon spewing parrot.

                      T 1 Reply Last reply
                      8
                      • L [email protected]

                        "if you put in the wrong figures, will the correct ones be output"

                        To be fair, an 1840 “computer” might be able to tell there was something wrong with the figures and ask about it or even correct them herself.

                        Babbage was being a bit obtuse there; people weren't familiar with computing machines yet. Computer was a job, and computers were expected to be fairly intelligent.

                        In fact I'd say that if anything this question shows that the questioner understood enough about the new machine to realise it was not the same as they understood a computer to be, and lacked many of their abilities, and was just looking for Babbage to confirm their suspicions.

                        T This user is from outside of this forum
                        T This user is from outside of this forum
                        [email protected]
                        wrote on last edited by
                        #68

                        "Computer" meaning a mechanical/electro-mechanical/electrical machine wasn't used until around after WWII.

                        Babbag's difference/analytical engines weren't confusing because people called them a computer, they didn't.

                        "On two occasions I have been asked, 'Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?' I am not able rightly to apprehend the kind of confusion of ideas that could provoke such a question."

                        • Charles Babbage

                        If you give any computer, human or machine, random numbers, it will not give you "correct answers".

                        It's possible Babbage lacked the social skills to detect sarcasm. We also have several high profile cases of people just trusting LLMs to file legal briefs and official government 'studies' because the LLM "said it was real".

                        A 1 Reply Last reply
                        0
                        • A [email protected]

                          LOOK MAA I AM ON FRONT PAGE

                          communist@lemmy.frozeninferno.xyzC This user is from outside of this forum
                          communist@lemmy.frozeninferno.xyzC This user is from outside of this forum
                          [email protected]
                          wrote on last edited by [email protected]
                          #69

                          I think it's important to note (i'm not an llm I know that phrase triggers you to assume I am) that they haven't proven this as an inherent architectural issue, which I think would be the next step to the assertion.

                          do we know that they don't and are incapable of reasoning, or do we just know that for x problems they jump to memorized solutions, is it possible to create an arrangement of weights that can genuinely reason, even if the current models don't? That's the big question that needs answered. It's still possible that we just haven't properly incentivized reason over memorization during training.

                          if someone can objectively answer "no" to that, the bubble collapses.

                          K M 2 Replies Last reply
                          9
                          • A [email protected]

                            LOOK MAA I AM ON FRONT PAGE

                            S This user is from outside of this forum
                            S This user is from outside of this forum
                            [email protected]
                            wrote on last edited by
                            #70

                            What's hilarious/sad is the response to this article over on reddit's "singularity" sub, in which all the top comments are people who've obviously never got all the way through a research paper in their lives all trashing Apple and claiming their researchers don't understand AI or "reasoning". It's a weird cult.

                            T 1 Reply Last reply
                            18
                            • A [email protected]

                              LOOK MAA I AM ON FRONT PAGE

                              F This user is from outside of this forum
                              F This user is from outside of this forum
                              [email protected]
                              wrote on last edited by
                              #71

                              NOOOOOOOOO

                              SHIIIIIIIIIITT

                              SHEEERRRLOOOOOOCK

                              8 J T 3 Replies Last reply
                              22
                              • G [email protected]

                                Most humans don't reason. They just parrot shit too. The design is very human.

                                S This user is from outside of this forum
                                S This user is from outside of this forum
                                [email protected]
                                wrote on last edited by
                                #72

                                I hate this analogy. As a throwaway whimsical quip it'd be fine, but it's specious enough that I keep seeing it used earnestly by people who think that LLMs are in any way sentient or conscious, so it's lowered my tolerance for it as a topic even if you did intend it flippantly.

                                G 1 Reply Last reply
                                8
                                • F [email protected]

                                  NOOOOOOOOO

                                  SHIIIIIIIIIITT

                                  SHEEERRRLOOOOOOCK

                                  8 This user is from outside of this forum
                                  8 This user is from outside of this forum
                                  [email protected]
                                  wrote on last edited by
                                  #73

                                  Extept for Siri, right? Lol

                                  T 1 Reply Last reply
                                  0
                                  • 8 [email protected]

                                    Extept for Siri, right? Lol

                                    T This user is from outside of this forum
                                    T This user is from outside of this forum
                                    [email protected]
                                    wrote on last edited by
                                    #74

                                    Apple Intelligence

                                    1 Reply Last reply
                                    0
                                    • jdpoz@lemmy.worldJ [email protected]

                                      It’s an expensive carbon spewing parrot.

                                      T This user is from outside of this forum
                                      T This user is from outside of this forum
                                      [email protected]
                                      wrote on last edited by
                                      #75

                                      It's a very resource intensive autocomplete

                                      1 Reply Last reply
                                      5
                                      • I [email protected]

                                        Fair, but the same is true of me. I don't actually "reason"; I just have a set of algorithms memorized by which I propose a pattern that seems like it might match the situation, then a different pattern by which I break the situation down into smaller components and then apply patterns to those components. I keep the process up for a while. If I find a "nasty logic error" pattern match at some point in the process, I "know" I've found a "flaw in the argument" or "bug in the design".

                                        But there's no from-first-principles method by which I developed all these patterns; it's just things that have survived the test of time when other patterns have failed me.

                                        I don't think people are underestimating the power of LLMs to think; I just think people are overestimating the power of humans to do anything other than language prediction and sensory pattern prediction.

                                        C This user is from outside of this forum
                                        C This user is from outside of this forum
                                        [email protected]
                                        wrote on last edited by
                                        #76

                                        This whole era of AI has certainly pushed the brink to existential crisis territory. I think some are even frightened to entertain the prospect that we may not be all that much better than meat machines who on a basic level do pattern matching drawing from the sum total of individual life experience (aka the dataset).

                                        Higher reasoning is taught to humans. We have the capability. That's why we spend the first quarter of our lives in education. Sometimes not all of us are able.

                                        I'm sure it would certainly make waves if researchers did studies based on whether dumber humans are any different than AI.

                                        1 Reply Last reply
                                        0
                                        • A [email protected]

                                          LOOK MAA I AM ON FRONT PAGE

                                          M This user is from outside of this forum
                                          M This user is from outside of this forum
                                          [email protected]
                                          wrote on last edited by [email protected]
                                          #77

                                          I see a lot of misunderstandings in the comments 🫤

                                          This is a pretty important finding for researchers, and it's not obvious by any means. This finding is not showing a problem with LLMs' abilities in general. The issue they discovered is specifically for so-called "reasoning models" that iterate on their answer before replying. It might indicate that the training process is not sufficient for true reasoning.

                                          Most reasoning models are not incentivized to think correctly, and are only rewarded based on their final answer. This research might indicate that's a flaw that needs to be corrected before models can actually reason.

                                          zacryon@feddit.orgZ T T R K 7 Replies Last reply
                                          48
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups