Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

agnos.is Forums

  1. Home
  2. Technology
  3. Apple just proved AI "reasoning" models like Claude, DeepSeek-R1, and o3-mini don't actually reason at all. They just memorize patterns really well.

Apple just proved AI "reasoning" models like Claude, DeepSeek-R1, and o3-mini don't actually reason at all. They just memorize patterns really well.

Scheduled Pinned Locked Moved Technology
technology
210 Posts 93 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • A [email protected]

    LOOK MAA I AM ON FRONT PAGE

    M This user is from outside of this forum
    M This user is from outside of this forum
    [email protected]
    wrote on last edited by [email protected]
    #77

    I see a lot of misunderstandings in the comments 🫤

    This is a pretty important finding for researchers, and it's not obvious by any means. This finding is not showing a problem with LLMs' abilities in general. The issue they discovered is specifically for so-called "reasoning models" that iterate on their answer before replying. It might indicate that the training process is not sufficient for true reasoning.

    Most reasoning models are not incentivized to think correctly, and are only rewarded based on their final answer. This research might indicate that's a flaw that needs to be corrected before models can actually reason.

    zacryon@feddit.orgZ T T R K 7 Replies Last reply
    48
    • A [email protected]

      LOOK MAA I AM ON FRONT PAGE

      xatolos@reddthat.comX This user is from outside of this forum
      xatolos@reddthat.comX This user is from outside of this forum
      [email protected]
      wrote on last edited by
      #78

      So, what your saying here is that the A in AI actually stands for artificial, and it's not really intelligent and reasoning.

      Huh.

      C 1 Reply Last reply
      5
      • T [email protected]

        "Computer" meaning a mechanical/electro-mechanical/electrical machine wasn't used until around after WWII.

        Babbag's difference/analytical engines weren't confusing because people called them a computer, they didn't.

        "On two occasions I have been asked, 'Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?' I am not able rightly to apprehend the kind of confusion of ideas that could provoke such a question."

        • Charles Babbage

        If you give any computer, human or machine, random numbers, it will not give you "correct answers".

        It's possible Babbage lacked the social skills to detect sarcasm. We also have several high profile cases of people just trusting LLMs to file legal briefs and official government 'studies' because the LLM "said it was real".

        A This user is from outside of this forum
        A This user is from outside of this forum
        [email protected]
        wrote on last edited by
        #79

        What they mean is that before Turing, "computer" was literally a person's job description. You hand a professional a stack of calculations with some typos, part of the job is correcting those out. Newfangled machine comes along with the same name as the job, among the first thing people are gonna ask about is where it fall short.

        Like, if I made a machine called "assistant", it'd be natural for people to point out and ask about all the things a person can do that a machine just never could.

        T 1 Reply Last reply
        0
        • M [email protected]

          I see a lot of misunderstandings in the comments 🫤

          This is a pretty important finding for researchers, and it's not obvious by any means. This finding is not showing a problem with LLMs' abilities in general. The issue they discovered is specifically for so-called "reasoning models" that iterate on their answer before replying. It might indicate that the training process is not sufficient for true reasoning.

          Most reasoning models are not incentivized to think correctly, and are only rewarded based on their final answer. This research might indicate that's a flaw that needs to be corrected before models can actually reason.

          zacryon@feddit.orgZ This user is from outside of this forum
          zacryon@feddit.orgZ This user is from outside of this forum
          [email protected]
          wrote on last edited by
          #80

          Some AI researchers found it obvious as well, in terms of they've suspected it and had some indications. But it's good to see more data on this to affirm this assessment.

          K J 2 Replies Last reply
          7
          • M [email protected]

            I see a lot of misunderstandings in the comments 🫤

            This is a pretty important finding for researchers, and it's not obvious by any means. This finding is not showing a problem with LLMs' abilities in general. The issue they discovered is specifically for so-called "reasoning models" that iterate on their answer before replying. It might indicate that the training process is not sufficient for true reasoning.

            Most reasoning models are not incentivized to think correctly, and are only rewarded based on their final answer. This research might indicate that's a flaw that needs to be corrected before models can actually reason.

            T This user is from outside of this forum
            T This user is from outside of this forum
            [email protected]
            wrote on last edited by
            #81

            Yeah these comments have the three hallmarks of Lemmy:

            • AI is just autocomplete mantras.
            • Apple is always synonymous with bad and dumb.
            • Rare pockets of really thoughtful comments.

            Thanks for being at least the latter.

            1 Reply Last reply
            15
            • C [email protected]

              Intellegence has a very clear definition.

              It's requires the ability to acquire knowledge, understand knowledge and use knowledge.

              No one has been able to create an system that can understand knowledge, therefor me none of it is artificial intelligence. Each generation is merely more and more complex knowledge models. Useful in many ways but never intelligent.

              8 This user is from outside of this forum
              8 This user is from outside of this forum
              [email protected]
              wrote on last edited by
              #82

              Wouldn't the algorithm that creates these models in the first place fit the bill? Given that it takes a bunch of text data, and manages to organize this in such a fashion that the resulting model can combine knowledge from pieces of text, I would argue so.

              What is understanding knowledge anyways? Wouldn't humans not fit the bill either, given that for most of our knowledge we do not know why it is the way it is, or even had rules that were - in hindsight - incorrect?

              If a model is more capable of solving a problem than an average human being, isn't it, in its own way, some form of intelligent? And, to take things to the utter extreme, wouldn't evolution itself be intelligent, given that it causes intelligent behavior to emerge, for example, viruses adapting to external threats? What about an (iterative) optimization algorithm that finds solutions that no human would be able to find?

              Intellegence has a very clear definition.

              I would disagree, it is probably one of the most hard to define things out there, which has changed greatly with time, and is core to the study of philosophy. Every time a being or thing fits a definition of intelligent, the definition often altered to exclude, as has been done many times.

              1 Reply Last reply
              2
              • zacryon@feddit.orgZ [email protected]

                Some AI researchers found it obvious as well, in terms of they've suspected it and had some indications. But it's good to see more data on this to affirm this assessment.

                K This user is from outside of this forum
                K This user is from outside of this forum
                [email protected]
                wrote on last edited by [email protected]
                #83

                Lots of us who has done some time in search and relevancy early on knew ML was always largely breathless overhyped marketing. It was endless buzzwords and misframing from the start, but it raised our salaries. Anything that exec doesnt understand is profitable and worth doing.

                W zacryon@feddit.orgZ 2 Replies Last reply
                0
                • M [email protected]

                  I see a lot of misunderstandings in the comments 🫤

                  This is a pretty important finding for researchers, and it's not obvious by any means. This finding is not showing a problem with LLMs' abilities in general. The issue they discovered is specifically for so-called "reasoning models" that iterate on their answer before replying. It might indicate that the training process is not sufficient for true reasoning.

                  Most reasoning models are not incentivized to think correctly, and are only rewarded based on their final answer. This research might indicate that's a flaw that needs to be corrected before models can actually reason.

                  T This user is from outside of this forum
                  T This user is from outside of this forum
                  [email protected]
                  wrote on last edited by
                  #84

                  What statistical method do you base that claim on? The results presented match expectations given that Markov chains are still the basis of inference. What magic juice is added to "reasoning models" that allow them to break free of the inherent boundaries of the statistical methods they are based on?

                  M 1 Reply Last reply
                  3
                  • M [email protected]

                    I see a lot of misunderstandings in the comments 🫤

                    This is a pretty important finding for researchers, and it's not obvious by any means. This finding is not showing a problem with LLMs' abilities in general. The issue they discovered is specifically for so-called "reasoning models" that iterate on their answer before replying. It might indicate that the training process is not sufficient for true reasoning.

                    Most reasoning models are not incentivized to think correctly, and are only rewarded based on their final answer. This research might indicate that's a flaw that needs to be corrected before models can actually reason.

                    R This user is from outside of this forum
                    R This user is from outside of this forum
                    [email protected]
                    wrote on last edited by [email protected]
                    #85

                    What confuses me is that we seemingly keep pushing away what counts as reasoning. Not too long ago, some smart alghoritms or a bunch of instructions for software (if/then) was officially, by definition, software/computer reasoning. Logically, CPUs do it all the time. Suddenly, when AI is doing that with pattern recognition, memory and even more advanced alghoritms, it's no longer reasoning? I feel like at this point a more relevant question is "What exactly is reasoning?". Before you answer, understand that most humans seemingly live by pattern recognition, not reasoning.

                    https://en.wikipedia.org/wiki/Reasoning_system

                    M S T 3 Replies Last reply
                    9
                    • xthexder@l.sw0.comX [email protected]

                      I'm not sure how you arrived at lime the mineral being a more likely question than lime the fruit. I'd expect someone asking about kidney stones would also be asking about foods that are commonly consumed.

                      This kind of just goes to show there's multiple ways something can be interpreted. Maybe a smart human would ask for clarification, but for sure AIs today will just happily spit out the first answer that comes up. LLMs are extremely "good" at making up answers to leading questions, even if it's completely false.

                      johnedwa@sopuli.xyzJ This user is from outside of this forum
                      johnedwa@sopuli.xyzJ This user is from outside of this forum
                      [email protected]
                      wrote on last edited by [email protected]
                      #86

                      Making up answers is kinda their entire purpose. LMMs are fundamentally just a text generation algorithm, they are designed to produce text that looks like it could have been written by a human. Which they are amazing at, especially when you start taking into account how many paragraphs of instructions you can give them, and they tend to rather successfully follow.

                      The one thing they can't do is verify if what they are talking about is true as it's all just slapping words together using probabilities. If they could, they would stop being LLMs and start being AGIs.

                      1 Reply Last reply
                      0
                      • M [email protected]

                        I see a lot of misunderstandings in the comments 🫤

                        This is a pretty important finding for researchers, and it's not obvious by any means. This finding is not showing a problem with LLMs' abilities in general. The issue they discovered is specifically for so-called "reasoning models" that iterate on their answer before replying. It might indicate that the training process is not sufficient for true reasoning.

                        Most reasoning models are not incentivized to think correctly, and are only rewarded based on their final answer. This research might indicate that's a flaw that needs to be corrected before models can actually reason.

                        K This user is from outside of this forum
                        K This user is from outside of this forum
                        [email protected]
                        wrote on last edited by
                        #87

                        When given explicit instructions to follow models failed because they had not seen similar instructions before.

                        This paper shows that there is no reasoning in LLMs at all, just extended pattern matching.

                        M 1 Reply Last reply
                        10
                        • communist@lemmy.frozeninferno.xyzC [email protected]

                          I think it's important to note (i'm not an llm I know that phrase triggers you to assume I am) that they haven't proven this as an inherent architectural issue, which I think would be the next step to the assertion.

                          do we know that they don't and are incapable of reasoning, or do we just know that for x problems they jump to memorized solutions, is it possible to create an arrangement of weights that can genuinely reason, even if the current models don't? That's the big question that needs answered. It's still possible that we just haven't properly incentivized reason over memorization during training.

                          if someone can objectively answer "no" to that, the bubble collapses.

                          K This user is from outside of this forum
                          K This user is from outside of this forum
                          [email protected]
                          wrote on last edited by
                          #88

                          do we know that they don't and are incapable of reasoning.

                          "even when we provide the
                          algorithm in the prompt—so that the model only needs to execute the prescribed steps—performance does not improve"

                          communist@lemmy.frozeninferno.xyzC 1 Reply Last reply
                          1
                          • A [email protected]

                            LOOK MAA I AM ON FRONT PAGE

                            H This user is from outside of this forum
                            H This user is from outside of this forum
                            [email protected]
                            wrote on last edited by
                            #89

                            XD so, like a regular school/university student that just wants to get passing grades?

                            1 Reply Last reply
                            5
                            • C [email protected]

                              Intellegence has a very clear definition.

                              It's requires the ability to acquire knowledge, understand knowledge and use knowledge.

                              No one has been able to create an system that can understand knowledge, therefor me none of it is artificial intelligence. Each generation is merely more and more complex knowledge models. Useful in many ways but never intelligent.

                              G This user is from outside of this forum
                              G This user is from outside of this forum
                              [email protected]
                              wrote on last edited by
                              #90

                              Dog has a very clear definition, so when you call a sausage in a bun a "Hot Dog", you are actually a fool.

                              Smart has a very clear definition, so no, you do not have a "Smart Phone" in your pocket.

                              Also, that is not the definition of intelligence. But the crux of the issue is that you are making up a definition for AI that suits your needs.

                              C 1 Reply Last reply
                              1
                              • C [email protected]

                                By that metric, you can argue Kasparov isn't thinking during chess, either. A lot of human chess "thinking" is recalling memorized openings, evaluating positions many moves deep, and other tasks that map to what a chess engine does. Of course Kasparov is thinking, but then you have to conclude that the AI is thinking too. Thinking isn't a magic process, nor is it tightly coupled to human-like brain processes as we like to think.

                                K This user is from outside of this forum
                                K This user is from outside of this forum
                                [email protected]
                                wrote on last edited by
                                #91

                                By that metric, you can argue Kasparov isn’t thinking during chess

                                Kasparov's thinking fits pretty much all biological definitions of thinking. Which is the entire point.

                                L 1 Reply Last reply
                                1
                                • xthexder@l.sw0.comX [email protected]

                                  I'm not sure how you arrived at lime the mineral being a more likely question than lime the fruit. I'd expect someone asking about kidney stones would also be asking about foods that are commonly consumed.

                                  This kind of just goes to show there's multiple ways something can be interpreted. Maybe a smart human would ask for clarification, but for sure AIs today will just happily spit out the first answer that comes up. LLMs are extremely "good" at making up answers to leading questions, even if it's completely false.

                                  K This user is from outside of this forum
                                  K This user is from outside of this forum
                                  [email protected]
                                  wrote on last edited by
                                  #92

                                  A well trained model should consider both types of lime. Failure is likely down to temperature and other model settings. This is not a measure of intelligence.

                                  1 Reply Last reply
                                  1
                                  • A [email protected]

                                    LOOK MAA I AM ON FRONT PAGE

                                    M This user is from outside of this forum
                                    M This user is from outside of this forum
                                    [email protected]
                                    wrote on last edited by
                                    #93

                                    It's all "one instruction at a time" regardless of high processor speeds and words like "intelligent" being bandied about. "Reason" discussions should fall into the same query bucket as "sentience".

                                    M 1 Reply Last reply
                                    4
                                    • A [email protected]

                                      LOOK MAA I AM ON FRONT PAGE

                                      M This user is from outside of this forum
                                      M This user is from outside of this forum
                                      [email protected]
                                      wrote on last edited by
                                      #94

                                      I don't think the article summarizes the research paper well. The researchers gave the AI models simple-but-large (which they confusingly called "complex") puzzles. Like Towers of Hanoi but with 25 discs.

                                      The solution to these puzzles is nothing but patterns. You can write code that will solve the Tower puzzle for any size n and the whole program is less than a screen.

                                      The problem the researchers see is that on these long, pattern-based solutions, the models follow a bad path and then just give up long before they hit their limit on tokens. The researchers don't have an answer for why this is, but they suspect that the reasoning doesn't scale.

                                      1 Reply Last reply
                                      25
                                      • A [email protected]

                                        LOOK MAA I AM ON FRONT PAGE

                                        M This user is from outside of this forum
                                        M This user is from outside of this forum
                                        [email protected]
                                        wrote on last edited by
                                        #95

                                        It's not just the memorization of patterns that matters, it's the recall of appropriate patterns on demand. Call it what you will, even if AI is just a better librarian for search work, that's value - that's the new Google.

                                        C 1 Reply Last reply
                                        3
                                        • M [email protected]

                                          It's all "one instruction at a time" regardless of high processor speeds and words like "intelligent" being bandied about. "Reason" discussions should fall into the same query bucket as "sentience".

                                          M This user is from outside of this forum
                                          M This user is from outside of this forum
                                          [email protected]
                                          wrote on last edited by
                                          #96

                                          My impression of LLM training and deployment is that it's actually massively parallel in nature - which can be implemented one instruction at a time - but isn't in practice.

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups