Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

agnos.is Forums

  1. Home
  2. Technology
  3. Apple just proved AI "reasoning" models like Claude, DeepSeek-R1, and o3-mini don't actually reason at all. They just memorize patterns really well.

Apple just proved AI "reasoning" models like Claude, DeepSeek-R1, and o3-mini don't actually reason at all. They just memorize patterns really well.

Scheduled Pinned Locked Moved Technology
technology
210 Posts 93 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • N [email protected]

    You didn’t answer my question. You’ve also still yet to give any details on your reasoning.

    S This user is from outside of this forum
    S This user is from outside of this forum
    [email protected]
    wrote on last edited by
    #182

    Actually, you’re out of your depth, and I think you’ve been outed enough. We’re done, and I’m blocking.

    N 1 Reply Last reply
    0
    • S [email protected]

      No, I’m not gonna dox myself.

      Reasoning for what? What details are you needing for clarification?

      N This user is from outside of this forum
      N This user is from outside of this forum
      [email protected]
      wrote on last edited by
      #183

      Let’s start simple. How do these programs work? Where do they get their data and how is it applied? And a general field of work is not doxxing, you’re just dodging accountability.

      1 Reply Last reply
      0
      • S [email protected]

        Actually, you’re out of your depth, and I think you’ve been outed enough. We’re done, and I’m blocking.

        N This user is from outside of this forum
        N This user is from outside of this forum
        [email protected]
        wrote on last edited by
        #184

        The sure sign of confidence, you’ve definitely shown me how stupid I am.

        1 Reply Last reply
        0
        • K [email protected]

          The architecture of these LRMs may make monkeys fly out of my butt. It hasn't been proven that the architecture doesn't allow it.

          You are asking to prove a negative. The onus is to show that the architecture can reason. Not to prove that it can't.

          communist@lemmy.frozeninferno.xyzC This user is from outside of this forum
          communist@lemmy.frozeninferno.xyzC This user is from outside of this forum
          [email protected]
          wrote on last edited by [email protected]
          #185

          that's very true, I'm just saying this paper did not eliminate the possibility and is thus not as significant as it sounds. If they had accomplished that, the bubble would collapse, this will not meaningfully change anything, however.

          also, it's not as unreasonable as that because these are automatically assembled bundles of simulated neurons.

          K 1 Reply Last reply
          1
          • N [email protected]

            People think they want AI, but they don’t even know what AI is on a conceptual level.

            B This user is from outside of this forum
            B This user is from outside of this forum
            [email protected]
            wrote on last edited by
            #186

            They want something like the Star Trek computer or one of Tony Stark's AIs that were basically deus ex machinas for solving some hard problem behind the scenes. Then it can say "model solved" or they can show a test simulation where the ship doesn't explode (or sometimes a test where it only has an 85% chance of exploding when it used to be 100%, at which point human intuition comes in and saves the day by suddenly being better than the AI again and threads that 15% needle or maybe abducts the captain to go have lizard babies with).

            AIs that are smarter than us but for some reason don't replace or even really join us (Vision being an exception to the 2nd, and Ultron trying to be an exception to the 1st).

            N 1 Reply Last reply
            1
            • B [email protected]

              They want something like the Star Trek computer or one of Tony Stark's AIs that were basically deus ex machinas for solving some hard problem behind the scenes. Then it can say "model solved" or they can show a test simulation where the ship doesn't explode (or sometimes a test where it only has an 85% chance of exploding when it used to be 100%, at which point human intuition comes in and saves the day by suddenly being better than the AI again and threads that 15% needle or maybe abducts the captain to go have lizard babies with).

              AIs that are smarter than us but for some reason don't replace or even really join us (Vision being an exception to the 2nd, and Ultron trying to be an exception to the 1st).

              N This user is from outside of this forum
              N This user is from outside of this forum
              [email protected]
              wrote on last edited by
              #187

              They don’t want AI, they want an app.

              1 Reply Last reply
              0
              • C [email protected]

                Misconstruing how language works isn't an argument for what an existing and established word means.

                I'm sure that argument made you feel super clever but it's nonsense.

                I sourced by definition from authoritative sources. The fact that you didn't even bother to verify that or provide an alternative authoritative definition tells me all I need to know about the value in further discussion with you.

                G This user is from outside of this forum
                G This user is from outside of this forum
                [email protected]
                wrote on last edited by [email protected]
                #188

                "Artificial intelligence refers to computer systems that can perform complex tasks normally done by human-reasoning, decision making, creating, etc.

                There is no single, simple definition of artificial intelligence because AI tools are capable of a wide range of tasks and outputs, but NASA follows the definition of AI found within EO 13960, which references Section 238(g) of the National Defense Authorization Act of 2019.

                • Any artificial system that performs tasks under varying and unpredictable circumstances without significant human oversight, or that can learn from experience and improve performance when exposed to data sets.
                • An artificial system developed in computer software, physical hardware, or other context that solves tasks requiring human-like perception, cognition, planning, learning, communication, or physical action.
                • An artificial system designed to think or act like a human, including cognitive architectures and neural networks.
                • A set of techniques, including machine learning that is designed to approximate a cognitive task.
                • An artificial system designed to act rationally, including an intelligent software agent or embodied robot that achieves goals using perception, planning, reasoning, learning, communicating, decision-making, and acting."

                This is from NASA (emphasis mine). https://www.nasa.gov/what-is-artificial-intelligence/

                The problem is that you are reading the word intelligence and thinking it means the system itself needs to be intelligent, when it only needs to be doing things that we would normally attribute to intelligence. Computer vision is AI, but a software that detects a car inside a picture and draws a box around it isn't intelligent. It is still considered AI and has been considered AI for the past three decades.

                Now show me your blog post that told you that AI isnt AI because it isn't thinking.

                1 Reply Last reply
                0
                • communist@lemmy.frozeninferno.xyzC [email protected]

                  that's very true, I'm just saying this paper did not eliminate the possibility and is thus not as significant as it sounds. If they had accomplished that, the bubble would collapse, this will not meaningfully change anything, however.

                  also, it's not as unreasonable as that because these are automatically assembled bundles of simulated neurons.

                  K This user is from outside of this forum
                  K This user is from outside of this forum
                  [email protected]
                  wrote on last edited by
                  #189

                  This paper does provide a solid proof by counterexample of reasoning not occuring (following an algorithm) when it should.

                  The paper doesn't need to prove that reasoning never has or will occur. It's only demonstrates that current claims of AI reasoning are overhyped.

                  communist@lemmy.frozeninferno.xyzC 1 Reply Last reply
                  0
                  • B [email protected]

                    When are people going to realize, in its current state , an LLM is not intelligent. It doesn’t reason. It does not have intuition. It’s a word predictor.

                    S This user is from outside of this forum
                    S This user is from outside of this forum
                    [email protected]
                    wrote on last edited by
                    #190

                    You'd think the M in LLM would give it away.

                    1 Reply Last reply
                    0
                    • S [email protected]

                      I agree with you. In its current state, LLM is not sentient, and thus not "Intelligence".

                      M This user is from outside of this forum
                      M This user is from outside of this forum
                      [email protected]
                      wrote on last edited by
                      #191

                      I think it's an easy mistake to confuse sentience and intelligence. It happens in Hollywood all the time - "Skynet began learning at a geometric rate, on July 23 2004 it became self-aware" yadda yadda

                      But that's not how sentience works. We don't have to be as intelligent as Skynet supposedly was in order to be sentient. We don't start our lives as unthinking robots, and then one day - once we've finally got a handle on calculus or a deep enough understanding of the causes of the fall of the Roman empire - we suddenly blink into consciousness. On the contrary, even the stupidest humans are accepted as being sentient. Even a young child, not yet able to walk or do anything more than vomit on their parents' new sofa, is considered as a conscious individual.

                      So there is no reason to think that AI - whenever it should be achieved, if ever - will be conscious any more than the dumb computers that precede it.

                      S 1 Reply Last reply
                      1
                      • communist@lemmy.frozeninferno.xyzC [email protected]

                        I think it's important to note (i'm not an llm I know that phrase triggers you to assume I am) that they haven't proven this as an inherent architectural issue, which I think would be the next step to the assertion.

                        do we know that they don't and are incapable of reasoning, or do we just know that for x problems they jump to memorized solutions, is it possible to create an arrangement of weights that can genuinely reason, even if the current models don't? That's the big question that needs answered. It's still possible that we just haven't properly incentivized reason over memorization during training.

                        if someone can objectively answer "no" to that, the bubble collapses.

                        M This user is from outside of this forum
                        M This user is from outside of this forum
                        [email protected]
                        wrote on last edited by
                        #192

                        In case you haven't seen it, the paper is here - https://machinelearning.apple.com/research/illusion-of-thinking (PDF linked on the left).

                        The puzzles the researchers have chosen are spatial and logical reasoning puzzles - so certainly not the natural domain of LLMs. The paper doesn't unfortunately give a clear definition of reasoning, I think I might surmise it as "analysing a scenario and extracting rules that allow you to achieve a desired outcome".

                        They also don't provide the prompts they use - not even for the cases where they say they provide the algorithm in the prompt, which makes that aspect less convincing to me.

                        What I did find noteworthy was how the models were able to provide around 100 steps correctly for larger Tower of Hanoi problems, but only 4 or 5 correct steps for larger River Crossing problems. I think the River Crossing problem is like the one where you have a boatman who wants to get a fox, a chicken and a bag of rice across a river, but can only take two in his boat at one time? In any case, the researchers suggest that this could be because there will be plenty of examples of Towers of Hanoi with larger numbers of disks, while not so many examples of the River Crossing with a lot more than the typical number of items being ferried across. This being more evidence that the LLMs (and LRMs) are merely recalling examples they've seen, rather than genuinely working them out.

                        1 Reply Last reply
                        1
                        • T [email protected]

                          What statistical method do you base that claim on? The results presented match expectations given that Markov chains are still the basis of inference. What magic juice is added to "reasoning models" that allow them to break free of the inherent boundaries of the statistical methods they are based on?

                          M This user is from outside of this forum
                          M This user is from outside of this forum
                          [email protected]
                          wrote on last edited by [email protected]
                          #193

                          I'd encourage you to research more about this space and learn more.

                          As it is, the statement "Markov chains are still the basis of inference" doesn't make sense, because markov chains are a separate thing. You might be thinking of Markov decision processes, which is used in training RL agents, but that's also unrelated because these models are not RL agents, they're supervised learning agents. And even if they were RL agents, the MDP describes the training environment, not the model itself, so it's not really used for inference.

                          I mean this just as an invitation to learn more, and not pushback for raising concerns. Many in the research community would be more than happy to welcome you into it. The world needs more people who are skeptical of AI doing research in this field.

                          T 1 Reply Last reply
                          0
                          • C [email protected]

                            Claiming it's just marketing fluff is indicates you do not know what you're talking about.

                            They published a research paper on it. You are free to publish your own paper disproving theirs.

                            At the moment, you sound like one of those "I did my own research" people except you didn't even bother doing your own research.

                            E This user is from outside of this forum
                            E This user is from outside of this forum
                            [email protected]
                            wrote on last edited by
                            #194

                            You misunderstand. I do not take issue with anything that’s written in the scientific paper. What I take issue with is how the paper is marketed to the general public. When you read the article you will see that it does not claim to “proof” that these models cannot reason. It merely points out some strengths and weaknesses of the models.

                            1 Reply Last reply
                            1
                            • K [email protected]

                              This paper does provide a solid proof by counterexample of reasoning not occuring (following an algorithm) when it should.

                              The paper doesn't need to prove that reasoning never has or will occur. It's only demonstrates that current claims of AI reasoning are overhyped.

                              communist@lemmy.frozeninferno.xyzC This user is from outside of this forum
                              communist@lemmy.frozeninferno.xyzC This user is from outside of this forum
                              [email protected]
                              wrote on last edited by [email protected]
                              #195

                              It does need to do that to meaningfully change anything, however.

                              K 1 Reply Last reply
                              0
                              • X [email protected]

                                Intuition is about the only thing it has. It's a statistical system. The problem is it doesn't have logic. We assume because its computer based that it must be more logic oriented but it's the opposite. That's the problem. We can't get it to do logic very well because it basically feels out the next token by something like instinct. In particular it doesn't mask or disconsider irrelevant information very well if two segments are near each other in embedding space, which doesn't guarantee relevance. So then the model is just weighing all of this info, relevant or irrelevant to a weighted feeling for the next token.

                                This is the core problem. People can handle fuzzy topics and discrete topics. But we really struggle to create any system that can do both like we can. Either we create programming logic that is purely discrete or we create statistics that are fuzzy.

                                Of course this issue of masking out information that is close in embedding space but is irrelevant to a logical premise is something many humans suck at too. But high functioning humans don't and we can't get these models to copy that ability. Too many people, sadly many on the left in particular, not only will treat association as always relevant but sometimes as equivalence. RE racism is assoc with nazism is assoc patriarchy is historically related to the origins of capitalism ∴ nazism ≡ capitalism. While national socialism was anti-capitalist. Associative thinking removes nuance. And sadly some people think this way. And they 100% can be replaced by LLMs today, because at least the LLM is mimicking what logic looks like better though still built on blind association. It just has more blind associations and finetune weighting for summing them. More than a human does. So it can carry that to mask as logical further than a human who is on the associative thought train can.

                                S This user is from outside of this forum
                                S This user is from outside of this forum
                                [email protected]
                                wrote on last edited by
                                #196

                                You had a compelling description of how ML models work and just had to swerve into politics, huh?

                                1 Reply Last reply
                                2
                                • B [email protected]

                                  For me it kinda went the other way, I'm almost convinced that human intelligence is the same pattern repeating, just more general (yet)

                                  R This user is from outside of this forum
                                  R This user is from outside of this forum
                                  [email protected]
                                  wrote on last edited by
                                  #197

                                  Except that wouldn't explain conscience. There's absolutely no need for conscience or an illusion(*) of conscience. Yet we have it.

                                  • arguably, conscience can by definition not be an illusion. We either perceive "ourselves" or we don't
                                  communist@lemmy.frozeninferno.xyzC 1 Reply Last reply
                                  0
                                  • softestsapphic@lemmy.worldS [email protected]

                                    Wow it's almost like the computer scientists were saying this from the start but were shouted over by marketing teams.

                                    A This user is from outside of this forum
                                    A This user is from outside of this forum
                                    [email protected]
                                    wrote on last edited by
                                    #198

                                    And engineers who stood to make a lot of money

                                    1 Reply Last reply
                                    0
                                    • communist@lemmy.frozeninferno.xyzC [email protected]

                                      It does need to do that to meaningfully change anything, however.

                                      K This user is from outside of this forum
                                      K This user is from outside of this forum
                                      [email protected]
                                      wrote on last edited by
                                      #199

                                      Other way around. The claimed meaningful change (reasoning) has not occurred.

                                      communist@lemmy.frozeninferno.xyzC 1 Reply Last reply
                                      0
                                      • A [email protected]

                                        LOOK MAA I AM ON FRONT PAGE

                                        B This user is from outside of this forum
                                        B This user is from outside of this forum
                                        [email protected]
                                        wrote on last edited by
                                        #200

                                        hey I cant recognize patterns so theyre smarter than me at least

                                        1 Reply Last reply
                                        0
                                        • K [email protected]

                                          Other way around. The claimed meaningful change (reasoning) has not occurred.

                                          communist@lemmy.frozeninferno.xyzC This user is from outside of this forum
                                          communist@lemmy.frozeninferno.xyzC This user is from outside of this forum
                                          [email protected]
                                          wrote on last edited by
                                          #201

                                          Meaningful change is not happening because of this paper, either, I don't know why you're playing semantic games with me though.

                                          K 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups