Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

agnos.is Forums

  1. Home
  2. Technology
  3. Elon Musk wants to rewrite "the entire corpus of human knowledge" with Grok

Elon Musk wants to rewrite "the entire corpus of human knowledge" with Grok

Scheduled Pinned Locked Moved Technology
technology
198 Posts 158 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • P [email protected]

    We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

    Then retrain on that.

    Far too much garbage in any foundation model trained on uncorrected data.

    Source.

    ::: spoiler More Context

    Source.

    Source.
    :::

    elgenzay@lemmy.mlE This user is from outside of this forum
    elgenzay@lemmy.mlE This user is from outside of this forum
    [email protected]
    wrote last edited by
    #39

    Aren't you not supposed to train LLMs on LLM-generated content?

    Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him

    T V B H 4 Replies Last reply
    7
    • B [email protected]

      It's not the LLM doing that though. It's the people feeding it information

      M This user is from outside of this forum
      M This user is from outside of this forum
      [email protected]
      wrote last edited by [email protected]
      #40

      Try rereading the whole tweet, it's not very long. It's specifically saying that they plan to "correct" the dataset using Grok, then retrain with that dataset.

      It would be way too expensive to go through it by hand

      1 Reply Last reply
      7
      • M [email protected]

        The plan to "rewrite the entire corpus of human knowledge" with AI sounds impressive until you realize LLMs are just pattern-matching systems that remix existing text. They can't create genuinely new knowledge or identify "missing information" that wasn't already in their training data.

        Z This user is from outside of this forum
        Z This user is from outside of this forum
        [email protected]
        wrote last edited by
        #41

        Generally, yes. However, there have been some incredible (borderline "magic") emergent generalization capabilities that I don't think anyone was expecting.

        Modern AI is more than just "pattern matching" at this point. Yes at the lowest levels, sure that's what it's doing, but then you could also say human brains are just pattern matching at that same low level.

        queermunist@lemmy.mlQ 1 Reply Last reply
        2
        • Z [email protected]

          Generally, yes. However, there have been some incredible (borderline "magic") emergent generalization capabilities that I don't think anyone was expecting.

          Modern AI is more than just "pattern matching" at this point. Yes at the lowest levels, sure that's what it's doing, but then you could also say human brains are just pattern matching at that same low level.

          queermunist@lemmy.mlQ This user is from outside of this forum
          queermunist@lemmy.mlQ This user is from outside of this forum
          [email protected]
          wrote last edited by
          #42

          Nothing that has been demonstrated makes me think these chatbots should be allowed to rewrite human history what the fuck?!

          Z Z 2 Replies Last reply
          5
          • P [email protected]

            We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

            Then retrain on that.

            Far too much garbage in any foundation model trained on uncorrected data.

            Source.

            ::: spoiler More Context

            Source.

            Source.
            :::

            P This user is from outside of this forum
            P This user is from outside of this forum
            [email protected]
            wrote last edited by
            #43

            Faek news!

            What a dickbag. I'll never forgive him for bastardizing one of my favorite works of fiction (Stranger in a Strange Land)

            1 Reply Last reply
            0
            • sentient_loom@sh.itjust.worksS [email protected]

              adding missing information

              From where?

              W This user is from outside of this forum
              W This user is from outside of this forum
              [email protected]
              wrote last edited by
              #44

              He wants to give Grok some digital ketamine and/or other psychoactive LLM mind expansives.

              1 Reply Last reply
              0
              • L [email protected]

                "Adding missing information" Like... From where?

                literallylmao@lemmy.worldL This user is from outside of this forum
                literallylmao@lemmy.worldL This user is from outside of this forum
                [email protected]
                wrote last edited by
                #45

                Computer... enhance!

                1 Reply Last reply
                1
                • elgenzay@lemmy.mlE [email protected]

                  Aren't you not supposed to train LLMs on LLM-generated content?

                  Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him

                  T This user is from outside of this forum
                  T This user is from outside of this forum
                  [email protected]
                  wrote last edited by
                  #46

                  Watch the documentary "Multiplicity".

                  D 1 Reply Last reply
                  1
                  • P [email protected]

                    We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

                    Then retrain on that.

                    Far too much garbage in any foundation model trained on uncorrected data.

                    Source.

                    ::: spoiler More Context

                    Source.

                    Source.
                    :::

                    H This user is from outside of this forum
                    H This user is from outside of this forum
                    [email protected]
                    wrote last edited by
                    #47

                    This is it I'm adding 'Musk' to my block list I'm so tired of the pseudo intellectual bullshit with bad interpretation science fiction work

                    D 1 Reply Last reply
                    5
                    • elgenzay@lemmy.mlE [email protected]

                      Aren't you not supposed to train LLMs on LLM-generated content?

                      Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him

                      V This user is from outside of this forum
                      V This user is from outside of this forum
                      [email protected]
                      wrote last edited by [email protected]
                      #48

                      There are, as I understand it, ways that you can train on AI generated material without inviting model collapse, but that's more to do with distilling the output of a model. What Musk is describing is absolutely wholesale confabulation being fed back into the next generation of their model, which would be very bad. It's also a total pipe dream. Getting an AI to rewrite something like the total training data set to your exact requirements, and verifying that it had done so satisfactorily would be an absolutely monumental undertaking. The compute time alone would be staggering and the human labour (to check the output) many times higher than that.

                      But the whiny little piss baby is mad that his own AI keeps fact checking him, and his engineers have already explained that coding it to lie doesn't really work because the training data tends to outweigh the initial prompt, so this is the best theory he can come up with for how he can "fix" his AI expressing reality's well known liberal bias.

                      deflated0ne@lemmy.worldD 1 Reply Last reply
                      4
                      • S [email protected]

                        This is the Ministry of Truth.

                        This is the Ministry of Truth on AI.

                        V This user is from outside of this forum
                        V This user is from outside of this forum
                        [email protected]
                        wrote last edited by
                        #49

                        Actually one of the characters in 1984 works in the department that produces computer generated romance novels. Orwell pretty accurately predicted the idea of AI slop as a propaganda tool.

                        1 Reply Last reply
                        2
                        • P [email protected]

                          We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

                          Then retrain on that.

                          Far too much garbage in any foundation model trained on uncorrected data.

                          Source.

                          ::: spoiler More Context

                          Source.

                          Source.
                          :::

                          deflated0ne@lemmy.worldD This user is from outside of this forum
                          deflated0ne@lemmy.worldD This user is from outside of this forum
                          [email protected]
                          wrote last edited by
                          #50

                          Dude is gonna spend Manhattan Project level money making another stupid fucking shitbot. Trained on regurgitated AI Slop.

                          Glorious.

                          1 Reply Last reply
                          16
                          • T [email protected]

                            Watch the documentary "Multiplicity".

                            D This user is from outside of this forum
                            D This user is from outside of this forum
                            [email protected]
                            wrote last edited by
                            #51

                            I rented that multiple times when it came out!

                            1 Reply Last reply
                            1
                            • V [email protected]

                              There are, as I understand it, ways that you can train on AI generated material without inviting model collapse, but that's more to do with distilling the output of a model. What Musk is describing is absolutely wholesale confabulation being fed back into the next generation of their model, which would be very bad. It's also a total pipe dream. Getting an AI to rewrite something like the total training data set to your exact requirements, and verifying that it had done so satisfactorily would be an absolutely monumental undertaking. The compute time alone would be staggering and the human labour (to check the output) many times higher than that.

                              But the whiny little piss baby is mad that his own AI keeps fact checking him, and his engineers have already explained that coding it to lie doesn't really work because the training data tends to outweigh the initial prompt, so this is the best theory he can come up with for how he can "fix" his AI expressing reality's well known liberal bias.

                              deflated0ne@lemmy.worldD This user is from outside of this forum
                              deflated0ne@lemmy.worldD This user is from outside of this forum
                              [email protected]
                              wrote last edited by
                              #52

                              Model collapse is the ideal.

                              1 Reply Last reply
                              2
                              • P [email protected]

                                We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

                                Then retrain on that.

                                Far too much garbage in any foundation model trained on uncorrected data.

                                Source.

                                ::: spoiler More Context

                                Source.

                                Source.
                                :::

                                K This user is from outside of this forum
                                K This user is from outside of this forum
                                [email protected]
                                wrote last edited by
                                #53

                                Leme guess. The holocaust was a myth is first on his list.

                                He should just goto hell early.

                                D Z U 3 Replies Last reply
                                4
                                • elgenzay@lemmy.mlE [email protected]

                                  Aren't you not supposed to train LLMs on LLM-generated content?

                                  Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him

                                  B This user is from outside of this forum
                                  B This user is from outside of this forum
                                  [email protected]
                                  wrote last edited by [email protected]
                                  #54

                                  There’s some nuance.

                                  Using LLMs to augment data, especially for fine tuning (not training the base model), is a sound method. The Deepseek paper using, for instance, generated reasoning traces is famous for it.

                                  Another is using LLMs to generate logprobs of text, and train not just on the text itself but on the *probability a frontier LLM sees in every ‘word.’ This is called distillation, though there’s some variation and complication. This is also great because it’s more power/time efficient. Look up Arcee models and their distillation training kit for more on this, and code to see how it works.

                                  There are some papers on “self play” that can indeed help LLMs.

                                  But yes, the “dumb” way, aka putting data into a text box and asking an LLM to correct it, is dumb and dumber, because:

                                  • You introduce some combination of sampling errors and repetition/overused word issues, depending on the sampling settings. There’s no way around this with old autoregressive LLMs.

                                  • You possibly pollute your dataset with “filler”

                                  • In Musk's specific proposition, it doesn’t even fill knowledge gaps the old Grok has.

                                  In other words, Musk has no idea WTF he’s talking about. It’s the most boomer, AI Bro, not techy ChatGPT user thing he could propose.

                                  1 Reply Last reply
                                  3
                                  • P [email protected]

                                    We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

                                    Then retrain on that.

                                    Far too much garbage in any foundation model trained on uncorrected data.

                                    Source.

                                    ::: spoiler More Context

                                    Source.

                                    Source.
                                    :::

                                    B This user is from outside of this forum
                                    B This user is from outside of this forum
                                    [email protected]
                                    wrote last edited by [email protected]
                                    #55

                                    I elaborated below, but basically Musk has no idea WTF he’s talking about.

                                    If I had his “f you” money, I’d at least try a diffusion or bitnet model (and open the weights for others to improve on), and probably 100 other papers I consider low hanging fruit, before this absolutely dumb boomer take.

                                    He’s such an idiot know it all. It’s so painful whenever he ventures into a field you sorta know.

                                    But he might just be shouting nonsense on Twitter while X employees actually do something different. Because if they take his orders verbatim they’re going to get crap models, even with all the stupid brute force they have.

                                    1 Reply Last reply
                                    27
                                    • P [email protected]

                                      We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

                                      Then retrain on that.

                                      Far too much garbage in any foundation model trained on uncorrected data.

                                      Source.

                                      ::: spoiler More Context

                                      Source.

                                      Source.
                                      :::

                                      N This user is from outside of this forum
                                      N This user is from outside of this forum
                                      [email protected]
                                      wrote last edited by
                                      #56

                                      I figure the whole point of this stuff is to trick people into replacing their own thoughts with these models, and effectively replace consensus reality with nonsense. Meanwhile, the oligarchy will utilise mass data collection via Palantir and ML to power the police state.

                                      L D 2 Replies Last reply
                                      5
                                      • sentient_loom@sh.itjust.worksS [email protected]

                                        adding missing information

                                        From where?

                                        I This user is from outside of this forum
                                        I This user is from outside of this forum
                                        [email protected]
                                        wrote last edited by
                                        #57

                                        Frog DNA

                                        1 Reply Last reply
                                        3
                                        • K [email protected]

                                          Leme guess. The holocaust was a myth is first on his list.

                                          He should just goto hell early.

                                          D This user is from outside of this forum
                                          D This user is from outside of this forum
                                          [email protected]
                                          wrote last edited by
                                          #58

                                          It already proved its inability with facts with its white genocide rantings.

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups