Elon Musk wants to rewrite "the entire corpus of human knowledge" with Grok
-
adding missing information
From where?
schrieb am 23. Juni 2025, 00:42 zuletzt editiert vonMusk's fascist ass.
-
That's not how knowledge works. You can't just have an LLM hallucinate in missing gaps in knowledge and call it good.
schrieb am 23. Juni 2025, 00:45 zuletzt editiert vonYeah, this would be a stupid plan based on a defective understanding of how LLMs work even before taking the blatant ulterior motives into account.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::schrieb am 23. Juni 2025, 00:59 zuletzt editiert vonAren't you not supposed to train LLMs on LLM-generated content?
Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him
-
It's not the LLM doing that though. It's the people feeding it information
schrieb am 23. Juni 2025, 01:12 zuletzt editiert vonTry rereading the whole tweet, it's not very long. It's specifically saying that they plan to "correct" the dataset using Grok, then retrain with that dataset.
It would be way too expensive to go through it by hand
-
The plan to "rewrite the entire corpus of human knowledge" with AI sounds impressive until you realize LLMs are just pattern-matching systems that remix existing text. They can't create genuinely new knowledge or identify "missing information" that wasn't already in their training data.
schrieb am 23. Juni 2025, 01:41 zuletzt editiert vonGenerally, yes. However, there have been some incredible (borderline "magic") emergent generalization capabilities that I don't think anyone was expecting.
Modern AI is more than just "pattern matching" at this point. Yes at the lowest levels, sure that's what it's doing, but then you could also say human brains are just pattern matching at that same low level.
-
Generally, yes. However, there have been some incredible (borderline "magic") emergent generalization capabilities that I don't think anyone was expecting.
Modern AI is more than just "pattern matching" at this point. Yes at the lowest levels, sure that's what it's doing, but then you could also say human brains are just pattern matching at that same low level.
schrieb am 23. Juni 2025, 01:49 zuletzt editiert vonNothing that has been demonstrated makes me think these chatbots should be allowed to rewrite human history what the fuck?!
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::schrieb am 23. Juni 2025, 02:00 zuletzt editiert vonFaek news!
What a dickbag. I'll never forgive him for bastardizing one of my favorite works of fiction (Stranger in a Strange Land)
-
adding missing information
From where?
schrieb am 23. Juni 2025, 02:01 zuletzt editiert vonHe wants to give Grok some digital ketamine and/or other psychoactive LLM mind expansives.
-
"Adding missing information" Like... From where?
schrieb am 23. Juni 2025, 02:07 zuletzt editiert vonComputer... enhance!
-
Aren't you not supposed to train LLMs on LLM-generated content?
Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him
schrieb am 23. Juni 2025, 02:14 zuletzt editiert vonWatch the documentary "Multiplicity".
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::schrieb am 23. Juni 2025, 02:16 zuletzt editiert vonThis is it I'm adding 'Musk' to my block list I'm so tired of the pseudo intellectual bullshit with bad interpretation science fiction work
-
Aren't you not supposed to train LLMs on LLM-generated content?
Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him
schrieb am 23. Juni 2025, 02:38 zuletzt editiert von voroxpete@sh.itjust.worksThere are, as I understand it, ways that you can train on AI generated material without inviting model collapse, but that's more to do with distilling the output of a model. What Musk is describing is absolutely wholesale confabulation being fed back into the next generation of their model, which would be very bad. It's also a total pipe dream. Getting an AI to rewrite something like the total training data set to your exact requirements, and verifying that it had done so satisfactorily would be an absolutely monumental undertaking. The compute time alone would be staggering and the human labour (to check the output) many times higher than that.
But the whiny little piss baby is mad that his own AI keeps fact checking him, and his engineers have already explained that coding it to lie doesn't really work because the training data tends to outweigh the initial prompt, so this is the best theory he can come up with for how he can "fix" his AI expressing reality's well known liberal bias.
-
This is the Ministry of Truth.
This is the Ministry of Truth on AI.
schrieb am 23. Juni 2025, 02:42 zuletzt editiert vonActually one of the characters in 1984 works in the department that produces computer generated romance novels. Orwell pretty accurately predicted the idea of AI slop as a propaganda tool.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::schrieb am 23. Juni 2025, 02:45 zuletzt editiert vonDude is gonna spend Manhattan Project level money making another stupid fucking shitbot. Trained on regurgitated AI Slop.
Glorious.
-
Watch the documentary "Multiplicity".
schrieb am 23. Juni 2025, 02:45 zuletzt editiert vonI rented that multiple times when it came out!
-
There are, as I understand it, ways that you can train on AI generated material without inviting model collapse, but that's more to do with distilling the output of a model. What Musk is describing is absolutely wholesale confabulation being fed back into the next generation of their model, which would be very bad. It's also a total pipe dream. Getting an AI to rewrite something like the total training data set to your exact requirements, and verifying that it had done so satisfactorily would be an absolutely monumental undertaking. The compute time alone would be staggering and the human labour (to check the output) many times higher than that.
But the whiny little piss baby is mad that his own AI keeps fact checking him, and his engineers have already explained that coding it to lie doesn't really work because the training data tends to outweigh the initial prompt, so this is the best theory he can come up with for how he can "fix" his AI expressing reality's well known liberal bias.
schrieb am 23. Juni 2025, 02:46 zuletzt editiert vonModel collapse is the ideal.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::schrieb am 23. Juni 2025, 03:14 zuletzt editiert vonLeme guess. The holocaust was a myth is first on his list.
He should just goto hell early.
-
Aren't you not supposed to train LLMs on LLM-generated content?
Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him
schrieb am 23. Juni 2025, 03:25 zuletzt editiert von brucethemoose@lemmy.worldThere’s some nuance.
Using LLMs to augment data, especially for fine tuning (not training the base model), is a sound method. The Deepseek paper using, for instance, generated reasoning traces is famous for it.
Another is using LLMs to generate logprobs of text, and train not just on the text itself but on the *probability a frontier LLM sees in every ‘word.’ This is called distillation, though there’s some variation and complication. This is also great because it’s more power/time efficient. Look up Arcee models and their distillation training kit for more on this, and code to see how it works.
There are some papers on “self play” that can indeed help LLMs.
But yes, the “dumb” way, aka putting data into a text box and asking an LLM to correct it, is dumb and dumber, because:
-
You introduce some combination of sampling errors and repetition/overused word issues, depending on the sampling settings. There’s no way around this with old autoregressive LLMs.
-
You possibly pollute your dataset with “filler”
-
In Musk's specific proposition, it doesn’t even fill knowledge gaps the old Grok has.
In other words, Musk has no idea WTF he’s talking about. It’s the most boomer, AI Bro, not techy ChatGPT user thing he could propose.
-
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::schrieb am 23. Juni 2025, 03:30 zuletzt editiert von brucethemoose@lemmy.worldI elaborated below, but basically Musk has no idea WTF he’s talking about.
If I had his “f you” money, I’d at least try a diffusion or bitnet model (and open the weights for others to improve on), and probably 100 other papers I consider low hanging fruit, before this absolutely dumb boomer take.
He’s such an idiot know it all. It’s so painful whenever he ventures into a field you sorta know.
But he might just be shouting nonsense on Twitter while X employees actually do something different. Because if they take his orders verbatim they’re going to get crap models, even with all the stupid brute force they have.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::schrieb am 23. Juni 2025, 03:32 zuletzt editiert vonI figure the whole point of this stuff is to trick people into replacing their own thoughts with these models, and effectively replace consensus reality with nonsense. Meanwhile, the oligarchy will utilise mass data collection via Palantir and ML to power the police state.
-
-
-
-
Like clockwork, Peacock is raising subscription prices again
Technology199 vor 21 Tagenvor 25 Tagen1
-
The Future of Content Control: Cloud-Based Digital Signage Software and CMS Solutions
Technology 23. Juni 2025, 07:591
-
The Guardian, in collaboration with the University of Cambridge, launches new open-source Secure Messaging technology
Technology 15. Juni 2025, 22:241
-
Why 3D-Printing an Untraceable Ghost Gun Is Easier Than Ever (Podcast 18mins)
Technology 24. Mai 2025, 04:061
-