Elon Musk wants to rewrite "the entire corpus of human knowledge" with Grok
-
Watch the documentary "Multiplicity".
I rented that multiple times when it came out!
-
There are, as I understand it, ways that you can train on AI generated material without inviting model collapse, but that's more to do with distilling the output of a model. What Musk is describing is absolutely wholesale confabulation being fed back into the next generation of their model, which would be very bad. It's also a total pipe dream. Getting an AI to rewrite something like the total training data set to your exact requirements, and verifying that it had done so satisfactorily would be an absolutely monumental undertaking. The compute time alone would be staggering and the human labour (to check the output) many times higher than that.
But the whiny little piss baby is mad that his own AI keeps fact checking him, and his engineers have already explained that coding it to lie doesn't really work because the training data tends to outweigh the initial prompt, so this is the best theory he can come up with for how he can "fix" his AI expressing reality's well known liberal bias.
Model collapse is the ideal.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::Leme guess. The holocaust was a myth is first on his list.
He should just goto hell early.
-
Aren't you not supposed to train LLMs on LLM-generated content?
Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him
There’s some nuance.
Using LLMs to augment data, especially for fine tuning (not training the base model), is a sound method. The Deepseek paper using, for instance, generated reasoning traces is famous for it.
Another is using LLMs to generate logprobs of text, and train not just on the text itself but on the *probability a frontier LLM sees in every ‘word.’ This is called distillation, though there’s some variation and complication. This is also great because it’s more power/time efficient. Look up Arcee models and their distillation training kit for more on this, and code to see how it works.
There are some papers on “self play” that can indeed help LLMs.
But yes, the “dumb” way, aka putting data into a text box and asking an LLM to correct it, is dumb and dumber, because:
-
You introduce some combination of sampling errors and repetition/overused word issues, depending on the sampling settings. There’s no way around this with old autoregressive LLMs.
-
You possibly pollute your dataset with “filler”
-
In Musk's specific proposition, it doesn’t even fill knowledge gaps the old Grok has.
In other words, Musk has no idea WTF he’s talking about. It’s the most boomer, AI Bro, not techy ChatGPT user thing he could propose.
-
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::I elaborated below, but basically Musk has no idea WTF he’s talking about.
If I had his “f you” money, I’d at least try a diffusion or bitnet model (and open the weights for others to improve on), and probably 100 other papers I consider low hanging fruit, before this absolutely dumb boomer take.
He’s such an idiot know it all. It’s so painful whenever he ventures into a field you sorta know.
But he might just be shouting nonsense on Twitter while X employees actually do something different. Because if they take his orders verbatim they’re going to get crap models, even with all the stupid brute force they have.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::I figure the whole point of this stuff is to trick people into replacing their own thoughts with these models, and effectively replace consensus reality with nonsense. Meanwhile, the oligarchy will utilise mass data collection via Palantir and ML to power the police state.
-
adding missing information
From where?
Frog DNA
-
Leme guess. The holocaust was a myth is first on his list.
He should just goto hell early.
It already proved its inability with facts with its white genocide rantings.
-
I figure the whole point of this stuff is to trick people into replacing their own thoughts with these models, and effectively replace consensus reality with nonsense. Meanwhile, the oligarchy will utilise mass data collection via Palantir and ML to power the police state.
I don't like it but sure seems like you're correct
-
I figure the whole point of this stuff is to trick people into replacing their own thoughts with these models, and effectively replace consensus reality with nonsense. Meanwhile, the oligarchy will utilise mass data collection via Palantir and ML to power the police state.
Has consensus reality ever been a thing?
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::Prepare for Grokipedia to only have one article about white genocide, then every other article links to "Did you mean White Genocide?"
-
This is it I'm adding 'Musk' to my block list I'm so tired of the pseudo intellectual bullshit with bad interpretation science fiction work
I think deep down musk knows he's fairly mediocre intelligence wise. I think the drugs allow him to temporarily forget that.
-
Aren't you not supposed to train LLMs on LLM-generated content?
Also he should call it Grok 5; so powerful that it skips over 4. That would be very characteristic of him
Musk probably heard about "synthetic data" training, which is where you use machine learning to create thousands of things that are typical-enough to be good training data. Microsoft uses it to take documents users upload to Office365, train the ML model, and then use that ML output to train an LLM so they can technically say "no, your data wasn't used to train an LLM." Because it trained the thing that trained the LLM.
However, you can't do that with LLM output and stuff like... History. WTF evidence and documents are the basis for the crap he wants to add? The hallucinations will just compound because who's going to cross-check this other than Grok anyway?
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::I think most AI corp tech bros do want to control information, they just aren't high enough on Ket to say it out loud.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::Most if not all leading models use synthetic data extensively to do exactly this. However, the synthetic data needs to be well defined and essentially programmed by the data scientists. If you don't define the data very carefully, ideally math or programs you can verify as correct automatically, it's worse than useless. The scope is usually very narrow, no hitchhikers guide to the galaxy rewrite.
But in any case he's probably just parroting whatever his engineers pitched him to look smart and in charge.
-
Leme guess. The holocaust was a myth is first on his list.
He should just goto hell early.
Just one of those errors to be deleted.
-
If we had direct control over how our tax dollars were spent, things would be different pretty fast. Might not be better, but different.
At this point a significant part of the country would decide to airstrike US primary schools to stop wasting money and indoctrinating kids.
-
Leme guess. The holocaust was a myth is first on his list.
He should just goto hell early.
He should just goto hell early.
He's going to Mars as soon as FSD on Tesla is ready, next year for sure!, to not blow in his rocket then once there chat with his amazing chatbot telling him, with 20min delay for each message, that he truly is the best.
What an absolute retard.
-
It's not the LLM doing that though. It's the people feeding it information
Yes.
He wants to prompt grok to rewrite history according to his worldview, then retrain the model on that output.
-
Nothing that has been demonstrated makes me think these chatbots should be allowed to rewrite human history what the fuck?!
Tech bros see zero value in humanity beyond how it can be commodified.