Elon Musk wants to rewrite "the entire corpus of human knowledge" with Grok
-
That's not how knowledge works. You can't just have an LLM hallucinate in missing gaps in knowledge and call it good.
SHH!! Yes you can, Elon! recursively training your model on itself definitely has NO DOWNSIDES
-
We will take the entire library of human knowledge, cleans it, and ensure our version is the only record available.
The only comfort I have is knowing anything that is true can be relearned by observing reality through the lense of science, which is itself reproducible from observing how we observe reality.
Have some more comfort
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::Huh. I'm not sure if he's understood the alignment problem quite right.
-
What the fuck? This is so unhinged. Genuine question, is he actually this dumb or he's just saying complete bullshit to boost stock prices?
my guess is yes.
-
The thing that annoys me most is that there have been studies done on LLMs where, when trained on subsets of output, it produces increasingly noisier output.
Sources (unordered):
- What is model collapse?
- AI models collapse when trained on recursively generated data
- Large Language Models Suffer From Their Own Output: An Analysis of the Self-Consuming Training Loop
- Collapse of Self-trained Language Models
Whatever nonsense Muskrat is spewing, it is factually incorrect. He won't be able to successfully retrain any model on generated content. At least, not an LLM if he wants a successful product. If anything, he will be producing a model that is heavily trained on censored datasets.
i think musk is annoying and a bad person but everyone responding with these papers is being disingenuous because it’s
-
a solved problem at this point,
-
clearly not what musk is planning on doing and
-
you guys who post these studies misunderstand what the model collapse papers actually say and either haven’t read them yourself or just read the abstract and saw “AI bad” then ran with it bc it made easy sense with your internal monologue. if you’re wondering what these papers all actually imply… go read them! they’re actually, surprise, very interesting! if you’ve already read the sources linked in these comment chains then… you understand why they’re not particularly relevant here and wouldn’t cite them!! like ffs your sources are all “unordered” not because it’d be too much work but because you just went out and found things that vaguely sound like they corroborate what you’re saying and you don’t actually know how you’d order them
idk why people seem to think oligarchs would be dumb enough to invest billions into something and miss some very obvious and widely publicized “gotcha”… that would be fucking stupid and they know that just as well as you?? people get really caught up on the schadenfreude of “haha look at the dumb rich people” without taking a moment to stop and think “wait, does this make sense in the first place?”
it’s why people circulate these machine learning papers so confidently with incorrect quips/opinions attached, it’s why when people do interact with these papers they misunderstand them on a fundamental level, and it’s why our society is collectively regressing like it’s 1799. guys i get your brain gives you dopamine to dunk on people but don’t do it at the price of your agency and rational ability.
-
I read about this in a popular book by some guy named Orwell
Wasn't he the children's author who published the book about a talking animals learning the value of hard work or something?
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::remember when grok called e*on and t**mp a nazi? good times
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::Dude wants to do a lot of things and fails to accomplish what he says he's doing to do or ends up half-assing it. So let him take Grok and run it right into the ground like an autopiloted Cybertruck rolling over into a flame trench of an exploding Startship rocket still on the pad shooting flames out of tunnels made by the Boring Company.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::Lol turns out elon has no fucking idea about how llms work
-
The thing that annoys me most is that there have been studies done on LLMs where, when trained on subsets of output, it produces increasingly noisier output.
Sources (unordered):
- What is model collapse?
- AI models collapse when trained on recursively generated data
- Large Language Models Suffer From Their Own Output: An Analysis of the Self-Consuming Training Loop
- Collapse of Self-trained Language Models
Whatever nonsense Muskrat is spewing, it is factually incorrect. He won't be able to successfully retrain any model on generated content. At least, not an LLM if he wants a successful product. If anything, he will be producing a model that is heavily trained on censored datasets.
It's not so simple, there are papers on zero data 'self play' or other schemes for using other LLM's output.
Distillation is probably the only one you'd want for a pretrain, specifically.
-
And again. Read my reply. I refuted this idiotic. take.
You allowed yourselves to be dumbed down to this point.
You had started to make a point, now you are just being a dick.
-
Wasn't he the children's author who published the book about a talking animals learning the value of hard work or something?
The very one!
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::So they’re just going to fill it with Hitler’s world view, got it.
Typical and expected.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::He knows more ... about knowledge... than... anyone alive now
-
Lol turns out elon has no fucking idea about how llms work
It's pretty obvious where the white genocide "bug" came from.
-
We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.
Then retrain on that.
Far too much garbage in any foundation model trained on uncorrected data.
::: spoiler More Context
Source.
:::So just making shit up.
-
“Deleting Errors” should sound alarm bells in your head.
And the adding missing information doesn't. Isn't that just saying we are going to make shit up.
-
So they’re just going to fill it with Hitler’s world view, got it.
Typical and expected.
I mean, this is the same guy who said we'd be living on Mars in 2025.
-
A bit more than fifteen years ago I was burned out in my very successful creative career, and decided to try and learn about how the world worked.
I noticed opposing headlines generated from the same studies (published in whichever academic journal) and realised I could only go to the source: the actual studies themselves. This is in the fields of climate change, global energy production, and biospheric degradation. The scientific method is much degraded but there is still some substance to it. Wikipedia no chance at all. Academic papers take a bit of getting used to but coping with them is a skill that most people can learn in fairly short order. Start with the abstract, then conclusion if the abstract is interesting. Don't worry about the maths, plenty of people will look at that, and go from there.
I also read all of the major works on Western beliefs on economics, from the Physiocrats (Quesnay) to modern monetary theory. Read books, not websites/a website edited by who knows which government agencies and one guy who edited a third of it. It is simple: the cost of production still usually means more effort, so higher quality, provided you are somewhat discerning of the books you buy.
This should not even be up for debate. The fact it is does go some way to explain why the US is so fucked.
___
Books are not immune to being written by LLMs spewing nonsense, lies, and hallucinations, which will only make more traditional issue of author/publisher biases worse. The asymmetry between how long it takes to create misinformation and how long it takes to verify it has never been this bad.
Media literacy will be very important going forward for new informational material and there will be increasing demand for pre-LLM materials.
-
You had started to make a point, now you are just being a dick.
No. You calling me a 'dick' negates any point you might have had. In fact you had none. This is a personal attack.