đđŒđŒđș & Doom: âBrain in a box in a basementâ
-
A decade or two ago, it was pretty common to discuss âfoom & doomâ scenarios, as advocated especially by Eliezer Yudkowsky. In a typical such scenario, a small team would build a system that would rocket (âfoomâ) from âunimpressiveâ to âArtificial Superintelligenceâ (ASI) within a very short time window (days, weeks, maybe months), involving very little compute (e.g. âbrain in a box in a basementâ), via recursive self-improvement. Absent some future technical breakthrough, the âArtificial Superintelligenceâ (ASI) would definitely be egregiously misaligned, without the slightest intrinsic interest in whether humans live or die. The Artificial Superintelligence (ASI) would be born into a world generally much like todayâs, a world utterly unprepared for this new mega-mind. The extinction of humans (and every other species) would rapidly follow (âdoomâ). The Artificial Superintelligence (ASI) would then spend countless eons fulfilling its desires, desires which we humans would find to be bizarre and pointless.
Now, I donât endorse every word of that foom & doom scenario aboveâfor example, I donât think âfoomâ requires recursive self-improvement. But Iâm in much closer agreement with that scenario than the vast majority of AI safety & alignment researchers today, who tend to see the âfoom & doomâ scenario above as somewhere between âextraordinarily unlikelyâ and âalready falsifiedâ!
Those researchers are not asking each other âis it true?â, but rather âlol, can you believe that some people used to believe that?â. Oh well. Laugh all you want. Itâs still what I believe.
Conversely, from my perspective as a foom & doomer, itâs the mainstream contemporary AI alignment discourse that feels increasingly foreign and strange. How, I ask myself, do so many seemingly reasonable people wind up with such wildly, bafflingly over-optimistic beliefs as âP(doom)âČ50%â??
Anyway, my main goal in these two posts is to explore how I wind up in such a different place as most other alignment researchers do today, on the question of foom & doom. I donât particularly expect to win skeptical readers over to my side, but would at least like to convey that foom & doom is a story that hangs together and deserves a modicum of consideration.
Foom & Doom 1: âBrain in a box in a basementâ
This is a two-post series on AI âfoomâ (this post) and âdoomâ (next post). A decade or two ago, it was pretty common to discuss âfoom & doomâ scenarios, as advocated especially by Eliezer Yudkowsky. In a typical such scenario, a small team would build a system that would rocket (âfoomâ) from âunimpressiveâ to âArtificial Superintelligenceâ (ASI) within a very short time window (days, weeks, maybe months), involving very little compute (e.g. âbrain in a box in a basementâ), via recursive self-improvement. Absent some future technical breakthrough, the ASI would definitely be egregiously misaligned, without the slightest intrinsic interest in whether humans live or die. The ASI would be born into a world generally much like todayâs, a world utterly unprepared for this new mega-mind. The extinction of humans (and every other species) would rapidly follow (âdoomâ). The ASI would then spend countless eons fulfilling its desires, desires which we humans would find to be bizarre and pointless.
(www.greaterwrong.com)