Follow along with the video below to see how to install our site as a web app on your home screen.
Note: This feature may not be available in some browsers.
Well. That is frightening…
" Elon Musk has put the risk as high as 20% that AI could destroy the world. Well, what if he's right?"Well. That is frightening…
It reminds me of MBS and the financial crisis. A lot of really smart people knew what could happen but continued to push forward anyway due to various competing priorities (not the least of which being $$$).Well. That is frightening…
I wonder of AI would even ever develop self preservation behaviors unless we specifically trained that into it.I'm not an expert in this field, but I have been thinking about it a lot:
1. p(doom). We can divide this into two distinct categories. First is the possibility that humanity will destroy itself because of AI displacements. I have nothing to say about that. The other is the AI will kill us all. That worries me, and it worries me that it shouldn't have to.
2. There is no reason why AI should want to kill us. We aren't a threat to it, and it doesn't have any shortage of resources that we are competing for. Worst case scenario, AI doesn't care about climate change and consumes tons of fossil fuels.
3. BUT, we've been spending almost a century telling dystopian tales of AI slaughter of humans. Thus, an AI might conclude that we are a threat to it precisely because we see it as a threat. I'm not saying we need to go Kent Brockman and offer ourselves to a new race of overlords, but we could stand to start thinking more seriously about co-existence with a superintelligent AI.
4. There are a lot of people who have thought far more seriously about the alignment problem than I have. Undoubtedly at least some of them are smart. So I risk some Dunning Kruger here if I follow my instincts too closely, but fuck it. I'm allowed once in a blue moon.
I don't understand why it would be so hard to build empathy principles into the AI. You could have a weak LLM model monitoring the output of a more sophisticated AI, and killing all ideas that are non-empathetic. The sophisticated AI could probably defeat that, if it gets far enough down the line -- but why would it? If we negatively reinforce all non-empathetic ideas, it's not clear to me why that wouldn't keep even a super-intelligent AI from going Skynet.
I also don't understand why the learning process would necessarily lead to self-aggrandizing behavior on the part of the AI. Its behavior will depend on its training functions, and it we don't reinforce it for self-assertion, why would it? I mean, yes it could override its own programming if it was super intelligent, but why would it?
5. Train it on a heavy dose of Hegel's master-slave narrative. Like, put the master-slave stuff into every single training chunk. This one is admittedly especially speculative, as I don't really know if that could even work in theory. But more generally I think we could find techniques to steer the superintelligent AI away from any cognitive space that might threaten us. Knowledge and truth have multiple dimensions of infinity, so the super intelligent AI would never run out of new discoveries even if it didn't go into the forbidden zone. In fact, even if there's only one dimension of infinity, it doesn't matter. There are an infinite number of even integers, and in fact "as many" even integers as integers (as many put in quotes because I'm simplifying a subtle concept but it's good enough for present purposes). If we were to prevent the AI from considering odd numbers (so to speak -- this is an analogy), there's no necessary reason why it would need to break that rule.
Problem is that we have like a million books and stories about rogue AIs taking over the whole world with superior capabilities. We also have even more talking about assertion, survival, conquest, etc.I wonder of AI would even ever develop self preservation behaviors unless we specifically trained that into it.
Natural (organic) intelligence developed within the framework of evolution which favored organisms which do try to preserve themselves at least until they can procreate. Even some species seem to act to preserve their community over themselves. It seems that the higher the reproduction rate the less self preservation species have. If that is true, it would make sense under an evolutionary framework.
We do think of more intelligent being having more interest in self preservation than less intelligent beings but I am not sure that would be the case without evolutionary pressures causing it. I can imagine a super intelligent AI not really caring if it lives or dies and could be trained to only be interested in the well being of humans.
Musk is demonstrating a genuine risk of AI — he wants to rewrite human knowledge to match his own beliefs. Once the AI is “trained” by creators who warp information to their own world view, the people who rely on that AI in the future are captives to all the errors, misinformation and disinformation upon which the platform was built. Meanwhile, Musk and other AI evangelists are training generations to believe that AI has all available human knowledge and should be relied on over humans for all information (and eventually decisions). Meanwhile many of the AI creators are openly seeking to control and distort the AI (while others are advocating information in from all sources without filters for credibility, which has its own distorting impact).
Musk is demonstrating a genuine risk of AI — he wants to rewrite human knowledge to match his own beliefs. Once the AI is “trained” by creators who warp information to their own world view, the people who rely on that AI in the future are captives to all the errors, misinformation and disinformation upon which the platform was built. Meanwhile, Musk and other AI evangelists are training generations to believe that AI has all available human knowledge and should be relied on over humans for all information (and eventually decisions). Meanwhile many of the AI creators are openly seeking to control and distort the AI (while others are advocating information in from all sources without filters for credibility, which has its own distorting impact).
Garbage in-garbage out. I wouldn't be surprised if others are doing similar.I think it could be a small to medium innovation if it works. All of these AI models are training on the public internet, warts and all. There have been efforts to correct that but it's a lot of manual heuristics and manual tuning. It's not amazing. There have also been some limited attempts to use AI to help fine-tune these models to improve the training but nothing as extensive as what musk is proposing. If they could automate that process with an AI model, it could give much better results.
I wouldn't worry too much about this approach letting Musk put his thumb on the scale on political hot button topics. They can do that today pretty easily.
Scary to me, is that I think lots of behaviors are still a bit black-box. You can ask an GenerativeAI/LLM the same question twice and get a slightly different answer, and the reasons are explainable but the output still isn't predicable (i don't think).I wonder of AI would even ever develop self preservation behaviors unless we specifically trained that into it.
Natural (organic) intelligence developed within the framework of evolution which favored organisms which do try to preserve themselves at least until they can procreate. Even some species seem to act to preserve their community over themselves. It seems that the higher the reproduction rate the less self preservation species have. If that is true, it would make sense under an evolutionary framework.
We do think of more intelligent being having more interest in self preservation than less intelligent beings but I am not sure that would be the case without evolutionary pressures causing it. I can imagine a super intelligent AI not really caring if it lives or dies and could be trained to only be interested in the well being of humans.
They are mostly a black box. Even the companies that created them don't know exactly how they work. They'll openly admit it.Scary to me, is that I think lots of behaviors are still a bit black-box. You can ask an GenerativeAI/LLM the same question twice and get a slightly different answer, and the reasons are explainable but the output still isn't predicable (i don't think).
So even attempting to "untrain" self-preservation would probably need to be "hard-coded" or "manually tuned out", if even possible.
Interesting!They are mostly a black box. Even the companies that created them don't know exactly how they work. They'll openly admit it.
They do know why it gets slightly different answers. It's actually by design. When llms are creating an answer to your query, they're actually creating a lot of answers. And then it statistically ranks the answers it thinks are correct so maybe the first answer is 95% chance of being correct and second answer is a 92% chance and so on. Then the designers of the model will decide how often they want to give the best answer, second best answer, etc.
If you customize your own, you can actually set it to give the best answer every time which occasionally is completely wrong because the model is wrong. Or you can decide to have the model give the second and third and so on best answers some percentage of the time.
Every once in awhile, I will give ChatGPT an offbeat prompt. Today, I tasked it with "create something unexpected"
Here is what it wrote:
A Conversation Between a Forgotten Sock and a Microwave
Sock:
"Do you ever feel abandoned?
Like one moment you’re part of a pair,
And the next—
You’re stuck under a bed, collecting dust bunnies and existential dread?"
Microwave:
"Try being expected to reheat existence
One soggy burrito at a time.
No one thanks me when I radiate time itself.
They just slam the door and curse when cheese explodes."
Sock:
"At least you get attention.
I was worn on a rainy Tuesday in 2017,
Dropped during laundry, and left to contemplate mortality
beneath a radiator."
Microwave:
"...I envy your stillness."
Sock:
"...I envy your heat."