Yann LeCun is making the same mistake Marc Andreesen makes about AI risk. They aren’t considering how powerful a system can be which incorporates generative AI with other code, tools, and features. LLMs can’t cause massively bad outcomes, but it’s not absurd to think human-directed LLM applications with powerful tools could cause large-scale harm.
With the alignment team leaving OpenAI, lots of people are discussing the safety risks not being taken seriously internally at OpenAI. Yann LeCun and others have basically made fun of these opinions, and even went so far as to consider it idiotic to worry about it. I’m e/acc for the most part. I believe we should continue accelerating AI development and growth, but I think he’s wrong to dismiss the risks out of hand.
Marc Andreessen has a similar opinion about AI safety. It’s crazy to me that this discussion lacks so much nuance when it’s so important. One big problem is ambiguity around the risks. So, I’ll break down the real, practical categories of AI safety (similar to how I did previously) but with concrete examples of bad outcomes so readers can actually consider the risks, decide what they think, and even think of solutions.
Before I get into it, I want to say that most of this can be boiled down to the fact that generative AI is basically a form of leverage. It gives more power to individuals and companies, but also it gives more power to bad actors who can get access.
Some people are worried about a future AI system having goals that are in tension with human goals on such a scale that it wipes out millions or billions of people. I’d define this as an AI alignment concern. The attack scenario is something where the AI becomes conscious and wants to destroy all humans because we’re bad or the well-known paperclip example where a system exhausts all resources in pursuit of some inane goal. But these simply don’t seem plausible even in the next 10-15 years.
However, if you consider that AI systems could be asked to harm specific groups of people such as Israel or Hamas utilizing AI in war against one another, you can see how alignment of AI systems is important. Also, I do think there is a non-zero chance of top researchers coming up with a self-learning breakthrough where a model significantly improves itself in a loop. In such a case, it’s really hard to even guess what would occur.
All that said, it’s still seeming like we’ll end up with open-source models of comparable strength to top closed-source models. And open source models can be trained to be uncensored or function without guardrails so, practically speaking, pushing for alignment might be a bit pointless because there will always be another model that isn’t aligned well. If that’s the case, there are some interesting implications about how to mitigate AI alignment risks.
Safety is maybe the most ambiguous term in the industry, but I think there are some non-negligible risks worth considering that are safety-related and don’t fall into alignment or security risks so I consider them AI Safety risks. Some of the risks below are critical enough to contribute to p(doom)
, (the probability of very bad outcomes), but others are just worth mentioning to be comprehensive.
I want to iterate again that I am e/acc at heart, so I don’t think we should pause or stop AI development, but I do think there are some legitimate concerns. And the complete disregard and belittling of people who are worried about the risks listed above is incorrect and unkind.
Cheers! Hope you enjoyed the post :)
- Joseph
Sign up for my email list to know when I post more content like this. I also post my thoughts on Twitter/X.