Discussion about this post

User's avatar
Karl von Wendt's avatar

Some thoughts:

On "Techno-Optimism": Like you, I believe technology has been strongly net-positive for humanity. At the same time, (almost) every new technology can cause damage that clearly outweighs its potential benefits: A knife can cut bread or kill a person. The reason why knives (and technology in general) are still net-positive is that we KNOW and CARE about their risks: we keep them out of the hands of children, don't allow them on board of planes, etc. With powerful AI, I think most decision makers either don't know or don't care sufficiently about the risks.

On the limits of intelligence: To become uncontrollable and destroy our future, an AI would neither have to be allknowing nor allmighty. All it needs to be is a superhuman manipulator that is able to take over power the same way a human dictator would, only 100x more effective. It could then make sure that we'll not get in the way of whatever goal it pursues, which is very likely not a goal we would want it to pursue. It will likely not destroy us on purpose, just as we don't destroy rhinos on purpose. We just change the world in a way that suits our goals, not theirs.

On alignment: Regardless how easy to solve problem 2 and 3 may or may not be, as long as problem 1 is not solved, we must not risk creating something that could get out of control. This does not mean we can't continue developing powerful AI - there are many safe ways to use narrow superhuman tool-AI or sub-human general-purpose AI that in combination I believe can help us achieve almost anything a superintelligent AGI could do for us.

Ben Schulz's avatar

Paul Christiano's prediction of gradual disempowerment seems much more likely. The debate between Yudkowsky and him was interesting. Computational irreducibility doesn't mean an AI couldn't build the future it wants. The best way to predict the future is by molding it.

1 more comment...

No posts

Ready for more?