Discussion about this post

User's avatar
Roro's avatar

Hi Steven, appreciate your article - particularly boiling down two issues of exfiltration along with just in time controls. That said, adjacent but distinct to Timothy Lee’s comment - what of the risks of intentional “rogue” AI deployment? Specifically, conversations tend to center on companies that have decent enmeshment with social contracts. They may or may not do an adequate job around risk controls. But what of shadow actors that are in fact motivated to release AIs with comprised motives designed into them? I see this as a likely scenario that is immune to ethical conversation.

Expand full comment
Timothy B. Lee's avatar

Hi, I really appreciate the thoughtful critique!

I really think that the fundamental disagreement here is about superintelligence. People who believe in superintelligence believe that once AI systems reach a certain level of intelligence, the default outcome will be that they take over the world. If that's true, we need a strategy for keeping AI systems under control—and for believers in superintelligence, that seems difficult.

As a superintelligence unbeliever, this doesn't seem right at all. I certainly think it's possible that AI systems will achieve human-level intelligence on most tasks and will go far beyond human capabilities on some of them. But because I do not think raw intelligence is all that important for gaining power, I do not think that a superintelligent AI is certain, or even all that likely, to gain control over the world.

So I 100 percent agree that at some point we're likely to see an AI system exfiltrate itself from an AI lab and spread across the Internet like a virus. And I'm sure that such a superintelligent AI system will cause a significant amount of damage. That just doesn't seem that different from the world we inhabit today, when computer viruses, botnets, ransomware, and North Korean hackers are regularly damaging our networks and computer systems.

Nobody think that the harms of computer viruses, ransomware, and North Korean hackers is so severe that we ought to shut down the Internet. And by the same token, I think that the harms from rogue human-level AI are likely to be significant but probably not so significant that we'll regret having developed human-level AI. And I think it's very unlikely that the harms will be so significant that they become existential.

Does that make sense?

Expand full comment
9 more comments...

No posts