Discussion about this post

User's avatar
Pjohn's avatar

If you create an AI whose capabilities are bounded as per. Graph Two and I create an AI whose capabilities are bounded as per. Graph One, my AI will outcompete your AI (and make my megacorp richer than your megacorp, or my country better at wars than your country, or whatever). The social problem (Moloch, essentially) of how we get people to restrict themselves to building Graph-Two-bounded AIs in this economic and political environment seems at least as hard, if not harder, than the problem of actually designing the Graph-Two-bounded AIs.

(nb. Personally I would consider D's "AI that wants what humans want" a massive failure; worse even than a rogue AI that wipes out civilisation. Human values are so flawed, and permit so much casual, unthinking cruelty, that the idea of having them superintelligently 'baked-in' to all future civilisation fills me with horror. I want an AI that *is* capable of thinking for itself about ethics, and *is* capable of disagreeing with humans about what is right: I would rather have the galaxy tiled with paperclips than with sweatshops and factory farms.)

Expand full comment
Paul Torek's avatar

Wanting in humans is, from a neuroscientific standpoint, probably a bizarre hodgepodge of electrical and chemical processes. But highly complex goal-directed behavior evolved multiple times in the course of evolution. It seems extremely unlikely to me that the neuroscience of octopus wanting and that of human wanting are very similar at a detailed level. And yet, both clearly want things, and will move heaven and earth to get them. I think this observation raises the probability that "wanting" would be an appropriate description for some possible AIs.

Expand full comment
2 more comments...

No posts