In 2000, AI researcher Eliezer Yudkowsky and entrepreneur Brian Atkins founded the Singularity Institute to work toward smarter-than-human intelligence by engaging in Artificial Intelligence and machine ethics research. On the Institute’s site, Yudkowsky states:
The Singularity is beyond huge, but it can begin with something small. If one smarter-than-human intelligence exists, that mind will find it easier to create still smarter minds. In this respect the dynamic of the Singularity resembles other cases where small causes can have large effects; toppling the first domino in a chain, starting an avalanche with a pebble, perturbing an upright object balanced on its tip. All it takes is one technology – Artificial Intelligence, brain-computer interfaces, or perhaps something unforeseen – that advances to the point of creating smarter-than-human minds. That one technological advance is the equivalent of the first self-replicating chemical that gave rise to life on Earth.