Thought Experiments: When the Singularity Is More than a Literary Device
An Interview with Futurist-Inventor Ray Kurzweil
Is the Singularity a spiritual or a technological belief system? Perhaps it is the melding of both, says science fiction author Cory Doctorow in this dialogue with Ray Kurzweil. "After all, this is a system of belief that dictates a means by which we can care for our bodies virtuously and live long enough to transcend them. It's no wonder that the Singularity has come to occupy so much of the science fiction narrative in these years. Science or spirituality, you could hardly ask for a subject better tailored to technological speculation and drama."
Originally published in Asimov's
Science Fiction, April 18, 2005. Published with permission
on KurzweilAI.net January 16, 2006.
It’s not clear to me whether the Singularity is a technical
belief system or a spiritual one.
The Singularity—a notion that’s crept into a lot of skiffy,
and whose most articulate in-genre spokesmodel is Vernor Vinge—describes
the black hole in history that will be created at the moment when
human intelligence can be digitized. When the speed and scope of
our cognition is hitched to the price-performance curve of microprocessors,
our "prog-ress" will double every eighteen months, and
then every twelve months, and then every ten, and eventually, every
five seconds.
Singularities are, literally, holes in space from whence no information
can emerge, and so SF writers occasionally mutter about how hard
it is to tell a story set after the information Singularity. Everything
will be different. What it means to be human will be so different
that what it means to be in danger, or happy, or sad, or any of
the other elements that make up the squeeze-and-release tension
in a good yarn will be unrecognizable to us pre-Singletons.
It’s a neat conceit to write around. I’ve committed Singularity
a couple of times, usually in collaboration with gonzo Singleton
Charlie Stross, the mad antipope of the Singularity. But those stories
have the same relation to futurism as romance novels do to love:
a shared jumping-off point, but radically different morphologies.
Of course, the Singularity isn’t just a conceit for noodling
with in the pages of the pulps: it’s the subject of serious-minded
punditry, futurism, and even science.
Ray Kurzweil is one such pundit-futurist-scientist. He’s a
serial entrepreneur who founded successful businesses that advanced
the fields of optical character recognition (machine-reading) software,
text-to-speech synthesis, synthetic musical instrument simulation,
computer-based speech recognition, and stock-market analysis. He
cured his own Type-II diabetes through a careful review of the literature
and the judicious application of first principles and reason. To
a casual observer, Kurzweil appears to be the star of some kind
of Heinlein novel, stealing fire from the gods and embarking on
a quest to bring his maverick ideas to the public despite the dismissals
of the establishment, getting rich in the process.
Kurzweil believes in the Singularity. In his 1990 manifesto, "The
Age of Intelligent Machines," Kurzweil persuasively argued
that we were on the brink of meaningful machine intelligence. A
decade later, he continued the argument in a book called The
Age of Spiritual Machines, whose most audacious claim is that
the world’s computational capacity has been slowly doubling
since the crust first cooled (and before!), and that the doubling
interval has been growing shorter and shorter with each passing
year, so that now we see it reflected in the computer industry’s
Moore’s Law, which predicts that microprocessors will get twice
as powerful for half the cost about every eighteen months. The breathtaking
sweep of this trend has an obvious conclusion: computers more powerful
than people; more powerful than we can comprehend.
Now Kurzweil has published two more books, The Singularity Is
Near, When Humans Transcend Biology (Viking, Spring 2005) and
Fantastic Voyage: Live Long Enough to Live Forever (with
Terry Grossman, Rodale, November 2004). The former is a technological
roadmap for creating the conditions necessary for ascent into Singularity;
the latter is a book about life-prolonging technologies that will
assist baby-boomers in living long enough to see the day when technological
immortality is achieved.
See what I meant about his being a Heinlein hero?
I still don’t know if the Singularity is a spiritual or a
technological belief system. It has all the trappings of spirituality,
to be sure. If you are pure and kosher, if you live right and if
your society is just, then you will live to see a moment of Rapture
when your flesh will slough away leaving nothing behind but your
ka, your soul, your consciousness, to ascend to an immortal and
pure state.
I wrote a novel called Down and Out in the Magic Kingdom
where characters could make backups of themselves and recover from
them if something bad happened, like catching a cold or being assassinated.
It raises a lot of existential questions: most prominently: are
you still you when you’ve been restored from backup?
The traditional AI answer is the Turing Test, invented by Alan
Turing, the gay pioneer of cryptography and artificial intelligence
who was forced by the British government to take hormone treatments
to "cure" him of his homosexuality, culminating in his
suicide in 1954. Turing cut through the existentialism about measuring
whether a machine is intelligent by proposing a parlor game: a computer
sits behind a locked door with a chat program, and a person sits
behind another locked door with his own chat program, and they both
try to convince a judge that they are real people. If the computer
fools a human judge into thinking that it’s a person, then
to all intents and purposes, it’s a person.
So how do you know if the backed-up you that you’ve restored
into a new body—or a jar with a speaker attached to it—is
really you? Well, you can ask it some questions, and if it answers
the same way that you do, you’re talking to a faithful copy
of yourself.
Sounds good. But the me who sent his first story into Asimov’s
seventeen years ago couldn’t answer the question, "Write
a story for Asimov’s" the same way the me of today
could. Does that mean I’m not me anymore?
Kurzweil has the answer.
"If you follow that logic, then if you were to take me ten
years ago, I could not pass for myself in a Ray Kurzweil Turing
Test. But once the requisite uploading technology becomes available
a few decades hence, you could make a perfect-enough copy
of me, and it would pass the Ray Kurzweil Turing Test. The
copy doesn’t have to match the quantum state of my every neuron,
either: if you meet me the next day, I’d pass the Ray Kurzweil
Turing Test. Nevertheless, none of the quantum states in my brain
would be the same. There are quite a few changes that each of us
undergo from day to day, we don’t examine the assumption that
we are the same person closely.
"We gradually change our pattern of atoms and neurons but
we very rapidly change the particles the pattern is made up of.
We used to think that in the brain—the physical part of us
most closely associated with our identity—cells change very
slowly, but it turns out that the components of the neurons, the
tubules and so forth, turn over in only days. I’m a
completely different set of particles from what I was a week ago.
"Consciousness is a difficult subject, and I’m always
surprised by how many people talk about consciousness routinely
as if it could be easily and readily tested scientifically. But
we can’t postulate a consciousness detector that does not have
some assumptions about consciousness built into it.
"Science is about objective third party observations and logical
deductions from them. Consciousness is about first-person, subjective
experience, and there’s a fundamental gap there. We live in
a world of assumptions about consciousness. We share the assumption
that other human beings are conscious, for example. But that breaks
down when we go outside of humans, when we consider, for example,
animals. Some say only humans are conscious and animals are instinctive
and machinelike. Others see humanlike behavior in an animal and
consider the animal conscious, but even these observers don’t
generally attribute consciousness to animals that aren’t humanlike.
"When machines are complex enough to have responses recognizable
as emotions, those machines will be more humanlike than animals."
The Kurzweil Singularity goes like this: computers get better and
smaller. Our ability to measure the world gains precision and grows
ever cheaper. Eventually, we can measure the world inside the brain
and make a copy of it in a computer that’s as fast and complex
as a brain, and voila, intelligence.
Here in the twenty-first century we like to view ourselves as ambulatory
brains, plugged into meat-puppets that lug our precious grey matter
from place to place. We tend to think of that grey matter as transcendently
complex, and we think of it as being the bit that makes us us.
But brains aren’t that complex, Kurzweil says. Already, we’re
starting to unravel their mysteries.
"We seem to have found one area of the brain closely associated
with higher-level emotions, the spindle cells, deeply embedded in
the brain. There are tens of thousands of them, spanning the whole
brain (maybe eighty thousand in total), which is an incredibly small
number. Babies don’t have any, most animals don’t have
any, and they likely only evolved over the last million years or
so. Some of the high-level emotions that are deeply human come from
these.
"Turing had the right insight: base the test for intelligence
on written language. Turing Tests really work. A novel is based
on language: with language you can conjure up any reality, much
more so than with images. Turing almost lived to see computers doing
a good job of performing in fields like math, medical diagnosis
and so on, but those tasks were easier for a machine than demonstrating
even a child’s mastery of language. Language is the true embodiment
of human intelligence."
If we’re not so complex, then it’s only a matter of time
until computers are more complex than us. When that comes, our brains
will be model-able in a computer and that’s when the fun begins.
That’s the thesis of Spiritual Machines, which even
includes a (Heinlein-style) timeline leading up to this day.
Now, it may be that a human brain contains n logic-gates
and runs at x cycles per second and stores z petabytes,
and that n and x and z are all within reach.
It may be that we can take a brain apart and record the position
and relationships of all the neurons and sub-neuronal elements that
constitute a brain.
But there are also a nearly infinite number of ways of modeling
a brain in a computer, and only a finite (or possibly nonexistent)
fraction of that space will yield a conscious copy of the original
meat-brain. Science fiction writers usually hand-wave this step:
in Heinlein’s "Man Who Sold the Moon," the gimmick
is that once the computer becomes complex enough, with enough "random
numbers," it just wakes up.
Computer programmers are a little more skeptical. Computers have
never been known for their skill at programming themselves—they
tend to be no smarter than the people who write their software.
But there are techniques for getting computers to program themselves,
based on evolution and natural selection. A programmer creates a
system that spits out lots—thousands or even millions—of
randomly generated programs. Each one is given the opportunity to
perform a computational task (say, sorting a list of numbers from
greatest to least) and the ones that solve the problem best are
kept aside while the others are erased. Now the survivors are used
as the basis for a new generation of randomly mutated descendants,
each based on elements of the code that preceded them. By running
many instances of a randomly varied program at once, and by culling
the least successful and regenerating the population from the winners
very quickly, it is possible to evolve effective software
that performs as well or better than the code written by human authors.
Indeed, evolutionary computing is a promising and exciting field
that’s realizing real returns through cool offshoots like "ant
colony optimization" and similar approaches that are showing
good results in fields as diverse as piloting military UAVs and
efficiently provisioning car-painting robots at automotive plants.
So if you buy Kurzweil’s premise that computation is getting
cheaper and more plentiful than ever, then why not just use evolutionary
algorithms to evolve the best way to model a scanned-in human
brain such that it "wakes up" like Heinlein’s Mike
computer?
Indeed, this is the crux of Kurzweil’s argument in Spiritual
Machines: if we have computation to spare and a detailed model
of a human brain, we need only combine them and out will pop the
mechanism whereby we may upload our consciousness to digital storage
media and transcend our weak and bothersome meat forever.
But it’s a cheat. Evolutionary algorithms depend on the same
mechanisms as real-world evolution: herit-able variation of candidates
and a system that culls the least-suitable candidates. This latter—the
fitness-factor that determines which individuals in a cohort breed
and which vanish—is the key to a successful evolutionary system.
Without it, there’s no pressure for the system to achieve the
desired goal: merely mutation and more mutation.
But how can a machine evaluate which of a trillion models of a
human brain is "most like" a conscious mind? Or better
still: which one is most like the individual whose brain is being
modeled?
"It is a sleight of hand in Spiritual Machines,"
Kurzweil admits. "But in The Singularity Is Near, I
have an in-depth discussion about what we know about the brain and
how to model it. Our tools for understanding the brain are subject
to the Law of Accelerating Returns, and we’ve made more progress
in reverse-engineering the human brain than most people realize."
This is a tasty Kurzweilism that observes that improvements in technology
yield tools for improving technology, round and round, so that the
thing that progress begets more than anything is more and yet faster
progress.
"Scanning resolution of human tissue—both spatial and
temporal—is doubling every year, and so is our knowledge of
the workings of the brain. The brain is not one big neural net,
the brain is several hundred different regions, and we can understand
each region, we can model the regions with mathematics, most of
which have some nexus with chaos and self-organizing systems. This
has already been done for a couple dozen regions out of the several
hundred.
"We have a good model of a dozen or so regions of the auditory
and visual cortex, how we strip images down to very low-resolution
movies based on pattern recognition. Interestingly, we don’t
actually see things, we essentially hallucinate them in detail from
what we see from these low resolution cues. Past the early phases
of the visual cortex, detail doesn’t reach the brain.
"We are getting exponentially more knowledge. We can
get detailed scans of neurons working in vivo, and are beginning
to understand the chaotic algorithms underlying human intelligence.
In some cases, we are getting comparable performance of brain regions
in simulation. These tools will continue to grow in detail and sophistication.
"We can have confidence of reverse-engineering the brain in
twenty years or so. The reason that brain reverse engineering has
not contributed much to artificial intelligence is that up until
recently we didn’t have the right tools. If I gave you a computer
and a few magnetic sensors and asked you to reverse-engineer it,
you might figure out that there’s a magnetic device spinning
when a file is saved, but you’d never get at the instruction
set. Once you reverse-engineer the computer fully, however, you
can express its principles of operation in just a few dozen pages.
"Now there are new tools that let us see the interneuronal
connections and their signaling, in vivo, and in real-time.
We’re just now getting these tools and there’s very rapid
application of the tools to obtain the data.
"Twenty years from now we will have realistic simulations
and models of all the regions of the brain and [we will] understand
how they work. We won’t blindly or mindlessly copy those methods,
we will understand them and use them to improve our AI toolkit.
So we’ll learn how the brain works and then apply the sophisticated
tools that we will obtain, as we discover how the brain works.
"Once we understand a subtle science principle, we can isolate,
amplify, and expand it. Air goes faster over a curved surface: from
that insight we isolated, amplified, and expanded the idea and invented
air travel. We’ll do the same with intelligence.
"Progress is exponential—not just a measure of power
of computation, number of Internet nodes, and magnetic spots on
a hard disk—the rate of paradigm shift is itself accelerating,
doubling every decade. Scientists look at a problem and they intuitively
conclude that since we’ve solved 1 percent over the last year,
it’ll therefore be one hundred years until the problem is exhausted:
but the rate of progress doubles every decade, and the power of
the information tools (in price-performance, resolution, bandwidth,
and so on) doubles every year. People, even scientists, don’t
grasp exponential growth. During the first decade of the human genome
project, we only solved 2 percent of the problem, but we solved
the remaining 98 percent in five years."
But Kurzweil doesn’t think that the future will arrive in
a rush. As William Gibson observed, "The future is here, it’s
just not evenly distributed."
"Sure, it’d be interesting to take a human brain, scan
it, reinstantiate the brain, and run it on another substrate. That
will ultimately happen."
"But the most salient scenario is that we’ll gradually
merge with our technology. We’ll use nanobots to kill pathogens,
then to kill cancer cells, and then they’ll go into our brain
and do benign things there like augment our memory, and very gradually
they’ll get more and more sophisticated. There’s no single
great leap, but there is ultimately a great leap comprised of many
small steps.
"In The Singularity Is Near, I describe the radically
different world of 2040, and how we’ll get there one benign
change at a time. The Singularity will be gradual, smooth.
"Really, this is about augmenting our biological thinking
with nonbiological thinking. We have a capacity of 1026 to 1029
calculations per second (cps) in the approximately 1010 biological
human brains on Earth and that number won’t change much in
fifty years, but nonbiological thinking will just crash through
that. By 2049, nonbiological thinking capacity will be on the order
of a billion times that. We’ll get to the point where bio thinking
is relatively insignificant.
"People didn’t throw their typewriters away when word-processing
started. There’s always an overlap—it’ll take time
before we realize how much more powerful nonbiological thinking
will ultimately be."
It’s well and good to talk about all the stuff we can do with
technology, but it’s a lot more important to talk about the
stuff we’ll be allowed to do with technology. Think
of the global freak-out caused by the relatively trivial advent
of peer-to-peer file-sharing tools: Universities are wiretapping
their campuses and disciplining computer science students for writing
legitimate, general purpose software; grandmothers and twelve-year-olds
are losing their life savings; privacy and due process have sailed
out the window without so much as a by-your-leave.
Even P2P’s worst enemies admit that this is a general-purpose
technology with good and bad uses, but when new tech comes along
it often engenders a response that countenances punishing an infinite
number of innocent people to get at the guilty.
What’s going to happen when the new technology paradigm isn’t
song-swapping, but transcendent super-intelligence? Will the reactionary
forces be justified in razing the whole ecosystem to eliminate a
few parasites who are doing negative things with the new tools?
"Complex ecosystems will always have parasites. Malware [malicious
software] is the most important battlefield today.
"Everything will become software—objects will
be malleable, we’ll spend lots of time in VR, and computhought
will be orders of magnitude more important than biothought.
"Software is already complex enough that we have an ecological
terrain that has emerged just as it did in the bioworld.
"That’s partly because technology is unregulated and
people have access to the tools to create malware and the medicine
to treat it. Today’s software viruses are clever and stealthy
and not simpleminded. Very clever.
"But here’s the thing: you don’t see people advocating
shutting down the Internet because malware is so destructive. I
mean, malware is potentially more than a nuisance—emergency
systems, air traffic control, and nuclear reactors all run on vulnerable
software. It’s an important issue, but the potential damage
is still a tiny fraction of the benefit we get from the Internet.
"I hope it’ll remain that way—that the Internet
won’t become a regulated space like medicine. Malware’s
not the most important issue facing human society today. Designer
bioviruses are. People are concerted about WMDs, but the most daunting
WMD would be a designed biological virus. The means exist in college
labs to create destructive viruses that erupt and spread silently
with long incubation periods.
"Importantly, a would-be bio-terrorist doesn’t have to
put malware through the FDA’s regulatory approval process,
but scientists working to fix bio-malware do.
"In Huxley’s Brave New World, the rationale for
the totalitarian system was that technology was too dangerous and
needed to be controlled. But that just pushes technology underground
where it becomes less stable. Regulation gives the edge of
power to the irresponsible who won’t listen to the regulators
anyway.
"The way to put more stones on the defense side of the scale
is to put more resources into defensive technologies, not create
a totalitarian regime of Draconian control.
"I advocate a one hundred billion dollar program to accelerate
the development of anti-biological virus technology. The way to
combat this is to develop broad tools to destroy viruses. We have
tools like RNA interference, just discovered in the past two years
to block gene expression. We could develop means to sequence the
genes of a new virus (SARS only took thirty-one days) and respond
to it in a matter of days.
"Think about it. There’s no FDA for software, no certification
for programmers. The government is thinking about it, though! The
reason the FCC is contemplating Trusted Computing mandates,"—a
system to restrict what a computer can do by means of hardware locks
embedded on the motherboard—"is that computing technology
is broadening to cover everything. So now you have communications
bureaucrats, biology bureaucrats, all wanting to regulate computers.
"Biology would be a lot more stable if we moved away from
regulation—which is extremely irrational and onerous and doesn’t
appropriately balance risks. Many medications are not available
today even though they should be. The FDA always wants to know what
happens if we approve this and will it turn into a thalidomide situation
that embarrasses us on CNN?
"Nobody asks about the harm that will certainly accrue from
delaying a treatment for one or more years. There’s no political
weight at all, people have been dying from diseases like heart disease
and cancer for as long as we’ve been alive. Attributable risks
get 100-1000 times more weight than unattributable risks."
Is this spirituality or science? Perhaps it is the melding of both—more
shades of Heinlein, this time the weird religions founded by people
who took Stranger in a Strange Land way too seriously.
After all, this is a system of belief that dictates a means by
which we can care for our bodies virtuously and live long enough
to transcend them. It is a system of belief that concerns itself
with the meddling of non-believers, who work to undermine its goals
through irrational systems predicated on their disbelief. It is
a system of belief that asks and answers the question of what it
means to be human.
It’s no wonder that the Singularity has come to occupy so
much of the science fiction narrative in these years. Science or
spirituality, you could hardly ask for a subject better tailored
to technological speculation and drama.
© 2005 Cory
Doctorow. Reprinted with permission.
|