Download The Inevitability of Artificial Intelligence (AI)

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Singularity Sky wikipedia , lookup

Human–computer interaction wikipedia , lookup

Ray Kurzweil wikipedia , lookup

History of artificial intelligence wikipedia , lookup

Embodied cognitive science wikipedia , lookup

Adaptive collaborative control wikipedia , lookup

How to Create a Mind wikipedia , lookup

Existential risk from artificial general intelligence wikipedia , lookup

Philosophy of artificial intelligence wikipedia , lookup

Ethics of artificial intelligence wikipedia , lookup

Intelligence explosion wikipedia , lookup

The Singularity Is Near wikipedia , lookup

Technological singularity wikipedia , lookup

Transcript
The
Inevitability
of
Artificial Intelligence (AI)
Slides by David Lanner
The Mechanistic View
of The Universe
Anything that is observable, measurable, and quantifiable can
(at least theoretically) be modeled or reproduced
But What If You Can’t?
:(
• What if there is some metaphysical “substance” in the brain that
prevents it from being fully modeled?
• Can you mathematically represent a soul?
• René Descartes
– The Pineal Gland (the “seat of the soul”)
Right Now, Human Brains Are
More Intelligent Than Computers
• Why?
– Neurons are highly combinatorial, whereas circuits
currently are not.
– Even though computers are capable of more
calculations per second (cps) than individual
neurons, neurons are still more concurrent/parallel
Why else?
• There is a serious limit to the hardware
that we can produce
– Specifically, the number of transistors able
to be fit onto an integrated circuit.
– Moore’s Law (Trend)
• This number doubles every two years
Moore’s Law
Moore’s Law won’t go on forever?
• Limit to the size of the transistors
The Law of Accelerating Returns
• Ray Kurzweil states that Moore’s Law is
only the fifth phase in a trend that has
existed at least as long as
electromechanical computers at the
beginning of the 20th century.
– This trend is the Law of Accelerating
Returns
Present-day computer hardware can only perform so many calculations per second
(cps). This turns out to fall far below the computing power of the human brain.
Kurzweil’s “estimate of brain capacity is 100 billion neurons times an average 1,000
connections per neuron (with the calculations taking place primarily in the
connections) times 200 calculations per second.” (Kurzweil). He predicts, based on
his law of accelerating returns, that
•We achieve one Human Brain capability (2 * 10^16 cps) for $1,000 around the year 2023,
•We achieve one Human Brain capability (2 * 10^16 cps) for one cent around the year 2037,
•We achieve one Human Race1 capability (2 * 10^26 cps) for $1,000 around the year 2049,
•We achieve one Human Race capability (2 * 10^26 cps) for one cent around the year 2059
(Kurzweil) [emphasis added].
1By “Human Race capability”, Kurzweil means a computer which is capable of performing n calculations per second,
where n is equal to the sum of the computing power (in cps) of every human brain in the world. At the time “The Law of
Accelerating Returns” was published, this amounted to roughly 6 billion people
The Technological Singularity
•The existence of AI whose intelligence far exceeds that of humans
•Machines that can build better versions of themselves (e.g. “thinking” speed)
•Which can build better versions of themselves
•Which can build better versions of themselves
•Which can build better versions of themselves
•Which can build better versions of themselves, etc.
•Humans will be “left behind”
•Is this good or bad?
•Can I still use my laptop for looking at the YouTube, or would it get mad at me?
Singularity Failure?
When this day comes - when machines are capable of
more calculations per second than human brains - well,
maybe nothing will happen.
[Cue Josh]
Vernor Vinge (popularizer of the term “Singularity” as it
applies to AI):
“A plausible explanation for 'Singularity failure' is that
we never figure out how to 'do the software' (or 'find the
soul in the hardware', if you're more mystically
inclined)”
Turning now to the more likely scenario,
in which the Singularity actually happens
•
Many Futurists like Kurzweil predicts that the machines will be able to
create more efficient versions of themselves, which, in turn, will be able
to improve on their own design, so on and so forth- potentially ad
infinitum.
•
With this incredible ability to improve themselves, the rate of change of
progress will become so fast that humans will be “left behind”, so to
speak, by technological and scientific advancements.
•
Saying that humans will be “left behind” evokes a sense of dread in
most people; it implies a loss of control, which is unacceptable to
many.
Who Will Have Control?
(it ain’t us)
•
•
•
•
The desirable outcome in a post-Singularity world, for these people, is
that humans still maintain some measure of control over these godlike
beings, while deriving utility and pleasure out of them.
This is, essentially, the desire to own and use an incredibly powerful,
super- intelligent PC.
This is understandable in the context of today's consumer-centric world
It is also understandable when considering the alternative scenario, in
which humans are unable for whatever reason to control these
machines.
When The Robots Have Control
•
In this scenario, we may expect to see one of two basic outcomes:
– Either humans continue to exist, despite not having any control over the new dominant type of life on earth
– Or they do not.
–
In the case of the former, the new rulers of earth do not see the human race as a threat (or, as in The Matrix, a source of
fuel or energy); humans would go on with their lives.
–
In the case of the latter, the reader can refer to any number of science fiction novels and Hollywood movies concerning
the various Apocalyptic possibilities, including such notables as:
•
•
•
•
•
•
•
Battlestar Galactica
The Cyberiad
The Matrix
I, Robot
The Terminator series
Do Androids Dream of Electric Sheep?
R.U.R. (Rossum's Universal Robots)
Man-Machine Merger (Mmmm)
•
A curious meeting between the scenario in which man goes extinction
and the one in which he survives is that man merges with the machine.
– How might this come about?
– One plausible explanation is that humans first start off simply using
intelligent machines for convenience, utility, and pleasure as is currently
being done with non-intelligent ones.
– Slowly, people begin to realize the benefits of relying more and more heavily
on these machines for augmentation of their human abilities, such as
amplifying their strength, speed, dexterity, intelligence, memory, longevity,
and so forth.
– It could come to the point, eventually, that augmentation leads to humans
transferring their identities ― their personalities and knowledge and
memories, and whatever else makes a human “human” ― (or at least
copies of their identities) into a machine.
– This is, many futurists believe, the only possible way for humans to survive
the Singularity. The only concern someone might have would be the
definition of “human” at that point.
•
•
•
But which of these can be expected?
Any number of events are coming up on the horizon:
whether or not the Singularity happens
–
whether or not humans will survive it
•
•
•
•
•
•
whether or not humans would even be humans at that point.
Certainly no one can be sure of their predictions.
This is what futurists like Kurzweil do, though- the prediction of future events
based on past performance (or trends).
The trend, in this case, is Moore's Law.
It is arguable that so long as the rate of technological improvement follows the
expected curve, then most futurists will be more or less accurate in their
predictions.
If this is true, then so is the one thing most futurists agree on:
AI
Is
Inevitable.
(so there)