Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated:
1
00:00:15,580 --> 00:00:25,547
♪♪
2
00:00:25,590 --> 00:00:35,035
♪♪
3
00:00:35,078 --> 00:00:41,302
What we're on the brink of is
a world of increasingly intense,
4
00:00:41,345 --> 00:00:45,219
sophisticated
artificial intelligence.
5
00:00:45,262 --> 00:00:48,396
Man: Technology is evolving
so much faster than our society
6
00:00:48,439 --> 00:00:51,181
has the ability
to protect us as citizens.
7
00:00:51,486 --> 00:00:55,707
The robots are coming, and they
will destroy our livelihoods.
8
00:00:55,751 --> 00:01:01,844
♪♪
9
00:01:01,887 --> 00:01:04,238
You have a networked
intelligence that watches us,
10
00:01:04,281 --> 00:01:08,590
knows everything about us,
and begins to try to change us.
11
00:01:08,633 --> 00:01:12,768
Man #2: Twitter has become the
world's number-one news site.
12
00:01:12,811 --> 00:01:15,205
Man #3:
Technology is never good or bad.
13
00:01:15,249 --> 00:01:18,948
It's what we do
with the technology.
14
00:01:18,991 --> 00:01:22,734
Eventually, millions of people
are gonna be thrown out of jobs
15
00:01:22,778 --> 00:01:25,737
because their skills
are going to be obsolete.
16
00:01:25,781 --> 00:01:27,435
Woman: Mass unemployment...
17
00:01:27,478 --> 00:01:32,527
greater inequalities,
even social unrest.
18
00:01:32,570 --> 00:01:35,530
Man #4: Regardless of whether
to be afraid or not afraid,
19
00:01:35,573 --> 00:01:38,185
the change is coming,
and nobody can stop it.
20
00:01:38,228 --> 00:01:44,539
♪♪
21
00:01:44,582 --> 00:01:46,323
Man #5: We've invested
huge amounts of money,
22
00:01:46,367 --> 00:01:49,283
and so it stands to reason
that the military,
23
00:01:49,326 --> 00:01:50,893
with their own desires,
24
00:01:50,936 --> 00:01:53,330
are gonna start to use
these technologies.
25
00:01:53,374 --> 00:01:55,419
Man #6:
Autonomous weapons systems
26
00:01:55,463 --> 00:01:57,552
could lead to a global arms race
27
00:01:57,595 --> 00:02:00,032
to rival the Nuclear Era.
28
00:02:00,076 --> 00:02:02,296
♪♪
29
00:02:02,339 --> 00:02:04,036
Man #7:
We know what the answer is.
30
00:02:04,080 --> 00:02:05,429
They'll eventually
be killing us.
31
00:02:05,473 --> 00:02:10,782
♪♪
32
00:02:10,826 --> 00:02:12,349
Man #8:
These technology leaps
33
00:02:12,393 --> 00:02:15,874
are gonna yield
incredible miracles...
34
00:02:15,918 --> 00:02:18,181
and incredible horrors.
35
00:02:18,225 --> 00:02:24,231
♪♪
36
00:02:24,274 --> 00:02:29,323
Man #9: We created it,
so I think, as we move forward,
37
00:02:29,366 --> 00:02:33,762
this intelligence
will contain parts of us.
38
00:02:33,805 --> 00:02:35,981
And I think the question is --
39
00:02:36,025 --> 00:02:39,463
Will it contain
the good parts...
40
00:02:39,507 --> 00:02:41,378
or the bad parts?
41
00:02:41,422 --> 00:02:47,079
♪♪
42
00:02:57,742 --> 00:03:04,793
♪♪
43
00:03:04,836 --> 00:03:08,840
Sarah: The survivors
called the war "Judgment Day."
44
00:03:08,884 --> 00:03:12,583
They lived only to face
a new nightmare --
45
00:03:12,627 --> 00:03:14,019
the war against the machines.
46
00:03:14,063 --> 00:03:15,412
Aah!
47
00:03:15,456 --> 00:03:18,023
Nolan: I think
we've completely fucked this up.
48
00:03:18,067 --> 00:03:21,549
I think Hollywood has managed
to inoculate the general public
49
00:03:21,592 --> 00:03:24,247
against this question --
50
00:03:24,291 --> 00:03:28,251
the idea of machines
that will take over the world.
51
00:03:28,295 --> 00:03:30,645
Open the pod bay doors, HAL.
52
00:03:30,688 --> 00:03:33,561
HAL: I'm sorry, Dave.
53
00:03:33,604 --> 00:03:35,911
I'm afraid I can't do that.
54
00:03:37,434 --> 00:03:38,696
HAL?
55
00:03:38,740 --> 00:03:40,437
Nolan:
We've cried wolf enough times...
56
00:03:40,481 --> 00:03:42,483
HAL?
...that the public
has stopped paying attention,
57
00:03:42,526 --> 00:03:43,962
because it feels like
science fiction.
58
00:03:44,006 --> 00:03:45,486
Even sitting here talking
about it right now,
59
00:03:45,529 --> 00:03:48,228
it feels a little bit silly,
a little bit like,
60
00:03:48,271 --> 00:03:51,666
"Oh, this is an artifact
of some cheeseball movie."
61
00:03:51,709 --> 00:03:56,584
The WOPR spends all its time
thinking about World War III.
62
00:03:56,627 --> 00:03:59,064
But it's not.
63
00:03:59,108 --> 00:04:02,111
The general public is about
to get blindsided by this.
64
00:04:02,154 --> 00:04:11,512
♪♪
65
00:04:11,555 --> 00:04:13,514
As a society and as individuals,
66
00:04:13,557 --> 00:04:18,954
we're increasingly surrounded
by machine intelligence.
67
00:04:18,997 --> 00:04:22,653
We carry this pocket device
in the palm of our hand
68
00:04:22,697 --> 00:04:24,829
that we use to make
a striking array
69
00:04:24,873 --> 00:04:26,831
of life decisions right now,
70
00:04:26,875 --> 00:04:29,007
aided by a set
of distant algorithms
71
00:04:29,051 --> 00:04:30,748
that we have no understanding.
72
00:04:30,792 --> 00:04:34,143
♪♪
73
00:04:34,186 --> 00:04:36,537
We're already pretty jaded
about the idea
74
00:04:36,580 --> 00:04:37,929
that we can talk to our phone,
75
00:04:37,973 --> 00:04:40,062
and it mostly understands us.
76
00:04:40,105 --> 00:04:42,456
Woman: I found quite a number
of action films.
77
00:04:42,499 --> 00:04:44,327
Five years ago -- no way.
78
00:04:44,371 --> 00:04:47,678
Markoff: Robotics.
Machines that see and speak...
79
00:04:47,722 --> 00:04:48,897
Woman: Hi, there....and listen.
80
00:04:48,940 --> 00:04:50,202
All that's real now.
81
00:04:50,246 --> 00:04:51,639
And these technologies
82
00:04:51,682 --> 00:04:55,686
are gonna fundamentally
change our society.
83
00:04:55,730 --> 00:05:00,212
Thrun: Now we have this great
movement of self-driving cars.
84
00:05:00,256 --> 00:05:01,953
Driving a car autonomously
85
00:05:01,997 --> 00:05:06,088
can move people's lives
into a better place.
86
00:05:06,131 --> 00:05:07,916
Horvitz: I've lost
a number of family members,
87
00:05:07,959 --> 00:05:09,570
including my mother,
88
00:05:09,613 --> 00:05:11,876
my brother and sister-in-law
and their kids,
89
00:05:11,920 --> 00:05:14,009
to automobile accidents.
90
00:05:14,052 --> 00:05:18,405
It's pretty clear we could
almost eliminate car accidents
91
00:05:18,448 --> 00:05:20,102
with automation.
92
00:05:20,145 --> 00:05:21,843
30,000 lives in the U.S. alone.
93
00:05:21,886 --> 00:05:25,455
About a million around the world
per year.
94
00:05:25,499 --> 00:05:27,501
Ferrucci:
In healthcare, early indicators
95
00:05:27,544 --> 00:05:29,503
are the name of the game
in that space,
96
00:05:29,546 --> 00:05:33,158
so that's another place where
it can save somebody's life.
97
00:05:33,202 --> 00:05:35,726
Dr. Herman: Here in
the breast-cancer center,
98
00:05:35,770 --> 00:05:38,381
all the things that
the radiologist's brain
99
00:05:38,425 --> 00:05:43,386
does in two minutes, the
computer does instantaneously.
100
00:05:43,430 --> 00:05:47,303
The computer has looked
at 1 billion mammograms,
101
00:05:47,347 --> 00:05:49,261
and it takes that data
and applies it
102
00:05:49,305 --> 00:05:51,438
to this image instantaneously,
103
00:05:51,481 --> 00:05:54,441
so the medical application
is profound.
104
00:05:56,399 --> 00:05:57,705
Zilis:
Another really exciting area
105
00:05:57,748 --> 00:05:59,402
that we're seeing
a lot of development in
106
00:05:59,446 --> 00:06:03,275
is actually understanding
our genetic code
107
00:06:03,319 --> 00:06:06,104
and using that
to both diagnose disease
108
00:06:06,148 --> 00:06:07,758
and create
personalized treatments.
109
00:06:07,802 --> 00:06:11,632
♪♪
110
00:06:11,675 --> 00:06:14,112
Kurzweil:
The primary application
of all these machines
111
00:06:14,156 --> 00:06:17,246
will be to extend
our own intelligence.
112
00:06:17,289 --> 00:06:19,422
We'll be able to make
ourselves smarter,
113
00:06:19,466 --> 00:06:22,643
and we'll be better
at solving problems.
114
00:06:22,686 --> 00:06:24,775
We don't have to age.
We'll actually understand aging.
115
00:06:24,819 --> 00:06:27,125
We'll be able to stop it.
116
00:06:27,169 --> 00:06:29,519
Man: There's really no limit
to what intelligent machines
117
00:06:29,563 --> 00:06:30,868
can do for the human race.
118
00:06:30,912 --> 00:06:36,265
♪♪
119
00:06:36,308 --> 00:06:39,399
How could a smarter machine
not be a better machine?
120
00:06:42,053 --> 00:06:44,708
It's hard to say exactly
when I began to think
121
00:06:44,752 --> 00:06:46,971
that that was a bit naive.
122
00:06:47,015 --> 00:06:56,459
♪♪
123
00:06:56,503 --> 00:06:59,288
Stuart Russell,
he's basically a god
124
00:06:59,331 --> 00:07:00,898
in the field
of artificial intelligence.
125
00:07:00,942 --> 00:07:04,380
He wrote the book that almost
every university uses.
126
00:07:04,424 --> 00:07:06,948
Russell: I used to say it's the
best-selling AI textbook.
127
00:07:06,991 --> 00:07:10,255
Now I just say "It's the PDF
that's stolen most often."
128
00:07:10,299 --> 00:07:13,650
♪♪
129
00:07:13,694 --> 00:07:17,306
Artificial intelligence is
about making computers smart,
130
00:07:17,349 --> 00:07:19,830
and from the point
of view of the public,
131
00:07:19,874 --> 00:07:21,484
what counts as AI
is just something
132
00:07:21,528 --> 00:07:23,268
that's surprisingly intelligent
133
00:07:23,312 --> 00:07:25,488
compared to what
we thought computers
134
00:07:25,532 --> 00:07:28,404
would typically be able to do.
135
00:07:28,448 --> 00:07:33,801
AI is a field of research
to try to basically simulate
136
00:07:33,844 --> 00:07:36,717
all kinds of human capabilities.
137
00:07:36,760 --> 00:07:38,719
We're in the AI era.
138
00:07:38,762 --> 00:07:40,503
Silicon Valley
has the ability to focus
139
00:07:40,547 --> 00:07:42,462
on one bright, shiny thing.
140
00:07:42,505 --> 00:07:43,767
It was social networking
141
00:07:43,811 --> 00:07:45,508
and social media
over the last decade,
142
00:07:45,552 --> 00:07:48,119
and it's pretty clear
that the bit has flipped.
143
00:07:48,163 --> 00:07:50,557
And it starts
with machine learning.
144
00:07:50,600 --> 00:07:54,343
Nolan: When we look back at this
moment, what was the first AI?
145
00:07:54,386 --> 00:07:55,736
It's not sexy,
and it isn't the thing
146
00:07:55,779 --> 00:07:57,389
we could see at the movies,
147
00:07:57,433 --> 00:08:00,741
but you'd make a great case
that Google created,
148
00:08:00,784 --> 00:08:03,395
not a search engine,
but a godhead.
149
00:08:03,439 --> 00:08:06,486
A way for people to ask
any question they wanted
150
00:08:06,529 --> 00:08:08,270
and get the answer they needed.
151
00:08:08,313 --> 00:08:11,273
Russell: Most people are not
aware that what Google is doing
152
00:08:11,316 --> 00:08:13,710
is actually a form of
artificial intelligence.
153
00:08:13,754 --> 00:08:16,234
They just go there,
they type in a thing.
154
00:08:16,278 --> 00:08:18,323
Google gives them the answer.
155
00:08:18,367 --> 00:08:21,544
Musk: With each search,
we train it to be better.
156
00:08:21,588 --> 00:08:23,851
Sometimes we're typing a search,
and it tell us the answer
157
00:08:23,894 --> 00:08:27,419
before you've finished
asking the question.
158
00:08:27,463 --> 00:08:29,944
You know, who is the president
of Kazakhstan?
159
00:08:29,987 --> 00:08:31,685
And it'll just tell you.
160
00:08:31,728 --> 00:08:33,600
You don't have to go to the
Kazakhstan national website
161
00:08:33,643 --> 00:08:34,818
to find out.
162
00:08:34,862 --> 00:08:37,081
You didn't used to be
able to do that.
163
00:08:37,125 --> 00:08:39,475
Nolan:
That is artificial intelligence.
164
00:08:39,519 --> 00:08:42,783
Years from now when we try
to understand, we will say,
165
00:08:42,826 --> 00:08:44,567
"How did we miss it?"
166
00:08:44,611 --> 00:08:47,527
Markoff: It's one of
the striking contradictions
167
00:08:47,570 --> 00:08:48,484
that we're facing.
168
00:08:48,528 --> 00:08:50,051
Google and Facebook, et al,
169
00:08:50,094 --> 00:08:52,053
have built businesses
on giving us,
170
00:08:52,096 --> 00:08:54,185
as a society, free stuff.
171
00:08:54,229 --> 00:08:56,013
But it's a Faustian bargain.
172
00:08:56,057 --> 00:09:00,017
They're extracting something
from us in exchange,
173
00:09:00,061 --> 00:09:01,628
but we don't know
174
00:09:01,671 --> 00:09:03,760
what code is running
on the other side and why.
175
00:09:03,804 --> 00:09:06,546
We have no idea.
176
00:09:06,589 --> 00:09:08,591
It does strike
right at the issue
177
00:09:08,635 --> 00:09:11,028
of how much we should
trust these machines.
178
00:09:14,162 --> 00:09:18,166
I use computers
literally for everything.
179
00:09:18,209 --> 00:09:21,386
There are so many
computer advancements now,
180
00:09:21,430 --> 00:09:23,824
and it's become such
a big part of our lives.
181
00:09:23,867 --> 00:09:26,174
It's just incredible
what a computer can do.
182
00:09:26,217 --> 00:09:29,090
You can actually carry
a computer in your purse.
183
00:09:29,133 --> 00:09:31,571
I mean, how awesome is that?
184
00:09:31,614 --> 00:09:35,052
I think most technology is meant
to make things easier
185
00:09:35,096 --> 00:09:37,315
and simpler for all of us,
186
00:09:37,359 --> 00:09:40,362
so hopefully that just
remains the focus.
187
00:09:40,405 --> 00:09:43,147
I think everybody loves
their computers.
188
00:09:43,191 --> 00:09:44,366
[ Chuckles ]
189
00:09:44,409 --> 00:09:51,678
♪♪
190
00:09:51,721 --> 00:09:53,810
People don't realize
they are constantly
191
00:09:53,854 --> 00:09:59,076
being negotiated with
by machines,
192
00:09:59,120 --> 00:10:02,993
whether that's the price
of products in your Amazon cart,
193
00:10:03,037 --> 00:10:05,517
whether you can get
on a particular flight,
194
00:10:05,561 --> 00:10:08,912
whether you can reserve
a room at a particular hotel.
195
00:10:08,956 --> 00:10:11,959
What you're experiencing
are machine-learning algorithms
196
00:10:12,002 --> 00:10:14,265
that have determined
that a person like you
197
00:10:14,309 --> 00:10:15,919
is willing to pay 2 cents more
198
00:10:15,963 --> 00:10:17,791
and is changing the price.
199
00:10:17,834 --> 00:10:21,751
♪♪
200
00:10:21,795 --> 00:10:24,014
Kosinski: Now, a computer looks
at millions of people
201
00:10:24,058 --> 00:10:28,105
simultaneously for
very subtle patterns.
202
00:10:28,149 --> 00:10:31,369
You can take seemingly
innocent digital footprints,
203
00:10:31,413 --> 00:10:34,677
such as someone's playlist
on Spotify,
204
00:10:34,721 --> 00:10:37,201
or stuff that they
bought on Amazon,
205
00:10:37,245 --> 00:10:40,291
and then use algorithms
to translate this
206
00:10:40,335 --> 00:10:44,513
into a very detailed and a
very accurate, intimate profile.
207
00:10:47,603 --> 00:10:50,911
Kaplan: There is a dossier on
each of us that is so extensive
208
00:10:50,954 --> 00:10:52,695
it would be possibly
accurate to say
209
00:10:52,739 --> 00:10:55,698
that they know more about you
than your mother does.
210
00:10:55,742 --> 00:11:04,054
♪♪
211
00:11:04,098 --> 00:11:06,883
Tegmark: The major cause
of the recent AI breakthrough
212
00:11:06,927 --> 00:11:08,580
isn't just that some dude
213
00:11:08,624 --> 00:11:11,583
had a brilliant insight
all of a sudden,
214
00:11:11,627 --> 00:11:14,325
but simply that we have
much bigger data
215
00:11:14,369 --> 00:11:18,242
to train them on
and vastly better computers.
216
00:11:18,286 --> 00:11:19,940
el Kaliouby:
The magic is in the data.
217
00:11:19,983 --> 00:11:21,463
It's a ton of data.
218
00:11:21,506 --> 00:11:23,726
I mean, it's data
that's never existed before.
219
00:11:23,770 --> 00:11:26,686
We've never had
this data before.
220
00:11:26,729 --> 00:11:30,733
We've created technologies
that allow us to capture
221
00:11:30,777 --> 00:11:33,040
vast amounts of information.
222
00:11:33,083 --> 00:11:35,738
If you think of a billion
cellphones on the planet
223
00:11:35,782 --> 00:11:38,393
with gyroscopes
and accelerometers
224
00:11:38,436 --> 00:11:39,786
and fingerprint readers...
225
00:11:39,829 --> 00:11:42,005
couple that with the GPS
and the photos they take
226
00:11:42,049 --> 00:11:43,964
and the tweets that you send,
227
00:11:44,007 --> 00:11:47,750
we're all giving off huge
amounts of data individually.
228
00:11:47,794 --> 00:11:50,274
Cars that drive as the cameras
on them suck up information
229
00:11:50,318 --> 00:11:52,059
about the world around them.
230
00:11:52,102 --> 00:11:54,844
The satellites that are now
in orbit the size of a toaster.
231
00:11:54,888 --> 00:11:57,629
The infrared about
the vegetation on the planet.
232
00:11:57,673 --> 00:11:59,109
The buoys that are out
in the oceans
233
00:11:59,153 --> 00:12:01,024
to feed into the climate models.
234
00:12:01,068 --> 00:12:05,028
♪♪
235
00:12:05,072 --> 00:12:08,902
And the NSA, the CIA,
as they collect information
236
00:12:08,945 --> 00:12:12,644
about the
geopolitical situations.
237
00:12:12,688 --> 00:12:15,604
The world today is literally
swimming in this data.
238
00:12:15,647 --> 00:12:20,565
♪♪
239
00:12:20,609 --> 00:12:22,480
Kosinski: Back in 2012,
240
00:12:22,524 --> 00:12:25,875
IBM estimated
that an average human being
241
00:12:25,919 --> 00:12:31,098
leaves 500 megabytes
of digital footprints every day.
242
00:12:31,141 --> 00:12:34,841
If you wanted to back up
on the one day worth of data
243
00:12:34,884 --> 00:12:36,494
that humanity produces
244
00:12:36,538 --> 00:12:39,062
and imprint it out
on a letter-sized paper,
245
00:12:39,106 --> 00:12:43,806
double-sided, font size 12,
and you stack it up,
246
00:12:43,850 --> 00:12:46,113
it would reach from
the surface of the Earth
247
00:12:46,156 --> 00:12:49,116
to the sun four times over.
248
00:12:49,159 --> 00:12:51,292
That's every day.
249
00:12:51,335 --> 00:12:53,816
Kaplan: The data itself
is not good or evil.
250
00:12:53,860 --> 00:12:55,470
It's how it's used.
251
00:12:55,513 --> 00:12:58,342
We're relying, really,
on the goodwill of these people
252
00:12:58,386 --> 00:13:01,171
and on the policies
of these companies.
253
00:13:01,215 --> 00:13:03,870
There is no legal requirement
for how they can
254
00:13:03,913 --> 00:13:06,307
and should use
that kind of data.
255
00:13:06,350 --> 00:13:09,266
That, to me, is at the heart
of the trust issue.
256
00:13:11,007 --> 00:13:13,793
Barrat: Right now there's a
giant race for creating machines
257
00:13:13,836 --> 00:13:15,751
that are as smart as humans.
258
00:13:15,795 --> 00:13:17,971
Google -- They're working on
what's really the kind of
259
00:13:18,014 --> 00:13:20,016
Manhattan Project
of artificial intelligence.
260
00:13:20,060 --> 00:13:22,671
They've got the most money.
They've got the most talent.
261
00:13:22,714 --> 00:13:27,067
They're buying up AI companies
and robotics companies.
262
00:13:27,110 --> 00:13:29,069
Urban: People still think
of Google as a search engine
263
00:13:29,112 --> 00:13:30,722
and their e-mail provider
264
00:13:30,766 --> 00:13:33,943
and a lot of other things
that we use on a daily basis,
265
00:13:33,987 --> 00:13:39,383
but behind that search box
are 10 million servers.
266
00:13:39,427 --> 00:13:42,299
That makes Google the most
powerful computing platform
267
00:13:42,343 --> 00:13:43,910
in the world.
268
00:13:43,953 --> 00:13:47,217
Google is now working
on an AI computing platform
269
00:13:47,261 --> 00:13:50,133
that will have
100 million servers.
270
00:13:52,179 --> 00:13:53,963
So when you're interacting
with Google,
271
00:13:54,007 --> 00:13:56,052
we're just seeing
the toenail of something
272
00:13:56,096 --> 00:13:58,881
that is a giant beast
in the making.
273
00:13:58,925 --> 00:14:00,622
And the truth is,
I'm not even sure
274
00:14:00,665 --> 00:14:02,798
that Google knows
what it's becoming.
275
00:14:02,842 --> 00:14:11,502
♪♪
276
00:14:11,546 --> 00:14:14,114
Phoenix: If you look inside of
what algorithms are being used
277
00:14:14,157 --> 00:14:15,811
at Google,
278
00:14:15,855 --> 00:14:20,076
it's technology
largely from the '80s.
279
00:14:20,120 --> 00:14:23,863
So these are models that you
train by showing them a 1, a 2,
280
00:14:23,906 --> 00:14:27,344
and a 3, and it learns not
what a 1 is or what a 2 is --
281
00:14:27,388 --> 00:14:30,434
It learns what the difference
between a 1 and a 2 is.
282
00:14:30,478 --> 00:14:32,436
It's just a computation.
283
00:14:32,480 --> 00:14:35,396
In the last half decade, where
we've made this rapid progress,
284
00:14:35,439 --> 00:14:38,268
it has all been
in pattern recognition.
285
00:14:38,312 --> 00:14:41,184
Tegmark: Most of
the good, old-fashioned AI
286
00:14:41,228 --> 00:14:44,057
was when we would tell
our computers
287
00:14:44,100 --> 00:14:46,798
how to play a game like chess...
288
00:14:46,842 --> 00:14:49,584
from the old paradigm where
you just tell the computer
289
00:14:49,627 --> 00:14:52,195
exactly what to do.
290
00:14:54,502 --> 00:14:57,505
Announcer:
This is "Jeopardy!"
291
00:14:57,548 --> 00:14:59,376
♪♪
292
00:14:59,420 --> 00:15:02,510
"The IBM Challenge"!
293
00:15:02,553 --> 00:15:05,730
Ferrucci: No one at the time
had thought that a machine
294
00:15:05,774 --> 00:15:08,298
could have the precision
and the confidence
295
00:15:08,342 --> 00:15:09,952
and the speed
to play "Jeopardy!"
296
00:15:09,996 --> 00:15:11,475
well enough against
the best humans.
297
00:15:11,519 --> 00:15:14,609
Let's play "Jeopardy!"
298
00:15:18,569 --> 00:15:20,354
Watson.Watson: What is "shoe"?
299
00:15:20,397 --> 00:15:21,877
You are right.
You get to pick.
300
00:15:21,921 --> 00:15:24,836
Literary Character APB
for $800.
301
00:15:24,880 --> 00:15:28,014
Answer --
the Daily Double.
302
00:15:28,057 --> 00:15:31,539
Watson actually got its
knowledge by reading Wikipedia
303
00:15:31,582 --> 00:15:34,672
and 200 million pages
of natural-language documents.
304
00:15:34,716 --> 00:15:36,674
Ferrucci:
You can't program every line
305
00:15:36,718 --> 00:15:38,502
of how the world works.
306
00:15:38,546 --> 00:15:40,722
The machine has to learn
by reading.
307
00:15:40,765 --> 00:15:42,202
Now we come to Watson.
308
00:15:42,245 --> 00:15:43,986
"Who is Bram Stoker?"
309
00:15:44,030 --> 00:15:45,988
And the wager?
310
00:15:46,032 --> 00:15:49,165
Hello! $17,973.
311
00:15:49,209 --> 00:15:50,993
$41,413.
312
00:15:51,037 --> 00:15:53,343
And a two-day total
of $77--
313
00:15:53,387 --> 00:15:56,694
Phoenix: Watson's trained
on huge amounts of text,
314
00:15:56,738 --> 00:15:59,828
but it's not like it
understands what it's saying.
315
00:15:59,871 --> 00:16:02,309
It doesn't know that water makes
things wet by touching water
316
00:16:02,352 --> 00:16:04,441
and by seeing the way
things behave in the world
317
00:16:04,485 --> 00:16:06,182
the way you and I do.
318
00:16:06,226 --> 00:16:10,143
A lot of language AI today
is not building logical models
319
00:16:10,186 --> 00:16:11,622
of how the world works.
320
00:16:11,666 --> 00:16:15,365
Rather, it's looking at
how the words appear
321
00:16:15,409 --> 00:16:18,238
in the context of other words.
322
00:16:18,281 --> 00:16:20,196
Barrat: David Ferrucci
developed IBM's Watson,
323
00:16:20,240 --> 00:16:23,547
and somebody asked him,
"Does Watson think?"
324
00:16:23,591 --> 00:16:27,160
And he said,
"Does a submarine swim?"
325
00:16:27,203 --> 00:16:29,031
And what they meant was,
when they developed submarines,
326
00:16:29,075 --> 00:16:32,992
they borrowed basic principles
of swimming from fish.
327
00:16:33,035 --> 00:16:35,037
But a submarine swims
farther and faster than fish
328
00:16:35,081 --> 00:16:36,125
and can carry a huge payload.
329
00:16:36,169 --> 00:16:39,911
It out-swims fish.
330
00:16:39,955 --> 00:16:41,870
Ng: Watson winning the game
of "Jeopardy!"
331
00:16:41,913 --> 00:16:43,741
will go down
in the history of AI
332
00:16:43,785 --> 00:16:46,570
as a significant milestone.
333
00:16:46,614 --> 00:16:49,269
We tend to be amazed
when the machine does so well.
334
00:16:49,312 --> 00:16:52,663
I'm even more amazed when the
computer beats humans at things
335
00:16:52,707 --> 00:16:55,188
that humans are
naturally good at.
336
00:16:55,231 --> 00:16:58,060
This is how we make progress.
337
00:16:58,104 --> 00:17:00,671
In the early days of
the Google Brain project,
338
00:17:00,715 --> 00:17:02,804
I gave the team a very
simple instruction,
339
00:17:02,847 --> 00:17:05,807
which was, "Build the biggest
neural network possible,
340
00:17:05,850 --> 00:17:08,157
like 1,000 computers."
341
00:17:08,201 --> 00:17:09,724
Musk: A neural net is
something very close
342
00:17:09,767 --> 00:17:12,161
to a simulation
of how the brain works.
343
00:17:12,205 --> 00:17:16,818
It's very probabilistic,
but with contextual relevance.
344
00:17:16,861 --> 00:17:18,298
Urban: In your brain,
you have long neurons
345
00:17:18,341 --> 00:17:20,256
that connect to thousands
of other neurons,
346
00:17:20,300 --> 00:17:22,519
and you have these pathways
that are formed and forged
347
00:17:22,563 --> 00:17:24,739
based on what
the brain needs to do.
348
00:17:24,782 --> 00:17:28,960
When a baby tries something and
it succeeds, there's a reward,
349
00:17:29,004 --> 00:17:32,312
and that pathway that created
the success is strengthened.
350
00:17:32,355 --> 00:17:34,662
If it fails at something,
the pathway is weakened,
351
00:17:34,705 --> 00:17:36,794
and so, over time,
the brain becomes honed
352
00:17:36,838 --> 00:17:40,320
to be good at
the environment around it.
353
00:17:40,363 --> 00:17:43,279
Ng: Really, it's just getting
machines to learn by themselves.
354
00:17:43,323 --> 00:17:45,238
This is called "deep learning,"
and "deep learning"
355
00:17:45,281 --> 00:17:48,676
and "neural networks"
mean roughly the same thing.
356
00:17:48,719 --> 00:17:52,375
Tegmark: Deep learning
is a totally different approach
357
00:17:52,419 --> 00:17:55,161
where the computer learns
more like a toddler,
358
00:17:55,204 --> 00:17:56,466
by just getting a lot of data
359
00:17:56,510 --> 00:18:00,340
and eventually
figuring stuff out.
360
00:18:00,383 --> 00:18:03,125
The computer just gets
smarter and smarter
361
00:18:03,169 --> 00:18:05,997
as it has more experiences.
362
00:18:06,041 --> 00:18:08,130
Ng: So, imagine, if you will,
a neural network, you know,
363
00:18:08,174 --> 00:18:09,697
like 1,000 computers.
364
00:18:09,740 --> 00:18:11,438
And it wakes up
not knowing anything.
365
00:18:11,481 --> 00:18:14,093
And we made it watch YouTube
for a week.
366
00:18:14,136 --> 00:18:16,704
♪♪
367
00:18:16,747 --> 00:18:18,662
[ Psy's "Gangnam Style" plays ]
368
00:18:18,706 --> 00:18:20,360
♪ Oppan Gangnam style
369
00:18:20,403 --> 00:18:23,189
Ow!
370
00:18:23,232 --> 00:18:25,365
[ Laughing ]
371
00:18:25,408 --> 00:18:28,194
Charlie!
That really hurt!
372
00:18:28,237 --> 00:18:30,152
♪♪
373
00:18:30,196 --> 00:18:31,327
♪ Gangnam style
374
00:18:31,371 --> 00:18:33,286
♪ Op, op, op, op
375
00:18:33,329 --> 00:18:36,202
♪ Oppan Gangnam style
376
00:18:36,245 --> 00:18:38,508
Ng: And so, after watching
YouTube for a week,
377
00:18:38,552 --> 00:18:39,988
what would it learn?
378
00:18:40,031 --> 00:18:41,903
We had a hypothesis that
it would learn to detect
379
00:18:41,946 --> 00:18:44,384
commonly occurring objects
in videos.
380
00:18:44,427 --> 00:18:47,517
And so, we know that human faces
appear a lot in videos,
381
00:18:47,561 --> 00:18:49,302
so we looked,
and, lo and behold,
382
00:18:49,345 --> 00:18:51,608
there was a neuron that had
learned to detect human faces.
383
00:18:51,652 --> 00:18:56,265
[ Voice breaking ]
Leave Britney alone!
384
00:18:56,309 --> 00:18:58,354
Well, what else
appears in videos a lot?
385
00:18:58,398 --> 00:19:00,051
[ Boing! Sloop!
Patter! Patter! ]
386
00:19:00,095 --> 00:19:01,792
So, we looked,
and to our surprise,
387
00:19:01,836 --> 00:19:04,882
there was actually a neuron
that had learned to detect cats.
388
00:19:04,926 --> 00:19:14,849
♪♪
389
00:19:14,892 --> 00:19:17,068
I still remember
seeing recognition.
390
00:19:17,112 --> 00:19:18,635
"Wow, that's a cat.
Okay, cool.
391
00:19:18,679 --> 00:19:20,071
Great."
[ Laughs ]
392
00:19:23,162 --> 00:19:24,859
Barrat:
It's all pretty innocuous
393
00:19:24,902 --> 00:19:26,295
when you're thinking
about the future.
394
00:19:26,339 --> 00:19:29,733
It all seems kind of
harmless and benign.
395
00:19:29,777 --> 00:19:31,605
But we're making
cognitive architectures
396
00:19:31,648 --> 00:19:33,520
that will fly farther
and faster than us
397
00:19:33,563 --> 00:19:35,086
and carry a bigger payload,
398
00:19:35,130 --> 00:19:37,437
and they won't be
warm and fuzzy.
399
00:19:37,480 --> 00:19:39,656
Ferrucci: I think that,
in three to five years,
400
00:19:39,700 --> 00:19:41,702
you will see a computer system
401
00:19:41,745 --> 00:19:45,401
that will be able
to autonomously learn
402
00:19:45,445 --> 00:19:49,013
how to understand,
how to build understanding,
403
00:19:49,057 --> 00:19:51,364
not unlike the way
the human mind works.
404
00:19:53,931 --> 00:19:56,891
Whatever that lunch was,
it was certainly delicious.
405
00:19:56,934 --> 00:19:59,807
Simply some of
Robby's synthetics.
406
00:19:59,850 --> 00:20:01,635
He's your cook, too?
407
00:20:01,678 --> 00:20:04,551
Even manufactures
the raw materials.
408
00:20:04,594 --> 00:20:06,944
Come around here, Robby.
409
00:20:06,988 --> 00:20:09,773
I'll show you
how this works.
410
00:20:11,122 --> 00:20:13,342
One introduces
a sample of human food
411
00:20:13,386 --> 00:20:15,344
through this aperture.
412
00:20:15,388 --> 00:20:17,738
Down here there's a small
built-in chemical laboratory,
413
00:20:17,781 --> 00:20:19,218
where he analyzes it.
414
00:20:19,261 --> 00:20:21,263
Later, he can reproduce
identical molecules
415
00:20:21,307 --> 00:20:22,482
in any shape or quantity.
416
00:20:22,525 --> 00:20:24,614
Why, it's
a housewife's dream.
417
00:20:24,658 --> 00:20:26,834
Announcer: Meet Baxter,
418
00:20:26,877 --> 00:20:29,445
revolutionary
new category of robots,
419
00:20:29,489 --> 00:20:30,490
with common sense.
420
00:20:30,533 --> 00:20:31,839
Baxter...
421
00:20:31,882 --> 00:20:33,449
Barrat: Baxter is
a really good example
422
00:20:33,493 --> 00:20:36,887
of the kind of competition
we face from machines.
423
00:20:36,931 --> 00:20:42,676
Baxter can do almost anything
we can do with our hands.
424
00:20:42,719 --> 00:20:45,722
Baxter costs about
what a minimum-wage worker
425
00:20:45,766 --> 00:20:47,507
makes in a year.
426
00:20:47,550 --> 00:20:48,769
But Baxter won't be
taking the place
427
00:20:48,812 --> 00:20:50,118
of one minimum-wage worker --
428
00:20:50,161 --> 00:20:51,772
He'll be taking
the place of three,
429
00:20:51,815 --> 00:20:55,515
because they never get tired,
they never take breaks.
430
00:20:55,558 --> 00:20:57,865
Gourley: That's probably the
first thing we're gonna see --
431
00:20:57,908 --> 00:20:59,475
displacement of jobs.
432
00:20:59,519 --> 00:21:01,651
They're gonna be done
quicker, faster, cheaper
433
00:21:01,695 --> 00:21:04,088
by machines.
434
00:21:04,132 --> 00:21:07,657
Our ability to even stay current
is so insanely limited
435
00:21:07,701 --> 00:21:10,138
compared to
the machines we build.
436
00:21:10,181 --> 00:21:13,446
For example, now we have this
great movement of Uber and Lyft
437
00:21:13,489 --> 00:21:15,056
kind of making
transportation cheaper
438
00:21:15,099 --> 00:21:16,405
and democratizing
transportation,
439
00:21:16,449 --> 00:21:17,711
which is great.
440
00:21:17,754 --> 00:21:19,321
The next step is gonna be
441
00:21:19,365 --> 00:21:21,149
that they're all gonna be
replaced by driverless cars,
442
00:21:21,192 --> 00:21:22,411
and then all the Uber
and Lyft drivers
443
00:21:22,455 --> 00:21:25,936
have to find
something new to do.
444
00:21:25,980 --> 00:21:28,156
Barrat: There are
4 million professional drivers
445
00:21:28,199 --> 00:21:29,723
in the United States.
446
00:21:29,766 --> 00:21:31,638
They're unemployed soon.
447
00:21:31,681 --> 00:21:34,075
7 million people
that do data entry.
448
00:21:34,118 --> 00:21:37,339
Those people
are gonna be jobless.
449
00:21:37,383 --> 00:21:40,342
A job isn't just about money,
right?
450
00:21:40,386 --> 00:21:42,605
On a biological level,
it serves a purpose.
451
00:21:42,649 --> 00:21:45,391
It becomes a defining thing.
452
00:21:45,434 --> 00:21:48,350
When the jobs went away
in any given civilization,
453
00:21:48,394 --> 00:21:50,787
it doesn't take long
until that turns into violence.
454
00:21:50,831 --> 00:21:53,312
[ Crowd chanting
in native language ]
455
00:21:53,355 --> 00:21:57,011
♪♪
456
00:21:57,054 --> 00:21:59,579
[ Gunshots ]
457
00:21:59,622 --> 00:22:02,016
We face a giant divide
between rich and poor,
458
00:22:02,059 --> 00:22:05,019
because that's what automation
and AI will provoke --
459
00:22:05,062 --> 00:22:08,588
a greater divide between
the haves and the have-nots.
460
00:22:08,631 --> 00:22:10,807
Right now, it's working
into the middle class,
461
00:22:10,851 --> 00:22:12,896
into white-collar jobs.
462
00:22:12,940 --> 00:22:15,334
IBM's Watson does
business analytics
463
00:22:15,377 --> 00:22:20,600
that we used to pay a business
analyst $300 an hour to do.
464
00:22:20,643 --> 00:22:23,037
Gourley: Today, you're going
to college to be a doctor,
465
00:22:23,080 --> 00:22:25,082
to be an accountant,
to be a journalist.
466
00:22:25,126 --> 00:22:28,608
It's unclear that there's
gonna be jobs there for you.
467
00:22:28,651 --> 00:22:32,612
Ng: If someone's planning for
a 40-year career in radiology,
468
00:22:32,655 --> 00:22:34,222
just reading images,
469
00:22:34,265 --> 00:22:35,745
I think that could be
a challenge
470
00:22:35,789 --> 00:22:36,920
to the new graduates of today.
471
00:22:36,964 --> 00:22:39,227
[ Monitor beeping ]
472
00:22:39,270 --> 00:22:49,193
♪♪
473
00:22:49,237 --> 00:22:50,804
[ Beeping continues ]
474
00:22:50,847 --> 00:22:58,464
♪♪
475
00:22:58,507 --> 00:23:02,729
Dr. Herman: The da Vinci robot
is currently utilized
476
00:23:02,772 --> 00:23:07,516
by a variety of surgeons
for its accuracy and its ability
477
00:23:07,560 --> 00:23:12,303
to avoid the inevitable
fluctuations of the human hand.
478
00:23:12,347 --> 00:23:17,787
♪♪
479
00:23:17,831 --> 00:23:23,358
♪♪
480
00:23:23,402 --> 00:23:28,494
Anybody who watches this
feels the amazingness of it.
481
00:23:30,931 --> 00:23:34,674
You look through the scope,
and you're seeing the claw hand
482
00:23:34,717 --> 00:23:36,893
holding that woman's ovary.
483
00:23:36,937 --> 00:23:42,638
Humanity was resting right here
in the hands of this robot.
484
00:23:42,682 --> 00:23:46,947
People say it's the future,
but it's not the future --
485
00:23:46,990 --> 00:23:50,516
It's the present.
486
00:23:50,559 --> 00:23:52,474
Zilis: If you think about
a surgical robot,
487
00:23:52,518 --> 00:23:54,737
there's often not a lot
of intelligence in these things,
488
00:23:54,781 --> 00:23:56,783
but over time, as we put
more and more intelligence
489
00:23:56,826 --> 00:23:58,567
into these systems,
490
00:23:58,611 --> 00:24:02,441
the surgical robots can actually
learn from each robot surgery.
491
00:24:02,484 --> 00:24:04,181
They're tracking the movements,
they're understanding
492
00:24:04,225 --> 00:24:05,966
what worked
and what didn't work.
493
00:24:06,009 --> 00:24:08,708
And eventually, the robot
for routine surgeries
494
00:24:08,751 --> 00:24:12,320
is going to be able to perform
that entirely by itself...
495
00:24:12,363 --> 00:24:13,756
or with human supervision.
496
00:24:32,558 --> 00:24:34,995
♪♪
497
00:24:35,038 --> 00:24:37,214
Dr. Herman: It seems that we're
feeding it and creating it,
498
00:24:37,258 --> 00:24:42,785
but, in a way, we are a slave
to the technology,
499
00:24:42,829 --> 00:24:45,701
because we can't go back.
500
00:24:45,745 --> 00:24:47,355
[ Beep ]
501
00:24:50,053 --> 00:24:52,882
Gourley: The machines are taking
bigger and bigger bites
502
00:24:52,926 --> 00:24:57,147
out of our skill set
at an ever-increasing speed.
503
00:24:57,191 --> 00:24:59,236
And so we've got to run
faster and faster
504
00:24:59,280 --> 00:25:00,890
to keep ahead of the machines.
505
00:25:02,675 --> 00:25:04,677
How do I look?
506
00:25:04,720 --> 00:25:06,374
Good.
507
00:25:10,030 --> 00:25:11,553
Are you attracted to me?
508
00:25:11,597 --> 00:25:14,251
What?Are you attracted to me?
509
00:25:14,295 --> 00:25:17,777
You give me indications
that you are.
510
00:25:17,820 --> 00:25:20,562
I do?
Yes.
511
00:25:20,606 --> 00:25:22,608
Nolan: This is the future
we're headed into.
512
00:25:22,651 --> 00:25:26,046
We want to design
our companions.
513
00:25:26,089 --> 00:25:29,266
We're gonna like to see
a human face on AI.
514
00:25:29,310 --> 00:25:33,967
Therefore, gaming our emotions
will be depressingly easy.
515
00:25:34,010 --> 00:25:35,272
We're not that complicated.
516
00:25:35,316 --> 00:25:38,101
We're simple.
Stimulus-response.
517
00:25:38,145 --> 00:25:43,063
I can make you like me basically
by smiling at you a lot.
518
00:25:43,106 --> 00:25:45,674
AIs are gonna be fantastic
at manipulating us.
519
00:25:45,718 --> 00:25:54,640
♪♪
520
00:25:54,683 --> 00:25:56,946
So, you've developed
a technology
521
00:25:56,990 --> 00:26:00,036
that can sense
what people are feeling.
522
00:26:00,080 --> 00:26:01,472
Right.
We've developed technology
523
00:26:01,516 --> 00:26:03,387
that can read
your facial expressions
524
00:26:03,431 --> 00:26:06,521
and map that to a number
of emotional states.
525
00:26:06,565 --> 00:26:08,697
el Kaliouby: 15 years ago,
I had just finished
526
00:26:08,741 --> 00:26:11,482
my undergraduate studies
in computer science,
527
00:26:11,526 --> 00:26:15,008
and it struck me that I was
spending a lot of time
528
00:26:15,051 --> 00:26:17,793
interacting with my laptops
and my devices,
529
00:26:17,837 --> 00:26:23,582
yet these devices had absolutely
no clue how I was feeling.
530
00:26:23,625 --> 00:26:26,802
I started thinking, "What if
this device could sense
531
00:26:26,846 --> 00:26:29,326
that I was stressed
or I was having a bad day?
532
00:26:29,370 --> 00:26:31,067
What would that open up?"
533
00:26:32,721 --> 00:26:34,418
Hi, first-graders!
534
00:26:34,462 --> 00:26:35,855
How are you?
535
00:26:35,898 --> 00:26:37,813
Can I get a hug?
536
00:26:37,857 --> 00:26:40,773
We had kids interact
with the technology.
537
00:26:40,816 --> 00:26:42,862
A lot of it
is still in development,
538
00:26:42,905 --> 00:26:44,472
but it was just amazing.
539
00:26:44,515 --> 00:26:46,648
Who likes robots?
Me!
540
00:26:46,692 --> 00:26:48,911
Who wants to have a robot
in their house?
541
00:26:48,955 --> 00:26:51,479
What would you use
a robot for, Jack?
542
00:26:51,522 --> 00:26:56,353
I would use it to ask my mom
very hard math questions.
543
00:26:56,397 --> 00:26:58,181
Okay.
What about you, Theo?
544
00:26:58,225 --> 00:27:02,272
I would use it
for scaring people.
545
00:27:02,316 --> 00:27:04,666
All right.
So, start by smiling.
546
00:27:04,710 --> 00:27:06,625
Nice.
547
00:27:06,668 --> 00:27:09,018
Brow furrow.
548
00:27:09,062 --> 00:27:10,890
Nice one.
Eyebrow raise.
549
00:27:10,933 --> 00:27:12,587
This generation, technology
550
00:27:12,631 --> 00:27:15,068
is just surrounding them
all the time.
551
00:27:15,111 --> 00:27:17,853
It's almost like they expect
to have robots in their homes,
552
00:27:17,897 --> 00:27:22,336
and they expect these robots
to be socially intelligent.
553
00:27:22,379 --> 00:27:25,252
What makes robots smart?
554
00:27:25,295 --> 00:27:29,648
Put them in, like, a math
or biology class.
555
00:27:29,691 --> 00:27:32,259
I think you would
have to train it.
556
00:27:32,302 --> 00:27:35,218
All right.
Let's walk over here.
557
00:27:35,262 --> 00:27:37,394
So, if you smile and you
raise your eyebrows,
558
00:27:37,438 --> 00:27:39,005
it's gonna run over to you.
559
00:27:39,048 --> 00:27:40,833
Woman: It's coming over!
It's coming over! Look.
560
00:27:40,876 --> 00:27:43,139
[ Laughter ]
561
00:27:43,183 --> 00:27:45,272
But if you look angry,
it's gonna run away.
562
00:27:45,315 --> 00:27:46,490
[ Child growls ]
563
00:27:46,534 --> 00:27:48,797
-Awesome!
-Oh, that was good.
564
00:27:48,841 --> 00:27:52,366
We're training computers to read
and recognize emotions.
565
00:27:52,409 --> 00:27:53,846
Ready? Set? Go!
566
00:27:53,889 --> 00:27:57,414
And the response so far
has been really amazing.
567
00:27:57,458 --> 00:27:59,590
People are integrating this
into health apps,
568
00:27:59,634 --> 00:28:04,465
meditation apps, robots, cars.
569
00:28:04,508 --> 00:28:06,728
We're gonna see
how this unfolds.
570
00:28:06,772 --> 00:28:09,426
♪♪
571
00:28:09,470 --> 00:28:11,602
Zilis:
Robots can contain AI,
572
00:28:11,646 --> 00:28:14,388
but the robot is just
a physical instantiation,
573
00:28:14,431 --> 00:28:16,782
and the artificial
intelligence is the brain.
574
00:28:16,825 --> 00:28:19,872
And so brains can exist purely
in software-based systems.
575
00:28:19,915 --> 00:28:22,483
They don't need to have
a physical form.
576
00:28:22,526 --> 00:28:25,094
Robots can exist without
any artificial intelligence.
577
00:28:25,138 --> 00:28:28,097
We have a lot of
dumb robots out there.
578
00:28:28,141 --> 00:28:31,753
But a dumb robot can be
a smart robot overnight,
579
00:28:31,797 --> 00:28:34,103
given the right software,
given the right sensors.
580
00:28:34,147 --> 00:28:38,629
Barrat: We can't help but impute
motive into inanimate objects.
581
00:28:38,673 --> 00:28:40,327
We do it with machines.
582
00:28:40,370 --> 00:28:41,502
We'll treat them like children.
583
00:28:41,545 --> 00:28:43,330
We'll treat them
like surrogates.
584
00:28:43,373 --> 00:28:45,027
-Goodbye!
-Goodbye!
585
00:28:45,071 --> 00:28:48,204
And we'll pay the price.
586
00:28:49,292 --> 00:28:58,998
♪♪
587
00:28:59,041 --> 00:29:08,572
♪♪
588
00:29:08,616 --> 00:29:10,792
Okay, welcome to ATR.
589
00:29:10,836 --> 00:29:18,060
♪♪
590
00:29:25,067 --> 00:29:30,594
♪♪
591
00:29:30,638 --> 00:29:36,122
♪♪
592
00:29:47,786 --> 00:29:51,485
♪♪
593
00:29:51,528 --> 00:29:52,791
Konnichiwa.
594
00:30:24,170 --> 00:30:29,436
♪♪
595
00:30:53,677 --> 00:30:56,942
♪♪
596
00:30:56,985 --> 00:30:58,682
Gourley: We build
artificial intelligence,
597
00:30:58,726 --> 00:31:02,948
and the very first thing
we want to do is replicate us.
598
00:31:02,991 --> 00:31:05,341
I think the key point will come
599
00:31:05,385 --> 00:31:09,258
when all the major senses
are replicated --
600
00:31:09,302 --> 00:31:11,130
sight...
601
00:31:11,173 --> 00:31:12,871
touch...
602
00:31:12,914 --> 00:31:14,611
smell.
603
00:31:14,655 --> 00:31:17,919
When we replicate our senses,
is that when it become alive?
604
00:31:17,963 --> 00:31:22,010
♪♪
605
00:31:22,054 --> 00:31:24,752
[ Indistinct conversations ]
606
00:31:24,795 --> 00:31:27,581
♪♪
607
00:31:27,624 --> 00:31:29,104
Nolan:
So many of our machines
608
00:31:29,148 --> 00:31:31,019
are being built
to understand us.
609
00:31:31,063 --> 00:31:32,803
[ Speaking Japanese ]
610
00:31:32,847 --> 00:31:34,805
But what happens when
an anthropomorphic creature
611
00:31:34,849 --> 00:31:37,417
discovers that they can
adjust their loyalty,
612
00:31:37,460 --> 00:31:40,028
adjust their courage,
adjust their avarice,
613
00:31:40,072 --> 00:31:42,291
adjust their cunning?
614
00:31:42,335 --> 00:31:44,815
♪♪
615
00:31:44,859 --> 00:31:47,166
Musk: The average person,
they don't see killer robots
616
00:31:47,209 --> 00:31:48,645
going down the streets.
617
00:31:48,689 --> 00:31:50,996
They're like, "What are
you talking about?"
618
00:31:51,039 --> 00:31:53,955
Man, we want to make sure
that we don't have killer robots
619
00:31:53,999 --> 00:31:57,045
going down the street.
620
00:31:57,089 --> 00:31:59,439
Once they're going down
the street, it is too late.
621
00:31:59,482 --> 00:32:05,010
♪♪
622
00:32:05,053 --> 00:32:07,099
Russell: The thing
that worries me right now,
623
00:32:07,142 --> 00:32:08,578
that keeps me awake,
624
00:32:08,622 --> 00:32:11,842
is the development
of autonomous weapons.
625
00:32:11,886 --> 00:32:19,850
♪♪
626
00:32:19,894 --> 00:32:27,771
♪♪
627
00:32:27,815 --> 00:32:32,733
Up to now, people have expressed
unease about drones,
628
00:32:32,776 --> 00:32:35,127
which are remotely
piloted aircraft.
629
00:32:35,170 --> 00:32:39,783
♪♪
630
00:32:39,827 --> 00:32:43,309
If you take a drone's camera
and feed it into the AI system,
631
00:32:43,352 --> 00:32:47,443
it's a very easy step from here
to fully autonomous weapons
632
00:32:47,487 --> 00:32:50,881
that choose their own targets
and release their own missiles.
633
00:32:50,925 --> 00:32:58,150
♪♪
634
00:32:58,193 --> 00:33:05,374
♪♪
635
00:33:05,418 --> 00:33:12,686
♪♪
636
00:33:12,729 --> 00:33:15,080
The expected life-span
of a human being
637
00:33:15,123 --> 00:33:16,516
in that kind of
battle environment
638
00:33:16,559 --> 00:33:20,520
would be measured in seconds.
639
00:33:20,563 --> 00:33:23,740
Singer: At one point,
drones were science fiction,
640
00:33:23,784 --> 00:33:28,832
and now they've become
the normal thing in war.
641
00:33:28,876 --> 00:33:33,402
There's over 10,000 in
U.S. military inventory alone.
642
00:33:33,446 --> 00:33:35,274
But they're not
just a U.S. phenomena.
643
00:33:35,317 --> 00:33:39,060
There's more than 80 countries
that operate them.
644
00:33:39,104 --> 00:33:41,932
Gourley: It stands to reason
that people making some
645
00:33:41,976 --> 00:33:44,587
of the most important and
difficult decisions in the world
646
00:33:44,631 --> 00:33:46,328
are gonna start to use
and implement
647
00:33:46,372 --> 00:33:48,591
artificial intelligence.
648
00:33:48,635 --> 00:33:50,724
♪♪
649
00:33:50,767 --> 00:33:53,596
The Air Force just designed
a $400-billion jet program
650
00:33:53,640 --> 00:33:55,555
to put pilots in the sky,
651
00:33:55,598 --> 00:34:01,300
and a $500 AI, designed by
a couple of graduate students,
652
00:34:01,343 --> 00:34:03,432
is beating the best human pilots
653
00:34:03,476 --> 00:34:05,782
with a relatively
simple algorithm.
654
00:34:05,826 --> 00:34:09,395
♪♪
655
00:34:09,438 --> 00:34:13,399
AI will have as big an impact
on the military
656
00:34:13,442 --> 00:34:17,490
as the combustion engine
had at the turn of the century.
657
00:34:17,533 --> 00:34:18,839
It will literally touch
658
00:34:18,882 --> 00:34:21,233
everything
that the military does,
659
00:34:21,276 --> 00:34:25,324
from driverless convoys
delivering logistical supplies,
660
00:34:25,367 --> 00:34:27,021
to unmanned drones
661
00:34:27,065 --> 00:34:30,764
delivering medical aid,
to computational propaganda,
662
00:34:30,807 --> 00:34:34,246
trying to win the hearts
and minds of a population.
663
00:34:34,289 --> 00:34:38,337
And so it stands to reason
that whoever has the best AI
664
00:34:38,380 --> 00:34:41,688
will probably achieve
dominance on this planet.
665
00:34:45,561 --> 00:34:47,650
At some point in
the early 21st century,
666
00:34:47,694 --> 00:34:51,219
all of mankind was
united in celebration.
667
00:34:51,263 --> 00:34:53,830
We marveled
at our own magnificence
668
00:34:53,874 --> 00:34:56,833
as we gave birth to AI.
669
00:34:56,877 --> 00:34:58,966
AI?
670
00:34:59,009 --> 00:35:00,489
You mean
artificial intelligence?
671
00:35:00,533 --> 00:35:01,751
A singular consciousness
672
00:35:01,795 --> 00:35:05,886
that spawned
an entire race of machines.
673
00:35:05,929 --> 00:35:09,716
We don't know
who struck first -- us or them,
674
00:35:09,759 --> 00:35:12,980
but we know that it was us
that scorched the sky.
675
00:35:13,023 --> 00:35:14,634
[ Thunder crashes ]
676
00:35:14,677 --> 00:35:16,766
Singer: There's a long history
of science fiction,
677
00:35:16,810 --> 00:35:19,987
not just predicting the future,
but shaping the future.
678
00:35:20,030 --> 00:35:26,820
♪♪
679
00:35:26,863 --> 00:35:30,389
Arthur Conan Doyle
writing before World War I
680
00:35:30,432 --> 00:35:34,393
on the danger of how
submarines might be used
681
00:35:34,436 --> 00:35:38,048
to carry out civilian blockades.
682
00:35:38,092 --> 00:35:40,399
At the time
he's writing this fiction,
683
00:35:40,442 --> 00:35:43,402
the Royal Navy made fun
of Arthur Conan Doyle
684
00:35:43,445 --> 00:35:45,230
for this absurd idea
685
00:35:45,273 --> 00:35:47,623
that submarines
could be useful in war.
686
00:35:47,667 --> 00:35:53,412
♪♪
687
00:35:53,455 --> 00:35:55,370
One of the things
we've seen in history
688
00:35:55,414 --> 00:35:58,243
is that our attitude
towards technology,
689
00:35:58,286 --> 00:36:01,942
but also ethics,
are very context-dependent.
690
00:36:01,985 --> 00:36:03,726
For example, the submarine...
691
00:36:03,770 --> 00:36:06,468
nations like Great Britain
and even the United States
692
00:36:06,512 --> 00:36:09,863
found it horrifying
to use the submarine.
693
00:36:09,906 --> 00:36:13,214
In fact, the German use of the
submarine to carry out attacks
694
00:36:13,258 --> 00:36:18,480
was the reason why the United
States joined World War I.
695
00:36:18,524 --> 00:36:20,613
But move the timeline forward.
696
00:36:20,656 --> 00:36:23,529
Man: The United States
of America was suddenly
697
00:36:23,572 --> 00:36:28,403
and deliberately attacked
by the empire of Japan.
698
00:36:28,447 --> 00:36:32,190
Five hours after Pearl Harbor,
the order goes out
699
00:36:32,233 --> 00:36:36,498
to commit unrestricted
submarine warfare against Japan.
700
00:36:39,936 --> 00:36:44,289
So Arthur Conan Doyle
turned out to be right.
701
00:36:44,332 --> 00:36:46,856
Nolan: That's the great old line
about science fiction --
702
00:36:46,900 --> 00:36:48,336
It's a lie that tells the truth.
703
00:36:48,380 --> 00:36:51,470
Fellow executives,
it gives me great pleasure
704
00:36:51,513 --> 00:36:54,821
to introduce you to the future
of law enforcement...
705
00:36:54,864 --> 00:36:56,562
ED-209.
706
00:36:56,605 --> 00:37:03,612
♪♪
707
00:37:03,656 --> 00:37:05,919
This isn't just a question
of science fiction.
708
00:37:05,962 --> 00:37:09,488
This is about what's next, about
what's happening right now.
709
00:37:09,531 --> 00:37:13,927
♪♪
710
00:37:13,970 --> 00:37:17,496
The role of intelligent systems
is growing very rapidly
711
00:37:17,539 --> 00:37:19,324
in warfare.
712
00:37:19,367 --> 00:37:22,152
Everyone is pushing
in the unmanned realm.
713
00:37:22,196 --> 00:37:26,374
♪♪
714
00:37:26,418 --> 00:37:28,898
Gourley: Today, the Secretary of
Defense is very, very clear --
715
00:37:28,942 --> 00:37:32,337
We will not create fully
autonomous attacking vehicles.
716
00:37:32,380 --> 00:37:34,643
Not everyone
is gonna hold themselves
717
00:37:34,687 --> 00:37:36,515
to that same set of values.
718
00:37:36,558 --> 00:37:40,693
And when China and Russia start
deploying autonomous vehicles
719
00:37:40,736 --> 00:37:45,611
that can attack and kill, what's
the move that we're gonna make?
720
00:37:45,654 --> 00:37:49,963
♪♪
721
00:37:50,006 --> 00:37:51,617
Russell: You can't say,
"Well, we're gonna use
722
00:37:51,660 --> 00:37:53,967
autonomous weapons
for our military dominance,
723
00:37:54,010 --> 00:37:56,796
but no one else
is gonna use them."
724
00:37:56,839 --> 00:38:00,495
If you make these weapons,
they're gonna be used to attack
725
00:38:00,539 --> 00:38:03,324
human populations
in large numbers.
726
00:38:03,368 --> 00:38:12,507
♪♪
727
00:38:12,551 --> 00:38:14,596
Autonomous weapons are,
by their nature,
728
00:38:14,640 --> 00:38:16,468
weapons of mass destruction,
729
00:38:16,511 --> 00:38:19,862
because it doesn't need a human
being to guide it or carry it.
730
00:38:19,906 --> 00:38:22,517
You only need one person,
to, you know,
731
00:38:22,561 --> 00:38:25,781
write a little program.
732
00:38:25,825 --> 00:38:30,220
It just captures
the complexity of this field.
733
00:38:30,264 --> 00:38:32,571
It is cool.
It is important.
734
00:38:32,614 --> 00:38:34,573
It is amazing.
735
00:38:34,616 --> 00:38:37,053
It is also frightening.
736
00:38:37,097 --> 00:38:38,968
And it's all about trust.
737
00:38:42,102 --> 00:38:44,583
It's an open letter about
artificial intelligence,
738
00:38:44,626 --> 00:38:47,063
signed by some of
the biggest names in science.
739
00:38:47,107 --> 00:38:48,413
What do they want?
740
00:38:48,456 --> 00:38:50,763
Ban the use of
autonomous weapons.
741
00:38:50,806 --> 00:38:52,373
Woman: The author stated,
742
00:38:52,417 --> 00:38:54,375
"Autonomous weapons
have been described
743
00:38:54,419 --> 00:38:56,595
as the third revolution
in warfare."
744
00:38:56,638 --> 00:38:58,553
Woman #2: ...thousand
artificial-intelligence
specialists
745
00:38:58,597 --> 00:39:01,817
calling for a global ban
on killer robots.
746
00:39:01,861 --> 00:39:04,342
Tegmark:
This open letter basically says
747
00:39:04,385 --> 00:39:06,344
that we should redefine the goal
748
00:39:06,387 --> 00:39:07,954
of the field of
artificial intelligence
749
00:39:07,997 --> 00:39:11,610
away from just creating pure,
undirected intelligence,
750
00:39:11,653 --> 00:39:13,655
towards creating
beneficial intelligence.
751
00:39:13,699 --> 00:39:16,092
The development of AI
is not going to stop.
752
00:39:16,136 --> 00:39:18,094
It is going to continue
and get better.
753
00:39:18,138 --> 00:39:19,835
If the international community
754
00:39:19,879 --> 00:39:21,968
isn't putting
certain controls on this,
755
00:39:22,011 --> 00:39:24,666
people will develop things
that can do anything.
756
00:39:24,710 --> 00:39:27,365
Woman: The letter says
that we are years, not decades,
757
00:39:27,408 --> 00:39:28,714
away from these weapons
being deployed.
758
00:39:28,757 --> 00:39:30,106
So first of all...
759
00:39:30,150 --> 00:39:32,413
We had 6,000 signatories
of that letter,
760
00:39:32,457 --> 00:39:35,155
including many of
the major figures in the field.
761
00:39:37,026 --> 00:39:39,942
I'm getting a lot of visits
from high-ranking officials
762
00:39:39,986 --> 00:39:42,989
who wish to emphasize that
American military dominance
763
00:39:43,032 --> 00:39:45,731
is very important,
and autonomous weapons
764
00:39:45,774 --> 00:39:50,083
may be part of
the Defense Department's plan.
765
00:39:50,126 --> 00:39:52,433
That's very, very scary,
because a value system
766
00:39:52,477 --> 00:39:54,479
of military developers
of technology
767
00:39:54,522 --> 00:39:57,307
is not the same as a value
system of the human race.
768
00:39:57,351 --> 00:40:00,746
♪♪
769
00:40:00,789 --> 00:40:02,922
Markoff: Out of the concerns
about the possibility
770
00:40:02,965 --> 00:40:06,665
that this technology might be
a threat to human existence,
771
00:40:06,708 --> 00:40:08,144
a number of the technologists
772
00:40:08,188 --> 00:40:09,972
have funded
the Future of Life Institute
773
00:40:10,016 --> 00:40:12,192
to try to grapple
with these problems.
774
00:40:13,193 --> 00:40:14,847
All of these guys are secretive,
775
00:40:14,890 --> 00:40:16,805
and so it's interesting
to me to see them,
776
00:40:16,849 --> 00:40:20,635
you know, all together.
777
00:40:20,679 --> 00:40:24,030
Everything we have is a result
of our intelligence.
778
00:40:24,073 --> 00:40:26,641
It's not the result
of our big, scary teeth
779
00:40:26,685 --> 00:40:29,470
or our large claws
or our enormous muscles.
780
00:40:29,514 --> 00:40:32,473
It's because we're actually
relatively intelligent.
781
00:40:32,517 --> 00:40:35,520
And among my generation,
we're all having
782
00:40:35,563 --> 00:40:37,086
what we call "holy cow,"
783
00:40:37,130 --> 00:40:39,045
or "holy something else"
moments,
784
00:40:39,088 --> 00:40:41,003
because we see
that the technology
785
00:40:41,047 --> 00:40:44,180
is accelerating faster
than we expected.
786
00:40:44,224 --> 00:40:46,705
I remember sitting
around the table there
787
00:40:46,748 --> 00:40:50,099
with some of the best and
the smartest minds in the world,
788
00:40:50,143 --> 00:40:52,058
and what really
struck me was,
789
00:40:52,101 --> 00:40:56,149
maybe the human brain
is not able to fully grasp
790
00:40:56,192 --> 00:40:58,673
the complexity of the world
that we're confronted with.
791
00:40:58,717 --> 00:41:01,415
Russell:
As it's currently constructed,
792
00:41:01,459 --> 00:41:04,766
the road that AI is following
heads off a cliff,
793
00:41:04,810 --> 00:41:07,595
and we need to change
the direction that we're going
794
00:41:07,639 --> 00:41:10,729
so that we don't take
the human race off the cliff.
795
00:41:10,772 --> 00:41:13,514
[ Applause ]
796
00:41:13,558 --> 00:41:17,126
Musk: Google acquired DeepMind
several years ago.
797
00:41:17,170 --> 00:41:18,737
DeepMind operates
798
00:41:18,780 --> 00:41:22,088
as a semi-independent
subsidiary of Google.
799
00:41:22,131 --> 00:41:24,960
The thing that makes
DeepMind unique
800
00:41:25,004 --> 00:41:26,919
is that DeepMind
is absolutely focused
801
00:41:26,962 --> 00:41:30,313
on creating digital
superintelligence --
802
00:41:30,357 --> 00:41:34,056
an AI that is vastly smarter
than any human on Earth
803
00:41:34,100 --> 00:41:36,624
and ultimately smarter than
all humans on Earth combined.
804
00:41:36,668 --> 00:41:40,715
This is from the DeepMind
reinforcement learning system.
805
00:41:40,759 --> 00:41:43,544
Basically wakes up
like a newborn baby
806
00:41:43,588 --> 00:41:46,852
and is shown the screen
of an Atari video game
807
00:41:46,895 --> 00:41:50,508
and then has to learn
to play the video game.
808
00:41:50,551 --> 00:41:55,600
It knows nothing about objects,
about motion, about time.
809
00:41:57,602 --> 00:41:59,604
It only knows that there's
an image on the screen
810
00:41:59,647 --> 00:42:02,563
and there's a score.
811
00:42:02,607 --> 00:42:06,436
So, if your baby woke up
the day it was born
812
00:42:06,480 --> 00:42:08,090
and, by late afternoon,
813
00:42:08,134 --> 00:42:11,093
was playing
40 different Atari video games
814
00:42:11,137 --> 00:42:15,315
at a superhuman level,
you would be terrified.
815
00:42:15,358 --> 00:42:19,101
You would say, "My baby
is possessed. Send it back."
816
00:42:19,145 --> 00:42:23,584
Musk: The DeepMind system
can win at any game.
817
00:42:23,628 --> 00:42:27,588
It can already beat all
the original Atari games.
818
00:42:27,632 --> 00:42:29,155
It is superhuman.
819
00:42:29,198 --> 00:42:31,636
It plays the games at superspeed
in less than a minute.
820
00:42:31,679 --> 00:42:33,507
[ Video-game beeping ]
821
00:42:35,640 --> 00:42:37,032
♪♪
822
00:42:37,076 --> 00:42:38,643
DeepMind turned
to another challenge,
823
00:42:38,686 --> 00:42:40,558
and the challenge
was the game of Go,
824
00:42:40,601 --> 00:42:42,603
which people
have generally argued
825
00:42:42,647 --> 00:42:45,084
has been beyond
the power of computers
826
00:42:45,127 --> 00:42:48,304
to play with
the best human Go players.
827
00:42:48,348 --> 00:42:51,264
First, they challenged
a European Go champion.
828
00:42:53,222 --> 00:42:55,834
Then they challenged
a Korean Go champion.
829
00:42:55,877 --> 00:42:57,836
Man:
Please start the game.
830
00:42:57,879 --> 00:42:59,838
And they were able
to win both times
831
00:42:59,881 --> 00:43:02,797
in kind of striking fashion.
832
00:43:02,841 --> 00:43:05,017
Nolan: You were reading articles
in New York Timesyears ago
833
00:43:05,060 --> 00:43:09,761
talking about how Go would take
100 years for us to solve.
834
00:43:09,804 --> 00:43:11,110
Urban:
People said, "Well, you know,
835
00:43:11,153 --> 00:43:13,460
but that's still just a board.
836
00:43:13,503 --> 00:43:15,027
Poker is an art.
837
00:43:15,070 --> 00:43:16,419
Poker involves reading people.
838
00:43:16,463 --> 00:43:18,073
Poker involves lying
and bluffing.
839
00:43:18,117 --> 00:43:19,553
It's not an exact thing.
840
00:43:19,597 --> 00:43:21,381
That will never be,
you know, a computer.
841
00:43:21,424 --> 00:43:22,861
You can't do that."
842
00:43:22,904 --> 00:43:24,732
They took the best
poker players in the world,
843
00:43:24,776 --> 00:43:27,387
and it took seven days
for the computer
844
00:43:27,430 --> 00:43:30,520
to start demolishing the humans.
845
00:43:30,564 --> 00:43:32,261
So it's the best poker player
in the world,
846
00:43:32,305 --> 00:43:34,655
it's the best Go player in the
world, and the pattern here
847
00:43:34,699 --> 00:43:37,440
is that AI might take
a little while
848
00:43:37,484 --> 00:43:40,443
to wrap its tentacles
around a new skill,
849
00:43:40,487 --> 00:43:44,883
but when it does, when it
gets it, it is unstoppable.
850
00:43:44,926 --> 00:43:51,977
♪♪
851
00:43:52,020 --> 00:43:55,110
DeepMind's AI has
administrator-level access
852
00:43:55,154 --> 00:43:57,156
to Google's servers
853
00:43:57,199 --> 00:44:00,768
to optimize energy usage
at the data centers.
854
00:44:00,812 --> 00:44:04,816
However, this could be
an unintentional Trojan horse.
855
00:44:04,859 --> 00:44:07,253
DeepMind has to have complete
control of the data centers,
856
00:44:07,296 --> 00:44:08,950
so with a little
software update,
857
00:44:08,994 --> 00:44:10,691
that AI could take
complete control
858
00:44:10,735 --> 00:44:12,214
of the whole Google system,
859
00:44:12,258 --> 00:44:13,607
which means
they can do anything.
860
00:44:13,651 --> 00:44:14,913
They could look
at all your data.
861
00:44:14,956 --> 00:44:16,131
They could do anything.
862
00:44:16,175 --> 00:44:18,917
♪♪
863
00:44:20,135 --> 00:44:23,051
We're rapidly heading towards
digital superintelligence
864
00:44:23,095 --> 00:44:24,313
that far exceeds any human.
865
00:44:24,357 --> 00:44:26,402
I think it's very obvious.
866
00:44:26,446 --> 00:44:27,708
Barrat:
The problem is, we're not gonna
867
00:44:27,752 --> 00:44:29,710
suddenly hit
human-level intelligence
868
00:44:29,754 --> 00:44:33,105
and say,
"Okay, let's stop research."
869
00:44:33,148 --> 00:44:34,715
It's gonna go beyond
human-level intelligence
870
00:44:34,759 --> 00:44:36,195
into what's called
"superintelligence,"
871
00:44:36,238 --> 00:44:39,459
and that's anything
smarter than us.
872
00:44:39,502 --> 00:44:41,287
Tegmark:
AI at the superhuman level,
873
00:44:41,330 --> 00:44:42,810
if we succeed with that,
will be
874
00:44:42,854 --> 00:44:46,553
by far the most powerful
invention we've ever made
875
00:44:46,596 --> 00:44:50,296
and the last invention
we ever have to make.
876
00:44:50,339 --> 00:44:53,168
And if we create AI
that's smarter than us,
877
00:44:53,212 --> 00:44:54,735
we have to be open
to the possibility
878
00:44:54,779 --> 00:44:57,520
that we might actually
lose control to them.
879
00:44:57,564 --> 00:45:00,741
♪♪
880
00:45:00,785 --> 00:45:02,612
Russell: Let's say
you give it some objective,
881
00:45:02,656 --> 00:45:04,745
like curing cancer,
and then you discover
882
00:45:04,789 --> 00:45:06,965
that the way
it chooses to go about that
883
00:45:07,008 --> 00:45:08,444
is actually in conflict
884
00:45:08,488 --> 00:45:12,405
with a lot of other things
you care about.
885
00:45:12,448 --> 00:45:16,496
Musk: AI doesn't have to be evil
to destroy humanity.
886
00:45:16,539 --> 00:45:20,674
If AI has a goal, and humanity
just happens to be in the way,
887
00:45:20,718 --> 00:45:22,894
it will destroy humanity
as a matter of course,
888
00:45:22,937 --> 00:45:25,113
without even thinking about it.
No hard feelings.
889
00:45:25,157 --> 00:45:27,072
It's just like
if we're building a road
890
00:45:27,115 --> 00:45:29,770
and an anthill happens
to be in the way...
891
00:45:29,814 --> 00:45:31,467
We don't hate ants.
892
00:45:31,511 --> 00:45:33,165
We're just building a road.
893
00:45:33,208 --> 00:45:34,557
And so goodbye, anthill.
894
00:45:34,601 --> 00:45:37,952
♪♪
895
00:45:37,996 --> 00:45:40,172
It's tempting
to dismiss these concerns,
896
00:45:40,215 --> 00:45:42,783
'cause it's, like,
something that might happen
897
00:45:42,827 --> 00:45:47,396
in a few decades or 100 years,
so why worry?
898
00:45:47,440 --> 00:45:50,704
Russell: But if you go back
to September 11, 1933,
899
00:45:50,748 --> 00:45:52,401
Ernest Rutherford,
900
00:45:52,445 --> 00:45:54,795
who was the most well-known
nuclear physicist of his time,
901
00:45:54,839 --> 00:45:56,318
said that the possibility
902
00:45:56,362 --> 00:45:58,668
of ever extracting
useful amounts of energy
903
00:45:58,712 --> 00:46:00,801
from the transmutation
of atoms, as he called it,
904
00:46:00,845 --> 00:46:03,151
was moonshine.
905
00:46:03,195 --> 00:46:04,849
The next morning, Leo Szilard,
906
00:46:04,892 --> 00:46:06,502
who was a much
younger physicist,
907
00:46:06,546 --> 00:46:09,984
read this and got really annoyed
and figured out
908
00:46:10,028 --> 00:46:11,943
how to make
a nuclear chain reaction
909
00:46:11,986 --> 00:46:13,379
just a few months later.
910
00:46:13,422 --> 00:46:20,560
♪♪
911
00:46:20,603 --> 00:46:23,693
We have spent more
than $2 billion
912
00:46:23,737 --> 00:46:27,523
on the greatest
scientific gamble in history.
913
00:46:27,567 --> 00:46:30,222
Russell: So when people say
that, "Oh, this is so far off
914
00:46:30,265 --> 00:46:32,528
in the future, we don't have
to worry about it,"
915
00:46:32,572 --> 00:46:36,271
it might only be three, four
breakthroughs of that magnitude
916
00:46:36,315 --> 00:46:40,275
that will get us from here
to superintelligent machines.
917
00:46:40,319 --> 00:46:42,974
Tegmark: If it's gonna take
20 years to figure out
918
00:46:43,017 --> 00:46:45,237
how to keep AI beneficial,
919
00:46:45,280 --> 00:46:48,849
then we should start today,
not at the last second
920
00:46:48,893 --> 00:46:51,460
when some dudes
drinking Red Bull
921
00:46:51,504 --> 00:46:53,332
decide to flip the switch
and test the thing.
922
00:46:53,375 --> 00:46:56,770
♪♪
923
00:46:56,814 --> 00:46:58,859
Musk:
We have five years.
924
00:46:58,903 --> 00:47:00,600
I think
digital superintelligence
925
00:47:00,643 --> 00:47:03,864
will happen in my lifetime.
926
00:47:03,908 --> 00:47:05,735
100%.
927
00:47:05,779 --> 00:47:07,215
Barrat: When this happens,
928
00:47:07,259 --> 00:47:09,696
it will be surrounded
by a bunch of people
929
00:47:09,739 --> 00:47:13,091
who are really just excited
about the technology.
930
00:47:13,134 --> 00:47:15,571
They want to see it succeed,
but they're not anticipating
931
00:47:15,615 --> 00:47:16,964
that it can get out of control.
932
00:47:17,008 --> 00:47:24,450
♪♪
933
00:47:25,494 --> 00:47:28,584
Oh, my God, I trust
my computer so much.
934
00:47:28,628 --> 00:47:30,195
That's an amazing question.
935
00:47:30,238 --> 00:47:31,457
I don't trust
my computer.
936
00:47:31,500 --> 00:47:32,937
If it's on,
I take it off.
937
00:47:32,980 --> 00:47:34,242
Like, even when it's off,
938
00:47:34,286 --> 00:47:35,896
I still think it's on.
Like, you know?
939
00:47:35,940 --> 00:47:37,637
Like, you really cannot tru--
Like, the webcams,
940
00:47:37,680 --> 00:47:39,595
you don't know if, like,
someone might turn it...
941
00:47:39,639 --> 00:47:41,249
You don't know, like.
942
00:47:41,293 --> 00:47:42,903
I don't trust my computer.
943
00:47:42,947 --> 00:47:46,907
Like, in my phone,
every time they ask me
944
00:47:46,951 --> 00:47:49,475
"Can we send your
information to Apple?"
945
00:47:49,518 --> 00:47:50,998
every time, I...
946
00:47:51,042 --> 00:47:53,087
So, I don't trust my phone.
947
00:47:53,131 --> 00:47:56,743
Okay. So, part of it is,
yes, I do trust it,
948
00:47:56,786 --> 00:48:00,660
because it would be really
hard to get through the day
949
00:48:00,703 --> 00:48:04,011
in the way our world is
set up without computers.
950
00:48:04,055 --> 00:48:05,360
♪♪
951
00:48:05,404 --> 00:48:07,232
[ Monitor beeping ]
952
00:48:10,975 --> 00:48:13,368
Dr. Herman: Trust is
such a human experience.
953
00:48:13,412 --> 00:48:21,246
♪♪
954
00:48:21,289 --> 00:48:25,119
I have a patient coming in
with an intracranial aneurysm.
955
00:48:25,163 --> 00:48:29,994
♪♪
956
00:48:30,037 --> 00:48:31,691
They want to look
in my eyes and know
957
00:48:31,734 --> 00:48:34,955
that they can trust
this person with their life.
958
00:48:34,999 --> 00:48:39,394
I'm not horribly concerned
about anything.
959
00:48:39,438 --> 00:48:40,830
Good.
Part of that
960
00:48:40,874 --> 00:48:42,920
is because
I have confidence in you.
961
00:48:42,963 --> 00:48:50,710
♪♪
962
00:48:50,753 --> 00:48:52,233
This procedure
we're doing today
963
00:48:52,277 --> 00:48:57,151
20 years ago
was essentially impossible.
964
00:48:57,195 --> 00:49:00,328
We just didn't have the
materials and the technologies.
965
00:49:04,202 --> 00:49:13,385
♪♪
966
00:49:13,428 --> 00:49:22,655
♪♪
967
00:49:22,698 --> 00:49:26,485
So, the coil is barely
in there right now.
968
00:49:26,528 --> 00:49:29,923
It's just a feather
holding it in.
969
00:49:29,967 --> 00:49:32,012
It's nervous time.
970
00:49:32,056 --> 00:49:36,147
♪♪
971
00:49:36,190 --> 00:49:37,626
We're just in purgatory,
972
00:49:37,670 --> 00:49:40,673
intellectual,
humanistic purgatory,
973
00:49:40,716 --> 00:49:43,632
and AI might know
exactly what to do here.
974
00:49:43,676 --> 00:49:50,596
♪♪
975
00:49:50,639 --> 00:49:52,554
We've got the coil
into the aneurysm.
976
00:49:52,598 --> 00:49:54,556
But it wasn't in
tremendously well
977
00:49:54,600 --> 00:49:56,428
that I knew that it would stay,
978
00:49:56,471 --> 00:50:01,041
so with a maybe 20% risk
of a very bad situation,
979
00:50:01,085 --> 00:50:04,436
I elected
to just bring her back.
980
00:50:04,479 --> 00:50:05,959
Because of my relationship
with her
981
00:50:06,003 --> 00:50:08,222
and knowing the difficulties
of coming in
982
00:50:08,266 --> 00:50:11,051
and having the procedure,
I consider things,
983
00:50:11,095 --> 00:50:14,272
when I should only consider
the safest possible route
984
00:50:14,315 --> 00:50:16,361
to achieve success.
985
00:50:16,404 --> 00:50:19,755
But I had to stand there for
10 minutes agonizing about it.
986
00:50:19,799 --> 00:50:21,757
The computer feels nothing.
987
00:50:21,801 --> 00:50:24,760
The computer just does
what it's supposed to do,
988
00:50:24,804 --> 00:50:26,284
better and better.
989
00:50:26,327 --> 00:50:30,288
♪♪
990
00:50:30,331 --> 00:50:32,551
I want to be AI in this case.
991
00:50:35,945 --> 00:50:38,861
But can AI be compassionate?
992
00:50:38,905 --> 00:50:43,040
♪♪
993
00:50:43,083 --> 00:50:47,827
I mean, it's everybody's
question about AI.
994
00:50:47,870 --> 00:50:51,961
We are the sole
embodiment of humanity,
995
00:50:52,005 --> 00:50:55,269
and it's a stretch for us
to accept that a machine
996
00:50:55,313 --> 00:50:58,794
can be compassionate
and loving in that way.
997
00:50:58,838 --> 00:51:05,105
♪♪
998
00:51:05,149 --> 00:51:07,281
Part of me
doesn't believe in magic,
999
00:51:07,325 --> 00:51:09,805
but part of me has faith
that there is something
1000
00:51:09,849 --> 00:51:11,546
beyond the sum of the parts,
1001
00:51:11,590 --> 00:51:15,637
that there is at least a oneness
in our shared ancestry,
1002
00:51:15,681 --> 00:51:20,338
our shared biology,
our shared history.
1003
00:51:20,381 --> 00:51:23,210
Some connection there
beyond machine.
1004
00:51:23,254 --> 00:51:30,304
♪♪
1005
00:51:30,348 --> 00:51:32,567
So, then, you have
the other side of that, is,
1006
00:51:32,611 --> 00:51:34,047
does the computer
know it's conscious,
1007
00:51:34,091 --> 00:51:37,137
or can it be conscious,
or does it care?
1008
00:51:37,181 --> 00:51:40,009
Does it need to be conscious?
1009
00:51:40,053 --> 00:51:42,011
Does it need to be aware?
1010
00:51:42,055 --> 00:51:47,365
♪♪
1011
00:51:47,408 --> 00:51:52,848
♪♪
1012
00:51:52,892 --> 00:51:56,417
I do not think that a robot
could ever be conscious.
1013
00:51:56,461 --> 00:51:58,376
Unless they programmed it
that way.
1014
00:51:58,419 --> 00:52:00,639
Conscious? No.
1015
00:52:00,682 --> 00:52:03,163
No.
No.
1016
00:52:03,207 --> 00:52:06,035
I mean, think a robot could be
programmed to be conscious.
1017
00:52:06,079 --> 00:52:09,648
How are they programmed
to do everything else?
1018
00:52:09,691 --> 00:52:12,390
That's another big part
of artificial intelligence,
1019
00:52:12,433 --> 00:52:15,741
is to make them conscious
and make them feel.
1020
00:52:17,003 --> 00:52:22,400
♪♪
1021
00:52:22,443 --> 00:52:26,230
Lipson: Back in 2005, we started
trying to build machines
1022
00:52:26,273 --> 00:52:27,709
with self-awareness.
1023
00:52:27,753 --> 00:52:33,062
♪♪
1024
00:52:33,106 --> 00:52:37,284
This robot, to begin with,
didn't know what it was.
1025
00:52:37,328 --> 00:52:40,244
All it knew was that it needed
to do something like walk.
1026
00:52:40,287 --> 00:52:44,073
♪♪
1027
00:52:44,117 --> 00:52:45,597
Through trial and error,
1028
00:52:45,640 --> 00:52:49,731
it figured out how to walk
using its imagination,
1029
00:52:49,775 --> 00:52:54,040
and then it walked away.
1030
00:52:54,083 --> 00:52:56,390
And then we did
something very cruel.
1031
00:52:56,434 --> 00:52:58,653
We chopped off a leg
and watched what happened.
1032
00:52:58,697 --> 00:53:03,005
♪♪
1033
00:53:03,049 --> 00:53:07,749
At the beginning, it didn't
quite know what had happened.
1034
00:53:07,793 --> 00:53:13,233
But over about a period
of a day, it then began to limp.
1035
00:53:13,277 --> 00:53:16,845
And then, a year ago,
we were training an AI system
1036
00:53:16,889 --> 00:53:20,240
for a live demonstration.
1037
00:53:20,284 --> 00:53:21,763
We wanted to show how we wave
1038
00:53:21,807 --> 00:53:24,113
all these objects
in front of the camera
1039
00:53:24,157 --> 00:53:27,334
and the AI could
recognize the objects.
1040
00:53:27,378 --> 00:53:29,031
And so, we're preparing
this demo,
1041
00:53:29,075 --> 00:53:31,251
and we had on a side screen
this ability
1042
00:53:31,295 --> 00:53:36,778
to watch what certain
neurons were responding to.
1043
00:53:36,822 --> 00:53:39,041
And suddenly we noticed
that one of the neurons
1044
00:53:39,085 --> 00:53:41,087
was tracking faces.
1045
00:53:41,130 --> 00:53:45,483
It was tracking our faces
as we were moving around.
1046
00:53:45,526 --> 00:53:48,616
Now, the spooky thing about this
is that we never trained
1047
00:53:48,660 --> 00:53:52,490
the system
to recognize human faces,
1048
00:53:52,533 --> 00:53:55,710
and yet, somehow,
it learned to do that.
1049
00:53:57,973 --> 00:53:59,584
Even though these robots
are very simple,
1050
00:53:59,627 --> 00:54:02,500
we can see there's
something else going on there.
1051
00:54:02,543 --> 00:54:05,851
It's not just programming.
1052
00:54:05,894 --> 00:54:08,462
So, this is just the beginning.
1053
00:54:10,377 --> 00:54:14,294
Horvitz: I often think about
that beach in Kitty Hawk,
1054
00:54:14,338 --> 00:54:18,255
the 1903 flight
by Orville and Wilbur Wright.
1055
00:54:21,214 --> 00:54:24,348
It was kind of a canvas plane,
and it's wood and iron,
1056
00:54:24,391 --> 00:54:26,828
and it gets off the ground for,
what, a minute and 20 seconds,
1057
00:54:26,872 --> 00:54:29,091
on this windy day
1058
00:54:29,135 --> 00:54:31,006
before touching back down again.
1059
00:54:33,270 --> 00:54:37,143
And it was
just around 65 summers or so
1060
00:54:37,186 --> 00:54:43,149
after that moment that you have
a 747 taking off from JFK...
1061
00:54:43,192 --> 00:54:50,156
♪♪
1062
00:54:50,199 --> 00:54:51,984
...where a major concern
of someone on the airplane
1063
00:54:52,027 --> 00:54:55,422
might be whether or not
their salt-free diet meal
1064
00:54:55,466 --> 00:54:56,902
is gonna be coming to them
or not.
1065
00:54:56,945 --> 00:54:58,469
We have a whole infrastructure,
1066
00:54:58,512 --> 00:55:01,385
with travel agents
and tower control,
1067
00:55:01,428 --> 00:55:03,778
and it's all casual,
and it's all part of the world.
1068
00:55:03,822 --> 00:55:07,042
♪♪
1069
00:55:07,086 --> 00:55:09,523
Right now, as far
as we've come with machines
1070
00:55:09,567 --> 00:55:12,134
that think and solve problems,
we're at Kitty Hawk now.
1071
00:55:12,178 --> 00:55:13,745
We're in the wind.
1072
00:55:13,788 --> 00:55:17,052
We have our tattered-canvas
planes up in the air.
1073
00:55:17,096 --> 00:55:20,882
♪♪
1074
00:55:20,926 --> 00:55:23,885
But what happens
in 65 summers or so?
1075
00:55:23,929 --> 00:55:27,889
We will have machines
that are beyond human control.
1076
00:55:27,933 --> 00:55:30,457
Should we worry about that?
1077
00:55:30,501 --> 00:55:32,590
♪♪
1078
00:55:32,633 --> 00:55:34,853
I'm not sure it's going to help.
1079
00:55:40,337 --> 00:55:44,036
Kaplan: Nobody has any idea
today what it means for a robot
1080
00:55:44,079 --> 00:55:46,430
to be conscious.
1081
00:55:46,473 --> 00:55:48,649
There is no such thing.
1082
00:55:48,693 --> 00:55:50,172
There are a lot of smart people,
1083
00:55:50,216 --> 00:55:53,088
and I have a great deal
of respect for them,
1084
00:55:53,132 --> 00:55:57,528
but the truth is, machines
are natural psychopaths.
1085
00:55:57,571 --> 00:55:59,225
Man:
Fear came back into the market.
1086
00:55:59,268 --> 00:56:01,706
Man #2: Went down 800,
nearly 1,000, in a heartbeat.
1087
00:56:01,749 --> 00:56:03,360
I mean,
it is classic capitulation.
1088
00:56:03,403 --> 00:56:04,796
There are some people
who are proposing
1089
00:56:04,839 --> 00:56:07,146
it was some kind
of fat-finger error.
1090
00:56:07,189 --> 00:56:09,583
Take the Flash Crash of 2010.
1091
00:56:09,627 --> 00:56:13,413
In a matter of minutes,
$1 trillion in value
1092
00:56:13,457 --> 00:56:15,415
was lost in the stock market.
1093
00:56:15,459 --> 00:56:18,984
Woman: The Dow dropped nearly
1,000 points in a half-hour.
1094
00:56:19,027 --> 00:56:22,553
Kaplan:
So, what went wrong?
1095
00:56:22,596 --> 00:56:26,644
By that point in time,
more than 60% of all the trades
1096
00:56:26,687 --> 00:56:29,124
that took place
on the stock exchange
1097
00:56:29,168 --> 00:56:32,693
were actually being
initiated by computers.
1098
00:56:32,737 --> 00:56:34,216
Man:
Panic selling on the way down,
1099
00:56:34,260 --> 00:56:35,783
and all of a sudden
it stopped on a dime.
1100
00:56:35,827 --> 00:56:37,611
Man #2: This is all happening
in real time, folks.
1101
00:56:37,655 --> 00:56:39,526
Wisz: The short story of what
happened in the Flash Crash
1102
00:56:39,570 --> 00:56:42,399
is that algorithms
responded to algorithms,
1103
00:56:42,442 --> 00:56:45,358
and it compounded upon itself
over and over and over again
1104
00:56:45,402 --> 00:56:47,012
in a matter of minutes.
1105
00:56:47,055 --> 00:56:50,972
Man: At one point, the market
fell as if down a well.
1106
00:56:51,016 --> 00:56:54,323
There is no regulatory body
that can adapt quickly enough
1107
00:56:54,367 --> 00:56:57,979
to prevent potentially
disastrous consequences
1108
00:56:58,023 --> 00:57:01,243
of AI operating
in our financial systems.
1109
00:57:01,287 --> 00:57:03,898
They are so prime
for manipulation.
1110
00:57:03,942 --> 00:57:05,639
Let's talk about the speed
with which
1111
00:57:05,683 --> 00:57:08,076
we are watching
this market deteriorate.
1112
00:57:08,120 --> 00:57:11,602
That's the type of AI-run-amuck
that scares people.
1113
00:57:11,645 --> 00:57:13,560
Kaplan:
When you give them a goal,
1114
00:57:13,604 --> 00:57:17,825
they will relentlessly
pursue that goal.
1115
00:57:17,869 --> 00:57:20,393
How many computer programs
are there like this?
1116
00:57:20,437 --> 00:57:23,483
Nobody knows.
1117
00:57:23,527 --> 00:57:27,444
Kosinski: One of the fascinating
aspects about AI in general
1118
00:57:27,487 --> 00:57:31,970
is that no one really
understands how it works.
1119
00:57:32,013 --> 00:57:36,975
Even the people who create AI
don't really fully understand.
1120
00:57:37,018 --> 00:57:39,804
Because it has millions
of elements,
1121
00:57:39,847 --> 00:57:41,675
it becomes completely impossible
1122
00:57:41,719 --> 00:57:45,113
for a human being
to understand what's going on.
1123
00:57:45,157 --> 00:57:52,512
♪♪
1124
00:57:52,556 --> 00:57:56,037
Grassegger: Microsoft had set up
this artificial intelligence
1125
00:57:56,081 --> 00:57:59,127
called Tay on Twitter,
which was a chatbot.
1126
00:58:00,912 --> 00:58:02,696
They started out in the morning,
1127
00:58:02,740 --> 00:58:06,526
and Tay was starting to tweet
and learning from stuff
1128
00:58:06,570 --> 00:58:10,835
that was being sent to him
from other Twitter people.
1129
00:58:10,878 --> 00:58:13,272
Because some people,
like trolls, attacked him,
1130
00:58:13,315 --> 00:58:18,582
within 24 hours, the Microsoft
bot became a terrible person.
1131
00:58:18,625 --> 00:58:21,367
They had to literally
pull Tay off the Net
1132
00:58:21,410 --> 00:58:24,718
because he had turned
into a monster.
1133
00:58:24,762 --> 00:58:30,550
A misanthropic, racist, horrible
person you'd never want to meet.
1134
00:58:30,594 --> 00:58:32,857
And nobody had foreseen this.
1135
00:58:35,337 --> 00:58:38,602
The whole idea of AI is that
we are not telling it exactly
1136
00:58:38,645 --> 00:58:42,780
how to achieve a given
outcome or a goal.
1137
00:58:42,823 --> 00:58:46,435
AI develops on its own.
1138
00:58:46,479 --> 00:58:48,829
Nolan: We're worried about
superintelligent AI,
1139
00:58:48,873 --> 00:58:52,790
the master chess player
that will outmaneuver us,
1140
00:58:52,833 --> 00:58:55,923
but AI won't have to
actually be that smart
1141
00:58:55,967 --> 00:59:00,145
to have massively disruptive
effects on human civilization.
1142
00:59:00,188 --> 00:59:01,886
We've seen over the last century
1143
00:59:01,929 --> 00:59:05,150
it doesn't necessarily take
a genius to knock history off
1144
00:59:05,193 --> 00:59:06,804
in a particular direction,
1145
00:59:06,847 --> 00:59:09,589
and it won't take a genius AI
to do the same thing.
1146
00:59:09,633 --> 00:59:13,158
Bogus election news stories
generated more engagement
1147
00:59:13,201 --> 00:59:17,075
on Facebook
than top real stories.
1148
00:59:17,118 --> 00:59:21,079
Facebook really is
the elephant in the room.
1149
00:59:21,122 --> 00:59:23,777
Kosinski:
AI running Facebook news feed --
1150
00:59:23,821 --> 00:59:28,347
The task for AI
is keeping users engaged,
1151
00:59:28,390 --> 00:59:29,827
but no one really understands
1152
00:59:29,870 --> 00:59:34,832
exactly how this AI
is achieving this goal.
1153
00:59:34,875 --> 00:59:38,792
Nolan: Facebook is building an
elegant mirrored wall around us.
1154
00:59:38,836 --> 00:59:41,665
A mirror that we can ask,
"Who's the fairest of them all?"
1155
00:59:41,708 --> 00:59:45,016
and it will answer, "You, you,"
time and again
1156
00:59:45,059 --> 00:59:48,193
and slowly begin
to warp our sense of reality,
1157
00:59:48,236 --> 00:59:53,502
warp our sense of politics,
history, global events,
1158
00:59:53,546 --> 00:59:57,028
until determining what's true
and what's not true,
1159
00:59:57,071 --> 00:59:58,943
is virtually impossible.
1160
01:00:01,032 --> 01:00:03,861
The problem is that AI
doesn't understand that.
1161
01:00:03,904 --> 01:00:08,039
AI just had a mission --
maximize user engagement,
1162
01:00:08,082 --> 01:00:10,041
and it achieved that.
1163
01:00:10,084 --> 01:00:13,653
Nearly 2 billion people
spend nearly one hour
1164
01:00:13,697 --> 01:00:17,831
on average a day
basically interacting with AI
1165
01:00:17,875 --> 01:00:21,530
that is shaping
their experience.
1166
01:00:21,574 --> 01:00:24,664
Even Facebook engineers,
they don't like fake news.
1167
01:00:24,708 --> 01:00:26,666
It's very bad business.
1168
01:00:26,710 --> 01:00:28,015
They want to get rid
of fake news.
1169
01:00:28,059 --> 01:00:29,974
It's just very difficult
to do because,
1170
01:00:30,017 --> 01:00:32,324
how do you recognize news
as fake
1171
01:00:32,367 --> 01:00:34,456
if you cannot read
all of those news personally?
1172
01:00:34,500 --> 01:00:39,418
There's so much
active misinformation
1173
01:00:39,461 --> 01:00:41,115
and it's packaged very well,
1174
01:00:41,159 --> 01:00:44,553
and it looks the same when
you see it on a Facebook page
1175
01:00:44,597 --> 01:00:47,426
or you turn on your television.
1176
01:00:47,469 --> 01:00:49,210
Nolan:
It's not terribly sophisticated,
1177
01:00:49,254 --> 01:00:51,691
but it is terribly powerful.
1178
01:00:51,735 --> 01:00:54,346
And what it means is
that your view of the world,
1179
01:00:54,389 --> 01:00:56,435
which, 20 years ago,
was determined,
1180
01:00:56,478 --> 01:01:00,004
if you watched the nightly news,
by three different networks,
1181
01:01:00,047 --> 01:01:02,528
the three anchors who endeavored
to try to get it right.
1182
01:01:02,571 --> 01:01:04,225
Might have had a little bias
one way or the other,
1183
01:01:04,269 --> 01:01:05,923
but, largely speaking,
we could all agree
1184
01:01:05,966 --> 01:01:08,273
on an objective reality.
1185
01:01:08,316 --> 01:01:10,754
Well, that objectivity is gone,
1186
01:01:10,797 --> 01:01:13,757
and Facebook has
completely annihilated it.
1187
01:01:13,800 --> 01:01:17,064
♪♪
1188
01:01:17,108 --> 01:01:19,197
If most of your understanding
of how the world works
1189
01:01:19,240 --> 01:01:20,807
is derived from Facebook,
1190
01:01:20,851 --> 01:01:23,418
facilitated
by algorithmic software
1191
01:01:23,462 --> 01:01:27,118
that tries to show you
the news you want to see,
1192
01:01:27,161 --> 01:01:28,815
that's a terribly
dangerous thing.
1193
01:01:28,859 --> 01:01:33,080
And the idea that we have not
only set that in motion,
1194
01:01:33,124 --> 01:01:37,258
but allowed bad-faith actors
access to that information...
1195
01:01:37,302 --> 01:01:39,565
I mean, this is a recipe
for disaster.
1196
01:01:39,608 --> 01:01:43,134
♪♪
1197
01:01:43,177 --> 01:01:45,876
Urban: I think that there will
definitely be lots of bad actors
1198
01:01:45,919 --> 01:01:48,922
trying to manipulate the world
with AI.
1199
01:01:48,966 --> 01:01:52,143
2016 was a perfect example
of an election
1200
01:01:52,186 --> 01:01:55,015
where there was lots of AI
producing lots of fake news
1201
01:01:55,059 --> 01:01:58,323
and distributing it
for a purpose, for a result.
1202
01:01:58,366 --> 01:01:59,846
[ Applause ]
1203
01:01:59,890 --> 01:02:02,283
Ladies and gentlemen,
honorable colleagues...
1204
01:02:02,327 --> 01:02:04,546
it's my privilege
to speak to you today
1205
01:02:04,590 --> 01:02:07,985
about the power of big data
and psychographics
1206
01:02:08,028 --> 01:02:09,682
in the electoral process
1207
01:02:09,726 --> 01:02:12,206
and, specifically,
to talk about the work
1208
01:02:12,250 --> 01:02:14,513
that we contributed
to Senator Cruz's
1209
01:02:14,556 --> 01:02:16,558
presidential primary campaign.
1210
01:02:16,602 --> 01:02:19,910
Nolan: Cambridge Analytica
emerged quietly as a company
1211
01:02:19,953 --> 01:02:21,563
that, according to its own hype,
1212
01:02:21,607 --> 01:02:26,307
has the ability to use
this tremendous amount of data
1213
01:02:26,351 --> 01:02:30,137
in order
to effect societal change.
1214
01:02:30,181 --> 01:02:33,358
In 2016, they had
three major clients.
1215
01:02:33,401 --> 01:02:34,794
Ted Cruz was one of them.
1216
01:02:34,838 --> 01:02:37,884
It's easy to forget
that, only 18 months ago,
1217
01:02:37,928 --> 01:02:41,148
Senator Cruz was one of
the less popular candidates
1218
01:02:41,192 --> 01:02:42,846
seeking nomination.
1219
01:02:42,889 --> 01:02:47,241
So, what was not possible maybe,
like, 10 or 15 years ago,
1220
01:02:47,285 --> 01:02:49,374
was that you can send fake news
1221
01:02:49,417 --> 01:02:52,420
to exactly the people
that you want to send it to.
1222
01:02:52,464 --> 01:02:56,685
And then you could actually see
how he or she reacts on Facebook
1223
01:02:56,729 --> 01:02:58,905
and then adjust that information
1224
01:02:58,949 --> 01:03:01,778
according to the feedback
that you got.
1225
01:03:01,821 --> 01:03:03,257
So you can start developing
1226
01:03:03,301 --> 01:03:06,130
kind of a real-time management
of a population.
1227
01:03:06,173 --> 01:03:08,697
In this case, we've zoned in
1228
01:03:08,741 --> 01:03:10,699
on a group
we've called "Persuasion."
1229
01:03:10,743 --> 01:03:13,746
These are people who are
definitely going to vote,
1230
01:03:13,790 --> 01:03:16,705
to caucus, but they need
moving from the center
1231
01:03:16,749 --> 01:03:18,490
a little bit more
towards the right.
1232
01:03:18,533 --> 01:03:19,708
in order to support Cruz.
1233
01:03:19,752 --> 01:03:22,059
They need a persuasion message.
1234
01:03:22,102 --> 01:03:23,800
"Gun rights," I've selected.
1235
01:03:23,843 --> 01:03:25,802
That narrows the field
slightly more.
1236
01:03:25,845 --> 01:03:29,066
And now we know that we need
a message on gun rights,
1237
01:03:29,109 --> 01:03:31,111
it needs to be
a persuasion message,
1238
01:03:31,155 --> 01:03:32,591
and it needs to be nuanced
1239
01:03:32,634 --> 01:03:34,201
according to
the certain personality
1240
01:03:34,245 --> 01:03:36,029
that we're interested in.
1241
01:03:36,073 --> 01:03:39,946
Through social media, there's an
infinite amount of information
1242
01:03:39,990 --> 01:03:42,514
that you can gather
about a person.
1243
01:03:42,557 --> 01:03:45,734
We have somewhere close
to 4,000 or 5,000 data points
1244
01:03:45,778 --> 01:03:48,563
on every adult
in the United States.
1245
01:03:48,607 --> 01:03:51,915
Grassegger: It's about targeting
the individual.
1246
01:03:51,958 --> 01:03:54,352
It's like a weapon,
which can be used
1247
01:03:54,395 --> 01:03:55,962
in the totally wrong direction.
1248
01:03:56,006 --> 01:03:58,051
That's the problem
with all of this data.
1249
01:03:58,095 --> 01:04:02,229
It's almost as if we built the
bullet before we built the gun.
1250
01:04:02,273 --> 01:04:04,362
Ted Cruz employed our data,
1251
01:04:04,405 --> 01:04:06,407
our behavioral insights.
1252
01:04:06,451 --> 01:04:09,541
He started from a base
of less than 5%
1253
01:04:09,584 --> 01:04:15,590
and had a very slow-and-steady-
but-firm rise to above 35%,
1254
01:04:15,634 --> 01:04:17,157
making him, obviously,
1255
01:04:17,201 --> 01:04:20,465
the second most threatening
contender in the race.
1256
01:04:20,508 --> 01:04:23,120
Now, clearly, the Cruz
campaign is over now,
1257
01:04:23,163 --> 01:04:24,904
but what I can tell you
1258
01:04:24,948 --> 01:04:28,168
is that of the two candidates
left in this election,
1259
01:04:28,212 --> 01:04:30,867
one of them is using
these technologies.
1260
01:04:32,564 --> 01:04:35,959
I, Donald John Trump,
do solemnly swear
1261
01:04:36,002 --> 01:04:38,222
that I will faithfully execute
1262
01:04:38,265 --> 01:04:42,226
the office of President
of the United States.
1263
01:04:42,269 --> 01:04:46,273
♪♪
1264
01:04:48,275 --> 01:04:50,234
Nolan: Elections are
a marginal exercise.
1265
01:04:50,277 --> 01:04:53,237
It doesn't take
a very sophisticated AI
1266
01:04:53,280 --> 01:04:57,719
in order to have
a disproportionate impact.
1267
01:04:57,763 --> 01:05:02,550
Before Trump, Brexit was
another supposed client.
1268
01:05:02,594 --> 01:05:04,726
Well, at 20 minutes to 5:00,
1269
01:05:04,770 --> 01:05:08,730
we can now say
the decision taken in 1975
1270
01:05:08,774 --> 01:05:10,950
by this country to join
the common market
1271
01:05:10,994 --> 01:05:15,999
has been reversed by this
referendum to leave the EU.
1272
01:05:16,042 --> 01:05:19,828
Nolan: Cambridge Analytica
allegedly uses AI
1273
01:05:19,872 --> 01:05:23,267
to push through two of
the most ground-shaking pieces
1274
01:05:23,310 --> 01:05:27,967
of political change
in the last 50 years.
1275
01:05:28,011 --> 01:05:30,709
These are epochal events,
and if we believe the hype,
1276
01:05:30,752 --> 01:05:33,755
they are connected directly
to a piece of software,
1277
01:05:33,799 --> 01:05:37,194
essentially, created
by a professor at Stanford.
1278
01:05:37,237 --> 01:05:41,415
♪♪
1279
01:05:41,459 --> 01:05:43,635
Kosinski:
Back in 2013, I described
1280
01:05:43,678 --> 01:05:45,593
that what they are doing
is possible
1281
01:05:45,637 --> 01:05:49,293
and warned against this
happening in the future.
1282
01:05:49,336 --> 01:05:51,382
Grassegger:
At the time, Michal Kosinski
1283
01:05:51,425 --> 01:05:52,949
was a young Polish researcher
1284
01:05:52,992 --> 01:05:54,994
working at the
Psychometrics Centre.
1285
01:05:55,038 --> 01:06:00,217
So, what Michal had done was to
gather the largest-ever data set
1286
01:06:00,260 --> 01:06:03,481
of how people
behave on Facebook.
1287
01:06:03,524 --> 01:06:07,789
Kosinski:
Psychometrics is trying
to measure psychological traits,
1288
01:06:07,833 --> 01:06:09,922
such as personality,
intelligence,
1289
01:06:09,966 --> 01:06:11,880
political views, and so on.
1290
01:06:11,924 --> 01:06:15,058
Now, traditionally,
those traits were measured
1291
01:06:15,101 --> 01:06:17,712
using tests and questions.
1292
01:06:17,756 --> 01:06:19,410
Nolan: Personality test --
the most benign thing
1293
01:06:19,453 --> 01:06:20,715
you could possibly think of.
1294
01:06:20,759 --> 01:06:22,065
Something that doesn't
necessarily have
1295
01:06:22,108 --> 01:06:24,197
a lot of utility, right?
1296
01:06:24,241 --> 01:06:27,331
Kosinski: Our idea was that
instead of tests and questions,
1297
01:06:27,374 --> 01:06:30,029
we could simply look at the
digital footprints of behaviors
1298
01:06:30,073 --> 01:06:32,553
that we are all leaving behind
1299
01:06:32,597 --> 01:06:34,903
to understand openness,
1300
01:06:34,947 --> 01:06:37,732
conscientiousness,
neuroticism.
1301
01:06:37,776 --> 01:06:39,560
Grassegger: You can easily buy
personal data,
1302
01:06:39,604 --> 01:06:43,129
such as where you live, what
club memberships you've tried,
1303
01:06:43,173 --> 01:06:45,044
which gym you go to.
1304
01:06:45,088 --> 01:06:47,873
There are actually marketplaces
for personal data.
1305
01:06:47,916 --> 01:06:49,918
Nolan: It turns out, we can
discover an awful lot
1306
01:06:49,962 --> 01:06:51,442
about what you're gonna do
1307
01:06:51,485 --> 01:06:55,750
based on a very, very tiny
set of information.
1308
01:06:55,794 --> 01:06:58,275
Kosinski: We are training
deep-learning networks
1309
01:06:58,318 --> 01:07:01,278
to infer intimate traits,
1310
01:07:01,321 --> 01:07:04,759
people's political views,
personality,
1311
01:07:04,803 --> 01:07:07,806
intelligence,
sexual orientation
1312
01:07:07,849 --> 01:07:10,504
just from an image
from someone's face.
1313
01:07:10,548 --> 01:07:17,033
♪♪
1314
01:07:17,076 --> 01:07:20,645
Now think about countries which
are not so free and open-minded.
1315
01:07:20,688 --> 01:07:23,300
If you can reveal people's
religious views
1316
01:07:23,343 --> 01:07:25,954
or political views
or sexual orientation
1317
01:07:25,998 --> 01:07:28,740
based on only profile pictures,
1318
01:07:28,783 --> 01:07:33,310
this could be literally
an issue of life and death.
1319
01:07:33,353 --> 01:07:36,965
♪♪
1320
01:07:37,009 --> 01:07:39,751
I think there's no going back.
1321
01:07:42,145 --> 01:07:44,321
Do you know what
the Turing test is?
1322
01:07:44,364 --> 01:07:48,977
It's when a human interacts
with a computer,
1323
01:07:49,021 --> 01:07:50,805
and if the human doesn't know
they're interacting
1324
01:07:50,849 --> 01:07:52,546
with a computer,
1325
01:07:52,590 --> 01:07:54,026
the test is passed.
1326
01:07:54,070 --> 01:07:57,247
And over the next few days,
1327
01:07:57,290 --> 01:07:59,684
you're gonna be the human
component in a Turing test.
1328
01:07:59,727 --> 01:08:02,295
Holy shit.Yeah, that's right, Caleb.
1329
01:08:02,339 --> 01:08:04,080
You got it.
1330
01:08:04,123 --> 01:08:06,865
'Cause if that test
is passed,
1331
01:08:06,908 --> 01:08:10,825
you are dead center of
the greatest scientific event
1332
01:08:10,869 --> 01:08:12,958
in the history of man.
1333
01:08:13,001 --> 01:08:14,612
If you've created
a conscious machine,
1334
01:08:14,655 --> 01:08:17,615
it's not the history
of man--
1335
01:08:17,658 --> 01:08:19,356
That's the history
of gods.
1336
01:08:19,399 --> 01:08:26,798
♪♪
1337
01:08:26,841 --> 01:08:28,452
Nolan: It's almost like
technology is a god
1338
01:08:28,495 --> 01:08:29,975
in and of itself.
1339
01:08:30,018 --> 01:08:33,152
♪♪
1340
01:08:33,196 --> 01:08:35,241
Like the weather.
We can't impact it.
1341
01:08:35,285 --> 01:08:39,593
We can't slow it down.
We can't stop it.
1342
01:08:39,637 --> 01:08:43,249
We feel powerless.
1343
01:08:43,293 --> 01:08:44,685
Kurzweil:
If we think of God
1344
01:08:44,729 --> 01:08:46,687
as an unlimited amount
of intelligence,
1345
01:08:46,731 --> 01:08:48,167
the closest we can get to that
1346
01:08:48,211 --> 01:08:50,474
is by evolving
our own intelligence
1347
01:08:50,517 --> 01:08:55,566
by merging with the artificial
intelligence we're creating.
1348
01:08:55,609 --> 01:08:58,003
Musk:
Today, our computers, phones,
1349
01:08:58,046 --> 01:09:01,615
applications give us
superhuman capability.
1350
01:09:01,659 --> 01:09:04,662
So, as the old maxim says,
if you can't beat 'em, join 'em.
1351
01:09:06,968 --> 01:09:09,971
el Kaliouby: It's about
a human-machine partnership.
1352
01:09:10,015 --> 01:09:11,669
I mean, we already see
how, you know,
1353
01:09:11,712 --> 01:09:14,933
our phones, for example, act
as memory prosthesis, right?
1354
01:09:14,976 --> 01:09:17,196
I don't have to remember
your phone number anymore
1355
01:09:17,240 --> 01:09:19,198
'cause it's on my phone.
1356
01:09:19,242 --> 01:09:22,070
It's about machines
augmenting our human abilities,
1357
01:09:22,114 --> 01:09:25,248
as opposed to, like,
completely displacing them.
1358
01:09:25,291 --> 01:09:27,380
Nolan: If you look at all the
objects that have made the leap
1359
01:09:27,424 --> 01:09:30,122
from analog to digital
over the last 20 years...
1360
01:09:30,166 --> 01:09:32,080
it's a lot.
1361
01:09:32,124 --> 01:09:35,388
We're the last analog object
in a digital universe.
1362
01:09:35,432 --> 01:09:36,911
And the problem with that,
of course,
1363
01:09:36,955 --> 01:09:40,567
is that the data input/output
is very limited.
1364
01:09:40,611 --> 01:09:42,613
It's this.
It's these.
1365
01:09:42,656 --> 01:09:45,355
Zilis:
Our eyes are pretty good.
1366
01:09:45,398 --> 01:09:48,445
We're able to take in a lot
of visual information.
1367
01:09:48,488 --> 01:09:52,536
But our information output
is very, very, very low.
1368
01:09:52,579 --> 01:09:55,669
The reason this is important --
If we envision a scenario
1369
01:09:55,713 --> 01:09:59,543
where AI's playing a more
prominent role in societies,
1370
01:09:59,586 --> 01:10:02,023
we want good ways to interact
with this technology
1371
01:10:02,067 --> 01:10:04,983
so that it ends up
augmenting us.
1372
01:10:05,026 --> 01:10:07,812
♪♪
1373
01:10:07,855 --> 01:10:09,553
Musk: I think
it's incredibly important
1374
01:10:09,596 --> 01:10:12,295
that AI not be "other."
1375
01:10:12,338 --> 01:10:14,862
It must be us.
1376
01:10:14,906 --> 01:10:18,605
And I could be wrong
about what I'm saying.
1377
01:10:18,649 --> 01:10:20,216
I'm certainly open to ideas
1378
01:10:20,259 --> 01:10:23,915
if anybody can suggest
a path that's better.
1379
01:10:23,958 --> 01:10:27,266
But I think we're gonna really
have to either merge with AI
1380
01:10:27,310 --> 01:10:28,963
or be left behind.
1381
01:10:29,007 --> 01:10:36,362
♪♪
1382
01:10:36,406 --> 01:10:38,756
Gourley: It's hard to kind of
think of unplugging a system
1383
01:10:38,799 --> 01:10:41,802
that's distributed
everywhere on the planet,
1384
01:10:41,846 --> 01:10:45,806
that's distributed now
across the solar system.
1385
01:10:45,850 --> 01:10:49,375
You can't just, you know,
shut that off.
1386
01:10:49,419 --> 01:10:51,290
Nolan:
We've opened Pandora's box.
1387
01:10:51,334 --> 01:10:55,642
We've unleashed forces that
we can't control, we can't stop.
1388
01:10:55,686 --> 01:10:57,296
We're in the midst
of essentially creating
1389
01:10:57,340 --> 01:10:59,516
a new life-form on Earth.
1390
01:10:59,559 --> 01:11:05,826
♪♪
1391
01:11:05,870 --> 01:11:07,611
Russell:
We don't know what happens next.
1392
01:11:07,654 --> 01:11:10,353
We don't know what shape
the intellect of a machine
1393
01:11:10,396 --> 01:11:14,531
will be when that intellect is
far beyond human capabilities.
1394
01:11:14,574 --> 01:11:17,360
It's just not something
that's possible.
1395
01:11:17,403 --> 01:11:24,715
♪♪
1396
01:11:24,758 --> 01:11:26,934
The least scary future
I can think of is one
1397
01:11:26,978 --> 01:11:29,633
where we have at least
democratized AI.
1398
01:11:31,548 --> 01:11:34,159
Because if one company
or small group of people
1399
01:11:34,202 --> 01:11:37,031
manages to develop godlike
digital superintelligence,
1400
01:11:37,075 --> 01:11:40,339
they can take over the world.
1401
01:11:40,383 --> 01:11:42,210
At least when there's
an evil dictator,
1402
01:11:42,254 --> 01:11:44,343
that human is going to die,
1403
01:11:44,387 --> 01:11:46,998
but, for an AI,
there would be no death.
1404
01:11:47,041 --> 01:11:49,392
It would live forever.
1405
01:11:49,435 --> 01:11:51,916
And then you have
an immortal dictator
1406
01:11:51,959 --> 01:11:53,570
from which we can never escape.
1407
01:11:53,613 --> 01:12:02,100
♪♪
1408
01:12:02,143 --> 01:12:10,587
♪♪
1409
01:12:10,630 --> 01:12:19,160
♪♪
1410
01:12:19,204 --> 01:12:27,647
♪♪
1411
01:12:27,691 --> 01:12:36,221
♪♪
1412
01:12:36,264 --> 01:12:44,838
♪♪
1413
01:12:51,845 --> 01:12:53,717
Woman on P.A.:
Alan. Macchiato.
1414
01:12:53,760 --> 01:12:55,806
[ Cellphone ringing ]
1415
01:12:58,112 --> 01:13:02,116
[ Computers, devices beeping ]
1416
01:13:05,816 --> 01:13:07,252
[ Cellphone dings ]
1417
01:13:07,295 --> 01:13:09,036
[ Beeping continues ]
1418
01:13:09,080 --> 01:13:10,908
[ Cellphone chirps ]
1419
01:13:10,951 --> 01:13:17,610
♪♪
1420
01:13:17,654 --> 01:13:24,269
♪♪
1421
01:13:24,312 --> 01:13:30,884
♪♪
1422
01:13:30,928 --> 01:13:37,543
♪♪
1423
01:13:37,587 --> 01:13:44,245
♪♪
1424
01:13:44,289 --> 01:13:50,817
♪♪
1425
01:13:50,861 --> 01:13:57,520
♪♪
1426
01:13:57,563 --> 01:14:04,178
♪♪
1427
01:14:04,222 --> 01:14:10,794
♪♪
1428
01:14:10,837 --> 01:14:17,496
♪♪
1429
01:14:17,540 --> 01:14:24,111
♪♪
1430
01:14:24,155 --> 01:14:30,727
♪♪
1431
01:14:30,770 --> 01:14:37,473
♪♪
1432
01:14:37,516 --> 01:14:44,044
♪♪
1433
01:14:44,088 --> 01:14:47,570
Woman: [ Echoing ]
Hello?
1434
01:14:47,613 --> 01:14:54,838
♪♪
1435
01:14:54,881 --> 01:15:02,236
♪♪
1436
01:15:02,280 --> 01:15:09,548
♪♪
1437
01:15:09,592 --> 01:15:16,773
♪♪
1438
01:15:16,816 --> 01:15:18,688
♪ Yeah, yeah
1439
01:15:18,731 --> 01:15:20,080
♪ Yeah, yeah
1440
01:15:20,124 --> 01:15:27,261
♪♪
1441
01:15:27,305 --> 01:15:34,442
♪♪
1442
01:15:34,486 --> 01:15:41,580
♪♪
1443
01:15:41,624 --> 01:15:43,234
♪ Yeah, yeah
1444
01:15:43,277 --> 01:15:45,541
♪ Yeah, yeah
1445
01:15:45,584 --> 01:15:51,764
♪♪
1446
01:15:51,808 --> 01:15:57,988
♪♪
1447
01:15:58,031 --> 01:16:04,342
♪♪
1448
01:16:04,385 --> 01:16:10,609
♪♪
1449
01:16:10,653 --> 01:16:13,046
[ Echoing ] Hello?
1450
01:16:13,090 --> 01:16:22,012
♪♪
1451
01:16:22,055 --> 01:16:31,021
♪♪
1452
01:16:31,064 --> 01:16:39,986
♪♪
1453
01:16:40,030 --> 01:16:48,952
♪♪
1454
01:16:48,995 --> 01:16:57,961
♪♪
1455
01:16:58,004 --> 01:17:06,926
♪♪
1456
01:17:06,970 --> 01:17:15,892
♪♪
1457
01:17:15,935 --> 01:17:24,901
♪♪
1458
01:17:24,944 --> 01:17:33,910
♪♪
1459
01:17:33,953 --> 01:17:40,960
♪♪
112150
Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.