All language subtitles for NOVA.S51E05.A.I.Revolution.1080p.WEB.h264-BAE_track3_[eng]

af Afrikaans
ak Akan
sq Albanian
am Amharic
ar Arabic Download
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English Download
eo Esperanto
et Estonian
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian Download
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranî)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian
pl Polish
pt-BR Portuguese (Brazil)
pt Portuguese (Portugal) Download
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese
sk Slovak
sl Slovenian
so Somali
es Spanish
es-419 Spanish (Latin American)
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:04,166 --> 00:00:08,500 ♪ ♪ 2 00:00:08,500 --> 00:00:10,766 MILES O'BRIEN: Machines that think like humans. 3 00:00:10,766 --> 00:00:14,300 Our dream to create machines in our own image 4 00:00:14,300 --> 00:00:17,000 that are smart and intelligent 5 00:00:17,000 --> 00:00:19,566 goes back to antiquity. 6 00:00:19,566 --> 00:00:21,666 Well, can it bring it to me? 7 00:00:21,666 --> 00:00:23,866 O'BRIEN: Is it possible that the dream of artificial intelligence 8 00:00:23,866 --> 00:00:26,233 has become reality? 9 00:00:27,333 --> 00:00:28,933 They're able to do things 10 00:00:28,933 --> 00:00:31,266 that we didn't think they could do. 11 00:00:32,866 --> 00:00:36,533 MANOLIS KELLIS: Go was thought to be a game where machines would never win. 12 00:00:36,533 --> 00:00:39,933 The number of choices for every move is enormous. 13 00:00:39,933 --> 00:00:44,033 O'BRIEN: And now, the possibilities seem endless. 14 00:00:44,033 --> 00:00:45,833 MUSTAFA SULEYMAN: And this is going to be 15 00:00:45,833 --> 00:00:47,866 one of the greatest boosts 16 00:00:47,866 --> 00:00:50,300 to productivity in the history of our species. 17 00:00:52,466 --> 00:00:55,966 That looks like just a hint of some type of smoke. 18 00:00:55,966 --> 00:00:58,333 O'BRIEN: Identifying problems before a human can... 19 00:00:58,333 --> 00:01:00,633 LECIA SEQUIST: We taught the model to recognize 20 00:01:00,633 --> 00:01:02,833 developing lung cancer. 21 00:01:02,833 --> 00:01:06,000 O'BRIEN: ...and inventing new drugs. 22 00:01:06,000 --> 00:01:08,100 PETRINA KAMYA: I never thought that we would be able 23 00:01:08,100 --> 00:01:10,033 to be doing the things we're doing with A.I.. 24 00:01:10,033 --> 00:01:12,400 O'BRIEN: But along with the hope... 25 00:01:12,400 --> 00:01:14,200 (imitating Obama): This is a dangerous time. 26 00:01:14,200 --> 00:01:17,100 O'BRIEN: ...comes deep concern. 27 00:01:17,100 --> 00:01:18,900 One of the first drops in the feared flood 28 00:01:18,900 --> 00:01:20,500 of A.I.-created disinformation. 29 00:01:20,500 --> 00:01:24,533 We have lowered barriers to entry to manipulate reality. 30 00:01:24,533 --> 00:01:27,266 We're going to live in a world where we don't know what's real. 31 00:01:27,266 --> 00:01:31,166 The risks are uncertain and potentially enormous. 32 00:01:31,166 --> 00:01:35,366 O'BRIEN: How powerful is A.I.? How does it work? 33 00:01:35,366 --> 00:01:37,933 And how can we reap its extraordinary benefits... 34 00:01:37,933 --> 00:01:39,400 Sybil looked here, 35 00:01:39,400 --> 00:01:42,100 and anticipated that there would be a problem. 36 00:01:42,100 --> 00:01:44,300 O'BRIEN: ...without jeopardizing our future? 37 00:01:44,300 --> 00:01:45,633 "A.I. Revolution" 38 00:01:45,633 --> 00:01:48,400 right now, on "NOVA!" 39 00:01:48,400 --> 00:01:51,533 (whirring) 40 00:01:51,533 --> 00:02:11,700 ♪ ♪ 41 00:02:12,966 --> 00:02:16,100 Tell me the backstory on inflection A.I.. 42 00:02:16,100 --> 00:02:20,300 (voiceover): Our story begins with the making of this story. 43 00:02:20,300 --> 00:02:25,266 PI (on computer): The story of Inflection A.I. is an exciting one. 44 00:02:25,266 --> 00:02:26,333 O'BRIEN (voiceover): I was researching 45 00:02:26,333 --> 00:02:27,500   an interview subject. 46 00:02:27,500 --> 00:02:30,033   Who is Mustafa Suleyman? 47 00:02:30,033 --> 00:02:31,633 (voiceover): Something I've done 48 00:02:31,633 --> 00:02:34,066 a thousand times in my 40-year career. 49 00:02:34,066 --> 00:02:35,900 PI (on computer): Mustafa Suleyman is a true pioneer 50 00:02:35,900 --> 00:02:38,366 in the field of artificial intelligence. 51 00:02:38,366 --> 00:02:40,566 (voiceover): But this time, it was different: 52 00:02:40,566 --> 00:02:43,700 I wasn't typing out search terms. 53 00:02:43,700 --> 00:02:46,500 What is machine learning? 54 00:02:46,500 --> 00:02:49,933 O'BRIEN (voiceover): I was having a conversation with a computer. 55 00:02:49,933 --> 00:02:52,066 PI: Sounds like an exciting project, Miles. 56 00:02:52,066 --> 00:02:55,500 (voiceover): It felt like something big had changed. 57 00:02:55,500 --> 00:02:59,166 PI: Machine learning, ML, is a type of artificial intelligence. 58 00:02:59,166 --> 00:03:00,766 O'BRIEN (voiceover): And as it happened, 59 00:03:00,766 --> 00:03:05,566 I was focused on one of the innovators of this revolution. 60 00:03:05,566 --> 00:03:07,100 Okay, so if I do this... 61 00:03:07,100 --> 00:03:10,100   (voiceover): Mustafa Suleyman is Co-Founder 62 00:03:10,100 --> 00:03:11,833 of a startup called Inflection. 63 00:03:11,833 --> 00:03:15,100 It makes an artificial intelligence assistant 64 00:03:15,100 --> 00:03:16,766 called Pi. 65 00:03:16,766 --> 00:03:18,166   So now you can speak... 66 00:03:18,166 --> 00:03:21,933 O'BRIEN (voiceover): I met them both in London. 67 00:03:21,933 --> 00:03:23,433 It's fundamentally different, isn't it? 68 00:03:23,433 --> 00:03:24,933 Yeah, it's a conversational style. 69 00:03:24,933 --> 00:03:27,433 All of us humans learn through stories, 70 00:03:27,433 --> 00:03:30,300 and through narrative, and through interactive dialogue. 71 00:03:30,300 --> 00:03:32,933 And now, the machine can kind of come alive, 72 00:03:32,933 --> 00:03:34,700 and talk to you about whatever it is 73 00:03:34,700 --> 00:03:36,100 that's on top of your mind. 74 00:03:36,100 --> 00:03:40,233 O'BRIEN: Tell me about the PBS program "NOVA." 75 00:03:40,233 --> 00:03:43,600 (voiceover): Chatbots can offer up quick answers, write poems, 76 00:03:43,600 --> 00:03:46,166 finish essays, and translate languages 77 00:03:46,166 --> 00:03:48,333 among many other things. 78 00:03:48,333 --> 00:03:51,033 PI (over phone): "NOVA" is a science documentary series... 79 00:03:51,033 --> 00:03:52,533 O'BRIEN (voiceover): They aren't perfect, 80 00:03:52,533 --> 00:03:54,800 but they have put artificial intelligence in our hands, 81 00:03:54,800 --> 00:03:57,233 and into the public consciousness. 82 00:03:57,233 --> 00:04:00,866 And it seems we're equal parts leery 83 00:04:00,866 --> 00:04:02,533 and intrigued. 84 00:04:02,533 --> 00:04:04,233 SULEYMAN: A.I. is a tool 85 00:04:04,233 --> 00:04:07,833 for helping us to understand the world around us, 86 00:04:07,833 --> 00:04:11,966 predict what's likely to happen, and then invent 87 00:04:11,966 --> 00:04:15,000   solutions that help improve the world around us. 88 00:04:15,000 --> 00:04:18,533 My motivation was to try to use A.I. tools 89 00:04:18,533 --> 00:04:20,800 to, uh, you know, invent the future. 90 00:04:20,800 --> 00:04:23,500 The rise in artificial intelligence... 91 00:04:23,500 --> 00:04:25,333 REPORTER: A.I. technology is developing... 92 00:04:25,333 --> 00:04:28,800 O'BRIEN (voiceover): Lately, it seems a dark future is already here... 93 00:04:28,800 --> 00:04:32,966 The technology could replace millions of jobs... 94 00:04:32,966 --> 00:04:34,733 O'BRIEN (voiceover): ...if you listen to the news reporting. 95 00:04:34,733 --> 00:04:37,966 The moment civilization was transformed. 96 00:04:37,966 --> 00:04:40,433 O'BRIEN (voiceover): So how can artificial intelligence help us, 97 00:04:40,433 --> 00:04:42,800 and how might it hurt us? 98 00:04:42,800 --> 00:04:45,966 At the center of the public handwringing: 99 00:04:45,966 --> 00:04:49,966 how should we put guardrails around it? 100 00:04:49,966 --> 00:04:52,900 We definitely need more regulations in place... 101 00:04:52,900 --> 00:04:54,300 O'BRIEN (voiceover): Artificial intelligence is moving fast 102 00:04:54,300 --> 00:04:56,200 and changing the world. 103 00:04:56,200 --> 00:04:57,766   Can we keep up? 104 00:04:57,766 --> 00:04:59,700 Non-human minds smarter than our own. 105 00:04:59,700 --> 00:05:02,133 O'BRIEN (voiceover): The news coverage may make it seem like 106 00:05:02,133 --> 00:05:04,866 artificial intelligence is something new. 107 00:05:04,866 --> 00:05:06,866 At a moment of revolution... 108 00:05:06,866 --> 00:05:09,300   O'BRIEN (voiceover): But human beings have been thinking about this 109 00:05:09,300 --> 00:05:12,433 for a very long time. 110 00:05:12,433 --> 00:05:16,300 I have a very fine brain. 111 00:05:16,300 --> 00:05:20,633 Our dream to create machines in our own image 112 00:05:20,633 --> 00:05:24,700 that are smart and intelligent goes back to antiquity. 113 00:05:24,700 --> 00:05:27,166   Uh, it's, it's something that has, 114 00:05:27,166 --> 00:05:31,966 has permeated the evolution of society and of science. 115 00:05:31,966 --> 00:05:34,400 (mortars firing) 116 00:05:34,400 --> 00:05:36,800 O'BRIEN (voiceover): The modern origins of artificial intelligence 117 00:05:36,800 --> 00:05:38,966 can be traced back to World War II, 118 00:05:38,966 --> 00:05:43,400 and the prodigious human brain of Alan Turing. 119 00:05:43,400 --> 00:05:46,100 The legendary British mathematician 120 00:05:46,100 --> 00:05:48,100 developed a machine 121 00:05:48,100 --> 00:05:52,233 capable of deciphering coded messages from the Nazis. 122 00:05:52,233 --> 00:05:56,300 After the war, he was among the first to predict computers 123 00:05:56,300 --> 00:05:59,566 might one day match the human brain. 124 00:05:59,566 --> 00:06:02,633 There are no surviving recordings of Turing's voice, 125 00:06:02,633 --> 00:06:08,233 but in 1951, he gave a short lecture on BBC radio. 126 00:06:08,233 --> 00:06:12,766 We asked an A.I.-generated voice to read a passage. 127 00:06:12,766 --> 00:06:15,066 TURING A.I. VOICE: I think it is probable, for instance, 128 00:06:15,066 --> 00:06:17,066 that at the end of the century, 129 00:06:17,066 --> 00:06:19,133 it will be possible to program a machine 130 00:06:19,133 --> 00:06:21,100 to answer questions in such a way 131 00:06:21,100 --> 00:06:23,200 that it will be extremely difficult to guess 132 00:06:23,200 --> 00:06:25,300 whether the answers are being given by a man 133 00:06:25,300 --> 00:06:27,400 or by the machine. 134 00:06:27,400 --> 00:06:30,300 O'BRIEN (voiceover): And so, the Turing test was born. 135 00:06:30,300 --> 00:06:32,433 Could anyone build a machine 136 00:06:32,433 --> 00:06:34,800 that could converse with a human in a way 137 00:06:34,800 --> 00:06:37,866 that is indistinguishable from another person? 138 00:06:37,866 --> 00:06:41,166 In 1956, 139 00:06:41,166 --> 00:06:43,466 a group of pioneering scientists spent the summer 140 00:06:43,466 --> 00:06:46,233 brainstorming at Dartmouth College. 141 00:06:47,266 --> 00:06:49,400 And they told the world that they have coined 142 00:06:49,400 --> 00:06:51,166 a new academic field of study. 143 00:06:51,166 --> 00:06:53,300 They called it artificial intelligence 144 00:06:53,300 --> 00:06:56,833 O'BRIEN (voiceover): For decades, their aspirations remained 145 00:06:56,833 --> 00:06:59,600 far ahead of the capabilities of computers. 146 00:07:01,333 --> 00:07:03,066 In 1978, 147 00:07:03,066 --> 00:07:07,933 "NOVA" released its first film on artificial intelligence. 148 00:07:07,933 --> 00:07:09,800 We have seen the first crude beginnings 149 00:07:09,800 --> 00:07:11,500 of artificial intelligence... 150 00:07:11,500 --> 00:07:13,300 O'BRIEN (voiceover): And the legendary science fiction writer, 151 00:07:13,300 --> 00:07:17,833 Arthur C. Clark was, as always, prescient. 152 00:07:17,833 --> 00:07:19,433 It doesn't really exist yet at any level, 153 00:07:19,433 --> 00:07:23,466 because our most complex computers are still morons, 154 00:07:23,466 --> 00:07:26,400 high-speed morons, but still morons. 155 00:07:26,400 --> 00:07:29,200 Nevertheless, we have the possibility of machines 156 00:07:29,200 --> 00:07:31,466 which can outpace their creators, 157 00:07:31,466 --> 00:07:35,966 and therefore, become more intelligent than us. 158 00:07:37,233 --> 00:07:41,333 At the time, researchers were developing "expert systems," 159 00:07:41,333 --> 00:07:46,500 purpose-built to perform specific tasks. 160 00:07:46,500 --> 00:07:48,100 So the thing that we need to do 161 00:07:48,100 --> 00:07:52,666 to make machine understand, um, you know, our world, 162 00:07:52,666 --> 00:07:55,600 is to put all our knowledge into a machine 163 00:07:55,600 --> 00:07:58,433   and then provide it with some rules. 164 00:07:58,433 --> 00:08:00,466 ♪ ♪ 165 00:08:00,466 --> 00:08:03,600 O'BRIEN (voiceover): Classic A.I. reached a pivotal moment in 1997 166 00:08:03,600 --> 00:08:07,766 when an artificial intelligence program devised by IBM, 167 00:08:07,766 --> 00:08:10,833 called "Deep Blue" defeated world chess champion 168 00:08:10,833 --> 00:08:14,133 and grandmaster Garry Kasparov. 169 00:08:14,133 --> 00:08:18,166 It searched about 200 million positions a second, 170 00:08:18,166 --> 00:08:20,600 navigating through a tree of possibilities 171 00:08:20,600 --> 00:08:23,100 to determine the best move. 172 00:08:23,100 --> 00:08:25,566 RUS: The program analyzed the board configuration, 173 00:08:25,566 --> 00:08:28,533 could project forward millions of moves 174 00:08:28,533 --> 00:08:31,133 to examine millions of possibilities, 175 00:08:31,133 --> 00:08:33,666 and then picked the best path. 176 00:08:33,666 --> 00:08:36,366 O'BRIEN (voiceover): Effective, but brittle, 177 00:08:36,366 --> 00:08:40,400 Deep Blue wasn't strategizing as a human does. 178 00:08:40,400 --> 00:08:43,300 From the outset, artificial intelligence researchers 179 00:08:43,300 --> 00:08:46,033 imagined making machines 180 00:08:46,033 --> 00:08:47,800 that think like us. 181 00:08:47,800 --> 00:08:51,100 The human brain, with more than 80 billion neurons, 182 00:08:51,100 --> 00:08:54,000 learns not by following rules, 183 00:08:54,000 --> 00:08:57,266 but rather by taking in a steady stream of data, 184 00:08:57,266 --> 00:08:59,633 and looking for patterns. 185 00:09:01,066 --> 00:09:03,400 KELLIS: The way that learning actually works 186 00:09:03,400 --> 00:09:06,266 in the human brain is by updating the weights 187 00:09:06,266 --> 00:09:07,933 of the synaptic connections 188 00:09:07,933 --> 00:09:09,800 that are underlying this neural network. 189 00:09:09,800 --> 00:09:13,633 O'BRIEN (voiceover): Manolis Kellis is a Professor of Computer Science 190 00:09:13,633 --> 00:09:17,966 at the Massachusetts Institute of Technology. 191 00:09:17,966 --> 00:09:19,966 So we have trillions of parameters in our brain 192 00:09:19,966 --> 00:09:22,100 that we can adjust based on experience. 193 00:09:22,100 --> 00:09:24,000 I'm getting a reward. 194 00:09:24,000 --> 00:09:25,833 I will update the strength of the connections 195 00:09:25,833 --> 00:09:27,900 that led to this reward-- I'm getting punished, 196 00:09:27,900 --> 00:09:29,666 I will diminish the strength of the connections 197 00:09:29,666 --> 00:09:31,333 that led to the punishment. 198 00:09:31,333 --> 00:09:33,500 So this is the original neural network. 199 00:09:33,500 --> 00:09:37,066 We did not invent it, we, you know, we inherited it. 200 00:09:37,066 --> 00:09:41,166 O'BRIEN (voiceover): But could an artificial neural network 201 00:09:41,166 --> 00:09:44,200 be made in our own image? Turing imagined it. 202 00:09:44,200 --> 00:09:46,500 But computers were nowhere near 203 00:09:46,500 --> 00:09:50,033 powerful enough to do it until recently. 204 00:09:51,666 --> 00:09:53,466 It's only with the advent of extraordinary data sets 205 00:09:53,466 --> 00:09:56,133 that we have, uh, since the early 2000s, 206 00:09:56,133 --> 00:09:59,133 that we were able to build up enough images, 207 00:09:59,133 --> 00:10:00,766 enough annotations, 208 00:10:00,766 --> 00:10:03,866 enough text to be able to finally train 209 00:10:03,866 --> 00:10:06,833 these sufficiently powerful models. 210 00:10:08,300 --> 00:10:10,733 O'BRIEN (voiceover): An artificial neural network is, in fact, 211 00:10:10,733 --> 00:10:13,266 modeled on the human brain. 212 00:10:13,266 --> 00:10:16,600 It uses interconnected nodes, or neurons, 213 00:10:16,600 --> 00:10:18,900 that communicate with each other. 214 00:10:18,900 --> 00:10:21,633 Each node receives inputs from other nodes 215 00:10:21,633 --> 00:10:25,566 and processes those inputs to produce outputs, 216 00:10:25,566 --> 00:10:29,133 which are then passed on to still other nodes. 217 00:10:29,133 --> 00:10:32,500 It learns by adjusting the strength of the connections 218 00:10:32,500 --> 00:10:37,033 between the nodes based on the data it is exposed to. 219 00:10:37,033 --> 00:10:39,500 This process of adjusting the connections 220 00:10:39,500 --> 00:10:41,400 is called training, 221 00:10:41,400 --> 00:10:43,866 and it allows an artificial neural network 222 00:10:43,866 --> 00:10:47,233 to recognize patterns and learn from its experiences 223 00:10:47,233 --> 00:10:49,466 like humans do. 224 00:10:51,166 --> 00:10:52,700 A child, how is it learning so fast? 225 00:10:52,700 --> 00:10:54,366 It is learning so fast 226 00:10:54,366 --> 00:10:56,700 because it's constantly predicting the future 227 00:10:56,700 --> 00:10:59,100 and then seeing what happens 228 00:10:59,100 --> 00:11:02,266 and updating their weights in their neural network 229 00:11:02,266 --> 00:11:04,266 based on what just happened. 230 00:11:04,266 --> 00:11:05,566 Now you can take this 231 00:11:05,566 --> 00:11:07,000 self-supervised learning paradigm 232 00:11:07,000 --> 00:11:09,166 and apply it to machines. 233 00:11:10,700 --> 00:11:13,933 O'BRIEN (voiceover): At first, some of these artificial neural networks 234 00:11:13,933 --> 00:11:16,733 were trained on vintage Atari video games 235 00:11:16,733 --> 00:11:18,700 like "Space Invaders" 236 00:11:18,700 --> 00:11:21,700 and "Breakout." 237 00:11:21,700 --> 00:11:24,933 Games reduce the complexity of the real world 238 00:11:24,933 --> 00:11:28,600 to a very narrow set of actions that can be taken. 239 00:11:28,600 --> 00:11:31,166 O'BRIEN (voiceover): Before he started Inflection, 240 00:11:31,166 --> 00:11:34,266 Mustafa Suleyman co-founded a company called 241 00:11:34,266 --> 00:11:36,766   DeepMind in 2010. 242 00:11:36,766 --> 00:11:40,766 It was acquired by Google four years later. 243 00:11:40,766 --> 00:11:42,166 When an A.I. plays a game, 244 00:11:42,166 --> 00:11:45,766 we show it frame-by-frame, every pixel 245 00:11:45,766 --> 00:11:47,966 in the moving image. 246 00:11:47,966 --> 00:11:49,900 And so the A.I. learns to associate pixels 247 00:11:49,900 --> 00:11:52,000 with actions that it can take 248 00:11:52,000 --> 00:11:55,800 moving left or right or pressing the fire button. 249 00:11:57,133 --> 00:12:00,433 O'BRIEN (voiceover): When it obliterates blocks or shoots aliens, 250 00:12:00,433 --> 00:12:03,800 the connections between the nodes that enabled that success 251 00:12:03,800 --> 00:12:05,566 are strengthened. 252 00:12:05,566 --> 00:12:07,933 In other words, it is rewarded. 253 00:12:07,933 --> 00:12:10,966 When it fails, no reward. 254 00:12:10,966 --> 00:12:13,566 Eventually, all those reinforced connections 255 00:12:13,566 --> 00:12:15,833 overrule the weaker ones. 256 00:12:15,833 --> 00:12:18,666 The program has learned how to win. 257 00:12:20,400 --> 00:12:22,833 This sort of repeated allocation of reward 258 00:12:22,833 --> 00:12:27,366 for repetitive behavior is a great way to train a dog. 259 00:12:27,366 --> 00:12:29,266 It's a great way to teach a kid. 260 00:12:29,266 --> 00:12:32,033 It's a great way for us as adults to adapt our behavior. 261 00:12:32,033 --> 00:12:34,433 And in fact, it's actually a good way 262 00:12:34,433 --> 00:12:37,100 to train machine learning algorithms to get better. 263 00:12:39,900 --> 00:12:43,133 O'BRIEN (voiceover): In 2014, DeepMind began work on an artificial neural network 264 00:12:43,133 --> 00:12:45,733 called "AlphaGo" 265 00:12:45,733 --> 00:12:47,200 that could play the ancient, 266 00:12:47,200 --> 00:12:50,133 and deceptively complex, board game of Go. 267 00:12:51,966 --> 00:12:55,400 KELLIS: Go was thought to be a game where machines would never win. 268 00:12:55,400 --> 00:12:58,800 The number of choices for every move is enormous. 269 00:12:58,800 --> 00:13:01,000 O'BRIEN (voiceover): But at DeepMind, 270 00:13:01,000 --> 00:13:02,766   they were counting on 271 00:13:02,766 --> 00:13:07,100 the astounding growth of compute power. 272 00:13:07,100 --> 00:13:09,666 And I think that's the key concept to try to grasp, 273 00:13:09,666 --> 00:13:14,133 is that we are massively, exponentially growing 274 00:13:14,133 --> 00:13:16,966 the amount of computation used, and in some sense, 275 00:13:16,966 --> 00:13:19,666 that computation is a proxy 276 00:13:19,666 --> 00:13:22,700 for how intelligent the model is. 277 00:13:23,800 --> 00:13:27,633 O'BRIEN (voiceover): AlphaGo was trained two ways. 278 00:13:27,633 --> 00:13:30,433 First, it was fed a large data set of expert Go games 279 00:13:30,433 --> 00:13:33,800 so that it could learn how to play the game. 280 00:13:33,800 --> 00:13:36,133 This is known as supervised learning. 281 00:13:36,133 --> 00:13:41,633 Then the software played against itself many millions of times, 282 00:13:41,633 --> 00:13:44,366 so-called reinforcement learning. 283 00:13:44,366 --> 00:13:47,800 This gradually improved its skills and strategies. 284 00:13:47,800 --> 00:13:50,600 In March 2016, 285 00:13:50,600 --> 00:13:52,366 AlphaGo faced Lee Sedol, 286 00:13:52,366 --> 00:13:54,333 one of the world's top-ranking players 287 00:13:54,333 --> 00:13:57,866 in a five-game match in Seoul, South Korea. 288 00:13:57,866 --> 00:13:59,933   AlphaGo not only won, 289 00:13:59,933 --> 00:14:04,266 but also made a move so novel, the Go cognoscenti 290 00:14:04,266 --> 00:14:07,166 thought it was a huge blunder. That's a very surprising move. 291 00:14:08,933 --> 00:14:11,066 There's no question to me that these A.I. models 292 00:14:11,066 --> 00:14:12,766 are creative. 293 00:14:12,766 --> 00:14:15,600 They're incredibly creative. 294 00:14:15,600 --> 00:14:19,700 O'BRIEN (voiceover): It turns out the move was a stroke of brilliance. 295 00:14:19,700 --> 00:14:21,700 And this emergent creative behavior 296 00:14:21,700 --> 00:14:23,733 was a hint of what was to come: 297 00:14:23,733 --> 00:14:26,900 generative A.I. 298 00:14:26,900 --> 00:14:28,633 Meanwhile, 299 00:14:28,633 --> 00:14:31,066 a company called OpenA.I. was creating 300 00:14:31,066 --> 00:14:33,066 a generative A.I. model 301 00:14:33,066 --> 00:14:36,200 that would become ChatGPT. 302 00:14:36,200 --> 00:14:38,333 It allows users to engage in a dialogue 303 00:14:38,333 --> 00:14:42,200 with a machine that seems uncannily human. 304 00:14:42,200 --> 00:14:44,866 It was first released in 2018, 305 00:14:44,866 --> 00:14:48,933 but it was a subsequent version that became a global sensation 306 00:14:48,933 --> 00:14:51,233 in late 2022. 307 00:14:51,233 --> 00:14:53,966 This promises to be the viral sensation 308 00:14:53,966 --> 00:14:56,633 that could completely reset how we do things. 309 00:14:56,633 --> 00:14:58,633 Cranking out entire essays 310 00:14:58,633 --> 00:15:00,500 in a matter of seconds. 311 00:15:00,500 --> 00:15:03,400 O'BRIEN (voiceover): Not only did it wow the public, it also caught 312 00:15:03,400 --> 00:15:06,666 artificial intelligence innovators off guard. 313 00:15:08,033 --> 00:15:10,133 YOSHUA BENGIO: It surprised me a lot 314 00:15:10,133 --> 00:15:12,200 that they're able to do things that 315 00:15:12,200 --> 00:15:15,700 we didn't think they could do simply by 316 00:15:15,700 --> 00:15:19,933 learning to imitate how humans respond. 317 00:15:19,933 --> 00:15:23,500   And I thought this kind of abilities would take 318 00:15:23,500 --> 00:15:26,300 many more years or decades. 319 00:15:26,300 --> 00:15:30,033 O'BRIEN (voiceover): ChatGPT is a large language model. 320 00:15:30,033 --> 00:15:34,266 LLMs start by consuming massive amounts of text: 321 00:15:34,266 --> 00:15:36,300 books, articles and websites, 322 00:15:36,300 --> 00:15:39,200 which are publicly available on the internet. 323 00:15:39,200 --> 00:15:42,833 By recognizing patterns in billions of words, 324 00:15:42,833 --> 00:15:46,066 they can make guesses at the next word in a sentence. 325 00:15:46,066 --> 00:15:49,333 That's how ChatGPT generates unique answers 326 00:15:49,333 --> 00:15:51,466 to your questions. 327 00:15:51,466 --> 00:15:54,033 If I ask for a haiku about the blue sky 328 00:15:54,033 --> 00:15:58,866 it writes something that seems completely original. 329 00:15:58,866 --> 00:16:00,833 KELLIS: If you're good at predicting 330 00:16:00,833 --> 00:16:02,766 this next word, 331 00:16:02,766 --> 00:16:04,800 it means you're understanding something about the sentence. 332 00:16:04,800 --> 00:16:07,500 What the style of the sentence is, 333 00:16:07,500 --> 00:16:10,233 what the feeling of the sentence is. 334 00:16:10,233 --> 00:16:13,766 And you can't tell whether this was a human or a machine. 335 00:16:13,766 --> 00:16:15,800 That's basically the definition of the Turing test. 336 00:16:15,800 --> 00:16:19,633 O'BRIEN (voiceover): So, how is this changing our world? 337 00:16:19,633 --> 00:16:23,333 Well, It might change my world-- as an arm amputee. 338 00:16:23,333 --> 00:16:25,366 Ready for my casting call, right? 339 00:16:25,366 --> 00:16:26,666 MONROE (chuckling): Yes. 340 00:16:26,666 --> 00:16:28,466 Let's do it. All right. 341 00:16:28,466 --> 00:16:30,266 O'BRIEN (voiceover): That's Brian Monroe of the Hanger Clinic. 342 00:16:30,266 --> 00:16:31,566 He's been my prosthetist 343 00:16:31,566 --> 00:16:34,500 since an injury took my arm above the elbow 344 00:16:34,500 --> 00:16:36,266 ten years ago. 345 00:16:36,266 --> 00:16:38,800 So what we're going to do today is take a mold of your arm. Uh-huh. 346 00:16:38,800 --> 00:16:41,200 Kind of is like a cast for a broken bone. 347 00:16:41,200 --> 00:16:45,266 O'BRIEN (voiceover): Up until now, I have used a body-powered prosthetic. 348 00:16:45,266 --> 00:16:48,400 Harness and a cable allow me to move it 349 00:16:48,400 --> 00:16:50,200 by shrugging my shoulders. 350 00:16:50,200 --> 00:16:54,333 The technology is more than a century old. 351 00:16:54,333 --> 00:16:56,133 But artificial intelligence, 352 00:16:56,133 --> 00:16:58,933 coupled with small electric motors, 353 00:16:58,933 --> 00:17:03,566 is finally pushing prosthetics into the 21st century. 354 00:17:05,100 --> 00:17:07,300 Which brings me to Chicago 355 00:17:07,300 --> 00:17:10,600 and the offices of a small company called Coapt. 356 00:17:10,600 --> 00:17:13,600 I met the C.E.O., Blair Locke, 357 00:17:13,600 --> 00:17:17,466 a pioneer in the push to apply artificial intelligence 358 00:17:17,466 --> 00:17:21,500 to artificial limbs. 359 00:17:21,500 --> 00:17:23,933 So, what do we have here? What are we going to do? 360 00:17:23,933 --> 00:17:27,000 This allows us to very easily test how your control would be 361 00:17:27,000 --> 00:17:30,100 using a pretty simple cuff; this has electrodes in it, 362 00:17:30,100 --> 00:17:32,033 and we'll let the power of the electronics 363 00:17:32,033 --> 00:17:33,600 that are doing the machine learning 364 00:17:33,600 --> 00:17:35,733 see what you're capable of. All right, let's give it a try. 365 00:17:35,733 --> 00:17:38,066 (voiceover): Like most amputees, 366 00:17:38,066 --> 00:17:42,133 I feel my missing hand almost as if it was still there-- 367 00:17:42,133 --> 00:17:43,600 a phantom. 368 00:17:43,600 --> 00:17:45,300 Everything will touch. Is that okay? 369 00:17:45,300 --> 00:17:46,366 Yeah. Not too tight? 370 00:17:46,366 --> 00:17:48,200 No. All good. Okay. 371 00:17:48,200 --> 00:17:50,400 O'BRIEN (voiceover): It's almost entirely immobile, stuck in molasses. 372 00:17:50,400 --> 00:17:52,933   Make a fist, not too hard. 373 00:17:52,933 --> 00:17:57,100 O'BRIEN (voiceover): But I am able to imagine moving it ever so slightly. 374 00:17:57,100 --> 00:17:58,766 And I'm gonna have you squeeze into that a little bit harder. 375 00:17:58,766 --> 00:18:01,566 Very good, and I see the pattern on the screen 376 00:18:01,566 --> 00:18:02,833 change a little bit. 377 00:18:02,833 --> 00:18:04,400 O'BRIEN (voiceover): And when I do, 378 00:18:04,400 --> 00:18:07,333 I generate an array of faint electrical signals in my stump. 379 00:18:07,333 --> 00:18:09,166   That's your muscle information. 380 00:18:09,166 --> 00:18:11,033 It feels, it feels like I'm overcoming 381 00:18:11,033 --> 00:18:12,866 something that's really stuck. 382 00:18:12,866 --> 00:18:14,366 I don't know, is that enough signal? 383 00:18:14,366 --> 00:18:16,300 Should be. Oh, okay. 384 00:18:16,300 --> 00:18:17,533 We don't need a lot of signal, 385 00:18:17,533 --> 00:18:18,866 we're going for information 386 00:18:18,866 --> 00:18:20,733 in the signal, not how loud it is. 387 00:18:20,733 --> 00:18:23,466 O'BRIEN (voiceover): And this is where artificial intelligence comes in. 388 00:18:25,266 --> 00:18:28,566 Using a virtual prosthetic depicted on a screen, 389 00:18:28,566 --> 00:18:32,300 I trained a machine learning algorithm to become fluent 390 00:18:32,300 --> 00:18:36,800 in the language of my nerves and muscles. 391 00:18:36,800 --> 00:18:38,433 We see eight different signals on the screen. 392 00:18:38,433 --> 00:18:40,866 All eight of those sensor sites 393 00:18:40,866 --> 00:18:42,400 are going to feed in together 394 00:18:42,400 --> 00:18:44,100 and let the algorithm sort out the data. 395 00:18:44,100 --> 00:18:46,166 What you are experiencing 396 00:18:46,166 --> 00:18:48,800 is your ability to teach the system 397 00:18:48,800 --> 00:18:50,633 what is hand-closed to you. 398 00:18:50,633 --> 00:18:52,633 And that's different than what it would be to me. 399 00:18:52,633 --> 00:18:57,033 O'BRIEN (voiceover): I told the software what motion I desired, 400 00:18:57,033 --> 00:18:59,433 open, close, or rotate, 401 00:18:59,433 --> 00:19:03,633 then imagined moving my phantom limb accordingly. 402 00:19:03,633 --> 00:19:05,833 This generates an array of electromyographic, 403 00:19:05,833 --> 00:19:08,700 or EMG, signals in my remaining muscles. 404 00:19:08,700 --> 00:19:12,333 I was training the A.I. to connect the pattern 405 00:19:12,333 --> 00:19:15,133 of these electrical signals with a specific movement. 406 00:19:17,366 --> 00:19:18,700 LOCK: The system adapts, 407 00:19:18,700 --> 00:19:21,300 and as you add more data and use it over time, 408 00:19:21,300 --> 00:19:23,300 it becomes more robust, 409 00:19:23,300 --> 00:19:27,100   and it learns to improve upon use. 410 00:19:27,100 --> 00:19:29,500 O'BRIEN: Is it me that's learning, or the algorithm that's learning? 411 00:19:29,500 --> 00:19:31,600 Or are we learning together? LOCK: You're learning together. 412 00:19:31,600 --> 00:19:32,633 Okay. 413 00:19:34,133 --> 00:19:37,300 O'BRIEN (voiceover): So, how does the Coapt pattern recognition system work? 414 00:19:37,300 --> 00:19:42,166 It's called a Bayesian classification model. 415 00:19:42,166 --> 00:19:43,966 As I train the software, 416 00:19:43,966 --> 00:19:46,600 it labels my various EMG patterns 417 00:19:46,600 --> 00:19:49,300 into corresponding classes of movement-- 418 00:19:49,300 --> 00:19:53,533 hand open, hand closed, wrist rotation, for example. 419 00:19:53,533 --> 00:19:56,166 As I use the arm, 420 00:19:56,166 --> 00:19:58,833 it compares the electrical signals I'm transmitting 421 00:19:58,833 --> 00:20:02,666 to the existing library of classifications I taught it. 422 00:20:02,666 --> 00:20:05,633 It relies on statistical probability 423 00:20:05,633 --> 00:20:08,466 to choose the best match. 424 00:20:08,466 --> 00:20:10,466 And this is just one way machine learning 425 00:20:10,466 --> 00:20:13,300 is quietly revolutionizing medicine. 426 00:20:16,100 --> 00:20:18,600 Computer scientist Regina Barzilay 427 00:20:18,600 --> 00:20:21,766 first started working on artificial intelligence 428 00:20:21,766 --> 00:20:26,133 in the 1990s, just as rule-based A.I. like Deep Blue 429 00:20:26,133 --> 00:20:28,633 was giving way to neural networks. 430 00:20:28,633 --> 00:20:30,766 She used the techniques 431 00:20:30,766 --> 00:20:32,733 to decipher dead languages. 432 00:20:32,733 --> 00:20:35,866 You might call it a small language model. 433 00:20:35,866 --> 00:20:38,433 Something that is fun and intellectually very challenging, 434 00:20:38,433 --> 00:20:40,466 but it's not like it's going to change our life. 435 00:20:41,966 --> 00:20:44,766 O'BRIEN (voiceover): And then her life changed in an instant. 436 00:20:44,766 --> 00:20:47,466 CONSTANCE LEHMAN: We see a spot there. 437 00:20:47,466 --> 00:20:51,000 O'BRIEN (voiceover): In 2014, she was diagnosed with breast cancer. 438 00:20:51,000 --> 00:20:52,833 BARZILAY (voiceover): When you go through the treatment, 439 00:20:52,833 --> 00:20:54,066 there are a lot of people who are suffering. 440 00:20:54,066 --> 00:20:55,600 I was interested in 441 00:20:55,600 --> 00:20:58,833 what I can do about it, and clearly it was not continuing 442 00:20:58,833 --> 00:21:00,633 deciphering dead languages, 443 00:21:00,633 --> 00:21:02,966 and it was quite a journey. 444 00:21:02,966 --> 00:21:07,500 O'BRIEN (voiceover): Not surprisingly, she began that journey with mammograms. 445 00:21:07,500 --> 00:21:09,166 LEHMAN: It's a little bit more prominent. 446 00:21:09,166 --> 00:21:10,966 O'BRIEN (voiceover): She and Constance Lehman, 447 00:21:10,966 --> 00:21:14,966 a radiologist at Massachusetts General Hospital, 448 00:21:14,966 --> 00:21:17,600 realized the Achilles heel in the diagnostic system 449 00:21:17,600 --> 00:21:20,400 is the human eye. 450 00:21:20,400 --> 00:21:22,600 BARZILAY (voiceover): So the question that we ask is, 451 00:21:22,600 --> 00:21:24,400 what is the likelihood of these patients 452 00:21:24,400 --> 00:21:27,666 to develop cancer within the next five years? 453 00:21:27,666 --> 00:21:29,466 We, with our human eyes, 454 00:21:29,466 --> 00:21:31,366 cannot really make these assertions 455 00:21:31,366 --> 00:21:33,966 because the patterns are so subtle. 456 00:21:33,966 --> 00:21:37,566 LEHMAN: Now, is that different from the surrounding tissue? 457 00:21:37,566 --> 00:21:40,066 O'BRIEN (voiceover): It's a perfect use case for pattern recognition 458 00:21:40,066 --> 00:21:43,733 using what is known as a convolutional neural network. 459 00:21:43,733 --> 00:21:45,566   ♪ ♪ 460 00:21:45,566 --> 00:21:48,900 Here's an example of how CNNs get smart: 461 00:21:48,900 --> 00:21:53,600 they comb through a picture with many virtual magnifying glasses. 462 00:21:53,600 --> 00:21:57,000 Each one is looking for a specific kind of puzzle piece, 463 00:21:57,000 --> 00:21:59,600 like an edge, a shape, or a texture. 464 00:21:59,600 --> 00:22:01,900 Then it makes simplified versions, 465 00:22:01,900 --> 00:22:05,533 repeating the process on larger and larger sections. 466 00:22:05,533 --> 00:22:08,233 Eventually the puzzle can be assembled. 467 00:22:08,233 --> 00:22:10,366 And it's time to make a guess. 468 00:22:10,366 --> 00:22:13,566 Is it a cat? A dog? A tree? 469 00:22:13,566 --> 00:22:17,633 Sometimes the guess is right, but sometimes it's wrong. 470 00:22:17,633 --> 00:22:19,900 And here's the learning part: 471 00:22:19,900 --> 00:22:22,366 with a process called backpropagation, 472 00:22:22,366 --> 00:22:27,233 labeled images are sent back to correct the previous operation. 473 00:22:27,233 --> 00:22:29,900 So the next time it plays the guessing game, 474 00:22:29,900 --> 00:22:31,933 it will be even better. 475 00:22:31,933 --> 00:22:35,233 To validate the model, Regina and her team gathered up 476 00:22:35,233 --> 00:22:38,233 more than 128,000 mammograms 477 00:22:38,233 --> 00:22:41,466 collected at seven sites in four countries. 478 00:22:41,466 --> 00:22:45,166 More than 3,800 of them led to a cancer diagnosis 479 00:22:45,166 --> 00:22:48,933 within five years. 480 00:22:48,933 --> 00:22:50,766 You just give to it the image, 481 00:22:50,766 --> 00:22:53,100 and then the five years of outcomes, 482 00:22:53,100 --> 00:22:57,400 and it can learn the likelihood of getting a cancer diagnosis. 483 00:22:57,400 --> 00:23:01,300 O'BRIEN (voiceover): The software, called Mirai, was a success. 484 00:23:01,300 --> 00:23:05,566 In fact, it is between 75% and 84% accurate 485 00:23:05,566 --> 00:23:08,966 in predicting future cancer diagnoses. 486 00:23:11,433 --> 00:23:16,133 Then, a friend of Regina's developed lung cancer. 487 00:23:16,133 --> 00:23:17,833 SEQUIST: In lung cancer, it's actually 488 00:23:17,833 --> 00:23:20,266 sort of mind boggling how much has changed. 489 00:23:20,266 --> 00:23:23,566 O'BRIEN (voiceover): Her friend saw oncologist Lecia Sequist. 490 00:23:24,800 --> 00:23:26,000 She and Regina wondered 491 00:23:26,000 --> 00:23:29,433 if artificial intelligence could be applied 492 00:23:29,433 --> 00:23:31,766 to CAT scans of patients' lungs. 493 00:23:31,766 --> 00:23:33,400 SEQUIST: We taught the model 494 00:23:33,400 --> 00:23:37,733 to recognize the patterns of developing lung cancer 495 00:23:37,733 --> 00:23:40,433 by using thousands of CAT scans 496 00:23:40,433 --> 00:23:41,566 from patients who were participating 497 00:23:41,566 --> 00:23:42,866 in a clinical trial. 498 00:23:42,866 --> 00:23:45,033 From the new study? Oh, interesting. Correct. 499 00:23:45,033 --> 00:23:47,133 SEQUIST (voiceover): We had a lot of information about them. 500 00:23:47,133 --> 00:23:48,800 We had demographic information, 501 00:23:48,800 --> 00:23:50,633 we had health information, 502 00:23:50,633 --> 00:23:52,266 and we had outcomes information. 503 00:23:52,266 --> 00:23:55,366 O'BRIEN (voiceover): They call the model Sibyl. 504 00:23:55,366 --> 00:23:56,766 In the retrospective study, right, 505 00:23:56,766 --> 00:23:58,366 so the retrospective data... 506 00:23:58,366 --> 00:23:59,900 O'BRIEN (voiceover): Radiologist Florian Fintelmann 507 00:23:59,900 --> 00:24:01,933 showed me what it can do. 508 00:24:01,933 --> 00:24:05,066 FINTELMANN: This is earlier, and this is later. 509 00:24:05,066 --> 00:24:06,766 There is nothing 510 00:24:06,766 --> 00:24:10,100 that I can perceive, pick up, or describe. 511 00:24:10,100 --> 00:24:12,666 There's no, what we call, a precursor lesion 512 00:24:12,666 --> 00:24:13,800 on this CT scan. 513 00:24:13,800 --> 00:24:15,366 Sibyl looked here 514 00:24:15,366 --> 00:24:17,566 and then anticipated that there would be a problem 515 00:24:17,566 --> 00:24:20,233 based on the baseline scan. What is it seeing? 516 00:24:20,233 --> 00:24:21,866 That's the million dollar question. 517 00:24:21,866 --> 00:24:24,066 And, and maybe not the million dollar question. 518 00:24:24,066 --> 00:24:26,500 Does it really matter? Does it? 519 00:24:26,500 --> 00:24:28,866 O'BRIEN (voiceover): When they compared the predictions 520 00:24:28,866 --> 00:24:33,233 to actual outcomes from previous cases, Sybil fared well. 521 00:24:33,233 --> 00:24:35,533 It correctly forecast cancer 522 00:24:35,533 --> 00:24:38,500 between 80% and 95% of the time, 523 00:24:38,500 --> 00:24:41,400 depending on the population it studied. 524 00:24:41,400 --> 00:24:44,100 The technique is still in the trial phase. 525 00:24:44,100 --> 00:24:46,133 But once it is deployed, 526 00:24:46,133 --> 00:24:49,766 it could provide a potent tool for prevention. 527 00:24:52,533 --> 00:24:54,933 The hope is that if you can predict very early on 528 00:24:54,933 --> 00:24:57,400 that the patient is in the wrong way, 529 00:24:57,400 --> 00:25:00,300 you can do clinical trials, you can develop the drugs 530 00:25:00,300 --> 00:25:05,100 that are doing the prevention, rather than treatment 531 00:25:05,100 --> 00:25:07,666 of very advanced disease that we are doing today. 532 00:25:09,000 --> 00:25:12,300 O'BRIEN (voiceover): Which takes us back to DeepMind and AlphaGo. 533 00:25:12,300 --> 00:25:14,733 The fun and games were just the beginning, 534 00:25:14,733 --> 00:25:17,300 a means to an end. 535 00:25:17,300 --> 00:25:20,900 We have always set out at DeepMind 536 00:25:20,900 --> 00:25:24,466 to, um, use our technologies to make the world a better place. 537 00:25:24,466 --> 00:25:27,400 O'BRIEN (voiceover): In 2021, 538 00:25:27,400 --> 00:25:29,400 the company released AlphaFold. 539 00:25:29,400 --> 00:25:31,733 It is pattern recognition software 540 00:25:31,733 --> 00:25:34,066 designed to make it easier for researchers 541 00:25:34,066 --> 00:25:35,800 to understand proteins, 542 00:25:35,800 --> 00:25:39,100 long chains of amino acids 543 00:25:39,100 --> 00:25:41,166 involved in nearly every function in our bodies. 544 00:25:41,166 --> 00:25:43,066 How a protein folds 545 00:25:43,066 --> 00:25:45,466 into a specific, three-dimensional shape 546 00:25:45,466 --> 00:25:50,133 determines how it interacts with other molecules. 547 00:25:50,133 --> 00:25:52,000 SULEYMAN: There's this correlation between 548 00:25:52,000 --> 00:25:55,233 what the protein does and how it's structured. 549 00:25:55,233 --> 00:25:58,733 So if we can predict how the protein folds, 550 00:25:58,733 --> 00:26:01,366 then say something about their function. 551 00:26:01,366 --> 00:26:04,766 O'BRIEN: If we know how a disease's protein is shaped, or folded, 552 00:26:04,766 --> 00:26:08,700 we can sometimes create a drug to disable it. 553 00:26:08,700 --> 00:26:12,800 But the shape of millions of proteins remained a mystery. 554 00:26:12,800 --> 00:26:15,333 DeepMind trained AlphaFold 555 00:26:15,333 --> 00:26:18,333 on thousands of known protein structures. 556 00:26:18,333 --> 00:26:20,600 It leveraged this knowledge to predict 557 00:26:20,600 --> 00:26:23,266 200 million protein structures, 558 00:26:23,266 --> 00:26:27,766 nearly all the proteins known to science. 559 00:26:27,766 --> 00:26:30,466 SULEYMAN: You take some high-quality known data, 560 00:26:30,466 --> 00:26:33,733 and you use that to, you know, 561 00:26:33,733 --> 00:26:37,800 make a prediction about how a similar piece of information 562 00:26:37,800 --> 00:26:40,100 is likely to unfold over some time series, 563 00:26:40,100 --> 00:26:42,166 and the structure of proteins is, 564 00:26:42,166 --> 00:26:44,366 you know, in that sense, no different to 565 00:26:44,366 --> 00:26:47,533 making a prediction in the game of Go or in Atari 566 00:26:47,533 --> 00:26:49,266 or in a mammography scan, 567 00:26:49,266 --> 00:26:51,800 or indeed, in a large language model. 568 00:26:51,800 --> 00:26:53,333 KAMYA: These thin sticks here? 569 00:26:53,333 --> 00:26:55,700 Yeah? They represent the amino acids 570 00:26:55,700 --> 00:26:57,233 that make up a protein. 571 00:26:57,233 --> 00:26:58,400 O'BRIEN (voiceover): Theoretical chemist 572 00:26:58,400 --> 00:27:01,266 Petrina Kamya works for a company called 573 00:27:01,266 --> 00:27:03,333 Insilico Medicine. 574 00:27:03,333 --> 00:27:05,133 It uses AlphaFold 575 00:27:05,133 --> 00:27:07,200 and its own deep-learning models 576 00:27:07,200 --> 00:27:12,533 to make accurate predictions about protein structures. 577 00:27:12,533 --> 00:27:14,766 What we're doing in drug design is we're designing a molecule 578 00:27:14,766 --> 00:27:17,966 that is analogous to the natural molecule 579 00:27:17,966 --> 00:27:19,166 that binds to the protein, 580 00:27:19,166 --> 00:27:20,966 but instead it will lock it, if this molecule 581 00:27:20,966 --> 00:27:23,533 is involved in a disease where it's hyperactive. 582 00:27:24,533 --> 00:27:26,333 O'BRIEN (voiceover): If the molecule fits well, 583 00:27:26,333 --> 00:27:29,266 it can inhibit the disease-causing proteins. 584 00:27:29,266 --> 00:27:30,800 So you're filtering it down 585 00:27:30,800 --> 00:27:33,333 like you're choosing an Airbnb or something to, 586 00:27:33,333 --> 00:27:35,300 you know, number of bedrooms, whatever. To suit your needs. 587 00:27:35,300 --> 00:27:36,466 (laughs) Exactly, right. 588 00:27:36,466 --> 00:27:38,100 Right, yeah. That's a very good analogy. 589 00:27:38,100 --> 00:27:39,966 It's sort of like Airbnb. 590 00:27:39,966 --> 00:27:42,033 So you are putting in your criteria, 591 00:27:42,033 --> 00:27:43,666 and then Airbnb will filter out 592 00:27:43,666 --> 00:27:44,900 all the different properties 593 00:27:44,900 --> 00:27:46,100 based on your criteria. 594 00:27:46,100 --> 00:27:47,433 So you can be very, very restrictive 595 00:27:47,433 --> 00:27:48,833 or you can be very, very free... Right. 596 00:27:48,833 --> 00:27:51,133 In terms of guiding the generative algorithms 597 00:27:51,133 --> 00:27:52,533 and telling them what types of molecules 598 00:27:52,533 --> 00:27:54,266 you want them to generate. 599 00:27:54,266 --> 00:27:59,266 O'BRIEN (voiceover): It will take 48 to 72 hours of computing time 600 00:27:59,266 --> 00:28:02,666 to identify the best candidates ranked in order. 601 00:28:02,666 --> 00:28:04,166 How long would it have taken you 602 00:28:04,166 --> 00:28:07,100 to figure that out as a computational chemist? 603 00:28:07,100 --> 00:28:08,700 I would have thought of some of these, 604 00:28:08,700 --> 00:28:09,733 but not all of them. Okay. 605 00:28:11,100 --> 00:28:13,600 O'BRIEN (voiceover): While there are no shortcuts for human trials, 606 00:28:13,600 --> 00:28:15,700 nor should we hope for that, 607 00:28:15,700 --> 00:28:20,100 this could greatly speed up the drug development pipeline. 608 00:28:21,500 --> 00:28:23,300 There will not be the need to invest so heavily 609 00:28:23,300 --> 00:28:25,500 in preclinical discovery, 610 00:28:25,500 --> 00:28:29,100 and so, drugs can therefore be cheaper. 611 00:28:29,100 --> 00:28:30,800 And you can go after those diseases 612 00:28:30,800 --> 00:28:33,566 that are otherwise neglected, 613 00:28:33,566 --> 00:28:35,266 because you don't have to invest so heavily 614 00:28:35,266 --> 00:28:36,433 in order for you to come up with a drug, 615 00:28:36,433 --> 00:28:38,666 a viable drug. 616 00:28:38,666 --> 00:28:40,766 O'BRIEN (voiceover): But medicine isn't the only place 617 00:28:40,766 --> 00:28:43,033 where A.I. is breaking new frontiers. 618 00:28:43,033 --> 00:28:46,400 It's conducting financial analysis, 619 00:28:46,400 --> 00:28:49,200 helps with fraud detection. 620 00:28:49,200 --> 00:28:50,666   (mechanical whirring) 621 00:28:50,666 --> 00:28:53,966 It's now being deployed to discover novel materials 622 00:28:53,966 --> 00:28:58,366 and could help us build clean energy technology. 623 00:28:58,366 --> 00:29:02,900 And It is even helping to save lives 624 00:29:02,900 --> 00:29:04,800 as the climate crisis boils over. 625 00:29:06,033 --> 00:29:07,600   (indistinct radio chatter) 626 00:29:07,600 --> 00:29:08,966 In St. Helena, California, 627 00:29:08,966 --> 00:29:10,300 dispatchers at the 628 00:29:10,300 --> 00:29:13,966 CAL FIRE Sonoma-Lake-Napa Command Center 629 00:29:13,966 --> 00:29:16,433 caught a break in 2023. 630 00:29:16,433 --> 00:29:22,100 Wildfires blackened nearly 700 acres of their territory. 631 00:29:22,100 --> 00:29:24,266 We were at 400,000 acres in 2020. 632 00:29:25,533 --> 00:29:27,300 Something like that would generate a response from us... 633 00:29:27,300 --> 00:29:30,500 O'BRIEN (voiceover): Chief Mike Marcucci has been fighting fires 634 00:29:30,500 --> 00:29:32,533 for more than 30 years. 635 00:29:32,533 --> 00:29:34,666 MARCUCCI (voiceover): Once we started having these devastating fires, 636 00:29:34,666 --> 00:29:35,766   we needed more intel. 637 00:29:35,766 --> 00:29:37,566 The need for intelligence 638 00:29:37,566 --> 00:29:40,200 is just overwhelming in today's fire service. 639 00:29:41,400 --> 00:29:43,200 O'BRIEN (voiceover): Over the past 20 years, 640 00:29:43,200 --> 00:29:45,166 California has installed a network 641 00:29:45,166 --> 00:29:47,366 of more than 1,000 remotely operated 642 00:29:47,366 --> 00:29:51,633 pan, tilt, zoom surveillance cameras on mountaintops. 643 00:29:53,000 --> 00:29:55,000 PETE AVANSINO: Vegetation fire, Highway 29 at Doton Road. 644 00:29:56,833 --> 00:29:59,933 O'BRIEN (voiceover): All those cameras generate petabytes of video. 645 00:30:00,966 --> 00:30:03,833 CAL FIRE partnered with scientists at U.C. San Diego 646 00:30:03,833 --> 00:30:05,766 to train a neural network 647 00:30:05,766 --> 00:30:08,200 to spot the early signs of trouble. 648 00:30:08,200 --> 00:30:11,500 It's called ALERT California. 649 00:30:11,500 --> 00:30:13,133 SeLEGUE: So here's one that just popped up. 650 00:30:13,133 --> 00:30:15,166 Here's an anomaly. 651 00:30:15,166 --> 00:30:19,166 O'BRIEN (voiceover): CAL FIRE Staff Chief of Fire and Intelligence Philip SeLegue 652 00:30:19,166 --> 00:30:22,533 showed me how it works while it was in action, 653 00:30:22,533 --> 00:30:24,433 detecting nascent fires, 654 00:30:24,433 --> 00:30:26,800 micro fires. 655 00:30:26,800 --> 00:30:28,266 That looks like just a little hint 656 00:30:28,266 --> 00:30:30,733 of some type of smoke that was there... 657 00:30:30,733 --> 00:30:32,433 O'BRIEN (voiceover): Based on this, dispatchers can orchestrate 658 00:30:32,433 --> 00:30:34,066 a fast response. 659 00:30:35,733 --> 00:30:40,500 A.I. has given us the ability to detect and to see 660 00:30:40,500 --> 00:30:42,333   where those fires are starting. 661 00:30:42,333 --> 00:30:45,100 AVANSINO: Transport 1447 responding via MDC. 662 00:30:45,100 --> 00:30:46,533 O'BRIEN (voiceover): For all they know, 663 00:30:46,533 --> 00:30:49,966 they have nipped some megafires in the bud. 664 00:30:49,966 --> 00:30:51,200   The success are the fires 665 00:30:51,200 --> 00:30:52,833 that you don't hear about in the news. 666 00:30:52,833 --> 00:30:55,200 O'BRIEN (voiceover): Artificial intelligence 667 00:30:55,200 --> 00:30:57,833 can't put out wildfires just yet. 668 00:30:57,833 --> 00:31:01,833 Human firefighters still need to do that job. 669 00:31:03,366 --> 00:31:05,600 But researchers are pushing hard 670 00:31:05,600 --> 00:31:07,700 to combine neural networks 671 00:31:07,700 --> 00:31:10,300 with mobility and dexterity. 672 00:31:11,766 --> 00:31:13,400 This is where people get nervous. 673 00:31:13,400 --> 00:31:15,133 Will they take our jobs? 674 00:31:15,133 --> 00:31:17,133 Or could they turn against us? 675 00:31:18,100 --> 00:31:19,766 But at M.I.T., 676 00:31:19,766 --> 00:31:22,433 they're exploring ideas to make robots 677 00:31:22,433 --> 00:31:24,100 good human partners. 678 00:31:25,900 --> 00:31:27,900 We are interested in making machines 679 00:31:27,900 --> 00:31:30,633 that help people with physical and cognitive tasks. 680 00:31:30,633 --> 00:31:32,266 So this is really great, 681 00:31:32,266 --> 00:31:35,266 it has the stiffness that we wanted... 682 00:31:35,266 --> 00:31:38,066 O'BRIEN (voiceover): Daniela Rus is director of M.I.T.'s Computer Science 683 00:31:38,066 --> 00:31:41,133 and Artificial Intelligence Lab. 684 00:31:41,133 --> 00:31:42,133 Oh, can you bring it to me? 685 00:31:42,133 --> 00:31:44,000 O'BRIEN (voiceover): CSAIL. 686 00:31:44,000 --> 00:31:46,066 They are different, like, kind of like muscles 687 00:31:46,066 --> 00:31:47,466 or actuators. 688 00:31:47,466 --> 00:31:49,066 RUS (voiceover): We can do so much more 689 00:31:49,066 --> 00:31:52,066 when we get people and machines working together. 690 00:31:53,233 --> 00:31:54,533 We can get better reach. 691 00:31:54,533 --> 00:31:55,533 We can get lift, 692 00:31:55,533 --> 00:31:58,700 precision, strength, vision. 693 00:31:58,700 --> 00:32:00,466 All of these are physical superpowers 694 00:32:00,466 --> 00:32:01,633 we can get through machines. 695 00:32:03,033 --> 00:32:04,066 O'BRIEN (voiceover): So, they're focusing 696 00:32:04,066 --> 00:32:05,633 on making it safe for humans 697 00:32:05,633 --> 00:32:08,866 to work in close proximity to machines. 698 00:32:08,866 --> 00:32:11,733 They're using some of the technology that's inside 699 00:32:11,733 --> 00:32:13,166 my prosthetic arm. 700 00:32:13,166 --> 00:32:15,233 Electrodes that can read 701 00:32:15,233 --> 00:32:17,666 the faint EMG signals generated 702 00:32:17,666 --> 00:32:19,000 as our nerves command 703 00:32:19,000 --> 00:32:20,466 our muscles to move. 704 00:32:22,866 --> 00:32:25,566 They have the capability to interact with a human, 705 00:32:25,566 --> 00:32:26,933 to understand the human, 706 00:32:26,933 --> 00:32:29,600 to step in and help the human as needed. 707 00:32:29,600 --> 00:32:33,466   I am at your disposal with 187 other languages, 708 00:32:33,466 --> 00:32:35,133 along with their various 709 00:32:35,133 --> 00:32:36,966 dialects and sub tongues. 710 00:32:36,966 --> 00:32:39,133 O'BRIEN (voiceover): But making robots as useful 711 00:32:39,133 --> 00:32:41,933 as they are in the movies is a big challenge. 712 00:32:41,933 --> 00:32:43,733   ♪ ♪ 713 00:32:43,733 --> 00:32:47,433 Most neural networks run on powerful supercomputers-- 714 00:32:47,433 --> 00:32:51,733 thousands of processors occupying entire buildings. 715 00:32:53,266 --> 00:32:54,966 RUS: We have brains that require 716 00:32:54,966 --> 00:32:58,800 massive computation, which you cannot include 717 00:32:58,800 --> 00:33:01,433 on a self-contained body. 718 00:33:01,433 --> 00:33:04,766 We address the size challenge by 719 00:33:04,766 --> 00:33:06,533 making liquid networks. 720 00:33:06,533 --> 00:33:08,533 O'BRIEN (voiceover): Liquid networks. 721 00:33:08,533 --> 00:33:10,033 So it looks like an autonomous vehicle 722 00:33:10,033 --> 00:33:11,266 like I've seen before, 723 00:33:11,266 --> 00:33:12,600 but it is a little different, right? 724 00:33:12,600 --> 00:33:13,966 ALEXANDER AMINI: Very different. 725 00:33:13,966 --> 00:33:15,300 This is an autonomous vehicle 726 00:33:15,300 --> 00:33:16,833 that can drive in brand-new environments 727 00:33:16,833 --> 00:33:19,466 that it has never seen before for the first time. 728 00:33:20,600 --> 00:33:22,700 O'BRIEN (voiceover): Most self-driving cars today rely, 729 00:33:22,700 --> 00:33:25,600 to some extent, on detailed databases 730 00:33:25,600 --> 00:33:28,500 that help them recognize their immediate environment. 731 00:33:28,500 --> 00:33:33,300 Those robot cars get lost in unfamiliar terrain. 732 00:33:34,733 --> 00:33:37,033 O'BRIEN: In this case, you're not relying on 733 00:33:37,033 --> 00:33:39,866 a huge, expansive neural network. 734 00:33:39,866 --> 00:33:41,366 You're running on 19 neurons, right? 735 00:33:41,366 --> 00:33:43,366 Correct. 736 00:33:43,366 --> 00:33:45,333 O'BRIEN (voiceover): Computer scientist Alexander Amini 737 00:33:45,333 --> 00:33:48,633 took me on a ride in an autonomous vehicle 738 00:33:48,633 --> 00:33:52,233 with a liquid neural network brain. 739 00:33:52,233 --> 00:33:54,366 AMINI: We've become very accustomed to relying on 740 00:33:54,366 --> 00:33:57,033 big, giant data centers and cloud compute. 741 00:33:57,033 --> 00:33:58,633 But in an autonomous vehicle, 742 00:33:58,633 --> 00:34:00,266 you cannot make such assumptions, right? 743 00:34:00,266 --> 00:34:01,833 You need to be able to operate, 744 00:34:01,833 --> 00:34:03,700   even if you lose internet connectivity 745 00:34:03,700 --> 00:34:06,200 and you cannot talk to the cloud anymore, 746 00:34:06,200 --> 00:34:07,766 your entire neural network, 747 00:34:07,766 --> 00:34:10,033 the brain of the car, needs to live on the car, 748 00:34:10,033 --> 00:34:12,800 and that imposes a lot of interesting constraints. 749 00:34:13,966 --> 00:34:15,366 O'BRIEN (voiceover): To build a brain smart enough 750 00:34:15,366 --> 00:34:17,333 and small enough to do this job, 751 00:34:17,333 --> 00:34:20,133 they took some inspiration from nature, 752 00:34:20,133 --> 00:34:24,333 a lowly worm called C. elegans. 753 00:34:24,333 --> 00:34:27,633 Its brain contains all of 300 neurons, 754 00:34:27,633 --> 00:34:30,100 but it's a very different kind of neuron. 755 00:34:32,133 --> 00:34:33,633 It can capture more complex behaviors 756 00:34:33,633 --> 00:34:35,166 in every single piece of that puzzle. 757 00:34:35,166 --> 00:34:36,366 And also the wiring, 758 00:34:36,366 --> 00:34:38,800 how a neuron talks to another neuron 759 00:34:38,800 --> 00:34:40,666 is completely different than what we see 760 00:34:40,666 --> 00:34:42,366 in today's neural networks. 761 00:34:43,800 --> 00:34:47,233 O'BRIEN (voiceover): Autonomous cars that tap into today's neural networks 762 00:34:47,233 --> 00:34:50,966 require huge amounts of compute power in the cloud. 763 00:34:52,500 --> 00:34:55,300 But this car is using just 19 liquid neurons. 764 00:34:56,433 --> 00:34:59,533 A worm at the wheel... sort of. 765 00:34:59,533 --> 00:35:01,000 AMINI (voiceover): Today's A.I. models 766 00:35:01,000 --> 00:35:02,700 are really pushing the boundaries 767 00:35:02,700 --> 00:35:05,300 of the scale of compute that we have. 768 00:35:05,300 --> 00:35:07,100 They're also pushing the boundaries 769 00:35:07,100 --> 00:35:08,433 of the data sets that we have. 770 00:35:08,433 --> 00:35:09,966 And that's not sustainable, 771 00:35:09,966 --> 00:35:11,900 because ultimately, we need to deploy A.I. 772 00:35:11,900 --> 00:35:13,500 onto the device itself, right? 773 00:35:13,500 --> 00:35:15,766 Onto the cars, onto the surgical robots. 774 00:35:15,766 --> 00:35:17,366 All of these edge devices 775 00:35:17,366 --> 00:35:20,733 that actually makes the decisions. 776 00:35:20,733 --> 00:35:23,366 O'BRIEN (voiceover): The A.I. worm may, in fact, 777 00:35:23,366 --> 00:35:24,666 turn. 778 00:35:27,500 --> 00:35:28,833 The portability of artificial intelligence 779 00:35:28,833 --> 00:35:32,100 was on my mind when it came time 780 00:35:32,100 --> 00:35:35,833 to pick up my new myoelectric arm... 781 00:35:35,833 --> 00:35:38,833 equipped with Coapt A.I. pattern recognition. 782 00:35:38,833 --> 00:35:40,666 All right, let's just check this 783 00:35:40,666 --> 00:35:42,133 real quick... 784 00:35:42,133 --> 00:35:43,500 O'BRIEN (voiceover): A few weeks after 785 00:35:43,500 --> 00:35:44,833 my trip to Chicago, 786 00:35:44,833 --> 00:35:46,266 I met Brian Monroe 787 00:35:46,266 --> 00:35:50,000 at his home office outside Washington, D.C. 788 00:35:50,000 --> 00:35:52,166 Are you happy with the way it came out? Yeah. 789 00:35:52,166 --> 00:35:54,133 Would you tell me otherwise? 790 00:35:54,133 --> 00:35:56,933 (laughing): Yeah, I would, yeah... 791 00:35:58,066 --> 00:35:59,366 O'BRIEN (voiceover): As usual, 792 00:35:59,366 --> 00:36:02,200 he did a great job making a tight socket. 793 00:36:03,566 --> 00:36:05,233 How's the socket feel? Does it feel like 794 00:36:05,233 --> 00:36:06,766 it's sliding down or 795 00:36:06,766 --> 00:36:09,200 falling out... No, it fits like a glove. 796 00:36:10,333 --> 00:36:11,900 O'BRIEN (voiceover): It's really important in this case, 797 00:36:11,900 --> 00:36:15,233 because the electrodes designed to read the signals 798 00:36:15,233 --> 00:36:17,600 from my muscles... 799 00:36:17,600 --> 00:36:19,266   ...have to stay in place snugly 800 00:36:19,266 --> 00:36:23,266 in order to generate accurate, reliable commands 801 00:36:23,266 --> 00:36:24,900 to the actuators in my new hand. 802 00:36:26,533 --> 00:36:28,200 Wait, is that you? That's me. 803 00:36:30,066 --> 00:36:32,366 (voiceover): He also provided me with 804 00:36:32,366 --> 00:36:34,366 a human-like bionic hand. 805 00:36:35,633 --> 00:36:37,400 But getting it to work just right 806 00:36:37,400 --> 00:36:39,333 took some time. 807 00:36:39,333 --> 00:36:41,566 That's open and it's closing. 808 00:36:41,566 --> 00:36:42,766 It's backwards? 809 00:36:42,766 --> 00:36:44,166 Yeah. Now try. 810 00:36:44,166 --> 00:36:45,533 If it's reversed, 811 00:36:45,533 --> 00:36:46,966 I can swap the electrodes. There we go. 812 00:36:46,966 --> 00:36:48,966 That's got it. Is it the right direction? 813 00:36:48,966 --> 00:36:50,333 Yeah. Uh-huh. Okay. 814 00:36:50,333 --> 00:36:53,766 O'BRIEN (voiceover): It's a long way from the movies, 815 00:36:53,766 --> 00:36:55,300 and I'm no Luke Skywalker. 816 00:36:55,300 --> 00:36:59,266 But my new arm and I are now together. 817 00:36:59,266 --> 00:37:01,033 And I'm heartened to know 818 00:37:01,033 --> 00:37:02,666 that I have the freedom and independence 819 00:37:02,666 --> 00:37:04,133 to teach and tweak it 820 00:37:04,133 --> 00:37:05,166 on my own. 821 00:37:05,166 --> 00:37:06,733 That's kind of cool. Yeah. 822 00:37:06,733 --> 00:37:09,266 (voiceover): Hopefully we will listen to each other. 823 00:37:09,266 --> 00:37:10,766 It's pretty awesome. 824 00:37:10,766 --> 00:37:12,466 O'BRIEN (voiceover): But we might want to listen 825 00:37:12,466 --> 00:37:14,533 with a skeptical ear. 826 00:37:15,766 --> 00:37:19,300 JORDAN PEELE (imitating Obama): You see, I would never say these things, 827 00:37:19,300 --> 00:37:21,833 at least not in a public address, 828 00:37:21,833 --> 00:37:23,766 but someone else would. 829 00:37:23,766 --> 00:37:26,100 Someone like Jordan Peele. 830 00:37:27,766 --> 00:37:29,900 This is a dangerous time. 831 00:37:29,900 --> 00:37:33,200 O'BRIEN (voiceover): It's even more dangerous now than it was in 2018 832 00:37:33,200 --> 00:37:35,333 when comedian Jordan Peele 833 00:37:35,333 --> 00:37:38,400 combined his pitch-perfect Obama impression 834 00:37:38,400 --> 00:37:44,233 with A.I. software to make this convincing fake video. 835 00:37:44,233 --> 00:37:47,166 ...or whether we become some kind of (bleep) up dystopia. 836 00:37:47,166 --> 00:37:49,166 ♪ ♪ 837 00:37:49,166 --> 00:37:51,233 O'BRIEN (voiceover): Fakes are about as old as 838 00:37:51,233 --> 00:37:53,200 photography itself. 839 00:37:53,200 --> 00:37:56,566 Mussolini, Hitler, and Stalin 840 00:37:56,566 --> 00:37:59,633 all ordered that pictures be doctored or redacted, 841 00:37:59,633 --> 00:38:03,100 erasing those who fell out of favor, 842 00:38:03,100 --> 00:38:05,233 consolidating power, 843 00:38:05,233 --> 00:38:08,333 manipulating their followers through images. 844 00:38:08,333 --> 00:38:09,600 HANY FARID: They've always been manipulated, 845 00:38:09,600 --> 00:38:12,000 throughout history, but-- 846 00:38:12,000 --> 00:38:14,266 there was literally, you can count on one hand, 847 00:38:14,266 --> 00:38:15,833 the number of people in the world who could do this. 848 00:38:15,833 --> 00:38:18,166 But now, you need almost no skill. 849 00:38:18,166 --> 00:38:19,900 And we said, "Give us an image 850 00:38:19,900 --> 00:38:21,100 "of a middle-aged woman, newscaster, 851 00:38:21,100 --> 00:38:22,833 sitting at her desk, reading the news." 852 00:38:22,833 --> 00:38:25,100 O'BRIEN (voiceover): Hany Farid is a professor of computer science 853 00:38:25,100 --> 00:38:27,033 at U.C. Berkeley. 854 00:38:27,033 --> 00:38:29,333 (on computer): And this is your daily dose of future flash. 855 00:38:29,333 --> 00:38:30,600 O'BRIEN (voiceover): He and his team 856 00:38:30,600 --> 00:38:32,933 are trying to navigate the house of mirrors 857 00:38:32,933 --> 00:38:36,266 that is the world of A.I.-enabled deepfake imagery. 858 00:38:37,266 --> 00:38:38,566   Not perfect. 859 00:38:38,566 --> 00:38:40,900 She's not blinking, but it's pretty good. 860 00:38:40,900 --> 00:38:43,633 And by the way, he did this in a day and a half. 861 00:38:43,633 --> 00:38:45,233 FARID (voiceover): It's the classic automation story. 862 00:38:45,233 --> 00:38:47,266 We have lowered barriers to entry 863 00:38:47,266 --> 00:38:49,133 to manipulate reality. 864 00:38:49,133 --> 00:38:50,666 And when you do that, 865 00:38:50,666 --> 00:38:52,066 more and more people will do it. 866 00:38:52,066 --> 00:38:53,200   Some good people will do it, 867 00:38:53,200 --> 00:38:54,433 but lots of bad people will do it. 868 00:38:54,433 --> 00:38:55,800 There'll be some interesting use cases, 869 00:38:55,800 --> 00:38:57,500 and there'll be a lot of nefarious use cases. 870 00:38:57,500 --> 00:39:00,766 Okay, so, um... 871 00:39:00,766 --> 00:39:02,866 Glasses off. How's the framing? 872 00:39:02,866 --> 00:39:03,966 Everything okay? 873 00:39:03,966 --> 00:39:05,333 (voiceover): About a week before 874 00:39:05,333 --> 00:39:06,966 I got on a plane to see him... Hold on. 875 00:39:06,966 --> 00:39:08,900 O'BRIEN (voiceover): He asked me to meet him on Zoom 876 00:39:08,900 --> 00:39:10,533 so he could get a good recording 877 00:39:10,533 --> 00:39:11,866 of my voice and mannerisms. 878 00:39:11,866 --> 00:39:14,366 And I assume you're recording, Miles. 879 00:39:14,366 --> 00:39:16,466 O'BRIEN (voiceover): And he turned the table on me a little bit, 880 00:39:16,466 --> 00:39:18,400 asking me a lot of questions 881 00:39:18,400 --> 00:39:20,100 to get a good sampling. 882 00:39:20,100 --> 00:39:21,700 FARID (on computer): How are you feeling about 883 00:39:21,700 --> 00:39:24,800 the role of A.I. as it enters into our world 884 00:39:24,800 --> 00:39:26,266 on a daily basis? 885 00:39:26,266 --> 00:39:28,233 I think it's very important, first of all, 886 00:39:28,233 --> 00:39:31,000 to calibrate the concern level. 887 00:39:31,000 --> 00:39:33,366 Let's take it away from the "Terminator" scenario... 888 00:39:34,566 --> 00:39:36,633 (voiceover): The "Terminator" scenario. 889 00:39:36,633 --> 00:39:38,066 Come with me if you want to live. 890 00:39:39,333 --> 00:39:42,400 O'BRIEN (voiceover): You know, a malevolent neural network 891 00:39:42,400 --> 00:39:44,133 hellbent on exterminating humanity. 892 00:39:44,133 --> 00:39:45,600 You're really real. 893 00:39:45,600 --> 00:39:47,533 O'BRIEN (voiceover): In the film series, 894 00:39:47,533 --> 00:39:48,866 the cyborg assassin 895 00:39:48,866 --> 00:39:51,966 is memorably played by Arnold Schwarzenegger. 896 00:39:51,966 --> 00:39:54,033 Hany thought it would be fun 897 00:39:54,033 --> 00:39:57,266 to use A.I. to turn Arnold into me. 898 00:39:57,266 --> 00:39:58,300   Okay. 899 00:39:59,500 --> 00:40:01,000 O'BRIEN (voiceover): A week later, I showed up at 900 00:40:01,000 --> 00:40:02,933 Berkeley's School of Information, 901 00:40:02,933 --> 00:40:06,800 ironically located in the oldest building on campus. 902 00:40:08,366 --> 00:40:10,300 So you had me do this strange thing on Zoom. 903 00:40:10,300 --> 00:40:12,666 Here I am. What did you do with me? 904 00:40:12,666 --> 00:40:14,300 Yeah, well, it's gonna teach you 905 00:40:14,300 --> 00:40:15,800 to let me record your Zoom call, isn't it? 906 00:40:15,800 --> 00:40:17,833 I did this with some trepidation. 907 00:40:17,833 --> 00:40:20,000 (voiceover): I was excited to see what tricks 908 00:40:20,000 --> 00:40:21,400 were up his sleeve. 909 00:40:21,400 --> 00:40:23,033 FARID (voiceover): I uploaded 90 seconds of audio, 910 00:40:23,033 --> 00:40:25,133   and I clicked a box saying 911 00:40:25,133 --> 00:40:27,500 "Miles has given me permission to use his voice," 912 00:40:27,500 --> 00:40:28,600 which I don't actually 913 00:40:28,600 --> 00:40:30,700 think you did. (chuckles) 914 00:40:30,700 --> 00:40:32,800 Um, and, I waited about, eh, maybe 20 seconds, 915 00:40:32,800 --> 00:40:35,600 and it said, "Okay, what would you like for Miles to say?" 916 00:40:35,600 --> 00:40:37,300 And I started typing, 917 00:40:37,300 --> 00:40:39,566 and I generated an audio of you saying 918 00:40:39,566 --> 00:40:40,933 whatever I wanted you to say. 919 00:40:40,933 --> 00:40:43,133 We are synthesizing, 920 00:40:43,133 --> 00:40:45,500 at much, much lower resolution. 921 00:40:45,500 --> 00:40:46,833 O'BRIEN (voiceover): You could have knocked me over 922 00:40:46,833 --> 00:40:49,566 with a feather when I watched this. 923 00:40:49,566 --> 00:40:50,966 A.I. O'BRIEN: Terminators were science fiction back then, 924 00:40:50,966 --> 00:40:54,266 but if you follow the recent A.I. media coverage, 925 00:40:54,266 --> 00:40:57,200 you might think that Terminators are just around the corner. 926 00:40:57,200 --> 00:40:58,966 The reality is... 927 00:40:58,966 --> 00:41:00,933 O'BRIEN (voiceover): The eyes and the mouth need some work, 928 00:41:00,933 --> 00:41:03,366 but it sure does sound like me. 929 00:41:04,366 --> 00:41:07,733 And consider what happened in May of 2023. 930 00:41:07,733 --> 00:41:10,833 Someone posted this A.I.-generated image 931 00:41:10,833 --> 00:41:13,300 of what appeared to be a terrorist bombing 932 00:41:13,300 --> 00:41:14,766 at the Pentagon. 933 00:41:14,766 --> 00:41:16,033 NEWS ANCHOR: Today we may have witnessed 934 00:41:16,033 --> 00:41:18,033 one of the first drops in the feared flood 935 00:41:18,033 --> 00:41:20,133 of A.I.-created disinformation. 936 00:41:20,133 --> 00:41:21,733 O'BRIEN (voiceover): It was shared on Twitter 937 00:41:21,733 --> 00:41:23,200 via what seemed to be 938 00:41:23,200 --> 00:41:26,500 a verified account from Bloomberg News. 939 00:41:26,500 --> 00:41:28,466 NEWS ANCHOR: It only took seconds to spread fast. 940 00:41:28,466 --> 00:41:31,766 The Dow now down about 200 points... 941 00:41:31,766 --> 00:41:33,633 Two minutes later, the stock market dropped 942 00:41:33,633 --> 00:41:36,000 a half a trillion dollars 943 00:41:36,000 --> 00:41:38,366 from a single fake image. 944 00:41:38,366 --> 00:41:39,933 Anybody could've made that image, 945 00:41:39,933 --> 00:41:41,833 whether it was intentionally manipulating the market 946 00:41:41,833 --> 00:41:43,000 or unintentionally, 947 00:41:43,000 --> 00:41:44,233 in some ways, it doesn't really matter. 948 00:41:45,333 --> 00:41:46,800 O'BRIEN (voiceover): So what are the technological 949 00:41:46,800 --> 00:41:50,166 innovations that make this tool widely available? 950 00:41:51,866 --> 00:41:53,800 One technique is called 951 00:41:53,800 --> 00:41:56,033 the generative adversarial network, 952 00:41:56,033 --> 00:41:57,266 or GAN. 953 00:41:57,266 --> 00:41:58,600 Two algorithms 954 00:41:58,600 --> 00:42:02,233 in a dizzying student-teacher back and forth. 955 00:42:02,233 --> 00:42:05,300 Let's say it's learning how to generate a cat. 956 00:42:05,300 --> 00:42:07,733 FARID: And it starts by just splatting down 957 00:42:07,733 --> 00:42:09,233 a bunch of pixels onto a canvas. 958 00:42:09,233 --> 00:42:12,266 And it sends it over to a discriminator. 959 00:42:12,266 --> 00:42:14,266 And the discriminator has access 960 00:42:14,266 --> 00:42:16,233 to millions and millions of images 961 00:42:16,233 --> 00:42:17,466 of the category that you want. 962 00:42:17,466 --> 00:42:18,933 And it says, 963 00:42:18,933 --> 00:42:20,833 "Nope, that doesn't look like all these other things." 964 00:42:20,833 --> 00:42:24,033 So it goes back to the generator and says, "Try again." 965 00:42:24,033 --> 00:42:25,233 Modifies some pixels, 966 00:42:25,233 --> 00:42:26,533 sends it back to the discriminator, 967 00:42:26,533 --> 00:42:28,000 and they do this in what's called 968 00:42:28,000 --> 00:42:29,233 an adversarial loop. 969 00:42:29,233 --> 00:42:30,633 O'BRIEN (voiceover): And eventually, 970 00:42:30,633 --> 00:42:33,066 after many thousands of volleys, 971 00:42:33,066 --> 00:42:36,000 the generator finally serves up a cat. 972 00:42:36,000 --> 00:42:38,000 And the discriminator says, 973 00:42:38,000 --> 00:42:40,233 "Do more like that." 974 00:42:40,233 --> 00:42:42,433 Today, we have a whole new way of doing these things. 975 00:42:42,433 --> 00:42:43,933 They're called diffusion-based. 976 00:42:44,966 --> 00:42:46,333 What diffusion does 977 00:42:46,333 --> 00:42:48,933 is it has vacuumed up billions of images 978 00:42:48,933 --> 00:42:51,466 with captions that are descriptive. 979 00:42:51,466 --> 00:42:54,133 O'BRIEN (voiceover): It starts by making those labeled images 980 00:42:54,133 --> 00:42:56,000 visually noisy on purpose. 981 00:42:57,833 --> 00:42:59,933 FARID: And then it corrupts it more, and it goes backwards 982 00:42:59,933 --> 00:43:01,466 and corrupts it more, and goes backwards 983 00:43:01,466 --> 00:43:02,533 and corrupts it more and goes backwards-- 984 00:43:02,533 --> 00:43:04,766 and it does that six billion times. 985 00:43:05,833 --> 00:43:07,200 O'BRIEN (voiceover): Eventually it corrupts it 986 00:43:07,200 --> 00:43:11,933 so it's unrecognizable from the original image. 987 00:43:11,933 --> 00:43:14,566 Now that it knows how to turn an image into nothing, 988 00:43:14,566 --> 00:43:16,466 it can reverse the process, 989 00:43:16,466 --> 00:43:20,166 turning seemingly nothing, into a beautiful image. 990 00:43:21,200 --> 00:43:22,933 FARID: What it's learned is how to take 991 00:43:22,933 --> 00:43:26,533 a completely indescript image, just pure noise, 992 00:43:26,533 --> 00:43:30,200 and go back to a coherent image, conditioned on a text prompt. 993 00:43:30,200 --> 00:43:33,466 You're basically reverse engineering an image 994 00:43:33,466 --> 00:43:34,866 down to the pixel. 995 00:43:34,866 --> 00:43:36,266 Yeah, exactly, yeah. 996 00:43:36,266 --> 00:43:38,333 And it's-- and by the way-- if you had asked me, 997 00:43:38,333 --> 00:43:39,866 "Will this work?" I would have said, 998 00:43:39,866 --> 00:43:41,333 "No, there's no way this system works." 999 00:43:41,333 --> 00:43:43,700 It just, it just doesn't seem like it should work. 1000 00:43:43,700 --> 00:43:45,500 And that's sort of the magic 1001 00:43:45,500 --> 00:43:47,133 of when you get this much data 1002 00:43:47,133 --> 00:43:49,466 and very powerful algorithms and very powerful computing 1003 00:43:49,466 --> 00:43:52,600 to be able to crunch these massive data sets. 1004 00:43:52,600 --> 00:43:54,466 I mean, we're not going to contain it. 1005 00:43:54,466 --> 00:43:55,566 That's done. 1006 00:43:55,566 --> 00:43:56,600 (voiceover): I sat down with Hany 1007 00:43:56,600 --> 00:43:57,766 and two of his grad students: 1008 00:43:57,766 --> 00:44:01,566 Justin Norman and Sarah Barrington. 1009 00:44:01,566 --> 00:44:04,066 We looked at some the A.I. trickery 1010 00:44:04,066 --> 00:44:05,700 they have seen and made. 1011 00:44:05,700 --> 00:44:08,100 Somebody else wrote some base code 1012 00:44:08,100 --> 00:44:09,533 and they got grew on to 1013 00:44:09,533 --> 00:44:11,300 and grow on to and grow on to and eventually... 1014 00:44:11,300 --> 00:44:12,733 O'BRIEN (voiceover): In a world where anything 1015 00:44:12,733 --> 00:44:14,800 can be manipulated with such ease 1016 00:44:14,800 --> 00:44:16,033 and seeming authenticity, 1017 00:44:16,033 --> 00:44:19,633 how are we to know what's real anymore? 1018 00:44:19,633 --> 00:44:20,800   How you look at the world, 1019 00:44:20,800 --> 00:44:22,200 how you interact with people in it, 1020 00:44:22,200 --> 00:44:24,033 and where you look for your threats of that change. 1021 00:44:24,033 --> 00:44:28,366 O'BRIEN (voiceover): Generative A.I. is now part of a larger ecosystem 1022 00:44:28,366 --> 00:44:31,666 that is built on mistrust. 1023 00:44:31,666 --> 00:44:32,900 We're going to live in a world where 1024 00:44:32,900 --> 00:44:34,433 we don't know what's real. 1025 00:44:34,433 --> 00:44:35,633 FARID (voiceover): There is distrust of governments, 1026 00:44:35,633 --> 00:44:37,266 there is distrust of media, 1027 00:44:37,266 --> 00:44:38,633 there is distrust of academics. 1028 00:44:38,633 --> 00:44:41,466 And now throw on top of that video evidence. 1029 00:44:41,466 --> 00:44:43,133 So-called video evidence. 1030 00:44:43,133 --> 00:44:44,900 I think this is the very definition 1031 00:44:44,900 --> 00:44:47,100 of throwing jet fuel onto a dumpster fire. 1032 00:44:47,100 --> 00:44:48,866 And it's already happening, 1033 00:44:48,866 --> 00:44:50,400 and I imagine we will see more of it. 1034 00:44:50,400 --> 00:44:52,033 (Arnold's voice): Come with me if you want to live. 1035 00:44:52,033 --> 00:44:53,700 O'BRIEN (voiceover): But it also can be 1036 00:44:53,700 --> 00:44:54,700 kind of fun. 1037 00:44:54,700 --> 00:44:55,866 As Hany promised, 1038 00:44:55,866 --> 00:44:58,000 here's my face 1039 00:44:58,000 --> 00:44:59,933 on the Terminator's body. 1040 00:44:59,933 --> 00:45:01,366   (gunfire blasting) 1041 00:45:01,366 --> 00:45:03,833 Long before A.I. might take 1042 00:45:03,833 --> 00:45:06,166 an existential turn against humanity, 1043 00:45:06,166 --> 00:45:08,966 we will need to reckon with the likes... 1044 00:45:08,966 --> 00:45:11,533 Go! Now! O'BRIEN (voiceover): Of the Milesinator. 1045 00:45:11,533 --> 00:45:13,566 TRAILER NARRATOR: This time, he's back. 1046 00:45:13,566 --> 00:45:15,200 (booming) 1047 00:45:15,200 --> 00:45:16,700 O'BRIEN (voiceover): Who will no doubt, be back. 1048 00:45:16,700 --> 00:45:18,266   Trust me. 1049 00:45:19,633 --> 00:45:20,800 O'BRIEN (voiceover): Trust, 1050 00:45:20,800 --> 00:45:23,100 but always verify. 1051 00:45:23,100 --> 00:45:26,366 So, what kind of A.I. magic 1052 00:45:26,366 --> 00:45:28,733 is readily available online? 1053 00:45:28,733 --> 00:45:30,600 It's pretty simple to make it look 1054 00:45:30,600 --> 00:45:33,066   like you're fluent in another language. 1055 00:45:33,066 --> 00:45:35,466 (speaking Mandarin): 1056 00:45:36,800 --> 00:45:37,966 It was pretty easy to do, 1057 00:45:37,966 --> 00:45:40,400 I just had to upload a video and wait. 1058 00:45:40,400 --> 00:45:43,333 (speaking German): 1059 00:45:44,766 --> 00:45:47,533 And, suddenly, I look pretty darn smart. 1060 00:45:47,533 --> 00:45:50,766 (speaking Greek): 1061 00:45:51,733 --> 00:45:53,900 Sure, it's fun, but I think you can see 1062 00:45:53,900 --> 00:45:55,600 where it leads to mischief 1063 00:45:55,600 --> 00:45:57,966 and possibly even mayhem. 1064 00:45:58,966 --> 00:46:03,466 (voiceover): Yoshua Bengio is an artificial intelligence pioneer. 1065 00:46:03,466 --> 00:46:05,166 He says he didn't spend much time 1066 00:46:05,166 --> 00:46:07,633 thinking about science fiction dystopia 1067 00:46:07,633 --> 00:46:10,566 as he was creating the technology. 1068 00:46:10,566 --> 00:46:13,500 But as his brilliant ideas became reality, 1069 00:46:13,500 --> 00:46:15,733 reality set in. 1070 00:46:15,733 --> 00:46:17,100 BENGIO: And the more I read, 1071 00:46:17,100 --> 00:46:19,000 the more I thought about it... 1072 00:46:19,000 --> 00:46:21,200 the more concerned I got. 1073 00:46:22,200 --> 00:46:25,233 If we are not honest with ourselves, 1074 00:46:25,233 --> 00:46:26,233 we're gonna fool ourselves. 1075 00:46:26,233 --> 00:46:28,233 We're gonna... lose. 1076 00:46:29,466 --> 00:46:30,866 O'BRIEN (voiceover): Avoiding that outcome 1077 00:46:30,866 --> 00:46:33,366 is now his main priority. 1078 00:46:33,366 --> 00:46:35,433 He has signed several public warnings 1079 00:46:35,433 --> 00:46:37,600 issued by A.I. thought leaders, 1080 00:46:37,600 --> 00:46:40,966 including this stark single-sentence statement 1081 00:46:40,966 --> 00:46:43,066 in May of 2023. 1082 00:46:43,066 --> 00:46:45,933 "Mitigating the risk of extinction from A.I. 1083 00:46:45,933 --> 00:46:47,900 "should be a global priority 1084 00:46:47,900 --> 00:46:50,633 "alongside other societal scale risks, 1085 00:46:50,633 --> 00:46:52,166 "such as pandemics 1086 00:46:52,166 --> 00:46:53,600 and nuclear war." 1087 00:46:56,666 --> 00:47:00,500 As we approach more and more capable A.I. systems 1088 00:47:00,500 --> 00:47:04,833 that might even become stronger than humans in many areas, 1089 00:47:04,833 --> 00:47:06,366   they become more and more dangerous. 1090 00:47:06,366 --> 00:47:07,766 Can't we just pull the plug on the thing? 1091 00:47:07,766 --> 00:47:09,366 Oh, that's the safest thing to do, 1092 00:47:09,366 --> 00:47:10,633 pull the plug. 1093 00:47:10,633 --> 00:47:13,066 Before it gets so powerful that 1094 00:47:13,066 --> 00:47:14,700 it prevents us from pulling the plug. 1095 00:47:14,700 --> 00:47:16,633 DAVE: Open the pod bay doors, Hal. 1096 00:47:16,633 --> 00:47:18,400 HAL: I'm sorry, Dave, 1097 00:47:18,400 --> 00:47:20,400 I'm afraid I can't do that. 1098 00:47:21,566 --> 00:47:23,366 O'BRIEN (voiceover): It may be some time 1099 00:47:23,366 --> 00:47:24,933 before computers are able 1100 00:47:24,933 --> 00:47:27,300 to act like movie supervillains... 1101 00:47:27,300 --> 00:47:28,333 HAL: Goodbye. 1102 00:47:29,766 --> 00:47:33,133 O'BRIEN (voiceover): But there are near-term dangers already emerging. 1103 00:47:33,133 --> 00:47:36,366 Besides deepfakes and misinformation, 1104 00:47:36,366 --> 00:47:40,300 A.I. can also supercharge bias and hate content, 1105 00:47:40,300 --> 00:47:43,133 replace human jobs... 1106 00:47:43,133 --> 00:47:44,766 This is why we're striking, everybody. (crowd exclaiming) 1107 00:47:45,900 --> 00:47:47,100 O'BRIEN (voiceover): And make it easier 1108 00:47:47,100 --> 00:47:50,500 for terrorists to create bioweapons. 1109 00:47:50,500 --> 00:47:53,300 And A.I. systems are so complex 1110 00:47:53,300 --> 00:47:56,000 that they are difficult to comprehend, 1111 00:47:56,000 --> 00:47:58,600 all but impossible to audit. 1112 00:47:58,600 --> 00:48:00,466 RUS (voiceover): Nobody really understands 1113 00:48:00,466 --> 00:48:03,133 how those systems reach their decisions. 1114 00:48:03,133 --> 00:48:05,300 So we have to be much more thoughtful 1115 00:48:05,300 --> 00:48:07,700 about how we test and evaluate them 1116 00:48:07,700 --> 00:48:09,100 before releasing them. 1117 00:48:09,100 --> 00:48:12,200 They're concerned whether machine will be able 1118 00:48:12,200 --> 00:48:14,400 to begin to think for itself. 1119 00:48:14,400 --> 00:48:17,133 O'BRIEN (voiceover): The U.S. and Europe have begun charting a strategy 1120 00:48:17,133 --> 00:48:18,866 to try to ensure safe, secure, 1121 00:48:18,866 --> 00:48:22,300 and trustworthy artificial intelligence. 1122 00:48:22,300 --> 00:48:24,766 RISHI SUNAK: ...in a way that will be safe for our communities... 1123 00:48:24,766 --> 00:48:26,233 O'BRIEN (voiceover): But how to do that 1124 00:48:26,233 --> 00:48:28,433 in the midst of a frenetic race 1125 00:48:28,433 --> 00:48:29,433 to dominate a technology 1126 00:48:29,433 --> 00:48:33,466 with a predicted economic impact 1127 00:48:33,466 --> 00:48:37,200 of 13 trillion dollars by 2030. 1128 00:48:37,200 --> 00:48:40,566 There is such a strong commercial incentive 1129 00:48:40,566 --> 00:48:43,066 to develop this and win the competition 1130 00:48:43,066 --> 00:48:44,333 against the other companies, 1131 00:48:44,333 --> 00:48:46,700 not to mention the other countries, 1132 00:48:46,700 --> 00:48:49,500 that it's hard to stop that train. 1133 00:48:50,600 --> 00:48:54,000 But that's what governments should be doing. 1134 00:48:54,000 --> 00:48:56,600 NEWS ANCHOR: The titans of social media 1135 00:48:56,600 --> 00:48:59,366 didn't want to come to Capitol Hill. 1136 00:48:59,366 --> 00:49:00,933 O'BRIEN (voiceover): Historically, the tech industry 1137 00:49:00,933 --> 00:49:03,833 has bridled against regulation. 1138 00:49:03,833 --> 00:49:06,900 You have an army of lawyers and lobbyists 1139 00:49:06,900 --> 00:49:08,166 that have fought us on this... 1140 00:49:08,166 --> 00:49:09,266 SULEYMAN (voiceover): There's no question that 1141 00:49:09,266 --> 00:49:10,566 guardrails will slow things down, 1142 00:49:10,566 --> 00:49:11,800 But, the risks are uncertain 1143 00:49:11,800 --> 00:49:15,200 and potentially enormous. 1144 00:49:15,200 --> 00:49:16,600 So, it makes sense for us 1145 00:49:16,600 --> 00:49:18,400 to start having the conversation right now. 1146 00:49:19,733 --> 00:49:21,300 O'BRIEN (voiceover): For me, the conversation 1147 00:49:21,300 --> 00:49:23,966 about A.I. is personal. 1148 00:49:23,966 --> 00:49:26,766 Okay, no network detected. 1149 00:49:26,766 --> 00:49:28,033 Okay, um... 1150 00:49:28,033 --> 00:49:30,366 Oh, here we go. Okay. 1151 00:49:30,366 --> 00:49:32,166 And now I'm going to open, open, open, open, open... 1152 00:49:33,633 --> 00:49:35,700 (voiceover): I used the Coapt app 1153 00:49:35,700 --> 00:49:38,900 to train the A.I. inside my new prosthetic. 1154 00:49:38,900 --> 00:49:41,900   ♪ ♪ 1155 00:49:41,900 --> 00:49:43,566 It says all of my training data is good, 1156 00:49:43,566 --> 00:49:44,866 it's four of five stars. 1157 00:49:44,866 --> 00:49:46,266 And now let's try to close. 1158 00:49:46,266 --> 00:49:47,866 (whirring) 1159 00:49:47,866 --> 00:49:49,033 All right. 1160 00:49:49,033 --> 00:49:53,900 Seems to be doing what it was told. 1161 00:49:53,900 --> 00:49:55,400 (voiceover): Was my new arm listening? 1162 00:49:55,400 --> 00:49:56,766 Maybe. 1163 00:49:56,766 --> 00:49:58,666 I decided to make things simpler. 1164 00:49:59,733 --> 00:50:03,733 I took off the hand and attached a myoelectric hook. 1165 00:50:03,733 --> 00:50:05,633 (quietly): All right. 1166 00:50:05,633 --> 00:50:08,266 (voiceover): Function over form. 1167 00:50:08,266 --> 00:50:10,966 Not a conversation piece necessarily at a cocktail party 1168 00:50:10,966 --> 00:50:12,933 like this thing is. 1169 00:50:12,933 --> 00:50:15,666 This looks more like Luke Skywalker, I suppose. 1170 00:50:15,666 --> 00:50:18,933 But this thing has a tremendous amount of function to it. 1171 00:50:18,933 --> 00:50:21,633 Although, right now, it wants to stay open. 1172 00:50:21,633 --> 00:50:23,600 (voiceover): And that problem persisted. 1173 00:50:23,600 --> 00:50:25,500   Find a tripod plate... 1174 00:50:25,500 --> 00:50:26,900 (voiceover): When I tried using it 1175 00:50:26,900 --> 00:50:28,800 to set up my basement studio 1176 00:50:28,800 --> 00:50:30,133 for a live broadcast. 1177 00:50:30,133 --> 00:50:32,800 Come on, close. 1178 00:50:32,800 --> 00:50:34,900 (voiceover): I was quickly frustrated. 1179 00:50:34,900 --> 00:50:37,266 (item drops, audio beep) 1180 00:50:37,266 --> 00:50:38,566 Really annoying. 1181 00:50:38,566 --> 00:50:41,100 Not useful. 1182 00:50:41,100 --> 00:50:44,433 (voiceover): The hook continuously opened on its own. 1183 00:50:44,433 --> 00:50:46,133 (clattering) Damn it! 1184 00:50:46,133 --> 00:50:48,600 (voiceover): So I completely reset 1185 00:50:48,600 --> 00:50:50,833 and retrained the arm. 1186 00:50:51,800 --> 00:50:53,800 And... reset, there we go. 1187 00:50:53,800 --> 00:50:56,600 Add data... 1188 00:50:56,600 --> 00:50:59,300 (voiceover): But the software was 1189 00:50:59,300 --> 00:51:00,766 artificially unhappy. 1190 00:51:02,700 --> 00:51:04,800 "Electrodes are not making good skin contact." 1191 00:51:04,800 --> 00:51:07,233 Maybe that is my problem, ultimately. 1192 00:51:08,533 --> 00:51:10,200 (voiceover): My problem really is 1193 00:51:10,200 --> 00:51:12,466 I haven't given this enough time. 1194 00:51:12,466 --> 00:51:14,733 Amputees tell me it can take 1195 00:51:14,733 --> 00:51:16,433 many months to really learn 1196 00:51:16,433 --> 00:51:18,400 how to use an arm like this one. 1197 00:51:19,400 --> 00:51:22,133 The choke point isn't artificial intelligence. 1198 00:51:22,133 --> 00:51:24,766   Dead as a doornail. 1199 00:51:24,766 --> 00:51:26,500 (voiceover): But rather, what is the best way 1200 00:51:26,500 --> 00:51:28,400 to communicate my intentions to it? 1201 00:51:29,766 --> 00:51:31,266 Little reboot there, I guess. 1202 00:51:31,266 --> 00:51:33,033 All right. 1203 00:51:33,033 --> 00:51:34,166 Close. 1204 00:51:34,166 --> 00:51:36,700 Open, close. 1205 00:51:36,700 --> 00:51:39,066   (voiceover): It turns out machine learning 1206 00:51:39,066 --> 00:51:42,100 isn't smart enough to give me a replacement arm 1207 00:51:42,100 --> 00:51:44,066 like Luke Skywalker got. 1208 00:51:44,066 --> 00:51:47,900 Nor is it capable of creating the Terminator. 1209 00:51:47,900 --> 00:51:51,833 Right now, it seems many hopes and fears 1210 00:51:51,833 --> 00:51:52,900 for artificial intelligence... 1211 00:51:52,900 --> 00:51:54,366 Oh! 1212 00:51:54,366 --> 00:51:57,100 (voiceover): ...are rooted in science fiction. 1213 00:51:59,066 --> 00:52:03,166 But we are walking down a road to the unknown. 1214 00:52:03,166 --> 00:52:06,066 The door is opening to a revolution. 1215 00:52:07,566 --> 00:52:08,600 (door closes) 1216 00:52:08,600 --> 00:52:12,633 ♪ ♪ 1217 00:52:31,766 --> 00:52:39,300 ♪ ♪ 1218 00:52:43,133 --> 00:52:50,666 ♪ ♪ 1219 00:52:52,300 --> 00:52:59,833 ♪ ♪ 1220 00:53:01,533 --> 00:53:09,066 ♪ ♪ 1221 00:53:14,800 --> 00:53:21,966 ♪ ♪ 95623

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.