All language subtitles for In the Age of AI (full film) _ FRONTLINE_HD

af Afrikaans
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bn Bengali
bs Bosnian
bg Bulgarian
ca Catalan
ceb Cebuano
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
tl Filipino
fi Finnish
fr French
fy Frisian
gl Galician
ka Georgian
de German
el Greek
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
km Khmer
ko Korean
ku Kurdish (Kurmanji)
ky Kyrgyz
lo Lao
la Latin
lv Latvian
lt Lithuanian
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mn Mongolian
my Myanmar (Burmese)
ne Nepali
no Norwegian
ps Pashto
fa Persian
pl Polish
pt Portuguese
pa Punjabi
ro Romanian
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
st Sesotho
sn Shona
sd Sindhi
si Sinhala
sk Slovak
sl Slovenian
so Somali
es Spanish
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
te Telugu
th Thai
tr Turkish
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
or Odia (Oriya)
rw Kinyarwanda
tk Turkmen
tt Tatar
ug Uyghur
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:16,630 --> 00:00:17,930 NARRATOR: Tonight-- 2 00:00:17,930 --> 00:00:20,470 The race to become an A.I.superpower is on... 3 00:00:20,470 --> 00:00:22,670 NARRATOR: The politics ofartificial intelligence... 4 00:00:22,670 --> 00:00:24,530 There will bea Chinese tech sector 5 00:00:24,530 --> 00:00:26,200 and there will bea American tech sector. 6 00:00:26,200 --> 00:00:27,730 NARRATOR: The new tech war. 7 00:00:27,730 --> 00:00:30,200 The more data,the better the A.I. works. 8 00:00:30,200 --> 00:00:33,930 So in the age of A.I.,where data is the new oil, 9 00:00:33,930 --> 00:00:36,170 China is the new Saudi Arabia. 10 00:00:36,170 --> 00:00:37,930 NARRATOR:The future of work... 11 00:00:37,930 --> 00:00:40,100 When I increase productivitythrough automation, 12 00:00:40,100 --> 00:00:42,070 jobs go away. 13 00:00:42,070 --> 00:00:46,100 I believe about 50% of jobswill be somewhat 14 00:00:46,100 --> 00:00:50,000 or extremely threatened by A.I.in the next 15 years or so. 15 00:00:50,000 --> 00:00:52,630 NARRATOR: A.I. and corporatesurveillance... 16 00:00:52,630 --> 00:00:55,500 We thought that we weresearching Google. 17 00:00:55,500 --> 00:00:57,930 We had no idea that Googlewas searching us. 18 00:00:57,930 --> 00:01:00,170 NARRATOR: And the threatto democracy. 19 00:01:00,170 --> 00:01:02,600 China is on its wayto building 20 00:01:02,600 --> 00:01:04,070 a total surveillance state. 21 00:01:04,070 --> 00:01:06,130 NARRATOR: Tonight on"Frontline"... 22 00:01:06,130 --> 00:01:09,300 It has pervaded so manyelements of everyday life. 23 00:01:09,300 --> 00:01:11,930 How do we make it transparentand accountable? 24 00:01:11,930 --> 00:01:13,770 NARRATOR:..."In the Age of A.I." 25 00:01:16,530 --> 00:01:21,200 ♪ ♪ 26 00:01:34,900 --> 00:01:37,970 ♪ ♪ 27 00:01:42,700 --> 00:01:46,330 NARRATOR: This is the world'smost complex board game. 28 00:01:48,070 --> 00:01:51,530 There are more possible movesin the game of Go 29 00:01:51,530 --> 00:01:55,800 than there are atomsin the universe. 30 00:01:55,800 --> 00:02:01,700 Legend has it that in 2300 BCE,Emperor Yao devised it 31 00:02:01,700 --> 00:02:08,000 to teach his son discipline,concentration, and balance. 32 00:02:08,000 --> 00:02:12,570 And, over 4,000 years later,this ancient Chinese game 33 00:02:12,570 --> 00:02:17,370 would signal the startof a new industrial age. 34 00:02:17,370 --> 00:02:18,430 ♪ ♪ 35 00:02:25,930 --> 00:02:31,400 It was 2016, in Seoul,South Korea. 36 00:02:31,400 --> 00:02:35,170 Can machines overtakehuman intelligence? 37 00:02:35,170 --> 00:02:37,730 A breakthrough moment when theworld champion 38 00:02:37,730 --> 00:02:40,630 of the Asian board game Gotakes on an A.I. program 39 00:02:40,630 --> 00:02:42,530 developed by Google. 40 00:02:42,530 --> 00:02:49,430 (speaking Korean): 41 00:02:55,300 --> 00:02:56,770 In countries whereit's very popular, 42 00:02:56,770 --> 00:03:00,570 like China and Japan and,and South Korea, to them, 43 00:03:00,570 --> 00:03:02,100 Go is not just a game, right? 44 00:03:02,100 --> 00:03:04,000 It's, like, how you learnstrategy. 45 00:03:04,000 --> 00:03:07,500 It has an almost spiritualcomponent. 46 00:03:07,500 --> 00:03:09,400 You know, if you talkto South Koreans, right, 47 00:03:09,400 --> 00:03:11,500 and Lee Sedol is the world'sgreatest Go player, 48 00:03:11,500 --> 00:03:13,730 he's a national heroin South Korea. 49 00:03:13,730 --> 00:03:18,630 They were sure that Lee Sedolwould beat AlphaGo hands down. 50 00:03:18,630 --> 00:03:23,030 ♪ ♪ 51 00:03:23,030 --> 00:03:26,130 NARRATOR: Google's AlphaGowas a computer program that, 52 00:03:26,130 --> 00:03:28,870 starting with the rules of Go 53 00:03:28,870 --> 00:03:31,330 and a databaseof historical games, 54 00:03:31,330 --> 00:03:34,630 had been designedto teach itself. 55 00:03:34,630 --> 00:03:38,700 I was one of the commentatorsat the Lee Sedol games. 56 00:03:38,700 --> 00:03:42,700 And yes, it was watched by tensof millions of people. 57 00:03:42,700 --> 00:03:44,300 (man speaking Korean) 58 00:03:44,300 --> 00:03:46,630 NARRATOR: ThroughoutSoutheast Asia, 59 00:03:46,630 --> 00:03:48,400 this was seen asa sports spectacle 60 00:03:48,400 --> 00:03:49,800 with national pride at stake. 61 00:03:49,800 --> 00:03:51,030 Wow, that was a player guess. 62 00:03:51,030 --> 00:03:53,530 NARRATOR: But much morewas in play. 63 00:03:53,530 --> 00:03:55,730 This was the public unveiling 64 00:03:55,730 --> 00:03:57,830 of a form of artificialintelligence 65 00:03:57,830 --> 00:04:00,400 called deep learning, 66 00:04:00,400 --> 00:04:03,400 that mimics the neural networksof the human brain. 67 00:04:03,400 --> 00:04:05,430 So what happens with machinelearning, 68 00:04:05,430 --> 00:04:08,300 or artificial intelligence--initially with AlphaGo-- 69 00:04:08,300 --> 00:04:12,130 is that the machine is fedall kinds of Go games, 70 00:04:12,130 --> 00:04:15,530 and then it studies them,learns from them, 71 00:04:15,530 --> 00:04:17,830 and figures out its own moves. 72 00:04:17,829 --> 00:04:19,699 And because it's an A.I.system-- 73 00:04:19,700 --> 00:04:21,570 it's not just followinginstructions, 74 00:04:21,570 --> 00:04:23,930 it's figuring out its owninstructions-- 75 00:04:23,930 --> 00:04:26,930 it comes up with moves thathumans hadn't thought of before. 76 00:04:26,930 --> 00:04:31,130 So, it studies games that humanshave played, it knows the rules, 77 00:04:31,130 --> 00:04:36,130 and then it comes upwith creative moves. 78 00:04:36,130 --> 00:04:38,030 (woman speaking Korean) 79 00:04:39,600 --> 00:04:42,370 (speaking Korean): 80 00:04:42,370 --> 00:04:44,800 That's a very...that's a very surprising move. 81 00:04:44,800 --> 00:04:47,670 I thought it was a mistake. 82 00:04:47,670 --> 00:04:51,470 NARRATOR: Game two, move 37. 83 00:04:51,470 --> 00:04:54,200 That move 37 was a move thathumans could not fathom, 84 00:04:54,200 --> 00:04:57,070 but yet it ended up beingbrilliant 85 00:04:57,070 --> 00:05:00,470 and woke people up to say, 86 00:05:00,470 --> 00:05:03,100 "Wow, after thousandsof years of playing, 87 00:05:03,100 --> 00:05:06,330 we never thought about makinga move like that." 88 00:05:06,330 --> 00:05:08,370 Oh, he resigned. 89 00:05:08,370 --> 00:05:12,300 It looks like... Lee Sedol hasjust resigned, actually. 90 00:05:12,300 --> 00:05:13,830 Yeah!Yes. 91 00:05:13,830 --> 00:05:15,530 NARRATOR: In the end, thescientists watched 92 00:05:15,530 --> 00:05:18,200 their algorithms win fourof the games. 93 00:05:18,200 --> 00:05:20,470 Lee Sedol took one. 94 00:05:20,470 --> 00:05:22,330 What happened with Go,first and foremost, 95 00:05:22,330 --> 00:05:25,830 was a huge victory for deep mindand for A.I., right? 96 00:05:25,830 --> 00:05:28,170 It wasn't that the computersbeat the humans, 97 00:05:28,170 --> 00:05:31,970 it was that, you know, one typeof intelligence beat another. 98 00:05:31,970 --> 00:05:34,230 NARRATOR: Artificialintelligence had proven 99 00:05:34,230 --> 00:05:36,770 it could marshal a vast amountof data, 100 00:05:36,770 --> 00:05:40,300 beyond anything any humancould handle, 101 00:05:40,300 --> 00:05:44,400 and use it to teach itself howto predict an outcome. 102 00:05:44,400 --> 00:05:48,400 The commercial implicationswere enormous. 103 00:05:48,400 --> 00:05:51,670 While AlphaGo is a,is a toy game, 104 00:05:51,670 --> 00:05:57,530 but its success and its wakingeveryone up, I think, 105 00:05:57,530 --> 00:06:03,770 is, is going to be rememberedas the pivotal moment 106 00:06:03,770 --> 00:06:07,230 where A.I. became mature 107 00:06:07,230 --> 00:06:09,100 and everybody jumpedon the bandwagon. 108 00:06:09,100 --> 00:06:10,570 ♪ ♪ 109 00:06:10,570 --> 00:06:14,270 NARRATOR: This is about theconsequences of that defeat. 110 00:06:14,270 --> 00:06:16,270 (man speaking local language) 111 00:06:16,270 --> 00:06:19,870 How the A.I. algorithms areushering in a new age 112 00:06:19,870 --> 00:06:24,200 of great potential andprosperity, 113 00:06:24,200 --> 00:06:29,170 but an age that will also deepeninequality, challenge democracy, 114 00:06:29,170 --> 00:06:35,200 and divide the worldinto two A.I. superpowers. 115 00:06:35,200 --> 00:06:39,130 Tonight, five stories about howartificial intelligence 116 00:06:39,130 --> 00:06:40,930 is changing our world. 117 00:06:40,930 --> 00:06:43,930 ♪ ♪ 118 00:06:51,800 --> 00:06:56,330 China has decided to chasethe A.I. future. 119 00:06:56,330 --> 00:06:58,770 The difference betweenthe internet mindset 120 00:06:58,770 --> 00:07:00,830 and the A.I. mindset... 121 00:07:00,830 --> 00:07:04,700 NARRATOR: A future made andembraced by a new generation. 122 00:07:07,000 --> 00:07:10,770 Well, it's hard not to feelthe kind of immense energy, 123 00:07:10,770 --> 00:07:15,570 and also the obvious factof the demographics. 124 00:07:15,570 --> 00:07:18,770 They're mostly very youngerpeople, 125 00:07:18,770 --> 00:07:22,830 so that this clearly istechnology which is being 126 00:07:22,830 --> 00:07:26,030 generated by a whole newgeneration. 127 00:07:26,030 --> 00:07:27,800 NARRATOR: Orville Schellis one of 128 00:07:27,800 --> 00:07:30,100 America's foremostChina scholars. 129 00:07:30,100 --> 00:07:31,730 (speaking Mandarin) 130 00:07:31,730 --> 00:07:34,830 NARRATOR: He first came here45 years ago. 131 00:07:34,830 --> 00:07:38,270 When I, when I first camehere, in 1975, 132 00:07:38,270 --> 00:07:40,770 Chairman Mao was still alive, 133 00:07:40,770 --> 00:07:43,300 the Cultural Revolutionwas coming on, 134 00:07:43,300 --> 00:07:47,830 and there wasn't a single whiffof anything 135 00:07:47,830 --> 00:07:49,170 of what you see here. 136 00:07:49,170 --> 00:07:50,770 It was unimaginable. 137 00:07:50,770 --> 00:07:54,500 In fact, in those years,one very much thought, 138 00:07:54,500 --> 00:08:00,530 "This is the way China is, thisis the way it's going to be." 139 00:08:00,530 --> 00:08:02,570 And the fact that it has gonethrough 140 00:08:02,570 --> 00:08:06,330 so many different changes sinceis quite extraordinary. 141 00:08:06,330 --> 00:08:08,270 (man giving instructions) 142 00:08:08,270 --> 00:08:11,770 NARRATOR: This extraordinaryprogress goes back 143 00:08:11,770 --> 00:08:14,370 to that game of Go. 144 00:08:14,370 --> 00:08:16,830 I think that the governmentrecognized 145 00:08:16,830 --> 00:08:18,300 that this was a sort of criticalthing for the future, 146 00:08:18,300 --> 00:08:20,270 and, "We need to catch upin this," that, you know, 147 00:08:20,270 --> 00:08:22,900 "We cannot have a foreigncompany showing us up 148 00:08:22,900 --> 00:08:24,300 at our own game. 149 00:08:24,300 --> 00:08:25,730 And this is going to besomething that is going to be 150 00:08:25,730 --> 00:08:27,100 critically importantin the future." 151 00:08:27,100 --> 00:08:29,230 So, you know, we called it theSputnik moment for, 152 00:08:29,230 --> 00:08:31,000 for the Chinese government-- 153 00:08:31,000 --> 00:08:33,970 the Chinese government kind ofwoke up. 154 00:08:33,969 --> 00:08:36,599 (translated): As we often sayin China, 155 00:08:36,599 --> 00:08:39,699 "The beginning is the mostdifficult part." 156 00:08:39,700 --> 00:08:42,630 NARRATOR: In 2017, Xi Jinpingannounced 157 00:08:42,630 --> 00:08:44,570 the government's bold new plans 158 00:08:44,570 --> 00:08:47,570 to an audienceof foreign diplomats. 159 00:08:47,570 --> 00:08:51,000 China would catch up with theU.S. in artificial intelligence 160 00:08:51,000 --> 00:08:55,170 by 2025 and lead the worldby 2030. 161 00:08:55,170 --> 00:08:57,500 (translated): ...andintensified cooperation 162 00:08:57,500 --> 00:09:00,270 in frontier areas such asdigital economy, 163 00:09:00,270 --> 00:09:02,930 artificial intelligence,nanotechnology, 164 00:09:02,930 --> 00:09:05,230 and accounting computing. 165 00:09:05,230 --> 00:09:08,730 ♪ ♪ 166 00:09:11,530 --> 00:09:15,200 NARRATOR: Today, China leadsthe world in e-commerce. 167 00:09:18,370 --> 00:09:22,070 Drones deliver to ruralvillages. 168 00:09:22,070 --> 00:09:25,070 And a society that bypassedcredit cards 169 00:09:25,070 --> 00:09:28,030 now shops in storeswithout cashiers, 170 00:09:28,030 --> 00:09:33,200 where the currencyis facial recognition. 171 00:09:33,200 --> 00:09:36,230 No country has ever movedthat fast. 172 00:09:36,230 --> 00:09:38,730 And in a short two-and-a-halfyears, 173 00:09:38,730 --> 00:09:43,400 China's A.I. implementationreally went from minimal amount 174 00:09:43,400 --> 00:09:47,230 to probably about17 or 18 unicorns, 175 00:09:47,230 --> 00:09:50,000 that is billion-dollarcompanies, in A.I. today. 176 00:09:50,000 --> 00:09:55,130 And that, that progress is,is hard to believe. 177 00:09:55,130 --> 00:09:57,830 NARRATOR: The progress waspowered by a new generation 178 00:09:57,830 --> 00:10:01,870 of ambitious young techs pouringout of Chinese universities, 179 00:10:01,870 --> 00:10:05,570 competing with each otherfor new ideas, 180 00:10:05,570 --> 00:10:11,630 and financed by a new cadre ofChinese venture capitalists. 181 00:10:11,630 --> 00:10:13,600 This is Sinovation, 182 00:10:13,600 --> 00:10:17,100 created by U.S.-educated A.I.scientist and businessman 183 00:10:17,100 --> 00:10:19,000 Kai-Fu Lee. 184 00:10:19,000 --> 00:10:24,170 These unicorns-- we've gotone, two, three, four, five, 185 00:10:24,170 --> 00:10:27,300 six, in the general A.I. area. 186 00:10:27,300 --> 00:10:29,630 And unicorn means abillion-dollar company, 187 00:10:29,630 --> 00:10:33,870 a company whose valuationor market capitalization 188 00:10:33,870 --> 00:10:36,870 is at $1 billion or higher. 189 00:10:36,870 --> 00:10:42,830 I think we put two unicornsto show $5 billion or higher. 190 00:10:42,830 --> 00:10:45,300 NARRATOR: Kai-Fu Lee was bornin Taiwan. 191 00:10:45,300 --> 00:10:48,530 His parents sent himto high school in Tennessee. 192 00:10:48,530 --> 00:10:51,270 His PhD thesisat Carnegie Mellon 193 00:10:51,270 --> 00:10:53,900 was on computer speechrecognition, 194 00:10:53,900 --> 00:10:55,570 which took him to Apple. 195 00:10:55,570 --> 00:10:57,900 Well, reality is a stepcloser to science fiction, 196 00:10:57,900 --> 00:11:00,770 with Apple Computers'new developed program... 197 00:11:00,770 --> 00:11:03,830 NARRATOR: And at 31,an early measure of fame. 198 00:11:03,830 --> 00:11:06,100 Kai-Fu Lee,the inventor of Apple's 199 00:11:06,100 --> 00:11:07,530 speech-recognition technology. 200 00:11:07,530 --> 00:11:10,400 Casper, copy thisto Make Write 2. 201 00:11:10,400 --> 00:11:12,730 Casper, paste. 202 00:11:12,730 --> 00:11:15,600 Casper, 72-point italic outline. 203 00:11:15,600 --> 00:11:18,970 NARRATOR: He would move on toMicrosoft research in Asia 204 00:11:18,970 --> 00:11:21,430 and became the headof Google China. 205 00:11:21,430 --> 00:11:26,530 Ten years ago, he startedSinovation in Beijing, 206 00:11:26,530 --> 00:11:30,700 and began looking for promisingstartups and A.I. talent. 207 00:11:30,700 --> 00:11:33,500 So, the Chineseentrepreneurial companies 208 00:11:33,500 --> 00:11:35,500 started as copycats. 209 00:11:35,500 --> 00:11:39,570 But over the last 15 years,China has developed its own form 210 00:11:39,570 --> 00:11:45,100 of entrepreneurship, and thatentrepreneurship is described 211 00:11:45,100 --> 00:11:50,130 as tenacious, very fast,winner-take-all, 212 00:11:50,130 --> 00:11:52,930 and incredible work ethic. 213 00:11:52,930 --> 00:11:57,430 I would say these few thousandChinese top entrepreneurs, 214 00:11:57,430 --> 00:11:59,230 they could take on anyentrepreneur 215 00:11:59,230 --> 00:12:01,400 anywhere in the world. 216 00:12:01,400 --> 00:12:04,170 NARRATOR: Entrepreneurs likeCao Xudong, 217 00:12:04,170 --> 00:12:10,100 the 33-year-old C.E.O. ofa new startup called Momenta. 218 00:12:10,100 --> 00:12:12,700 This is a ring road aroundBeijing. 219 00:12:12,700 --> 00:12:15,470 The car is driving itself. 220 00:12:15,470 --> 00:12:18,730 ♪ ♪ 221 00:12:21,230 --> 00:12:24,470 You see, another cutting,another cutting-in. 222 00:12:24,470 --> 00:12:26,670 Another cut-in, yeah, yeah. 223 00:12:26,670 --> 00:12:29,470 NARRATOR: Cao has no doubtabout the inevitability 224 00:12:29,470 --> 00:12:33,430 of autonomous vehicles. 225 00:12:33,430 --> 00:12:39,130 Just like AlphaGo can beatthe human player in, in Go, 226 00:12:39,130 --> 00:12:43,570 I think the machine willdefinitely surpass 227 00:12:43,570 --> 00:12:47,370 the human driver, in the end. 228 00:12:47,370 --> 00:12:48,730 NARRATOR: Recently, therehave been cautions 229 00:12:48,730 --> 00:12:53,530 about how soon autonomousvehicles will be deployed, 230 00:12:53,530 --> 00:12:55,700 but Cao and his team areconfident 231 00:12:55,700 --> 00:12:58,730 they're in for the long haul. 232 00:12:58,730 --> 00:13:01,030 U.S. will be the firstto deploy, 233 00:13:01,030 --> 00:13:03,830 but China may be the firstto popularize. 234 00:13:03,830 --> 00:13:05,270 It is 50-50 right now. 235 00:13:05,270 --> 00:13:07,000 U.S. is ahead in technology. 236 00:13:07,000 --> 00:13:10,030 China has a larger market,and the Chinese government 237 00:13:10,030 --> 00:13:12,870 is helping with infrastructureefforts-- 238 00:13:12,870 --> 00:13:16,100 for example, building a new citythe size of Chicago 239 00:13:16,100 --> 00:13:18,670 with autonomous driving enabled, 240 00:13:18,670 --> 00:13:21,700 and also a new highway that hassensors built in 241 00:13:21,700 --> 00:13:24,230 to help autonomous vehiclebe safer. 242 00:13:24,230 --> 00:13:27,470 NARRATOR: Their earlyinvestors included 243 00:13:27,470 --> 00:13:29,470 Mercedes-Benz. 244 00:13:29,470 --> 00:13:33,430 I feel very lucky and veryinspiring 245 00:13:33,430 --> 00:13:38,000 and very exciting that we'reliving in this era. 246 00:13:38,000 --> 00:13:40,800 ♪ ♪ 247 00:13:40,800 --> 00:13:42,630 NARRATOR: Life in China islargely conducted 248 00:13:42,630 --> 00:13:45,000 on smartphones. 249 00:13:45,000 --> 00:13:48,270 A billion people use WeChat,the equivalent of Facebook, 250 00:13:48,270 --> 00:13:51,030 Messenger, and PayPal,and much more, 251 00:13:51,030 --> 00:13:54,200 combined into just onesuper-app. 252 00:13:54,200 --> 00:13:55,870 And there are many more. 253 00:13:55,870 --> 00:14:00,070 China is the best placefor A.I. implementation today, 254 00:14:00,070 --> 00:14:04,230 because the vast amount of datathat's available in China. 255 00:14:04,230 --> 00:14:07,670 China has a lot more users thanany other country, 256 00:14:07,670 --> 00:14:10,700 three to four times more thanthe U.S. 257 00:14:10,700 --> 00:14:14,900 There are 50 times more mobilepayments than the U.S. 258 00:14:14,900 --> 00:14:17,300 There are ten times more fooddeliveries, 259 00:14:17,300 --> 00:14:21,030 which serve as data to learnmore about user behavior 260 00:14:21,030 --> 00:14:22,870 than the U.S. 261 00:14:22,870 --> 00:14:26,570 300 times more shared bicyclerides, 262 00:14:26,570 --> 00:14:30,400 and each shared bicycle ridehas all kinds of sensors 263 00:14:30,400 --> 00:14:32,730 submitting data up to the cloud. 264 00:14:32,730 --> 00:14:36,230 We're talking about maybe tentimes more data than the U.S., 265 00:14:36,230 --> 00:14:41,230 and A.I. is basically run ondata and fueled by data. 266 00:14:41,230 --> 00:14:44,400 The more data, the betterthe A.I. works, 267 00:14:44,400 --> 00:14:47,500 more importantly than howbrilliant the researcher is 268 00:14:47,500 --> 00:14:49,000 working on the problem. 269 00:14:49,000 --> 00:14:54,100 So, in the age of A.I.,where data is the new oil, 270 00:14:54,100 --> 00:14:57,370 China is the new Saudi Arabia. 271 00:14:57,370 --> 00:14:59,570 NARRATOR: And access to allthat data 272 00:14:59,570 --> 00:15:02,870 means that the deep-learningalgorithm can quickly predict 273 00:15:02,870 --> 00:15:05,370 behavior, like thecreditworthiness of someone 274 00:15:05,370 --> 00:15:06,870 wanting a short-term loan. 275 00:15:06,870 --> 00:15:09,270 Here is our application. 276 00:15:09,270 --> 00:15:13,900 And customer can choose how manymoney they want to borrow 277 00:15:13,900 --> 00:15:16,830 and how long they wantto borrow, 278 00:15:16,830 --> 00:15:21,130 and they can inputtheir datas here. 279 00:15:21,130 --> 00:15:27,530 And after, after that, you canjust borrow very quickly. 280 00:15:27,530 --> 00:15:31,030 NARRATOR: The C.E.O. shows ushow quickly you can get a loan. 281 00:15:31,030 --> 00:15:33,100 It is, it has done. 282 00:15:33,100 --> 00:15:35,430 NARRATOR: It takes an averageof eight seconds. 283 00:15:35,430 --> 00:15:38,400 It has passed to banks.Wow. 284 00:15:38,400 --> 00:15:40,170 NARRATOR:In the eight seconds, 285 00:15:40,170 --> 00:15:42,930 the algorithm has assessed5,000 personal features 286 00:15:42,930 --> 00:15:44,630 from all your data. 287 00:15:44,630 --> 00:15:50,500 5,000 features that isrelated with the delinquency, 288 00:15:50,500 --> 00:15:57,800 when maybe the banks only usefew, maybe, maybe ten features 289 00:15:57,800 --> 00:16:02,170 when they are doingtheir risk amendment. 290 00:16:02,170 --> 00:16:03,630 NARRATOR: Processing millionsof transactions, 291 00:16:03,630 --> 00:16:06,930 it'll dig up features that wouldnever be apparent 292 00:16:06,930 --> 00:16:11,870 to a human loan officer,like how confidently you type 293 00:16:11,870 --> 00:16:15,600 your loan application,or, surprisingly, 294 00:16:15,600 --> 00:16:18,670 if you keep your cell phonebattery charged. 295 00:16:18,670 --> 00:16:21,300 It's very interesting, thebattery of the phone 296 00:16:21,300 --> 00:16:24,170 is related with theirdelinquency rate. 297 00:16:24,170 --> 00:16:26,530 Someone who has much morelower battery, 298 00:16:26,530 --> 00:16:31,400 they get much more dangerousthan others. 299 00:16:31,400 --> 00:16:34,370 It's probably unfathomableto an American 300 00:16:34,370 --> 00:16:39,600 how a country can dramaticallyevolve itself 301 00:16:39,600 --> 00:16:43,470 from a copycat laggard to,all of a sudden, 302 00:16:43,470 --> 00:16:48,030 to nearly as good as the U.S. intechnology. 303 00:16:48,030 --> 00:16:50,300 NARRATOR: Like thisfacial-recognition startup 304 00:16:50,300 --> 00:16:51,800 he invested in. 305 00:16:51,800 --> 00:16:56,470 Megvii was started by threeyoung graduates in 2011. 306 00:16:56,470 --> 00:17:00,700 It's now a world leader in usingA.I. to identify people. 307 00:17:03,530 --> 00:17:05,000 It's pretty fast. 308 00:17:05,000 --> 00:17:07,530 For example,on the mobile device, 309 00:17:07,530 --> 00:17:10,670 we have timed thefacial-recognition speed. 310 00:17:10,670 --> 00:17:13,830 It's actually lessthan 100 milliseconds. 311 00:17:13,829 --> 00:17:15,829 So, that's very, very fast. 312 00:17:15,829 --> 00:17:19,969 So 0.1 second that we can, wewill be able to recognize you, 313 00:17:19,970 --> 00:17:24,200 even on a mobile device. 314 00:17:24,199 --> 00:17:26,299 NARRATOR: The company claimsthe system is better 315 00:17:26,300 --> 00:17:30,170 than any human at identifyingpeople in its database. 316 00:17:30,170 --> 00:17:33,770 And for those who aren't,it can describe them. 317 00:17:33,770 --> 00:17:36,530 Like our director--what he's wearing, 318 00:17:36,530 --> 00:17:42,230 and a good guess at his age,missing it by only a few months. 319 00:17:42,230 --> 00:17:46,970 We are the first one toreally take facial recognition 320 00:17:46,970 --> 00:17:50,570 to commercial quality. 321 00:17:50,570 --> 00:17:52,070 NARRATOR: That's why inBeijing today, 322 00:17:52,070 --> 00:17:57,630 you can pay for your KFCwith a smile. 323 00:17:57,630 --> 00:17:59,070 You know, it's not sosurprising, 324 00:17:59,070 --> 00:18:01,230 we've seen Chinese companiescatching up to the U.S. 325 00:18:01,230 --> 00:18:02,630 in technology for a long time. 326 00:18:02,630 --> 00:18:05,200 And so, if particular effortand attention is paid 327 00:18:05,200 --> 00:18:07,700 in a specific sector,it's not so surprising 328 00:18:07,700 --> 00:18:09,600 that they would surpassthe rest of the world. 329 00:18:09,600 --> 00:18:12,000 And facial recognition is one ofthe, really the first places 330 00:18:12,000 --> 00:18:15,300 we've seen that start to happen. 331 00:18:15,300 --> 00:18:18,230 NARRATOR: It's a technologyprized by the government, 332 00:18:18,230 --> 00:18:23,370 like this program in Shenzhento discourage jaywalking. 333 00:18:23,370 --> 00:18:27,400 Offenders are shamed in public--and with facial recognition, 334 00:18:27,400 --> 00:18:31,330 can be instantly fined. 335 00:18:31,330 --> 00:18:34,570 Critics warn that the governmentand some private companies 336 00:18:34,570 --> 00:18:37,170 have been building a nationaldatabase 337 00:18:37,170 --> 00:18:41,330 from dozens of experimentalsocial-credit programs. 338 00:18:41,330 --> 00:18:43,500 The government wants tointegrate 339 00:18:43,500 --> 00:18:48,670 all these individual behaviors,or corporations' records, 340 00:18:48,670 --> 00:18:55,670 into some kind of metrics andcompute out a single number 341 00:18:55,670 --> 00:18:59,030 or set of number associatedwith a individual, 342 00:18:59,030 --> 00:19:04,670 a citizen, and using that,to implement a incentive 343 00:19:04,670 --> 00:19:06,130 or punishment system. 344 00:19:06,130 --> 00:19:07,400 NARRATOR: A highsocial-credit number 345 00:19:07,400 --> 00:19:11,100 can be rewarded with discountson bus fares. 346 00:19:11,100 --> 00:19:15,800 A low number can leadto a travel ban. 347 00:19:15,800 --> 00:19:18,430 Some say it's very popularwith a Chinese public 348 00:19:18,430 --> 00:19:21,500 that wants to punishbad behavior. 349 00:19:21,500 --> 00:19:25,070 Others see a future that rewardsparty loyalty 350 00:19:25,070 --> 00:19:28,570 and silences criticism. 351 00:19:28,570 --> 00:19:32,970 Right now, there is no finalsystem being implemented. 352 00:19:32,970 --> 00:19:41,070 And from those experiments, wealready see that the possibility 353 00:19:41,070 --> 00:19:44,400 of what this social-creditsystem can do to individual. 354 00:19:44,400 --> 00:19:48,400 It's very powerful--Orwellian-like-- 355 00:19:48,400 --> 00:19:56,170 and it's extremely troublesomein terms of civil liberty. 356 00:19:56,170 --> 00:19:58,270 NARRATOR: Every eveningin Shanghai, 357 00:19:58,270 --> 00:20:01,270 ever-present cameras record thecrowds 358 00:20:01,270 --> 00:20:03,430 as they surge down to the Bund, 359 00:20:03,430 --> 00:20:07,530 the promenade along the banksof the Huangpu River. 360 00:20:07,530 --> 00:20:10,830 Once the great trading houses ofEurope came here to do business 361 00:20:10,830 --> 00:20:12,570 with the Middle Kingdom. 362 00:20:12,570 --> 00:20:15,600 In the last century,they were all shut down 363 00:20:15,600 --> 00:20:18,200 by Mao's revolution. 364 00:20:18,200 --> 00:20:20,530 But now, in the age of A.I., 365 00:20:20,530 --> 00:20:22,670 people come here to takein a spectacle 366 00:20:22,670 --> 00:20:26,100 that reflects China'sremarkable progress. 367 00:20:26,100 --> 00:20:28,630 (spectators gasp) 368 00:20:28,630 --> 00:20:32,070 And illuminates the greatpolitical paradox of capitalism 369 00:20:32,070 --> 00:20:37,200 taken rootin the communist state. 370 00:20:37,200 --> 00:20:40,800 People have called itmarket Leninism, 371 00:20:40,800 --> 00:20:43,570 authoritarian capitalism. 372 00:20:43,570 --> 00:20:46,970 We are watching a kindof a Petri dish 373 00:20:46,970 --> 00:20:54,270 in which an experiment of, youknow, extraordinary importance 374 00:20:54,270 --> 00:20:55,970 to the world isbeing carried out. 375 00:20:55,970 --> 00:20:59,030 Whether you can combine thesethings 376 00:20:59,030 --> 00:21:02,170 and get somethingthat's more powerful, 377 00:21:02,170 --> 00:21:04,870 that's coherent,that's durable in the world. 378 00:21:04,870 --> 00:21:07,600 Whether you can bring togethera one-party state 379 00:21:07,600 --> 00:21:12,370 with an innovative sector,both economically 380 00:21:12,370 --> 00:21:14,400 and technologically innovative, 381 00:21:14,400 --> 00:21:20,700 and that's something we thoughtcould not coexist. 382 00:21:20,700 --> 00:21:23,070 NARRATOR:As China reinvents itself, 383 00:21:23,070 --> 00:21:25,170 it has set its sightson leading the world 384 00:21:25,170 --> 00:21:29,170 in artificial intelligenceby 2030. 385 00:21:29,170 --> 00:21:32,230 But that means taking on theworld's most innovative 386 00:21:32,230 --> 00:21:34,100 A.I. culture. 387 00:21:34,100 --> 00:21:37,600 ♪ ♪ 388 00:21:46,900 --> 00:21:49,930 On an interstatein the U.S. Southwest, 389 00:21:49,930 --> 00:21:52,830 artificial intelligence is atwork solving the problem 390 00:21:52,830 --> 00:21:56,030 that's become emblematicof the new age, 391 00:21:56,030 --> 00:21:58,800 replacing a human driver. 392 00:21:58,800 --> 00:22:04,370 ♪ ♪ 393 00:22:04,370 --> 00:22:08,730 This is the company's C.E.O.,24-year-old Alex Rodrigues. 394 00:22:11,630 --> 00:22:13,800 The more things we buildsuccessfully, 395 00:22:13,800 --> 00:22:15,970 the less people ask questions 396 00:22:15,970 --> 00:22:18,870 about how old you are when youhave working trucks. 397 00:22:18,870 --> 00:22:21,800 NARRATOR: And this is whathe's built. 398 00:22:21,800 --> 00:22:24,670 Commercial goods are beingdriven from California 399 00:22:24,670 --> 00:22:29,170 to Arizona on Interstate 10. 400 00:22:29,170 --> 00:22:34,270 There is a driver in the cab,but he's not driving. 401 00:22:34,270 --> 00:22:40,630 It's a path set by a C.E.O.with an unusual CV. 402 00:22:40,630 --> 00:22:42,930 Are we ready, Henry? 403 00:22:42,930 --> 00:22:47,730 The aim is to score these pucksinto the scoring area. 404 00:22:47,730 --> 00:22:51,400 So I, I did competitive roboticsstarting when I was 11, 405 00:22:51,400 --> 00:22:53,130 and I took it very, veryseriously. 406 00:22:53,130 --> 00:22:55,900 To, to give you a sense, I wonthe Robotics World Championships 407 00:22:55,900 --> 00:22:57,830 for the first timewhen I was 13. 408 00:22:57,830 --> 00:22:59,430 I've been to worlds seven times 409 00:22:59,430 --> 00:23:02,200 between the ages of 13and 20-ish. 410 00:23:02,200 --> 00:23:04,330 I eventually founded a team, 411 00:23:04,330 --> 00:23:07,000 did a lot of work at avery high competitive level. 412 00:23:07,000 --> 00:23:08,470 Things looking pretty good. 413 00:23:08,470 --> 00:23:10,930 NARRATOR: This was aprototype of sorts, 414 00:23:10,930 --> 00:23:15,130 from which he has built hismulti-million-dollar company. 415 00:23:15,130 --> 00:23:18,100 I hadn't built a robot in awhile, wanted to get back to it, 416 00:23:18,100 --> 00:23:21,030 and felt that this was by farthe most exciting piece 417 00:23:21,030 --> 00:23:22,930 of robotics technology that wasup and coming. 418 00:23:22,930 --> 00:23:25,170 A lot of people told us wewouldn't be able to build it. 419 00:23:25,170 --> 00:23:28,570 But knew roughly the techniquesthat you would use. 420 00:23:28,570 --> 00:23:30,370 And I was pretty confident thatif you put them together, 421 00:23:30,370 --> 00:23:32,330 you would get somethingthat worked. 422 00:23:32,330 --> 00:23:35,900 Took the summer off, built in myparents' garage a golf cart 423 00:23:35,900 --> 00:23:40,470 that could drive itself. 424 00:23:40,470 --> 00:23:42,430 NARRATOR: That golf cartgot the attention 425 00:23:42,430 --> 00:23:45,400 of Silicon Valley,and the first of several rounds 426 00:23:45,400 --> 00:23:47,570 of venture capital. 427 00:23:47,570 --> 00:23:50,670 He formed a team and thendecided the business opportunity 428 00:23:50,670 --> 00:23:53,700 was in self-driving trucks. 429 00:23:53,700 --> 00:23:56,470 He says there's alsoa human benefit. 430 00:23:56,470 --> 00:23:58,630 If we can build a truckthat's ten times safer 431 00:23:58,630 --> 00:24:02,770 than a human driver, then notmuch else actually matters. 432 00:24:02,770 --> 00:24:05,770 When we talk to regulators,especially, 433 00:24:05,770 --> 00:24:08,930 everyone agrees that the onlyway that we're going to get 434 00:24:08,930 --> 00:24:11,770 to zero highway deaths,which is everyone's objective, 435 00:24:11,770 --> 00:24:13,800 is to use self-driving. 436 00:24:13,800 --> 00:24:17,030 And so, I'm sure you've heardthe statistic, 437 00:24:17,030 --> 00:24:19,230 more than 90% of all crashes 438 00:24:19,230 --> 00:24:20,870 have a human driveras the cause. 439 00:24:20,870 --> 00:24:24,230 So if you want to solvetraffic fatalities, 440 00:24:24,230 --> 00:24:28,170 which, in my opinion, are thesingle biggest tragedy 441 00:24:28,170 --> 00:24:30,970 that happens year after yearin the United States, 442 00:24:30,970 --> 00:24:33,800 this is the only solution. 443 00:24:33,800 --> 00:24:36,230 NARRATOR:It's an ambitious goal, 444 00:24:36,230 --> 00:24:38,430 but only possible becauseof the recent breakthroughs 445 00:24:38,430 --> 00:24:40,170 in deep learning. 446 00:24:40,170 --> 00:24:42,300 Artificial intelligence isone of those key pieces 447 00:24:42,300 --> 00:24:46,530 that has made it possible nowto do driverless vehicles 448 00:24:46,530 --> 00:24:49,070 where it wasn't possibleten years ago, 449 00:24:49,070 --> 00:24:53,870 particularly in the abilityto see and understand scenes. 450 00:24:53,870 --> 00:24:57,130 A lot of people don't know this,but it's remarkably hard 451 00:24:57,130 --> 00:24:58,870 for computers,until very, very recently, 452 00:24:58,870 --> 00:25:02,800 to do even the most basicvisual tasks, 453 00:25:02,800 --> 00:25:04,530 like seeing a pictureof a person 454 00:25:04,530 --> 00:25:06,070 and knowing that it's a person. 455 00:25:06,070 --> 00:25:09,270 And we've made gigantic strideswith artificial intelligence 456 00:25:09,270 --> 00:25:11,530 in being able to see andunderstanding tasks, 457 00:25:11,530 --> 00:25:14,000 and that's obviously fundamentalto being able to understand 458 00:25:14,000 --> 00:25:15,770 the world around youwith the sensors that, 459 00:25:15,770 --> 00:25:19,870 that you have available. 460 00:25:19,870 --> 00:25:21,270 NARRATOR: That's now possible 461 00:25:21,270 --> 00:25:23,970 because of the algorithmswritten by Yoshua Bengio 462 00:25:23,970 --> 00:25:28,070 and a small group of scientists. 463 00:25:28,070 --> 00:25:30,630 There are many aspectsof the world 464 00:25:30,630 --> 00:25:34,200 which we can't explainwith words. 465 00:25:34,200 --> 00:25:36,400 And that part of our knowledgeis actually 466 00:25:36,400 --> 00:25:39,170 probably the majority of it. 467 00:25:39,170 --> 00:25:41,400 So, like, the stuff we cancommunicate verbally 468 00:25:41,400 --> 00:25:43,370 is the tip of the iceberg. 469 00:25:43,370 --> 00:25:48,600 And so to get at the bottom ofthe iceberg, the solution was, 470 00:25:48,600 --> 00:25:53,000 the computers have to acquirethat knowledge by themselves 471 00:25:53,000 --> 00:25:54,500 from data, from examples. 472 00:25:54,500 --> 00:25:58,400 Just like children learn,most not from their teachers, 473 00:25:58,400 --> 00:26:01,370 but from interactingwith the world, 474 00:26:01,370 --> 00:26:03,500 and playing around, and, andtrying things 475 00:26:03,500 --> 00:26:05,570 and seeing what worksand what doesn't work. 476 00:26:05,570 --> 00:26:07,870 NARRATOR: This is an earlydemonstration. 477 00:26:07,870 --> 00:26:12,470 In 2013, deep-mind scientistsset a machine-learning program 478 00:26:12,470 --> 00:26:16,070 on the Atari video gameBreakout. 479 00:26:16,070 --> 00:26:19,870 The computer was only toldthe goal-- to win the game. 480 00:26:19,870 --> 00:26:24,230 After 100 games, it learned touse the bat at the bottom 481 00:26:24,230 --> 00:26:27,770 to hit the ball and breakthe bricks at the top. 482 00:26:27,770 --> 00:26:33,030 After 300, it could do thatbetter than a human player. 483 00:26:33,030 --> 00:26:37,970 After 500 games, it came up witha creative way to win the game-- 484 00:26:37,970 --> 00:26:40,730 by digging a tunnel on the side 485 00:26:40,730 --> 00:26:42,270 and sending the ballaround the top 486 00:26:42,270 --> 00:26:44,730 to break many brickswith one hit. 487 00:26:44,730 --> 00:26:48,170 That was deep learning. 488 00:26:48,170 --> 00:26:50,630 That's the A.I. program basedon learning, 489 00:26:50,630 --> 00:26:52,430 really, that has beenso successful 490 00:26:52,430 --> 00:26:54,870 in the last few years and has... 491 00:26:54,870 --> 00:26:57,430 It wasn't clear ten years agothat it would work, 492 00:26:57,430 --> 00:27:00,600 but it has completely changedthe map 493 00:27:00,600 --> 00:27:06,570 and is now used in almostevery sector of society. 494 00:27:06,570 --> 00:27:08,970 Even the best and brightestamong us, 495 00:27:08,970 --> 00:27:11,000 we just don't have enoughcompute power 496 00:27:11,000 --> 00:27:13,530 inside of our heads. 497 00:27:13,530 --> 00:27:16,000 NARRATOR: Amy Webb is aprofessor at N.Y.U. 498 00:27:16,000 --> 00:27:19,970 and founder of the Future TodayInstitute. 499 00:27:19,970 --> 00:27:26,270 As A.I. progresses, the greatpromise is that they... 500 00:27:26,270 --> 00:27:30,700 they, these, these machines,alongside of us, 501 00:27:30,700 --> 00:27:34,330 are able to think and imagineand see things 502 00:27:34,330 --> 00:27:36,730 in ways that we never havebefore, 503 00:27:36,730 --> 00:27:40,370 which means that maybe we havesome kind of new, 504 00:27:40,370 --> 00:27:45,330 weird, seemingly implausiblesolution to climate change. 505 00:27:45,330 --> 00:27:49,530 Maybe we have some radicallydifferent approach 506 00:27:49,530 --> 00:27:52,930 to dealing withincurable cancers. 507 00:27:52,930 --> 00:27:58,330 The real practical and wonderfulpromise is that machines help us 508 00:27:58,330 --> 00:28:02,170 be more creative, and,using that creativity, 509 00:28:02,170 --> 00:28:06,430 we get to terrific solutions. 510 00:28:06,430 --> 00:28:09,500 NARRATOR: Solutions thatcould come unexpectedly 511 00:28:09,500 --> 00:28:11,770 to urgent problems. 512 00:28:11,770 --> 00:28:13,700 It's going to changethe face of breast cancer. 513 00:28:13,700 --> 00:28:16,870 Right now, 40,000 womenin the U.S. alone 514 00:28:16,870 --> 00:28:19,670 die from breast cancerevery single year. 515 00:28:19,670 --> 00:28:21,870 NARRATOR: Dr. Connie Lehmanis head 516 00:28:21,870 --> 00:28:23,230 of the breast imaging center 517 00:28:23,230 --> 00:28:26,470 at Massachusetts GeneralHospital in Boston. 518 00:28:26,470 --> 00:28:28,930 We've become so complacentabout it, 519 00:28:28,930 --> 00:28:31,070 we almost don't think it canreally be changed. 520 00:28:31,070 --> 00:28:33,470 We, we somehow think we shouldput all of our energy 521 00:28:33,470 --> 00:28:36,670 into chemotherapiesto save women 522 00:28:36,670 --> 00:28:38,370 with metastatic breast cancer, 523 00:28:38,370 --> 00:28:41,430 and yet, you know, when we findit early, we cure it, 524 00:28:41,430 --> 00:28:44,730 and we cure it without havingthe ravages to the body 525 00:28:44,730 --> 00:28:46,830 when we diagnose it late. 526 00:28:46,830 --> 00:28:51,700 This shows the progression of asmall, small spot from one year 527 00:28:51,700 --> 00:28:54,530 to the next,and then to the diagnosis 528 00:28:54,530 --> 00:28:57,730 of the small cancer here. 529 00:28:57,730 --> 00:28:59,830 NARRATOR: This is whathappened when a woman 530 00:28:59,830 --> 00:29:02,100 who had been diagnosedwith breast cancer 531 00:29:02,100 --> 00:29:04,170 started to ask questions 532 00:29:04,170 --> 00:29:07,370 about why it couldn't have beendiagnosed earlier. 533 00:29:07,370 --> 00:29:10,170 It really brings a lot ofanxiety, 534 00:29:10,170 --> 00:29:12,270 and you're asking the questions,you know, 535 00:29:12,270 --> 00:29:13,530 "Am I going to survive? 536 00:29:13,530 --> 00:29:15,200 What's going to happento my son?" 537 00:29:15,200 --> 00:29:19,230 And I start askingother questions. 538 00:29:19,230 --> 00:29:21,700 NARRATOR: She was used toasking questions. 539 00:29:21,700 --> 00:29:24,970 At M.I.T.'sartificial-intelligence lab, 540 00:29:24,970 --> 00:29:27,970 Professor Regina Barzilay usesdeep learning 541 00:29:27,970 --> 00:29:31,100 to teach the computer tounderstand language, 542 00:29:31,100 --> 00:29:34,270 as well as read text and data. 543 00:29:34,270 --> 00:29:37,600 I was really surprisedthat the very basic question 544 00:29:37,600 --> 00:29:39,970 that I ask my physicians, 545 00:29:39,970 --> 00:29:43,330 which were really excellentphysicians here at MGH, 546 00:29:43,330 --> 00:29:47,030 they couldn't give me answersthat I was looking for. 547 00:29:47,030 --> 00:29:50,770 NARRATOR: She was convincedthat if you analyze enough data, 548 00:29:50,770 --> 00:29:53,530 from mammogramsto diagnostic notes, 549 00:29:53,530 --> 00:29:56,800 the computer could predictearly-stage conditions. 550 00:29:56,800 --> 00:30:02,830 If we fast-forward from 2012to '13 to 2014, 551 00:30:02,830 --> 00:30:05,900 we then see when Reginawas diagnosed, 552 00:30:05,900 --> 00:30:10,170 because of this spot on hermammogram. 553 00:30:10,170 --> 00:30:14,830 Is it possible, with moreelegant computer applications, 554 00:30:14,830 --> 00:30:19,100 that we might have identifiedthis spot the year before, 555 00:30:19,100 --> 00:30:21,200 or even back here? 556 00:30:21,200 --> 00:30:22,830 So, those are standardprediction problems 557 00:30:22,830 --> 00:30:26,870 in machine learning-- there isnothing special about them. 558 00:30:26,870 --> 00:30:29,900 And to my big surprise,none of the technologies 559 00:30:29,900 --> 00:30:33,130 that we are developingat M.I.T., 560 00:30:33,130 --> 00:30:38,730 even in the most simple form,doesn't penetrate the hospital. 561 00:30:38,730 --> 00:30:41,670 NARRATOR: Regina and Conniebegan the slow process 562 00:30:41,670 --> 00:30:45,070 of getting access to thousandsof mammograms and records 563 00:30:45,070 --> 00:30:46,770 from MGH's breast-imagingprogram. 564 00:30:49,470 --> 00:30:53,130 So, our first foray was justto take all of the patients 565 00:30:53,130 --> 00:30:56,130 we had at MGH duringa period of time, 566 00:30:56,130 --> 00:30:58,530 who had had breast surgeryfor a certain type 567 00:30:58,530 --> 00:31:00,430 of high-risk lesion. 568 00:31:00,430 --> 00:31:03,700 And we found that most of themdidn't really need the surgery. 569 00:31:03,700 --> 00:31:05,070 They didn't have cancer. 570 00:31:05,070 --> 00:31:07,670 But about ten percentdid have cancer. 571 00:31:07,670 --> 00:31:10,570 With Regina's techniquesin deep learning 572 00:31:10,570 --> 00:31:13,370 and machine learning, we wereable to predict the women 573 00:31:13,370 --> 00:31:15,600 that truly needed the surgeryand separate out 574 00:31:15,600 --> 00:31:19,470 those that really could avoidthe unnecessary surgery. 575 00:31:19,470 --> 00:31:23,030 What machine can do, it cantake hundreds of thousands 576 00:31:23,030 --> 00:31:25,730 of images where the outcomeis known 577 00:31:25,730 --> 00:31:30,700 and learn, based on how, youknow, pixels are distributed, 578 00:31:30,700 --> 00:31:35,170 what are the very uniquepatterns that correlate highly 579 00:31:35,170 --> 00:31:38,370 with future occurrenceof the disease. 580 00:31:38,370 --> 00:31:40,900 So, instead of using humancapacity 581 00:31:40,900 --> 00:31:44,770 to kind of recognize pattern,formalize pattern-- 582 00:31:44,770 --> 00:31:48,700 which is inherently limitedby our cognitive capacity 583 00:31:48,700 --> 00:31:50,800 and how much we can seeand remember-- 584 00:31:50,800 --> 00:31:53,700 we're providing machine with alot of data 585 00:31:53,700 --> 00:31:57,630 and make it learnthis prediction. 586 00:31:57,630 --> 00:32:02,370 So, we are using technologynot only to be better 587 00:32:02,370 --> 00:32:04,770 at assessing the breast density, 588 00:32:04,770 --> 00:32:07,200 but to get more to the point ofwhat we're trying to predict. 589 00:32:07,200 --> 00:32:10,930 "Does this woman havea cancer now, 590 00:32:10,930 --> 00:32:13,170 and will she develop a cancerin five years? " 591 00:32:13,170 --> 00:32:16,770 And that's, again, wherethe artificial intelligence, 592 00:32:16,770 --> 00:32:18,700 machine and deep learning canreally help us 593 00:32:18,700 --> 00:32:20,770 and our patients. 594 00:32:20,770 --> 00:32:22,830 NARRATOR: In the age of A.I., 595 00:32:22,830 --> 00:32:26,330 the algorithms are transportingus into a universe 596 00:32:26,330 --> 00:32:29,970 of vast potential andtransforming almost every aspect 597 00:32:29,970 --> 00:32:34,200 of human endeavor andexperience. 598 00:32:34,200 --> 00:32:38,000 Andrew McAfee is a researchscientist at M.I.T. 599 00:32:38,000 --> 00:32:42,000 who co-authored"The Second Machine Age." 600 00:32:42,000 --> 00:32:45,070 The great compliment that asongwriter gives another one is, 601 00:32:45,070 --> 00:32:46,600 "Gosh, I wish I had writtenthat one." 602 00:32:46,600 --> 00:32:49,100 The great compliment a geekgives another one is, 603 00:32:49,100 --> 00:32:50,900 "Wow, I wish I had drawnthat graph." 604 00:32:50,900 --> 00:32:53,630 So, I wish I had drawnthis graph. 605 00:32:53,630 --> 00:32:55,500 NARRATOR:The graph uses a formula 606 00:32:55,500 --> 00:32:59,400 to show human development andgrowth since 2000 BCE. 607 00:32:59,400 --> 00:33:01,570 The state of humancivilization 608 00:33:01,570 --> 00:33:04,970 is not very advanced, and it'snot getting better 609 00:33:04,970 --> 00:33:07,130 very quickly at all,and this is true for thousands 610 00:33:07,130 --> 00:33:08,970 and thousands of years. 611 00:33:08,970 --> 00:33:12,470 When we, when we formed empiresand empires got overturned, 612 00:33:12,470 --> 00:33:16,530 when we tried democracy,when we invented zero 613 00:33:16,530 --> 00:33:19,630 and mathematics and fundamentaldiscoveries about the universe, 614 00:33:19,630 --> 00:33:21,400 big deal. 615 00:33:21,400 --> 00:33:23,300 It just, the numbers don'tchange very much. 616 00:33:23,300 --> 00:33:26,900 What's weird is that the numberschange essentially in the blink 617 00:33:26,900 --> 00:33:28,370 of an eye at one point in time. 618 00:33:28,370 --> 00:33:32,030 And it goes from reallyhorizontal, unchanging, 619 00:33:32,030 --> 00:33:36,600 uninteresting, to, holy Toledo,crazy vertical. 620 00:33:36,600 --> 00:33:39,200 And then the question is,what on Earth happened 621 00:33:39,200 --> 00:33:40,570 to cause that change? 622 00:33:40,570 --> 00:33:42,770 And the answeris the Industrial Revolution. 623 00:33:42,770 --> 00:33:44,800 There were other things thathappened, 624 00:33:44,800 --> 00:33:46,830 but really what fundamentallyhappened is 625 00:33:46,830 --> 00:33:49,530 we overcame the limitationsof our muscle power. 626 00:33:49,530 --> 00:33:52,400 Something equally interesting ishappening right now. 627 00:33:52,400 --> 00:33:55,330 We are overcoming thelimitations of our minds. 628 00:33:55,330 --> 00:33:56,930 We're not getting rid of them, 629 00:33:56,930 --> 00:33:58,970 we're not making themunnecessary, 630 00:33:58,970 --> 00:34:02,500 but, holy cow, can we leveragethem and amplify them now. 631 00:34:02,500 --> 00:34:04,170 You have to be a huge pessimist 632 00:34:04,170 --> 00:34:06,730 not to find that profoundlygood news. 633 00:34:06,730 --> 00:34:09,370 I really do think the worldhas entered a new era. 634 00:34:09,370 --> 00:34:12,830 Artificial intelligence holds somuch promise, 635 00:34:12,830 --> 00:34:15,730 but it's going to reshape everyaspect of the economy, 636 00:34:15,730 --> 00:34:17,370 so many aspects of our lives. 637 00:34:17,370 --> 00:34:20,770 Because A.I. is a little bitlike electricity. 638 00:34:20,770 --> 00:34:22,670 Everybody's going to use it. 639 00:34:22,669 --> 00:34:26,399 Every company is going to beincorporating A.I., 640 00:34:26,399 --> 00:34:28,299 integrating it intowhat they do, 641 00:34:28,300 --> 00:34:29,630 governments are going to beusing it, 642 00:34:29,629 --> 00:34:33,599 nonprofit organizations aregoing to be using it. 643 00:34:33,600 --> 00:34:37,200 It's going to create all kindsof benefits 644 00:34:37,200 --> 00:34:41,070 in ways large and small,and challenges for us, as well. 645 00:34:41,070 --> 00:34:44,730 NARRATOR: The challenges,the benefits-- 646 00:34:44,730 --> 00:34:47,000 the autonomous truckrepresents both 647 00:34:47,000 --> 00:34:50,070 as it maneuversinto the marketplace. 648 00:34:50,070 --> 00:34:53,070 The engineers are confidentthat, in spite of questions 649 00:34:53,070 --> 00:34:55,370 about when this will happen, 650 00:34:55,370 --> 00:34:57,330 they can get it working safelysooner 651 00:34:57,330 --> 00:34:58,770 than most people realize. 652 00:34:58,770 --> 00:35:02,130 I think that you will see thefirst vehicles operating 653 00:35:02,130 --> 00:35:05,570 with no one inside them movingfreight in the next few years, 654 00:35:05,570 --> 00:35:07,700 and then you're going to seethat expanding to more freight, 655 00:35:07,700 --> 00:35:11,030 more geographies,more weather over time as, 656 00:35:11,030 --> 00:35:12,530 as that capability builds up. 657 00:35:12,530 --> 00:35:16,600 We're talking, like,less than half a decade. 658 00:35:16,600 --> 00:35:19,370 NARRATOR: He already has aFortune 500 company 659 00:35:19,370 --> 00:35:23,830 as a client, shipping appliancesacross the Southwest. 660 00:35:23,830 --> 00:35:27,330 He says the sales pitchis straightforward. 661 00:35:27,330 --> 00:35:30,070 They spend hundreds ofmillions of dollars a year 662 00:35:30,070 --> 00:35:31,670 shipping parts aroundthe country. 663 00:35:31,670 --> 00:35:34,100 We can bring that cost in half. 664 00:35:34,100 --> 00:35:36,930 And they're really excited to beable to start working with us, 665 00:35:36,930 --> 00:35:39,800 both because of the potential, 666 00:35:39,800 --> 00:35:42,100 the potential savings fromdeploying self-driving, 667 00:35:42,100 --> 00:35:44,470 and also because of all theoperational efficiencies 668 00:35:44,470 --> 00:35:47,830 that they see, the biggest onebeing able to operate 669 00:35:47,830 --> 00:35:49,800 24 hours a day. 670 00:35:49,800 --> 00:35:51,970 So, right now, human drivers arelimited to 11 hours 671 00:35:51,970 --> 00:35:55,470 by federal law,and a driverless truck 672 00:35:55,470 --> 00:35:57,000 obviously wouldn't havethat limitation. 673 00:35:57,000 --> 00:36:02,530 ♪ ♪ 674 00:36:02,530 --> 00:36:05,330 NARRATOR: The idea of adriverless truck comes up often 675 00:36:05,330 --> 00:36:11,430 in discussions about artificialintelligence. 676 00:36:11,430 --> 00:36:14,800 Steve Viscelli is a sociologistwho drove a truck 677 00:36:14,800 --> 00:36:20,330 while researching his book "TheBig Rig" about the industry. 678 00:36:20,330 --> 00:36:23,000 This is one of the mostremarkable stories 679 00:36:23,000 --> 00:36:25,830 in, in U.S. labor history,I think, 680 00:36:25,830 --> 00:36:30,400 is, you know, the decline of,of unionized trucking. 681 00:36:30,400 --> 00:36:33,600 The industry was deregulatedin 1980, 682 00:36:33,600 --> 00:36:37,400 and at that time, you know,truck drivers were earning 683 00:36:37,400 --> 00:36:41,400 the equivalent of over$100,000 in today's dollars. 684 00:36:41,400 --> 00:36:45,500 And today the typical truckdriver will earn 685 00:36:45,500 --> 00:36:50,370 a little over $40,000 a year. 686 00:36:50,370 --> 00:36:52,630 And I think it'san important part 687 00:36:52,630 --> 00:36:54,230 of the automation story, right? 688 00:36:54,230 --> 00:36:56,900 Why are they so afraid ofautomation? 689 00:36:56,900 --> 00:37:00,670 Because we've had four decadesof rising inequality in wages. 690 00:37:00,670 --> 00:37:03,330 And if anybody is going to takeit on the chin 691 00:37:03,330 --> 00:37:05,330 from automationin the trucking industry, 692 00:37:05,330 --> 00:37:07,630 the, the first in line is goingto be the driver, 693 00:37:07,630 --> 00:37:12,300 without a doubt. 694 00:37:12,300 --> 00:37:14,730 NARRATOR: For his research,Viscelli tracked down truckers 695 00:37:14,730 --> 00:37:17,600 and their families,like Shawn and Hope Cumbee 696 00:37:17,600 --> 00:37:19,530 of Beaverton, Michigan.Hi. 697 00:37:19,530 --> 00:37:20,870 Hey, Hope,I'm Steve Viscelli. 698 00:37:20,870 --> 00:37:21,870 Hi, Steve, nice to meet you.Come on in. 699 00:37:21,870 --> 00:37:24,800 Great to meet you, too,thanks. 700 00:37:24,800 --> 00:37:26,430 NARRATOR: And their sonCharlie. 701 00:37:26,430 --> 00:37:31,730 This is Daddy, me,Daddy, and Mommy. 702 00:37:31,730 --> 00:37:34,230 NARRATOR: But Daddy's nothere. 703 00:37:34,230 --> 00:37:38,900 Shawn Cumbee's truck has brokendown in Tennessee. 704 00:37:38,900 --> 00:37:43,470 Hope, who drove a truck herself,knows the business well. 705 00:37:43,470 --> 00:37:46,870 We made $150,000, right,in a year. 706 00:37:46,870 --> 00:37:48,070 That sounds great, right? 707 00:37:48,070 --> 00:37:50,400 That's, like, good money. 708 00:37:50,400 --> 00:37:53,870 We paid $100,000 in fuel, okay? 709 00:37:53,870 --> 00:37:57,030 So, right there,now I made $50,000. 710 00:37:57,030 --> 00:37:59,030 But I didn't really, because,you know, 711 00:37:59,030 --> 00:38:00,600 you get an oil change everymonth, 712 00:38:00,600 --> 00:38:02,200 so that's $300 a month. 713 00:38:02,200 --> 00:38:04,170 You still have to doall the maintenance. 714 00:38:04,170 --> 00:38:06,500 We had a motor blow out, right? 715 00:38:06,500 --> 00:38:09,170 $13,000. Right? 716 00:38:09,170 --> 00:38:11,800 I know, I mean, I choke up alittle just thinking about it, 717 00:38:11,800 --> 00:38:13,770 because it was... 718 00:38:13,770 --> 00:38:17,470 And it was 13,000, and we wereoff work for two weeks. 719 00:38:17,470 --> 00:38:19,670 So, by the end of the year,with that $150,000, 720 00:38:19,670 --> 00:38:22,670 by the end of the year,we'd made about 20... 721 00:38:22,670 --> 00:38:26,030 About $22,000. 722 00:38:26,030 --> 00:38:28,400 NARRATOR: In a truck stopin Tennessee, 723 00:38:28,400 --> 00:38:31,500 Shawn has been sidelinedwaiting for a new part. 724 00:38:31,500 --> 00:38:35,300 The garage owner is letting himstay in the truck to save money. 725 00:38:37,870 --> 00:38:39,770 Hi, baby. 726 00:38:39,770 --> 00:38:41,330 (on phone): Hey, how's itgoing? 727 00:38:41,330 --> 00:38:42,730 It's going.Chunky-butt! 728 00:38:42,730 --> 00:38:44,600 Hi, Daddy!Hi, Chunky-butt. 729 00:38:44,600 --> 00:38:47,300 What're you doing?(talking inaudibly) 730 00:38:47,300 --> 00:38:49,600 Believe it or not,I do it because I love it. 731 00:38:49,600 --> 00:38:51,330 I mean, you know,it's in the blood. 732 00:38:51,330 --> 00:38:52,900 Third-generation driver. 733 00:38:52,900 --> 00:38:55,230 And my granddaddy told me a longtime ago, 734 00:38:55,230 --> 00:38:58,630 when I was probably11, 12 years old, probably, 735 00:38:58,630 --> 00:39:01,500 he said, "The world meets nobodyhalfway. 736 00:39:01,500 --> 00:39:02,930 Nobody." 737 00:39:02,930 --> 00:39:07,030 He said, "If you want it,you have to earn it." 738 00:39:07,030 --> 00:39:09,870 And that's what I do every day. 739 00:39:09,870 --> 00:39:11,330 I live by that creed. 740 00:39:11,330 --> 00:39:16,100 And I've lived by thatsince it was told to me. 741 00:39:16,100 --> 00:39:18,300 So, if you're down for a weekin a truck, 742 00:39:18,300 --> 00:39:19,870 you still have to pay yourbills. 743 00:39:19,870 --> 00:39:22,100 I have enough money in mychecking account at all times 744 00:39:22,100 --> 00:39:23,470 to pay a month's worth of bills. 745 00:39:23,470 --> 00:39:25,070 That does not include my food. 746 00:39:25,070 --> 00:39:27,630 That doesn't include field tripsfor my son's school. 747 00:39:27,630 --> 00:39:31,700 My son and I just went to ouryearly doctor appointment. 748 00:39:31,700 --> 00:39:36,270 I took, I took money out of myson's piggy bank to pay for it, 749 00:39:36,270 --> 00:39:40,600 because it's not...it's not scheduled in. 750 00:39:40,600 --> 00:39:43,430 It's, it's not something thatyou can, you know, afford. 751 00:39:43,430 --> 00:39:45,500 I mean, like, when... 752 00:39:45,500 --> 00:39:46,900 (sighs): Sorry. 753 00:39:46,900 --> 00:39:48,970 It's okay. 754 00:39:48,970 --> 00:39:52,600 ♪ ♪ 755 00:39:57,230 --> 00:39:59,170 Have you guys ever talked aboutself-driving trucks? 756 00:39:59,170 --> 00:40:00,500 Is he... 757 00:40:00,500 --> 00:40:03,130 (laughing): So, kind of. 758 00:40:03,130 --> 00:40:05,830 Um, I asked him once, you know. 759 00:40:05,830 --> 00:40:07,230 And he laughed so hard. 760 00:40:07,230 --> 00:40:10,330 He said, "No way will theyever have a truck 761 00:40:10,330 --> 00:40:12,970 that can drive itself." 762 00:40:12,970 --> 00:40:15,230 It's kind of interesting whenyou think about it, you know, 763 00:40:15,230 --> 00:40:17,730 they're putting all this newtechnology into things, 764 00:40:17,730 --> 00:40:19,570 but, you know,it's still man-made. 765 00:40:19,570 --> 00:40:22,970 And man, you know,does make mistakes. 766 00:40:22,970 --> 00:40:26,170 I really don't see it beinga problem with the industry, 767 00:40:26,170 --> 00:40:28,770 'cause, one, you still got tohave a driver in it, 768 00:40:28,770 --> 00:40:30,330 because I don't see itdoing city. 769 00:40:30,330 --> 00:40:32,600 I don't see it doing,you know, main things. 770 00:40:32,600 --> 00:40:34,700 I don't see it backing intoa dock. 771 00:40:34,700 --> 00:40:37,870 I don't see the automation part,you know, doing... 772 00:40:37,870 --> 00:40:39,900 maybe the box-trailer side,you know, I can see that, 773 00:40:39,900 --> 00:40:41,400 but not stuff like I do. 774 00:40:41,400 --> 00:40:44,830 So, I ain't really worried aboutthe automation of trucks. 775 00:40:44,830 --> 00:40:46,230 How near of a future is it? 776 00:40:46,230 --> 00:40:49,300 Yeah, self-driving, um... 777 00:40:49,300 --> 00:40:52,600 So, some, you know, somecompanies are already operating. 778 00:40:52,600 --> 00:40:56,170 Embark, for instance, is onethat has been doing 779 00:40:56,170 --> 00:40:59,030 driverless truckson the interstate. 780 00:40:59,030 --> 00:41:01,930 And what's called exit-to-exitself-driving. 781 00:41:01,930 --> 00:41:04,830 And they're currently runningreal freight. 782 00:41:04,830 --> 00:41:07,530 Really?Yeah, on I-10. 783 00:41:07,530 --> 00:41:10,530 ♪ ♪ 784 00:41:10,530 --> 00:41:15,170 (on P.A.): Shower guest 100,your shower is now ready. 785 00:41:15,170 --> 00:41:18,430 NARRATOR: Over time, it hasbecome harder and harder 786 00:41:18,430 --> 00:41:21,230 for veteran independent driverslike the Cumbees 787 00:41:21,230 --> 00:41:23,070 to make a living. 788 00:41:23,070 --> 00:41:25,070 They've been replaced byyounger, 789 00:41:25,070 --> 00:41:28,200 less experienced drivers. 790 00:41:28,200 --> 00:41:32,630 So, the, the truckingindustry's $740 billion a year, 791 00:41:32,630 --> 00:41:34,770 and, again, in, in manyof these operations, 792 00:41:34,770 --> 00:41:37,470 labor's a third of that cost. 793 00:41:37,470 --> 00:41:40,500 By my estimate, I, you know,I think we're in the range 794 00:41:40,500 --> 00:41:42,970 of 300,000 or so jobsin the foreseeable future 795 00:41:42,970 --> 00:41:47,930 that could be automated to somesignificant extent. 796 00:41:47,930 --> 00:41:50,630 ♪ ♪ 797 00:41:50,630 --> 00:41:53,530 (groans) 798 00:41:53,530 --> 00:41:57,070 ♪ ♪ 799 00:42:03,000 --> 00:42:06,130 NARRATOR: The A.I. futurewas built with great optimism 800 00:42:06,130 --> 00:42:09,100 out here in the West. 801 00:42:09,100 --> 00:42:12,630 In 2018, many of the peoplewho invented it 802 00:42:12,630 --> 00:42:16,170 gathered in San Francisco tocelebrate the 25th anniversary 803 00:42:16,170 --> 00:42:18,700 of the industry magazine. 804 00:42:18,700 --> 00:42:22,300 Howdy, welcome to WIRED25. 805 00:42:22,300 --> 00:42:24,200 NARRATOR: It is acelebration, for sure, 806 00:42:24,200 --> 00:42:27,070 but there's also a growing senseof caution 807 00:42:27,070 --> 00:42:28,670 and even skepticism. 808 00:42:31,130 --> 00:42:33,330 We're having a really goodweekend here. 809 00:42:33,330 --> 00:42:37,030 NARRATOR: Nick Thompson iseditor-in-chief of "Wired." 810 00:42:37,030 --> 00:42:40,030 When it started,it was very much a magazine 811 00:42:40,030 --> 00:42:44,100 about what's coming and why youshould be excited about it. 812 00:42:44,100 --> 00:42:47,730 Optimism was the definingfeature of "Wired" 813 00:42:47,730 --> 00:42:49,400 for many, many years. 814 00:42:49,400 --> 00:42:53,130 Or, as our slogan used to be,"Change Is Good." 815 00:42:53,130 --> 00:42:55,070 And over time,it shifted a little bit. 816 00:42:55,070 --> 00:42:59,170 And now it's more,"We love technology, 817 00:42:59,170 --> 00:43:00,630 but let's look at someof the big issues, 818 00:43:00,630 --> 00:43:03,400 and let's look at some of themcritically, 819 00:43:03,400 --> 00:43:05,730 and let's look at the wayalgorithms are changing 820 00:43:05,730 --> 00:43:07,930 the way we behave,for good and for ill." 821 00:43:07,930 --> 00:43:12,030 So, the whole nature of "Wired"has gone from a champion 822 00:43:12,030 --> 00:43:14,830 of technological change to moreof a observer 823 00:43:14,830 --> 00:43:16,700 of technological change. 824 00:43:16,700 --> 00:43:18,570 So, um, before we start... 825 00:43:18,570 --> 00:43:20,530 NARRATOR: Thereare 25 speakers, 826 00:43:20,530 --> 00:43:23,700 all named as iconsof the last 25 years 827 00:43:23,700 --> 00:43:25,500 of technological progress. 828 00:43:25,500 --> 00:43:27,770 So, why is Apple sosecretive? 829 00:43:27,770 --> 00:43:29,470 (chuckling) 830 00:43:29,470 --> 00:43:31,630 NARRATOR: Jony Ive, whodesigned Apple's iPhone. 831 00:43:31,630 --> 00:43:34,300 It would be bizarrenot to be. 832 00:43:34,300 --> 00:43:36,670 There's this question of,like, 833 00:43:36,670 --> 00:43:39,000 what are we doing here in thislife, in this reality? 834 00:43:39,000 --> 00:43:43,170 NARRATOR: Jaron Lanier, whopioneered virtual reality. 835 00:43:43,170 --> 00:43:46,500 And Jeff Bezos,the founder of Amazon. 836 00:43:46,500 --> 00:43:47,870 Amazon was a garage startup. 837 00:43:47,870 --> 00:43:49,370 Now it's a very large company. 838 00:43:49,370 --> 00:43:50,570 Two kids in a dorm... 839 00:43:50,570 --> 00:43:52,070 NARRATOR: His message is, 840 00:43:52,070 --> 00:43:54,730 "All will be wellin the new world." 841 00:43:54,730 --> 00:43:58,470 I guess, first of all, Iremain incredibly optimistic 842 00:43:58,470 --> 00:43:59,630 about technology, 843 00:43:59,630 --> 00:44:01,830 and technologies alwaysare two-sided. 844 00:44:01,830 --> 00:44:03,230 But that's not new. 845 00:44:03,230 --> 00:44:05,400 That's always been the case. 846 00:44:05,400 --> 00:44:07,830 And, and we will figure it out. 847 00:44:07,830 --> 00:44:10,570 The last thing we would everwant to do is stop the progress 848 00:44:10,570 --> 00:44:16,630 of new technologies,even when they are dual-use. 849 00:44:16,630 --> 00:44:19,800 NARRATOR: But, says Thompson,beneath the surface, 850 00:44:19,800 --> 00:44:22,530 there's a worry most of themdon't like to talk about. 851 00:44:22,530 --> 00:44:26,630 There are some people inSilicon Valley who believe that, 852 00:44:26,630 --> 00:44:29,900 "You just have to trustthe technology. 853 00:44:29,900 --> 00:44:32,870 Throughout history, there's beena complicated relationship 854 00:44:32,870 --> 00:44:34,470 between humans and machines, 855 00:44:34,470 --> 00:44:36,770 we've always worried aboutmachines, 856 00:44:36,770 --> 00:44:38,130 and it's always been fine. 857 00:44:38,130 --> 00:44:41,000 And we don't know how A.I. willchange the labor force, 858 00:44:41,000 --> 00:44:42,300 but it will be okay." 859 00:44:42,300 --> 00:44:44,070 So, that argument exists. 860 00:44:44,070 --> 00:44:45,700 There's another argument, 861 00:44:45,700 --> 00:44:48,170 which is what I think most ofthem believe deep down, 862 00:44:48,170 --> 00:44:51,100 which is, "This is different. 863 00:44:51,100 --> 00:44:52,930 We're going to have labor-forcedisruption 864 00:44:52,930 --> 00:44:55,030 like we've never seen before. 865 00:44:55,030 --> 00:44:59,370 And if that happens,will they blame us?" 866 00:44:59,370 --> 00:45:02,600 NARRATOR: There is, however,one of the WIRED25 icons 867 00:45:02,600 --> 00:45:05,800 willing to take on the issue. 868 00:45:05,800 --> 00:45:09,470 Onstage, Kai-Fu Lee dispenseswith one common fear. 869 00:45:09,470 --> 00:45:11,670 Well, I think there are somany myths out there. 870 00:45:11,670 --> 00:45:14,530 I think one, one myth is that 871 00:45:14,530 --> 00:45:17,570 because A.I. is so good at asingle task, 872 00:45:17,570 --> 00:45:21,600 that one day we'll wake up, andwe'll all be enslaved 873 00:45:21,600 --> 00:45:24,100 or forced to plug our brainsto the A.I. 874 00:45:24,100 --> 00:45:28,800 But it is nowhere closeto displacing humans. 875 00:45:28,800 --> 00:45:32,130 NARRATOR: But in interviewsaround the event and beyond, 876 00:45:32,130 --> 00:45:37,430 he takes a decidedly contrarianposition on A.I. and job loss. 877 00:45:37,430 --> 00:45:41,270 The A.I. giants want to paintthe rosier picture 878 00:45:41,270 --> 00:45:43,500 because they're happilymaking money. 879 00:45:43,500 --> 00:45:47,330 So, I think they prefer not totalk about the negative side. 880 00:45:47,330 --> 00:45:53,070 I believe about 50% of jobswill be 881 00:45:53,070 --> 00:45:56,900 somewhat or extremelythreatened by A.I. 882 00:45:56,900 --> 00:46:00,500 in the next 15 years or so. 883 00:46:00,500 --> 00:46:02,570 NARRATOR: Kai-Fu Lee alsomakes a great deal 884 00:46:02,570 --> 00:46:04,900 of money from A.I. 885 00:46:04,900 --> 00:46:06,800 What separates him from most ofhis colleagues 886 00:46:06,800 --> 00:46:09,930 is that he's frankabout its downside. 887 00:46:09,930 --> 00:46:13,900 Yes, yes, we, we've madeabout 40 investments in A.I. 888 00:46:13,900 --> 00:46:16,930 I think, based on these 40investments, 889 00:46:16,930 --> 00:46:20,000 most of them are not impactinghuman jobs. 890 00:46:20,000 --> 00:46:21,970 They're creating value,making high margins, 891 00:46:21,970 --> 00:46:24,300 inventing a new model. 892 00:46:24,300 --> 00:46:27,730 But I could list seven or eight 893 00:46:27,730 --> 00:46:32,670 that would lead to a very cleardisplacement of human jobs. 894 00:46:32,670 --> 00:46:34,370 NARRATOR: He says that A.I.is coming, 895 00:46:34,370 --> 00:46:36,470 whether we like it or not. 896 00:46:36,470 --> 00:46:38,300 And he wants to warn society 897 00:46:38,300 --> 00:46:41,030 about what he sees asinevitable. 898 00:46:41,030 --> 00:46:43,600 You have a view which I thinkis different than many others, 899 00:46:43,600 --> 00:46:48,670 which is that A.I. is not goingto take blue-collar jobs 900 00:46:48,670 --> 00:46:51,230 so quickly, but is actuallygoing to take white-collar jobs. 901 00:46:51,230 --> 00:46:53,770 Yeah.Well, both will happen. 902 00:46:53,770 --> 00:46:57,000 A.I. will be, at the same time,a replacement for blue-collar, 903 00:46:57,000 --> 00:47:00,630 white-collar jobs, and bea great symbiotic tool 904 00:47:00,630 --> 00:47:03,630 for doctors, lawyers, and you,for example. 905 00:47:03,630 --> 00:47:05,700 But the white-collar jobs areeasier to take, 906 00:47:05,700 --> 00:47:10,030 because they're a purequantitative analytical process. 907 00:47:10,030 --> 00:47:15,370 Let's say reporters, traders,telemarketing, 908 00:47:15,370 --> 00:47:17,270 telesales, customer service... 909 00:47:17,270 --> 00:47:18,730 Analysts? 910 00:47:18,730 --> 00:47:23,170 Analysts, yes, these can allbe replaced just by a software. 911 00:47:23,170 --> 00:47:26,330 To do blue-collar, some of thework requires, you know, 912 00:47:26,330 --> 00:47:30,030 hand-eye coordination, thingsthat machines are not yet 913 00:47:30,030 --> 00:47:32,300 good enough to do. 914 00:47:32,300 --> 00:47:36,400 Today, there are many peoplewho are ringing the alarm, 915 00:47:36,400 --> 00:47:37,600 "Oh, my God, what are we goingto do? 916 00:47:37,600 --> 00:47:39,830 Half the jobs are going away." 917 00:47:39,830 --> 00:47:43,430 I believe that's true, buthere's the missing fact. 918 00:47:43,430 --> 00:47:46,400 I've done the research on this,and if you go back 20, 30, 919 00:47:46,400 --> 00:47:50,930 or 40 years ago, you will findthat 50% of the jobs 920 00:47:50,930 --> 00:47:54,400 that people performed back thenare gone today. 921 00:47:54,400 --> 00:47:56,900 You know, where are all thetelephone operators, 922 00:47:56,900 --> 00:48:00,600 bowling-pin setters,elevator operators? 923 00:48:00,600 --> 00:48:04,270 You used to have seas ofsecretaries in corporations 924 00:48:04,270 --> 00:48:06,070 that have now been eliminated--travel agents. 925 00:48:06,070 --> 00:48:08,770 You can just go through fieldafter field after field. 926 00:48:08,770 --> 00:48:12,100 That same pattern has recurredmany times throughout history, 927 00:48:12,100 --> 00:48:14,230 with each new waveof automation. 928 00:48:14,230 --> 00:48:20,270 But I would argue thathistory is only trustable 929 00:48:20,270 --> 00:48:24,670 if it is multiple repetitionsof similar events, 930 00:48:24,670 --> 00:48:28,670 not once-in-a-blue-moonoccurrence. 931 00:48:28,670 --> 00:48:33,070 So, over the history of manytech inventions, 932 00:48:33,070 --> 00:48:34,770 most are small things. 933 00:48:34,770 --> 00:48:41,330 Only maybe three are at themagnitude of A.I. revolution-- 934 00:48:41,330 --> 00:48:44,730 the steam, steam engine,electricity, 935 00:48:44,730 --> 00:48:46,570 and the computer revolution. 936 00:48:46,570 --> 00:48:48,970 I'd say everything elseis too small. 937 00:48:48,970 --> 00:48:52,670 And the reason I think it mightbe something brand-new 938 00:48:52,670 --> 00:48:58,930 is that A.I. is fundamentallyreplacing our cognitive process 939 00:48:58,930 --> 00:49:03,670 in doing a job in itssignificant entirety, 940 00:49:03,670 --> 00:49:06,400 and it can do it dramaticallybetter. 941 00:49:06,400 --> 00:49:08,570 NARRATOR: This argumentabout job loss 942 00:49:08,570 --> 00:49:11,470 in the age of A.I. was ignitedsix years ago 943 00:49:11,470 --> 00:49:15,830 amid the gargoyles and spiresof Oxford University. 944 00:49:15,830 --> 00:49:19,970 Two researchers had been poringthrough U.S. labor statistics, 945 00:49:19,970 --> 00:49:25,270 identifying jobs that could bevulnerable to A.I. automation. 946 00:49:25,270 --> 00:49:27,300 Well, vulnerable toautomation, 947 00:49:27,300 --> 00:49:30,730 in the context that we discussedfive years ago now, 948 00:49:30,730 --> 00:49:34,430 essentially meant that thosejobs are potentially automatable 949 00:49:34,430 --> 00:49:36,900 over an unspecified number ofyears. 950 00:49:36,900 --> 00:49:41,530 And the figure we came up withwas 47%. 951 00:49:41,530 --> 00:49:43,330 NARRATOR: 47%. 952 00:49:43,330 --> 00:49:46,470 That number quickly traveledthe world in headlines 953 00:49:46,470 --> 00:49:47,830 and news bulletins. 954 00:49:47,830 --> 00:49:51,030 But authors Carl Freyand Michael Osborne 955 00:49:51,030 --> 00:49:52,770 offered a caution. 956 00:49:52,770 --> 00:49:57,670 They can't predict how many jobswill be lost, or how quickly. 957 00:49:57,670 --> 00:50:02,430 But Frey believes that there arelessons in history. 958 00:50:02,430 --> 00:50:04,830 And what worries me the mostis that there is actually 959 00:50:04,830 --> 00:50:08,830 one episode that looks quitefamiliar to today, 960 00:50:08,830 --> 00:50:12,270 which is the BritishIndustrial Revolution, 961 00:50:12,270 --> 00:50:16,400 where wages didn't growfor nine decades, 962 00:50:16,400 --> 00:50:20,530 and a lot of people actuallysaw living standards decline 963 00:50:20,530 --> 00:50:23,870 as technology progressed. 964 00:50:23,870 --> 00:50:25,630 ♪ ♪ 965 00:50:25,630 --> 00:50:28,370 NARRATOR: Saginaw, Michigan,knows about decline 966 00:50:28,370 --> 00:50:31,170 in living standards. 967 00:50:31,170 --> 00:50:34,900 Harry Cripps, an auto workerand a local union president, 968 00:50:34,900 --> 00:50:40,730 has witnessed what 40 years ofautomation can do to a town. 969 00:50:40,730 --> 00:50:43,470 You know, we're one of thecities in the country that, 970 00:50:43,470 --> 00:50:47,170 I think we were left behind inthis recovery. 971 00:50:47,170 --> 00:50:51,670 And I just... I don't know howwe get on the bandwagon now. 972 00:50:54,770 --> 00:50:57,030 NARRATOR: Once, this was theU.A.W. hall 973 00:50:57,030 --> 00:50:59,230 for one local union. 974 00:50:59,230 --> 00:51:03,670 Now, with falling membership,it's shared by five locals. 975 00:51:03,670 --> 00:51:05,730 Rudy didn't get his shift. 976 00:51:05,730 --> 00:51:07,330 NARRATOR: This day,it's the center 977 00:51:07,330 --> 00:51:09,570 for a Christmas food drive. 978 00:51:09,570 --> 00:51:12,030 Even in a growth economy, 979 00:51:12,030 --> 00:51:14,830 unemployment here is nearsix percent. 980 00:51:14,830 --> 00:51:18,930 Poverty in Saginaw is over 30%. 981 00:51:21,830 --> 00:51:25,130 Our factory has about1.9 million square feet. 982 00:51:25,130 --> 00:51:29,100 Back in the '70s, that 1.9million square feet 983 00:51:29,100 --> 00:51:32,330 had about 7,500 U.A.W.automotive workers 984 00:51:32,330 --> 00:51:34,300 making middle-class wage withdecent benefits 985 00:51:34,300 --> 00:51:36,770 and able to send their kids tocollege and do all the things 986 00:51:36,770 --> 00:51:39,000 that the middle-class familyshould be able to do. 987 00:51:39,000 --> 00:51:42,270 Our factory today, withautomation, 988 00:51:42,270 --> 00:51:46,300 would probably be about700 United Auto Workers. 989 00:51:46,300 --> 00:51:50,130 That's a dramatic change. 990 00:51:50,130 --> 00:51:52,230 Lot of union brothers usedto work there, buddy. 991 00:51:52,230 --> 00:51:55,130 The TRW plant, that wasunfortunate. 992 00:51:55,130 --> 00:51:57,830 Delphi... looks like they'restarting to tear it down now. 993 00:51:57,830 --> 00:51:59,300 Wow. 994 00:51:59,300 --> 00:52:02,770 Automations is, is definitelytaking away a lot of jobs. 995 00:52:02,770 --> 00:52:05,530 Robots, I don't know how theybuy cars, 996 00:52:05,530 --> 00:52:07,300 I don't know howthey buy sandwiches, 997 00:52:07,300 --> 00:52:09,100 I don't know how they go to thegrocery store. 998 00:52:09,100 --> 00:52:11,430 They definitely don't pay taxes,which serves the infrastructure. 999 00:52:11,430 --> 00:52:15,300 So, you don't have the sheriffsand the police and the firemen, 1000 00:52:15,300 --> 00:52:18,830 and anybody else that supportsthe city is gone, 1001 00:52:18,830 --> 00:52:19,900 'cause there's no tax base. 1002 00:52:19,900 --> 00:52:23,770 Robots don't pay taxes. 1003 00:52:23,770 --> 00:52:25,900 NARRATOR: The averagepersonal income in Saginaw 1004 00:52:25,900 --> 00:52:29,570 is $16,000 a year. 1005 00:52:29,570 --> 00:52:32,600 A lot of the families that Iwork with here in the community, 1006 00:52:32,600 --> 00:52:33,830 both parents are working. 1007 00:52:33,830 --> 00:52:35,470 They're working two jobs. 1008 00:52:35,470 --> 00:52:38,370 Mainly, it's the wages,you know, 1009 00:52:38,370 --> 00:52:43,270 people not making a decent wageto be able to support a family. 1010 00:52:43,270 --> 00:52:46,930 Like, back in the day, my dadeven worked at the plant. 1011 00:52:46,930 --> 00:52:49,300 My mom stayed home,raised the children. 1012 00:52:49,300 --> 00:52:52,000 And that give us the opportunityto put food on the table, 1013 00:52:52,000 --> 00:52:53,370 and things of that nature. 1014 00:52:53,370 --> 00:52:56,000 And, and them times are gone. 1015 00:52:56,000 --> 00:52:57,930 If you look at this graph ofwhat's been happening 1016 00:52:57,930 --> 00:52:59,670 to America since the endof World War II, 1017 00:52:59,670 --> 00:53:03,000 you see a line for ourproductivity, 1018 00:53:03,000 --> 00:53:05,730 and our productivitygets better over time. 1019 00:53:05,730 --> 00:53:08,830 It used to be the casethat our pay, our income, 1020 00:53:08,830 --> 00:53:12,700 would increase in lockstep withthose productivity increases. 1021 00:53:12,700 --> 00:53:17,570 The weird part about this graphis how the income has decoupled, 1022 00:53:17,570 --> 00:53:21,900 is not going up the same waythat productivity is anymore. 1023 00:53:21,900 --> 00:53:24,170 NARRATOR: As automation hastaken over, 1024 00:53:24,170 --> 00:53:27,770 workers are either laid off orleft with less-skilled jobs 1025 00:53:27,770 --> 00:53:31,400 for less pay,while productivity goes up. 1026 00:53:31,400 --> 00:53:33,100 There are still plentyof factories in America. 1027 00:53:33,100 --> 00:53:35,430 We are a manufacturingpowerhouse, 1028 00:53:35,430 --> 00:53:37,670 but if you go walk aroundan American factory, 1029 00:53:37,670 --> 00:53:40,070 you do not see long linesof people 1030 00:53:40,070 --> 00:53:42,470 doing repetitive manual labor. 1031 00:53:42,470 --> 00:53:44,600 You see a whole lotof automation. 1032 00:53:44,600 --> 00:53:46,230 If you go upstairs in thatfactory 1033 00:53:46,230 --> 00:53:47,830 and look at the payrolldepartment, 1034 00:53:47,830 --> 00:53:51,130 you see one or two peoplelooking into a screen all day. 1035 00:53:51,130 --> 00:53:53,800 So, the activity is still there, 1036 00:53:53,800 --> 00:53:56,000 but the number of jobsis very, very low, 1037 00:53:56,000 --> 00:53:58,330 because of automationand tech progress. 1038 00:53:58,330 --> 00:54:01,130 Now, dealing withthat challenge, 1039 00:54:01,130 --> 00:54:02,900 and figuring out whatthe next generation 1040 00:54:02,900 --> 00:54:05,700 of the American middle classshould be doing, 1041 00:54:05,700 --> 00:54:07,700 is a really important challenge, 1042 00:54:07,700 --> 00:54:10,530 because I am pretty confidentthat we are never again 1043 00:54:10,530 --> 00:54:13,330 going to have this large,stable, prosperous 1044 00:54:13,330 --> 00:54:15,730 middle class doing routine work. 1045 00:54:15,730 --> 00:54:19,430 ♪ ♪ 1046 00:54:19,430 --> 00:54:21,970 NARRATOR: Evidence of howA.I. is likely to bring 1047 00:54:21,970 --> 00:54:25,530 accelerated change to the U.S.workforce can be found 1048 00:54:25,530 --> 00:54:27,970 not far from Saginaw. 1049 00:54:27,970 --> 00:54:29,600 This is the U.S. headquarters 1050 00:54:29,600 --> 00:54:34,070 for one of the world's largestbuilders of industrial robots, 1051 00:54:34,070 --> 00:54:38,030 a Japanese-owned company calledFanuc Robotics. 1052 00:54:38,030 --> 00:54:41,230 We've been producing robotsfor well over 35 years. 1053 00:54:41,230 --> 00:54:42,770 And you can imagine,over the years, 1054 00:54:42,770 --> 00:54:45,330 they've changed quite a bit. 1055 00:54:45,330 --> 00:54:48,230 We're utilizing the artificialintelligence 1056 00:54:48,230 --> 00:54:49,800 to really make the robotseasier to use 1057 00:54:49,800 --> 00:54:54,400 and be able to handle a broaderspectrum of opportunities. 1058 00:54:54,400 --> 00:54:57,770 We see a huge growth potentialin robotics. 1059 00:54:57,770 --> 00:55:00,330 And we see that growth potentialas being, really, 1060 00:55:00,330 --> 00:55:03,230 there's 90% of the market left. 1061 00:55:03,230 --> 00:55:05,230 NARRATOR: The industry saysoptimistically 1062 00:55:05,230 --> 00:55:09,270 that with that growth,they can create more jobs. 1063 00:55:09,270 --> 00:55:11,630 Even if there were fivepeople on a job, 1064 00:55:11,630 --> 00:55:12,870 and we reduced that down to twopeople, 1065 00:55:12,870 --> 00:55:15,800 because we automatedsome level of it, 1066 00:55:15,800 --> 00:55:18,570 we might produce two times moreparts than we did before, 1067 00:55:18,570 --> 00:55:20,170 because we automated it. 1068 00:55:20,170 --> 00:55:26,430 So now, there might be the needfor two more fork-truck drivers, 1069 00:55:26,430 --> 00:55:29,900 or two more quality-inspectionpersonnel. 1070 00:55:29,900 --> 00:55:31,870 So, although we reducesome of the people, 1071 00:55:31,870 --> 00:55:36,100 we grow in other areas as weproduce more things. 1072 00:55:36,100 --> 00:55:41,070 When I increase productivitythrough automation, I lose jobs. 1073 00:55:41,070 --> 00:55:42,370 Jobs go away. 1074 00:55:42,370 --> 00:55:45,170 And I don't care what the robotmanufacturers say, 1075 00:55:45,170 --> 00:55:47,830 you aren't replacing those tenproduction people 1076 00:55:47,830 --> 00:55:51,570 that that robot is now doingthat job, with ten people. 1077 00:55:51,570 --> 00:55:54,830 You can increase productivity toa level to stay competitive 1078 00:55:54,830 --> 00:55:58,970 with the global market-- that'swhat they're trying to do. 1079 00:55:58,970 --> 00:56:00,530 ♪ ♪ 1080 00:56:00,530 --> 00:56:02,900 NARRATOR:In the popular telling, 1081 00:56:02,900 --> 00:56:06,800 blame for widespread job losshas been aimed overseas, 1082 00:56:06,800 --> 00:56:08,900 at what's called offshoring. 1083 00:56:08,900 --> 00:56:11,200 We want to keepour factories here, 1084 00:56:11,200 --> 00:56:13,100 we want to keepour manufacturing here. 1085 00:56:13,100 --> 00:56:17,470 We don't want them movingto China, to Mexico, to Japan, 1086 00:56:17,470 --> 00:56:21,630 to India, to Vietnam. 1087 00:56:21,630 --> 00:56:23,770 NARRATOR: But it turns outmost of the job loss 1088 00:56:23,770 --> 00:56:26,370 isn't because of offshoring. 1089 00:56:26,370 --> 00:56:27,700 There's been offshoring. 1090 00:56:27,700 --> 00:56:32,300 And I think offshoring isresponsible for maybe 20% 1091 00:56:32,300 --> 00:56:34,000 of the jobs that have been lost. 1092 00:56:34,000 --> 00:56:36,270 I would say most of the jobsthat have been lost, 1093 00:56:36,270 --> 00:56:38,830 despite what most Americansthinks, was due to automation 1094 00:56:38,830 --> 00:56:41,830 or productivity growth. 1095 00:56:41,830 --> 00:56:43,570 NARRATOR:Mike Hicks is an economist 1096 00:56:43,570 --> 00:56:46,600 at Ball State Universityin Muncie, Indiana. 1097 00:56:46,600 --> 00:56:50,300 He and sociologist Emily Wornellhave been documenting 1098 00:56:50,300 --> 00:56:52,670 employment trendsin Middle America. 1099 00:56:52,670 --> 00:56:57,130 Hicks says that automation hasbeen a mostly silent job killer, 1100 00:56:57,130 --> 00:56:59,200 lowering the standard of living. 1101 00:56:59,200 --> 00:57:02,400 So, in the last 15 years, thestandard of living has dropped 1102 00:57:02,400 --> 00:57:04,600 by 15, ten to 15 percent. 1103 00:57:04,600 --> 00:57:07,100 So, that's unusualin a developed world. 1104 00:57:07,100 --> 00:57:08,600 A one-year declineis a recession. 1105 00:57:08,600 --> 00:57:12,470 A 15-year decline givesan entirely different sense 1106 00:57:12,470 --> 00:57:14,830 about the prospectsof a community. 1107 00:57:14,830 --> 00:57:18,500 And so that is commonfrom the Canadian border 1108 00:57:18,500 --> 00:57:20,970 to the Gulf of Mexico 1109 00:57:20,970 --> 00:57:23,300 in the middle swathof the United States. 1110 00:57:23,300 --> 00:57:26,130 This is something we're gonnado for you guys. 1111 00:57:26,130 --> 00:57:30,730 These were left over from oursuggestion drive that we did, 1112 00:57:30,730 --> 00:57:32,200 and we're going to give themeach two. 1113 00:57:32,200 --> 00:57:33,300 That is awesome.I mean, 1114 00:57:33,300 --> 00:57:35,070 that is going to go a long ways,right? 1115 00:57:35,070 --> 00:57:37,070 I mean, that'll really help thatfamily out during the holidays. 1116 00:57:37,070 --> 00:57:39,800 Yes, well, with the kids homefrom school, 1117 00:57:39,800 --> 00:57:41,430 the families have three mealsa day that they got 1118 00:57:41,430 --> 00:57:43,170 to put on the table. 1119 00:57:43,170 --> 00:57:45,130 So, it's going to make a bigdifference. 1120 00:57:45,130 --> 00:57:47,130 So, thank you, guys.You're welcome. 1121 00:57:47,130 --> 00:57:48,830 This is wonderful.Let them know Merry Christmas 1122 00:57:48,830 --> 00:57:50,370 on behalf of us hereat the local, okay? 1123 00:57:50,370 --> 00:57:52,930 Absolutely, you guys arejust, just amazing, thank you. 1124 00:57:52,930 --> 00:57:56,270 And please, tell, tell all theworkers how grateful 1125 00:57:56,270 --> 00:57:57,900 these families will be.We will. 1126 00:57:57,900 --> 00:58:00,870 I mean, this is not a smallproblem. 1127 00:58:00,870 --> 00:58:02,700 The need is so great. 1128 00:58:02,700 --> 00:58:05,830 And I can tell youthat it's all races, 1129 00:58:05,830 --> 00:58:08,070 it's all income classes 1130 00:58:08,070 --> 00:58:09,700 that you might think someonemight be from. 1131 00:58:09,700 --> 00:58:11,900 But I can tell you that when yousee it, 1132 00:58:11,900 --> 00:58:15,000 and you deliver this typeof gift to somebody 1133 00:58:15,000 --> 00:58:18,600 who is in need, just thegratitude that they show you 1134 00:58:18,600 --> 00:58:22,470 is incredible. 1135 00:58:22,470 --> 00:58:26,470 We actually know that peopleare at greater risk of mortality 1136 00:58:26,470 --> 00:58:30,130 for over 20 years after theylose their job due to, 1137 00:58:30,130 --> 00:58:32,670 due to no fault of their own, sosomething like automation 1138 00:58:32,670 --> 00:58:34,770 or offshoring. 1139 00:58:34,770 --> 00:58:36,970 They're at higher riskfor cardiovascular disease, 1140 00:58:36,970 --> 00:58:42,500 they're at higher riskfor depression and suicide. 1141 00:58:42,500 --> 00:58:44,630 But then with theintergenerational impacts, 1142 00:58:44,630 --> 00:58:48,230 we also see their childrenare more likely-- 1143 00:58:48,230 --> 00:58:50,300 children of parents who havelost their job 1144 00:58:50,300 --> 00:58:53,670 due to automation-- are morelikely to repeat a grade, 1145 00:58:53,670 --> 00:58:55,570 they're more likely to drop outof school, 1146 00:58:55,570 --> 00:58:57,700 they're more likely to besuspended from school, 1147 00:58:57,700 --> 00:58:59,470 and they have lower educationalattainment 1148 00:58:59,470 --> 00:59:03,200 over their entire lifetimes. 1149 00:59:03,200 --> 00:59:06,200 It's the future of this,not the past, that scares me. 1150 00:59:06,200 --> 00:59:08,700 Because I think we're in theearly decades 1151 00:59:08,700 --> 00:59:11,170 of what is a multi-decadeadjustment period. 1152 00:59:11,170 --> 00:59:14,000 ♪ ♪ 1153 00:59:14,000 --> 00:59:18,170 NARRATOR: The world is beingre-imagined. 1154 00:59:18,170 --> 00:59:20,370 This is a supermarket. 1155 00:59:20,370 --> 00:59:24,800 Robots, guided by A.I., packeverything from soap powder 1156 00:59:24,800 --> 00:59:29,530 to cantaloupes for onlineconsumers. 1157 00:59:29,530 --> 00:59:31,600 Machines that pick groceries, 1158 00:59:31,600 --> 00:59:35,170 machines that can also readreports, learn routines, 1159 00:59:35,170 --> 00:59:38,730 and comprehend are reaching deepinto factories, 1160 00:59:38,730 --> 00:59:41,870 stores, and offices. 1161 00:59:41,870 --> 00:59:43,800 At a college in Goshen, Indiana, 1162 00:59:43,800 --> 00:59:47,030 a group of local business andpolitical leaders come together 1163 00:59:47,030 --> 00:59:52,830 to try to understand the impactof A.I. and the new machines. 1164 00:59:52,830 --> 00:59:54,870 Molly Kinder studiesthe future of work 1165 00:59:54,870 --> 00:59:56,470 at a Washington think tank. 1166 00:59:56,470 --> 00:59:58,970 How many people have goneinto a fast-food restaurant 1167 00:59:58,970 --> 01:00:01,370 and done a self-ordering? 1168 01:00:01,370 --> 01:00:02,530 Anyone, yes? 1169 01:00:02,530 --> 01:00:04,400 Panera, for instance,is doing this. 1170 01:00:04,400 --> 01:00:08,270 Cashier was my first job,and in, in, where I live, 1171 01:00:08,270 --> 01:00:10,830 in Washington, DC, it's actuallythe number-one occupation 1172 01:00:10,830 --> 01:00:12,300 for the greater DC region. 1173 01:00:12,300 --> 01:00:14,670 There are millions of people whowork in cashier positions. 1174 01:00:14,670 --> 01:00:17,000 This is not a futuristicchallenge, 1175 01:00:17,000 --> 01:00:19,800 this is something that'shappening sooner than we think. 1176 01:00:19,800 --> 01:00:24,770 In the popular discussions aboutrobots and automation and work, 1177 01:00:24,770 --> 01:00:28,600 almost every image is of a manon a factory floor 1178 01:00:28,600 --> 01:00:29,770 or a truck driver. 1179 01:00:29,770 --> 01:00:32,900 And yet, in our data, when welooked, 1180 01:00:32,900 --> 01:00:35,900 women disproportionately holdthe jobs that today 1181 01:00:35,900 --> 01:00:37,900 are at highest riskof automation. 1182 01:00:37,900 --> 01:00:40,800 And that's not really beingtalked about, 1183 01:00:40,800 --> 01:00:43,700 and that's in part because womenare over-represented 1184 01:00:43,700 --> 01:00:45,570 in some of these marginalizedoccupations, 1185 01:00:45,570 --> 01:00:48,230 like a cashieror a fast-food worker. 1186 01:00:48,230 --> 01:00:53,670 And also in a large numbersin clerical jobs in offices-- 1187 01:00:53,670 --> 01:00:57,400 HR departments,payroll, finance, 1188 01:00:57,400 --> 01:01:00,900 a lot of that is more routineprocessing information, 1189 01:01:00,900 --> 01:01:03,530 processing paper,transferring data. 1190 01:01:03,530 --> 01:01:08,000 That has huge potential forautomation. 1191 01:01:08,000 --> 01:01:11,000 A.I. is going to dosome of that, software, 1192 01:01:11,000 --> 01:01:12,900 robots are going to dosome of that. 1193 01:01:12,900 --> 01:01:14,830 So how many people are stillworking 1194 01:01:14,830 --> 01:01:16,300 as switchboard operators? 1195 01:01:16,300 --> 01:01:18,170 Probably none in this country. 1196 01:01:18,170 --> 01:01:20,470 NARRATOR: The workplace ofthe future will demand 1197 01:01:20,470 --> 01:01:24,230 different skills, and gainingthem, says Molly Kinder, 1198 01:01:24,230 --> 01:01:26,300 will depend on whocan afford them. 1199 01:01:26,300 --> 01:01:28,570 I mean it's not a goodsituation in the United States. 1200 01:01:28,570 --> 01:01:30,330 There's been some excellentresearch that says 1201 01:01:30,330 --> 01:01:32,800 that half of Americanscouldn't afford 1202 01:01:32,800 --> 01:01:35,300 a $400 unexpected expense. 1203 01:01:35,300 --> 01:01:38,630 And if you want to get to a$1,000, there's even less. 1204 01:01:38,630 --> 01:01:41,270 So imagine you're going to goout without a month's pay, 1205 01:01:41,270 --> 01:01:43,330 two months' pay, a year. 1206 01:01:43,330 --> 01:01:47,030 Imagine you want to put savingstoward a course 1207 01:01:47,030 --> 01:01:49,670 to, to redevelop your career. 1208 01:01:49,670 --> 01:01:52,330 People can't afford to take timeoff of work. 1209 01:01:52,330 --> 01:01:56,600 They don't have a cushion, sothis lack of economic stability, 1210 01:01:56,600 --> 01:01:59,500 married with the disruptions inpeople's careers, 1211 01:01:59,500 --> 01:02:01,230 is a really toxic mix. 1212 01:02:01,230 --> 01:02:03,630 (blowing whistle) 1213 01:02:03,630 --> 01:02:05,530 NARRATOR: The new machineswill penetrate every sector 1214 01:02:05,530 --> 01:02:08,600 of the economy:from insurance companies 1215 01:02:08,600 --> 01:02:11,130 to human resource departments; 1216 01:02:11,130 --> 01:02:14,030 from law firms to the tradingfloors of Wall Street. 1217 01:02:14,030 --> 01:02:15,470 Wall Street'sgoing through it, 1218 01:02:15,470 --> 01:02:16,970 but every industry is goingthrough it. 1219 01:02:16,970 --> 01:02:19,630 Every company is looking at allof the disruptive technologies, 1220 01:02:19,630 --> 01:02:23,630 could be robotics or dronesor blockchain. 1221 01:02:23,630 --> 01:02:27,130 And whatever it is, everycompany's using everything 1222 01:02:27,130 --> 01:02:29,570 that's developed, everythingthat's disruptive, 1223 01:02:29,570 --> 01:02:32,370 in thinking about, "How doI apply that to my business 1224 01:02:32,370 --> 01:02:35,000 to make myself more efficient?" 1225 01:02:35,000 --> 01:02:37,400 And what efficiency means is,mostly, 1226 01:02:37,400 --> 01:02:40,670 "How do I do thiswith fewer workers?" 1227 01:02:43,900 --> 01:02:47,330 And I do think that when we lookat some of the studies 1228 01:02:47,330 --> 01:02:50,700 about opportunityin this country, 1229 01:02:50,700 --> 01:02:53,030 and the inequalityof opportunity, 1230 01:02:53,030 --> 01:02:55,830 the likelihood that you won't beable to advance 1231 01:02:55,830 --> 01:02:59,300 from where your parents were, Ithink that's, that's, 1232 01:02:59,300 --> 01:03:02,000 is very serious and getsto the heart of the way 1233 01:03:02,000 --> 01:03:06,430 we like to think of America asthe land of opportunity. 1234 01:03:06,430 --> 01:03:08,970 NARRATOR: Inequality has beenrising in America. 1235 01:03:08,970 --> 01:03:13,270 It used to be the top 1%of earners-- here in red-- 1236 01:03:13,270 --> 01:03:16,670 owned a relatively small portionof the country's wealth. 1237 01:03:16,670 --> 01:03:20,100 Middle and lower earners--in blue-- had the largest share. 1238 01:03:20,100 --> 01:03:24,970 Then, 15 years ago,the lines crossed. 1239 01:03:24,970 --> 01:03:29,500 And inequality has beenincreasing ever since. 1240 01:03:29,500 --> 01:03:31,830 There's many factors that aredriving inequality today, 1241 01:03:31,830 --> 01:03:33,330 and unfortunately,artificial intelligence-- 1242 01:03:33,330 --> 01:03:38,270 without being thoughtfulabout it-- 1243 01:03:38,270 --> 01:03:41,430 is a driver for increasedinequality 1244 01:03:41,430 --> 01:03:43,900 because it's a form ofautomation, 1245 01:03:43,900 --> 01:03:47,000 and automation is thesubstitution of capital 1246 01:03:47,000 --> 01:03:49,070 for labor. 1247 01:03:49,070 --> 01:03:52,800 And when you do that,the people with the capital win. 1248 01:03:52,800 --> 01:03:55,800 So Karl Marx was right, 1249 01:03:55,800 --> 01:03:58,100 it's a struggle between capitaland labor, 1250 01:03:58,100 --> 01:03:59,600 and with artificialintelligence, 1251 01:03:59,600 --> 01:04:02,830 we're putting our finger on thescale on the side of capital, 1252 01:04:02,830 --> 01:04:05,770 and how we wish to distributethe benefits, 1253 01:04:05,770 --> 01:04:07,230 the economic benefits, 1254 01:04:07,230 --> 01:04:09,130 that that will create is goingto be a major 1255 01:04:09,130 --> 01:04:13,430 moral consideration for societyover the next several decades. 1256 01:04:13,430 --> 01:04:19,370 This is really an outgrowthof the increasing gaps 1257 01:04:19,370 --> 01:04:23,600 of haves and have-nots--the wealthy getting wealthier, 1258 01:04:23,600 --> 01:04:24,870 the poor getting poorer. 1259 01:04:24,870 --> 01:04:28,400 It may not be specificallyrelated to A.I., 1260 01:04:28,400 --> 01:04:30,770 but as... but A.I. willexacerbate that. 1261 01:04:30,770 --> 01:04:36,430 And that, I think, will tearthe society apart, 1262 01:04:36,430 --> 01:04:38,930 because the rich will have justtoo much, 1263 01:04:38,930 --> 01:04:44,200 and those who are have-nots willhave perhaps very little way 1264 01:04:44,200 --> 01:04:46,700 of digging themselvesout of the hole. 1265 01:04:46,700 --> 01:04:50,800 And with A.I. making its impact,it, it'll be worse, I think. 1266 01:04:50,800 --> 01:04:56,170 ♪ ♪ 1267 01:04:56,170 --> 01:05:01,870 (crowd cheering and applauding) 1268 01:05:01,870 --> 01:05:05,000 (speaking on P.A.) 1269 01:05:05,000 --> 01:05:08,830 I'm here today for one mainreason. 1270 01:05:08,830 --> 01:05:12,630 To say thank you to Ohio. 1271 01:05:12,630 --> 01:05:17,400 (crowd cheering and applauding) 1272 01:05:17,400 --> 01:05:20,300 I think the Trump votewas a protest. 1273 01:05:20,300 --> 01:05:22,030 I mean that for whatever reason, 1274 01:05:22,030 --> 01:05:25,000 whatever the hot button wasthat, you know, 1275 01:05:25,000 --> 01:05:28,800 that really hit home with theseAmericans who voted for him 1276 01:05:28,800 --> 01:05:30,800 were, it was a protest vote. 1277 01:05:30,800 --> 01:05:34,530 They didn't like the directionthings were going. 1278 01:05:34,530 --> 01:05:38,270 (crowd booing and shouting) 1279 01:05:39,170 --> 01:05:40,700 I'm scared. 1280 01:05:40,700 --> 01:05:42,900 I'm gonna be quite honest withyou, I worry about the future 1281 01:05:42,900 --> 01:05:47,100 of not just this country,but the, the entire globe. 1282 01:05:47,100 --> 01:05:51,100 If we continue to go in anautomated system, 1283 01:05:51,100 --> 01:05:52,730 what are we going to do? 1284 01:05:52,730 --> 01:05:54,870 Now I've got a group of peopleat the top 1285 01:05:54,870 --> 01:05:57,170 that are making all the moneyand I don't have anybody 1286 01:05:57,170 --> 01:06:00,070 in the middlethat can support a family. 1287 01:06:00,070 --> 01:06:05,330 So do we have to go to the pointwhere we crash to come back? 1288 01:06:05,330 --> 01:06:06,630 And in this case, 1289 01:06:06,630 --> 01:06:08,030 the automation's already gonnabe there, 1290 01:06:08,030 --> 01:06:09,630 so I don't know howyou come back. 1291 01:06:09,630 --> 01:06:11,730 I'm really worriedabout where this, 1292 01:06:11,730 --> 01:06:13,700 where this leads usin the future. 1293 01:06:13,700 --> 01:06:17,000 ♪ ♪ 1294 01:06:27,200 --> 01:06:28,730 NARRATOR: The future islargely being shaped 1295 01:06:28,730 --> 01:06:32,370 by a few hugely successfultech companies. 1296 01:06:32,370 --> 01:06:35,700 They're constantly buying upsuccessful smaller companies 1297 01:06:35,700 --> 01:06:37,900 and recruiting talent. 1298 01:06:37,900 --> 01:06:39,830 Between the U.S. and China, 1299 01:06:39,830 --> 01:06:42,800 they employ a great majority ofthe leading A.I. researchers 1300 01:06:42,800 --> 01:06:46,070 and scientists. 1301 01:06:46,070 --> 01:06:48,370 In the course of amassingsuch power, 1302 01:06:48,370 --> 01:06:51,930 they've also become among therichest companies in the world. 1303 01:06:51,930 --> 01:06:58,130 A.I. really is the ultimatetool of wealth creation. 1304 01:06:58,130 --> 01:07:03,730 Think about the massive datathat, you know, Facebook has 1305 01:07:03,730 --> 01:07:08,270 on user preferences, and howit can very smartly target 1306 01:07:08,270 --> 01:07:10,400 an ad that you might buysomething 1307 01:07:10,400 --> 01:07:16,430 and get a much bigger cut thata smaller company couldn't do. 1308 01:07:16,430 --> 01:07:18,970 Same with Google,same with Amazon. 1309 01:07:18,970 --> 01:07:23,300 So it's... A.I. is a set oftools 1310 01:07:23,300 --> 01:07:26,500 that helps you maximize anobjective function, 1311 01:07:26,500 --> 01:07:32,200 and that objective functioninitially will simply be, 1312 01:07:32,200 --> 01:07:34,200 make more money. 1313 01:07:34,200 --> 01:07:36,400 NARRATOR: And it is how thesecompanies make that money, 1314 01:07:36,400 --> 01:07:41,230 and how their algorithms reachdeeper and deeper into our work, 1315 01:07:41,230 --> 01:07:42,870 our daily lives,and our democracy, 1316 01:07:42,870 --> 01:07:47,870 that makes many peopleincreasingly uncomfortable. 1317 01:07:47,870 --> 01:07:52,330 Pedro Domingos wrote the book"The Master Algorithm." 1318 01:07:52,330 --> 01:07:55,470 Everywhere you go,you generate a cloud of data. 1319 01:07:55,470 --> 01:07:58,500 You're trailing data, everythingthat you do is producing data. 1320 01:07:58,500 --> 01:07:59,900 And then there are computerslooking at that data 1321 01:07:59,900 --> 01:08:02,970 that are learning, and thesecomputers are essentially 1322 01:08:02,970 --> 01:08:05,100 trying to serve you better. 1323 01:08:05,100 --> 01:08:07,270 They're trying to personalizethings to you. 1324 01:08:07,270 --> 01:08:08,900 They're trying to adaptthe world to you. 1325 01:08:08,900 --> 01:08:10,800 So on the one hand,this is great, 1326 01:08:10,800 --> 01:08:12,430 because the world will getadapted to you 1327 01:08:12,430 --> 01:08:15,900 without you even having toexplicitly adapt it. 1328 01:08:15,900 --> 01:08:18,600 There's also a danger, becausethe entities in the companies 1329 01:08:18,600 --> 01:08:20,100 that are in control of thosealgorithms 1330 01:08:20,100 --> 01:08:22,000 don't necessarily have the samegoals as you, 1331 01:08:22,000 --> 01:08:24,800 and this is where I think peopleneed to be aware that, 1332 01:08:24,799 --> 01:08:29,999 what's going on, so they canhave more control over it. 1333 01:08:30,000 --> 01:08:31,630 You know, we came into thisnew world thinking 1334 01:08:31,630 --> 01:08:35,530 that we were usersof social media. 1335 01:08:35,529 --> 01:08:37,799 It didn't occur to usthat social media 1336 01:08:37,799 --> 01:08:40,429 was actually using us. 1337 01:08:40,430 --> 01:08:43,830 We thought that we weresearching Google. 1338 01:08:43,830 --> 01:08:48,700 We had no idea that Googlewas searching us. 1339 01:08:48,700 --> 01:08:50,800 NARRATOR: Shoshana Zuboffis a Harvard Business School 1340 01:08:50,799 --> 01:08:52,869 professor emerita. 1341 01:08:52,870 --> 01:08:55,930 In 1988, she wrote a definitivebook called 1342 01:08:55,930 --> 01:08:58,370 "In the Age ofthe Smart Machine." 1343 01:08:58,370 --> 01:09:01,970 For the last seven years,she has worked on a new book, 1344 01:09:01,970 --> 01:09:04,730 making the case that we have nowentered a new phase 1345 01:09:04,730 --> 01:09:09,970 of the economy, which she calls"surveillance capitalism." 1346 01:09:09,970 --> 01:09:16,330 So, famously, industrialcapitalism claimed nature. 1347 01:09:16,330 --> 01:09:20,300 Innocent rivers, and meadows,and forests, and so forth, 1348 01:09:20,299 --> 01:09:25,129 for the market dynamic to bereborn as real estate, 1349 01:09:25,130 --> 01:09:27,970 as land that could be soldand purchased. 1350 01:09:27,970 --> 01:09:32,230 Industrial capitalism claimedwork for the market dynamic 1351 01:09:32,230 --> 01:09:35,170 to reborn, to be reborn as labor 1352 01:09:35,170 --> 01:09:38,700 that could be soldand purchased. 1353 01:09:38,700 --> 01:09:40,830 Now, here comes surveillancecapitalism, 1354 01:09:40,830 --> 01:09:47,700 following this pattern, but witha dark and startling twist. 1355 01:09:47,700 --> 01:09:51,700 What surveillance capitalismclaims is private, 1356 01:09:51,700 --> 01:09:53,930 human experience. 1357 01:09:53,930 --> 01:09:58,970 Private, human experience isclaimed as a free source 1358 01:09:58,970 --> 01:10:05,800 of raw material, fabricated intopredictions of human behavior. 1359 01:10:05,800 --> 01:10:09,370 And it turns out that there area lot of businesses 1360 01:10:09,370 --> 01:10:14,270 that really want to know whatwe will do now, soon, and later. 1361 01:10:17,800 --> 01:10:19,430 NARRATOR: Like most people, 1362 01:10:19,430 --> 01:10:21,470 Alastair Mactaggarthad know idea 1363 01:10:21,470 --> 01:10:23,600 about this new surveillancebusiness, 1364 01:10:23,600 --> 01:10:27,370 until one evening in 2015. 1365 01:10:27,370 --> 01:10:30,230 I had a conversation with afellow who's an engineer, 1366 01:10:30,230 --> 01:10:33,870 and I was just talking to himone night at a, 1367 01:10:33,870 --> 01:10:35,270 you know, a dinner,at a cocktail party. 1368 01:10:35,270 --> 01:10:37,470 And I... there had beensomething in the press that day 1369 01:10:37,470 --> 01:10:39,900 about privacy in the paper,and I remember asking him-- 1370 01:10:39,900 --> 01:10:41,870 he worked for Google-- "What'sthe big deal about all, 1371 01:10:41,870 --> 01:10:44,270 why are people so worked upabout it?" 1372 01:10:44,270 --> 01:10:45,670 And I thought it was gonna beone of those conversations, 1373 01:10:45,670 --> 01:10:49,330 like, with, you know, if youever ask an airline pilot, 1374 01:10:49,330 --> 01:10:50,570 "Should I be worried aboutflying?" 1375 01:10:50,570 --> 01:10:52,070 and they say,"Oh, the most dangerous part 1376 01:10:52,070 --> 01:10:55,030 is coming to the airport,you know, in the car." 1377 01:10:55,030 --> 01:10:57,730 And he said, "Oh, you'd behorrified 1378 01:10:57,730 --> 01:10:59,900 if you knew how much we knewabout you." 1379 01:10:59,900 --> 01:11:02,100 And I remember that kind ofstuck in my head, 1380 01:11:02,100 --> 01:11:04,530 because it was notwhat I expected. 1381 01:11:04,530 --> 01:11:08,400 NARRATOR: That questionwould change his life. 1382 01:11:08,400 --> 01:11:09,930 A successful California realestate developer, 1383 01:11:09,930 --> 01:11:15,730 Mactaggart began researchingthe new business model. 1384 01:11:15,730 --> 01:11:17,870 What I've learned since isthat their entire business 1385 01:11:17,870 --> 01:11:20,630 is learning as much about youas they can. 1386 01:11:20,630 --> 01:11:21,970 Everything about your thoughts,and your desires, 1387 01:11:21,970 --> 01:11:25,730 and your dreams,and who your friends are, 1388 01:11:25,730 --> 01:11:27,700 and what you're thinking, whatyour private thoughts are. 1389 01:11:27,700 --> 01:11:29,770 And with that,that's true power. 1390 01:11:29,770 --> 01:11:33,430 And so, I think...I didn't know that at the time. 1391 01:11:33,430 --> 01:11:35,470 That their entire businessis basically mining 1392 01:11:35,470 --> 01:11:37,200 the data of your life. 1393 01:11:37,200 --> 01:11:39,070 ♪ ♪ 1394 01:11:39,070 --> 01:11:43,200 NARRATOR: Shoshana Zuboff hadbeen doing her own research. 1395 01:11:43,200 --> 01:11:45,970 You know, I'd been readingand reading and reading. 1396 01:11:45,970 --> 01:11:48,370 From patents, to transcriptsof earnings calls, 1397 01:11:48,370 --> 01:11:50,430 research reports. 1398 01:11:50,430 --> 01:11:52,130 And, you know,just literally everything, 1399 01:11:52,130 --> 01:11:56,730 for years and years and years. 1400 01:11:56,730 --> 01:11:57,930 NARRATOR: Her studiesincluded the early days 1401 01:11:57,930 --> 01:12:00,400 of Google, started in 1998 1402 01:12:00,400 --> 01:12:02,570 by two young Stanford gradstudents, 1403 01:12:02,570 --> 01:12:06,330 Sergey Brin and Larry Page. 1404 01:12:06,330 --> 01:12:10,000 In the beginning, they had noclear business model. 1405 01:12:10,000 --> 01:12:13,900 Their unofficial motto was,"Don't Be Evil." 1406 01:12:13,900 --> 01:12:16,270 Right from the start,the founders, 1407 01:12:16,270 --> 01:12:20,170 Larry Page and Sergey Brin,they had been very public 1408 01:12:20,170 --> 01:12:26,200 about their antipathytoward advertising. 1409 01:12:26,200 --> 01:12:31,300 Advertising would distortthe internet 1410 01:12:31,300 --> 01:12:37,800 and it would distort anddisfigure the, the purity 1411 01:12:37,800 --> 01:12:41,700 of any search engine,including their own. 1412 01:12:41,700 --> 01:12:43,070 Once in love with e-commerce, 1413 01:12:43,070 --> 01:12:46,300 Wall Street has turned its backon the dotcoms. 1414 01:12:46,300 --> 01:12:49,400 NARRATOR: Then came thedotcom crash of the early 2000s. 1415 01:12:49,400 --> 01:12:51,400 ...has left hundreds ofunprofitable internet companies 1416 01:12:51,400 --> 01:12:54,830 begging for love and money. 1417 01:12:54,830 --> 01:12:56,730 NARRATOR: While Google hadrapidly become the default 1418 01:12:56,730 --> 01:12:58,830 search engine for tens ofmillions of users, 1419 01:12:58,830 --> 01:13:04,070 their investors were pressuringthem to make more money. 1420 01:13:04,070 --> 01:13:06,100 Without a new business model, 1421 01:13:06,100 --> 01:13:10,530 the founders knew that the youngcompany was in danger. 1422 01:13:10,530 --> 01:13:14,470 In this state of emergency,the founders decided, 1423 01:13:14,470 --> 01:13:19,300 "We've simply got to find a wayto save this company." 1424 01:13:19,300 --> 01:13:25,630 And so, parallel to this wereanother set of discoveries, 1425 01:13:25,630 --> 01:13:30,970 where it turns out that wheneverwe search or whenever we browse, 1426 01:13:30,970 --> 01:13:35,030 we're leaving behind traces--digital traces-- 1427 01:13:35,030 --> 01:13:37,230 of our behavior. 1428 01:13:37,230 --> 01:13:39,330 And those traces,back in these days, 1429 01:13:39,330 --> 01:13:43,500 were called digital exhaust. 1430 01:13:43,500 --> 01:13:45,400 NARRATOR: They realized howvaluable this data could be 1431 01:13:45,400 --> 01:13:47,700 by applying machine learningalgorithms 1432 01:13:47,700 --> 01:13:52,570 to predict users' interests. 1433 01:13:52,570 --> 01:13:54,670 What happened was,they decided to turn 1434 01:13:54,670 --> 01:13:57,630 to those data logsin a systematic way, 1435 01:13:57,630 --> 01:14:01,670 and to begin to use thesesurplus data 1436 01:14:01,670 --> 01:14:06,970 as a way to come up withfine-grained predictions 1437 01:14:06,970 --> 01:14:11,330 of what a user would click on,what kind of ad 1438 01:14:11,330 --> 01:14:14,230 a user would click on. 1439 01:14:14,230 --> 01:14:18,700 And inside Google, they startedseeing these revenues 1440 01:14:18,700 --> 01:14:22,830 pile up at a startling rate. 1441 01:14:22,830 --> 01:14:26,000 They realized that they had tokeep it secret. 1442 01:14:26,000 --> 01:14:28,930 They didn't want anyone to knowhow much money they were making, 1443 01:14:28,930 --> 01:14:31,500 or how they were making it. 1444 01:14:31,500 --> 01:14:35,700 Because users had no idea thatthese extra-behavioral data 1445 01:14:35,700 --> 01:14:39,070 that told so much about them,you know, was just out there, 1446 01:14:39,070 --> 01:14:43,700 and now it was being usedto predict their future. 1447 01:14:43,700 --> 01:14:46,100 NARRATOR: When Google'sI.P.O. took place 1448 01:14:46,100 --> 01:14:47,170 just a few years later, 1449 01:14:47,170 --> 01:14:49,700 the company had a marketcapitalization 1450 01:14:49,700 --> 01:14:53,230 of around $23 billion. 1451 01:14:53,230 --> 01:14:56,000 Google's stock was now asvaluable as General Motors. 1452 01:14:56,000 --> 01:14:59,100 ♪ ♪ 1453 01:14:59,100 --> 01:15:02,330 And it was only when Googlewent public in 2004 1454 01:15:02,330 --> 01:15:05,600 that the numbers were released. 1455 01:15:05,600 --> 01:15:10,600 And it's at that point that welearn that between the year 2000 1456 01:15:10,600 --> 01:15:14,500 and the year 2004, Google'srevenue line increased 1457 01:15:14,500 --> 01:15:20,400 by 3,590%. 1458 01:15:20,400 --> 01:15:22,470 Let's talk a little aboutinformation, and search, 1459 01:15:22,470 --> 01:15:24,630 and how people consume it. 1460 01:15:24,630 --> 01:15:27,230 NARRATOR: By 2010, the C.E.O.of Google, Eric Schmidt, 1461 01:15:27,230 --> 01:15:29,570 would tell "The Atlantic"magazine... 1462 01:15:29,570 --> 01:15:33,230 ...is, we don't need you totype at all. 1463 01:15:33,230 --> 01:15:35,930 Because we know where you are,with your permission, 1464 01:15:35,930 --> 01:15:39,630 we know where you've been,with your permission. 1465 01:15:39,630 --> 01:15:41,700 We can more or less guess whatyou're thinking about. 1466 01:15:41,700 --> 01:15:44,200 (audience laughing)Now, is that over the line? 1467 01:15:44,200 --> 01:15:45,800 NARRATOR: Eric Schmidtand Google declined 1468 01:15:45,800 --> 01:15:49,370 to be interviewedfor this program. 1469 01:15:49,370 --> 01:15:52,670 Google's new business model forpredicting users' profiles 1470 01:15:52,670 --> 01:15:58,100 had migrated to other companies,particularly Facebook. 1471 01:15:58,100 --> 01:16:00,130 Roger McNamee was an earlyinvestor 1472 01:16:00,130 --> 01:16:02,330 and adviser to Facebook. 1473 01:16:02,330 --> 01:16:05,900 He's now a critic, and wrotea book about the company. 1474 01:16:05,900 --> 01:16:08,930 He says he's concerned about howwidely companies like Facebook 1475 01:16:08,930 --> 01:16:11,770 and Google have been castingthe net for data. 1476 01:16:11,770 --> 01:16:13,530 And then they realized,"Wait a minute, 1477 01:16:13,530 --> 01:16:16,470 there's all this data inthe economy we don't have." 1478 01:16:16,470 --> 01:16:18,700 So they went to credit cardprocessors, 1479 01:16:18,700 --> 01:16:20,430 and credit rating services, 1480 01:16:20,430 --> 01:16:23,030 and said, "We wantto buy your data." 1481 01:16:23,030 --> 01:16:25,100 They go to health and wellnessapps and say, 1482 01:16:25,100 --> 01:16:26,600 "Hey, you got women'smenstrual cycles? 1483 01:16:26,600 --> 01:16:28,530 We want all that stuff." 1484 01:16:28,530 --> 01:16:30,830 Why are they doing that? 1485 01:16:30,830 --> 01:16:34,430 They're doing that becausebehavioral prediction 1486 01:16:34,430 --> 01:16:38,270 is about taking uncertaintyout of life. 1487 01:16:38,270 --> 01:16:40,670 Advertising and marketingare all about uncertainty-- 1488 01:16:40,670 --> 01:16:43,530 you never really know who'sgoing to buy your product. 1489 01:16:43,530 --> 01:16:45,500 Until now. 1490 01:16:45,500 --> 01:16:49,870 We have to recognize that wegave technology a place 1491 01:16:49,870 --> 01:16:55,700 in our livesthat it had not earned. 1492 01:16:55,700 --> 01:17:00,530 That essentially, becausetechnology always made things 1493 01:17:00,530 --> 01:17:03,530 better in the '50s, '60s, '70s,'80s, and '90s, 1494 01:17:03,530 --> 01:17:07,030 we developed a sense ofinevitability 1495 01:17:07,030 --> 01:17:10,000 that it will always make thingsbetter. 1496 01:17:10,000 --> 01:17:13,630 We developed a trust, and theindustry earned good will 1497 01:17:13,630 --> 01:17:20,330 that Facebook and Google havecashed in. 1498 01:17:20,330 --> 01:17:23,500 NARRATOR: The model is simplythis: provide a free service-- 1499 01:17:23,500 --> 01:17:26,630 like Facebook-- and in exchange,you collect the data 1500 01:17:26,630 --> 01:17:28,870 of the millions who use it. 1501 01:17:28,870 --> 01:17:31,800 ♪ ♪ 1502 01:17:31,800 --> 01:17:37,470 And every sliver of informationis valuable. 1503 01:17:37,470 --> 01:17:41,370 It's not just what you post,it's that you post. 1504 01:17:41,370 --> 01:17:44,800 It's not just that you makeplans to see your friends later. 1505 01:17:44,800 --> 01:17:47,470 It's whether you say,"I'll see you later," 1506 01:17:47,470 --> 01:17:51,000 or, "I'll see you at 6:45." 1507 01:17:51,000 --> 01:17:54,030 It's not just that you talkabout the things 1508 01:17:54,030 --> 01:17:56,330 that you have to do today. 1509 01:17:56,330 --> 01:17:59,230 It's whether you simply rattlethem on in a, 1510 01:17:59,230 --> 01:18:04,500 in a rambling paragraph,or list them as bullet points. 1511 01:18:04,500 --> 01:18:09,030 All of these tiny signals arethe behavioral surplus 1512 01:18:09,030 --> 01:18:13,630 that turns out to have immensepredictive value. 1513 01:18:13,630 --> 01:18:16,300 NARRATOR: In 2010, Facebookexperimented 1514 01:18:16,300 --> 01:18:19,070 with A.I.'s predictive powersin what they called 1515 01:18:19,070 --> 01:18:21,830 a "social contagion" experiment. 1516 01:18:21,830 --> 01:18:25,470 They wanted to see if, throughonline messaging, 1517 01:18:25,470 --> 01:18:30,070 they could influence real-worldbehavior. 1518 01:18:30,070 --> 01:18:32,670 The aim was to get more peopleto the polls 1519 01:18:32,670 --> 01:18:34,530 in the 2010 midterm elections. 1520 01:18:34,530 --> 01:18:38,000 Cleveland, I need you to keepon fighting. 1521 01:18:38,000 --> 01:18:41,030 I need you to keep on believing. 1522 01:18:41,030 --> 01:18:42,530 NARRATOR: They offered61 million users 1523 01:18:42,530 --> 01:18:45,470 an "I voted" button togetherwith faces of friends 1524 01:18:45,470 --> 01:18:47,330 who had voted. 1525 01:18:47,330 --> 01:18:52,000 A subset of users receivedjust the button. 1526 01:18:52,000 --> 01:18:56,130 In the end, they claimed to havenudged 340,000 people to vote. 1527 01:19:00,270 --> 01:19:03,170 They would conduct other"massive contagion" experiments. 1528 01:19:03,170 --> 01:19:06,600 Among them, one showing that byadjusting their feeds, 1529 01:19:06,600 --> 01:19:12,000 they could make usershappy or sad. 1530 01:19:12,000 --> 01:19:13,300 When they went to write upthese findings, 1531 01:19:13,300 --> 01:19:16,270 they boasted about two things. 1532 01:19:16,270 --> 01:19:19,770 One was, "Oh, my goodness. 1533 01:19:19,770 --> 01:19:24,770 Now we know that we can use cuesin the online environment 1534 01:19:24,770 --> 01:19:28,830 to change real-world behavior. 1535 01:19:28,830 --> 01:19:31,770 That's big news." 1536 01:19:31,770 --> 01:19:35,600 The second thing that theyunderstood, and they celebrated, 1537 01:19:35,600 --> 01:19:39,030 was that, "We can do this in away that bypasses 1538 01:19:39,030 --> 01:19:43,370 the users' awareness." 1539 01:19:43,370 --> 01:19:47,500 Private corporations havebuilt a corporate surveillance 1540 01:19:47,500 --> 01:19:52,370 state without our awarenessor permission. 1541 01:19:52,370 --> 01:19:55,230 And the systems necessary tomake it work 1542 01:19:55,230 --> 01:19:58,430 are getting a lot better,specifically with what are known 1543 01:19:58,430 --> 01:20:01,500 as internet of things,smart appliances, you know, 1544 01:20:01,500 --> 01:20:04,430 powered by the Alexa voicerecognition system, 1545 01:20:04,430 --> 01:20:06,870 or the Google Home system. 1546 01:20:06,870 --> 01:20:09,700 Okay, Google,play the morning playlist. 1547 01:20:09,700 --> 01:20:12,200 Okay, playing morningplaylist. 1548 01:20:12,200 --> 01:20:14,300 ♪ ♪ 1549 01:20:14,300 --> 01:20:16,370 Okay, Google,play music in all rooms. 1550 01:20:16,370 --> 01:20:18,030 ♪ ♪ 1551 01:20:18,030 --> 01:20:21,000 And those will put thesurveillance in places 1552 01:20:21,000 --> 01:20:22,270 we've never had it before-- 1553 01:20:22,270 --> 01:20:24,800 living rooms, kitchens,bedrooms. 1554 01:20:24,800 --> 01:20:27,300 And I find all of thatterrifying. 1555 01:20:27,300 --> 01:20:29,630 Okay, Google, I'm listening. 1556 01:20:29,630 --> 01:20:31,400 NARRATOR: The companies saythey're not using the data 1557 01:20:31,400 --> 01:20:36,770 to target ads, but helping A.I.improve the user experience. 1558 01:20:36,770 --> 01:20:40,030 Alexa, turn on the fan. 1559 01:20:40,030 --> 01:20:41,500 (fan clicks on) 1560 01:20:41,500 --> 01:20:42,670 Okay. 1561 01:20:42,670 --> 01:20:43,830 NARRATOR: Meanwhile, they areresearching 1562 01:20:43,830 --> 01:20:45,930 and applying for patents 1563 01:20:45,930 --> 01:20:48,900 to expand their reachinto homes and lives. 1564 01:20:48,900 --> 01:20:51,230 Alexa, take a video. 1565 01:20:51,230 --> 01:20:52,670 (camera chirps) 1566 01:20:52,670 --> 01:20:54,570 The more and more that youuse spoken interfaces-- 1567 01:20:54,570 --> 01:20:57,800 so smart speakers-- they'rebeing trained 1568 01:20:57,800 --> 01:21:00,770 not just to recognizewho you are, 1569 01:21:00,770 --> 01:21:03,970 but they're starting to takebaselines 1570 01:21:03,970 --> 01:21:09,770 and comparing changes over time. 1571 01:21:09,770 --> 01:21:12,970 So does your cadence increaseor decrease? 1572 01:21:12,970 --> 01:21:15,600 Are you sneezingwhile you're talking? 1573 01:21:15,600 --> 01:21:18,730 Is your voice a little wobbly? 1574 01:21:18,730 --> 01:21:21,570 The purpose of doing this isto understand 1575 01:21:21,570 --> 01:21:24,430 more about you in real time. 1576 01:21:24,430 --> 01:21:27,830 So that a system could makeinferences, perhaps, 1577 01:21:27,830 --> 01:21:30,600 like, do you have a cold? 1578 01:21:30,600 --> 01:21:33,370 Are you in a manic phase? 1579 01:21:33,370 --> 01:21:35,100 Are you feeling depressed? 1580 01:21:35,100 --> 01:21:38,700 So that is an extraordinaryamount of information 1581 01:21:38,700 --> 01:21:41,670 that can be gleaned by yousimply waking up 1582 01:21:41,670 --> 01:21:45,330 and asking your smart speaker,"What's the weather today?" 1583 01:21:45,330 --> 01:21:47,430 Alexa, what's the weatherfor tonight? 1584 01:21:47,430 --> 01:21:50,630 Currently, in Pasadena, it's58 degrees with cloudy skies. 1585 01:21:50,630 --> 01:21:52,900 Inside it is, then. 1586 01:21:52,900 --> 01:21:54,700 Dinner! 1587 01:21:54,700 --> 01:21:57,630 The point is that thisis the same 1588 01:21:57,630 --> 01:22:01,800 micro-behavioral targeting thatis directed 1589 01:22:01,800 --> 01:22:08,670 toward individuals based onintimate, detailed understanding 1590 01:22:08,670 --> 01:22:11,000 of personalities. 1591 01:22:11,000 --> 01:22:15,600 So this is precisely whatCambridge Analytica did, 1592 01:22:15,600 --> 01:22:19,300 simply pivoting fromthe advertisers 1593 01:22:19,300 --> 01:22:23,500 to the political outcomes. 1594 01:22:23,500 --> 01:22:26,170 NARRATOR: The CambridgeAnalytica scandal of 2018 1595 01:22:26,170 --> 01:22:29,970 engulfed Facebook, forcingMark Zuckerberg to appear 1596 01:22:29,970 --> 01:22:32,900 before Congress to explain howthe data 1597 01:22:32,900 --> 01:22:36,300 of up to 87 million Facebookusers had been harvested 1598 01:22:36,300 --> 01:22:42,870 by a political consultingcompany based in the U.K. 1599 01:22:42,870 --> 01:22:45,500 The purpose was to targetand manipulate voters 1600 01:22:45,500 --> 01:22:48,170 in the 2016 presidentialcampaign, 1601 01:22:48,170 --> 01:22:51,800 as well as the Brexitreferendum. 1602 01:22:51,800 --> 01:22:53,870 Cambridge Analytica had beenlargely funded 1603 01:22:53,870 --> 01:22:58,830 by conservative hedge fundbillionaire Robert Mercer. 1604 01:22:58,830 --> 01:23:02,370 And now we know that anybillionaire with enough money, 1605 01:23:02,370 --> 01:23:04,200 who can buy the data, 1606 01:23:04,200 --> 01:23:07,230 buy the machine intelligencecapabilities, 1607 01:23:07,230 --> 01:23:10,700 buy the skilled data scientists, 1608 01:23:10,700 --> 01:23:16,330 you know, they too cancommandeer the public, 1609 01:23:16,330 --> 01:23:23,370 and infect and infiltrate andupend our democracy 1610 01:23:23,370 --> 01:23:27,770 with the same methodologies thatsurveillance capitalism 1611 01:23:27,770 --> 01:23:32,270 uses every single day. 1612 01:23:32,270 --> 01:23:35,070 We didn't take a broad enoughview of our responsibility, 1613 01:23:35,070 --> 01:23:37,230 and that was a big mistake. 1614 01:23:37,230 --> 01:23:40,830 And it was my mistake,and I'm sorry. 1615 01:23:40,830 --> 01:23:41,770 NARRATOR:Zuckerberg has apologized 1616 01:23:41,770 --> 01:23:44,370 for numerous violations ofprivacy, 1617 01:23:44,370 --> 01:23:47,130 and his company was recentlyfined $5 billion 1618 01:23:47,130 --> 01:23:50,300 by the Federal Trade Commission. 1619 01:23:50,300 --> 01:23:53,230 He has said Facebook will nowmake data protection a priority, 1620 01:23:53,230 --> 01:23:56,800 and the company has suspendedtens of thousands 1621 01:23:56,800 --> 01:23:59,400 of third-party apps from itsplatform 1622 01:23:59,400 --> 01:24:02,930 as a result of an internalinvestigation. 1623 01:24:02,930 --> 01:24:06,870 You know, I wish I could saythat after Cambridge Analytica, 1624 01:24:06,870 --> 01:24:09,030 we've learned our lesson andthat everything will be much 1625 01:24:09,030 --> 01:24:12,930 better after that, but I'mafraid the opposite is true. 1626 01:24:12,930 --> 01:24:14,900 In some ways, CambridgeAnalytica was using tools 1627 01:24:14,900 --> 01:24:16,830 that were ten years old. 1628 01:24:16,830 --> 01:24:18,570 It was really, in some ways,old-school, 1629 01:24:18,570 --> 01:24:20,670 first-wave data science. 1630 01:24:20,670 --> 01:24:22,270 What we're looking at now,with current tools 1631 01:24:22,270 --> 01:24:26,270 and machine learning, is thatthe ability for manipulation, 1632 01:24:26,270 --> 01:24:28,930 both in terms of electionsand opinions, 1633 01:24:28,930 --> 01:24:31,700 but more broadly,just how information travels, 1634 01:24:31,700 --> 01:24:34,630 That is a much bigger problem, 1635 01:24:34,630 --> 01:24:36,200 and certainly much more seriousthan what we faced 1636 01:24:36,200 --> 01:24:40,070 with Cambridge Analytica. 1637 01:24:40,070 --> 01:24:43,670 NARRATOR: A.I. pioneer YoshuaBengio also has concerns 1638 01:24:43,670 --> 01:24:48,470 about how his algorithmsare being used. 1639 01:24:48,470 --> 01:24:51,600 So the A.I.s are tools. 1640 01:24:51,600 --> 01:24:56,330 And they will serve the peoplewho control those tools. 1641 01:24:56,330 --> 01:25:01,700 If those people's interests goagainst the, the values 1642 01:25:01,700 --> 01:25:04,670 of democracy, then democracy isin danger. 1643 01:25:04,670 --> 01:25:10,430 So I believe that scientistswho contribute to science, 1644 01:25:10,430 --> 01:25:14,670 when that science can or willhave an impact on society, 1645 01:25:14,670 --> 01:25:17,670 those scientists have aresponsibility. 1646 01:25:17,670 --> 01:25:19,700 It's a little bit like thephysicists of, 1647 01:25:19,700 --> 01:25:21,800 around the Second World War, 1648 01:25:21,800 --> 01:25:25,100 who rose up to tellthe governments, 1649 01:25:25,100 --> 01:25:29,130 "Wait, nuclear powercan be dangerous 1650 01:25:29,130 --> 01:25:31,930 and nuclear war can be really,really destructive." 1651 01:25:31,930 --> 01:25:36,400 And today, the equivalent of aphysicist of the '40s and '50s 1652 01:25:36,400 --> 01:25:38,730 and '60s are,are the computer scientists 1653 01:25:38,730 --> 01:25:41,430 who are doing machine learningand A.I. 1654 01:25:41,430 --> 01:25:45,000 ♪ ♪ 1655 01:25:45,000 --> 01:25:46,300 NARRATOR: One person whowanted to do something 1656 01:25:46,300 --> 01:25:49,330 about the dangers was nota computer scientist, 1657 01:25:49,330 --> 01:25:53,330 but an ordinary citizen. 1658 01:25:53,330 --> 01:25:55,600 Alastair Mactaggart was alarmed. 1659 01:25:55,600 --> 01:25:58,800 Voting is, for me,the most alarming one. 1660 01:25:58,800 --> 01:26:00,570 If less than 100,000 votesseparated 1661 01:26:00,570 --> 01:26:03,330 the last two candidates in thelast presidential election, 1662 01:26:03,330 --> 01:26:06,900 in three states... 1663 01:26:06,900 --> 01:26:10,430 NARRATOR: He began a solitarycampaign. 1664 01:26:10,430 --> 01:26:12,100 We're talking aboutconvincing a relatively tiny 1665 01:26:12,100 --> 01:26:14,900 fraction of the votersin a very... 1666 01:26:14,900 --> 01:26:17,700 in a handful of statesto either come out and vote 1667 01:26:17,700 --> 01:26:18,970 or stay home. 1668 01:26:18,970 --> 01:26:21,200 And remember, these companiesknow everybody intimately. 1669 01:26:21,200 --> 01:26:24,470 They know who's a racist,who's a misogynist, 1670 01:26:24,470 --> 01:26:26,770 who's a homophobe,who's a conspiracy theorist. 1671 01:26:26,770 --> 01:26:28,770 They know the lazy people andthe gullible people. 1672 01:26:28,770 --> 01:26:31,130 They have access to the greatesttrove of personal information 1673 01:26:31,130 --> 01:26:32,670 that's ever been assembled. 1674 01:26:32,670 --> 01:26:35,370 They have the world's best datascientists. 1675 01:26:35,370 --> 01:26:37,430 And they have essentiallya frictionless way 1676 01:26:37,430 --> 01:26:39,670 of communicating with you. 1677 01:26:39,670 --> 01:26:43,070 This is power. 1678 01:26:43,070 --> 01:26:44,800 NARRATOR: Mactaggart starteda signature drive 1679 01:26:44,800 --> 01:26:47,000 for a California ballotinitiative, 1680 01:26:47,000 --> 01:26:51,230 for a law to give consumerscontrol of their digital data. 1681 01:26:51,230 --> 01:26:54,670 In all, he would spend$4 million of his own money 1682 01:26:54,670 --> 01:26:58,600 in an effort to rein in thegoliaths of Silicon Valley. 1683 01:26:58,600 --> 01:27:02,570 Google, Facebook, AT&T,and Comcast 1684 01:27:02,570 --> 01:27:06,170 all opposed his initiative. 1685 01:27:06,170 --> 01:27:09,200 I'll tell you, I was scared.Fear. 1686 01:27:09,200 --> 01:27:12,000 Fear of looking likea world-class idiot. 1687 01:27:12,000 --> 01:27:14,930 The market cap of all the firmsarrayed against me were, 1688 01:27:14,930 --> 01:27:19,800 was over $6 trillion. 1689 01:27:19,800 --> 01:27:21,600 NARRATOR: He needed 500,000signatures 1690 01:27:21,600 --> 01:27:25,070 to get his initiativeon the ballot. 1691 01:27:25,070 --> 01:27:27,370 He got well over 600,000. 1692 01:27:27,370 --> 01:27:33,530 Polls showed 80% approvalfor a privacy law. 1693 01:27:33,530 --> 01:27:37,670 That made the politicians inSacramento pay attention. 1694 01:27:37,670 --> 01:27:40,030 So Mactaggart decided thatbecause he was holding 1695 01:27:40,030 --> 01:27:44,100 a strong hand, it was worthnegotiating with them. 1696 01:27:44,100 --> 01:27:46,530 And if AB-375 passesby tomorrow 1697 01:27:46,530 --> 01:27:48,230 and is signed into lawby the governor, 1698 01:27:48,230 --> 01:27:49,770 we will withdraw the initiative. 1699 01:27:49,770 --> 01:27:51,270 Our deadline to do so istomorrow at 5:00. 1700 01:27:51,270 --> 01:27:53,470 NARRATOR:At the very last moment, 1701 01:27:53,470 --> 01:27:55,600 a new law was rushed to thefloor of the state house. 1702 01:27:55,600 --> 01:27:57,470 Everyone take their seats,please. 1703 01:27:57,470 --> 01:28:01,800 Mr. Secretary,please call the roll. 1704 01:28:01,800 --> 01:28:05,900 The voting starts.Alan, aye. 1705 01:28:05,900 --> 01:28:07,570 And the first guy,I think, was a Republican, 1706 01:28:07,570 --> 01:28:08,870 and he voted for it. 1707 01:28:08,870 --> 01:28:10,600 And everybody had said theRepublicans won't vote for it 1708 01:28:10,600 --> 01:28:11,670 because it has this privateright of action, 1709 01:28:11,670 --> 01:28:13,830 where consumers can sue. 1710 01:28:13,830 --> 01:28:15,530 And the guy in the Senate,he calls the name. 1711 01:28:15,530 --> 01:28:16,770 Aye, Roth. 1712 01:28:16,770 --> 01:28:17,830 Aye, Skinner. 1713 01:28:17,830 --> 01:28:19,000 Aye, Stern. 1714 01:28:19,000 --> 01:28:20,800 Aye, Stone. 1715 01:28:20,800 --> 01:28:23,600 You can see down below,and everyone went green, 1716 01:28:23,600 --> 01:28:26,270 and then it passed unanimously. 1717 01:28:26,270 --> 01:28:29,770 Ayes 36; No zero,the measure passes. 1718 01:28:29,770 --> 01:28:32,200 Immediate transmittal to the... 1719 01:28:32,200 --> 01:28:34,530 So I was blown away. 1720 01:28:34,530 --> 01:28:36,500 It was, it was a day I willnever forget. 1721 01:28:41,770 --> 01:28:43,630 So in January, next year,you as a California resident 1722 01:28:43,630 --> 01:28:45,900 will have the right to go to anycompany and say, 1723 01:28:45,900 --> 01:28:47,270 "What have you collected on mein the last 12 years... 1724 01:28:47,270 --> 01:28:48,700 12 months? 1725 01:28:48,700 --> 01:28:51,370 What of my personal informationdo you have?" 1726 01:28:51,370 --> 01:28:52,300 So that's the first right. 1727 01:28:52,300 --> 01:28:54,200 It's right of... we call thatthe right to know. 1728 01:28:54,200 --> 01:28:56,230 The second is the rightto say no. 1729 01:28:56,230 --> 01:28:59,030 And that's the right to go toany company and click a button, 1730 01:28:59,030 --> 01:29:00,930 on any page where they'recollecting your information, 1731 01:29:00,930 --> 01:29:03,200 and say, "Do not sellmy information." 1732 01:29:03,200 --> 01:29:06,430 More importantly, we requirethat they honor 1733 01:29:06,430 --> 01:29:09,370 what's called a third-partyopt-out. 1734 01:29:09,370 --> 01:29:11,130 You will click oncein your browser, 1735 01:29:11,130 --> 01:29:13,830 "Don't sell my information," 1736 01:29:13,830 --> 01:29:18,230 and it will then send the signalto every single website 1737 01:29:18,230 --> 01:29:21,400 that you visit: "Don't sellthis person's information." 1738 01:29:21,400 --> 01:29:22,970 And that's gonna have a hugeimpact on the spread 1739 01:29:22,970 --> 01:29:25,530 of your informationacross the internet. 1740 01:29:25,530 --> 01:29:27,900 NARRATOR: The tech companieshad been publicly cautious, 1741 01:29:27,900 --> 01:29:31,500 but privately alarmedabout regulation. 1742 01:29:31,500 --> 01:29:34,400 Then one tech giant came onboard in support 1743 01:29:34,400 --> 01:29:37,000 of Mactaggart's efforts. 1744 01:29:37,000 --> 01:29:39,670 I find the reaction amongother tech companies to, 1745 01:29:39,670 --> 01:29:42,730 at this point, be pretty muchall over the place. 1746 01:29:42,730 --> 01:29:45,970 Some people are saying,"You're right to raise this. 1747 01:29:45,970 --> 01:29:47,430 These are good ideas." 1748 01:29:47,430 --> 01:29:49,100 Some people say, "We're not surethese are good ideas, 1749 01:29:49,100 --> 01:29:50,870 but you're right to raise it," 1750 01:29:50,870 --> 01:29:54,300 and some people are saying,"We don't want regulation." 1751 01:29:54,300 --> 01:29:56,970 And so, you know, we haveconversations with people 1752 01:29:56,970 --> 01:30:00,030 where we point out that the autoindustry is better 1753 01:30:00,030 --> 01:30:03,130 because there aresafety standards. 1754 01:30:03,130 --> 01:30:05,430 Pharmaceuticals,even food products, 1755 01:30:05,430 --> 01:30:08,200 all of these industries arebetter because the public 1756 01:30:08,200 --> 01:30:11,000 has confidence in the products, 1757 01:30:11,000 --> 01:30:14,930 in part because of a mixtureof responsible companies 1758 01:30:14,930 --> 01:30:19,030 and responsible regulation. 1759 01:30:19,030 --> 01:30:21,400 NARRATOR: But the lobbyistsfor big tech have been working 1760 01:30:21,400 --> 01:30:24,270 the corridors in Washington. 1761 01:30:24,270 --> 01:30:26,300 They're looking fora more lenient 1762 01:30:26,300 --> 01:30:29,670 national privacy standard,one that could perhaps override 1763 01:30:29,670 --> 01:30:33,300 the California lawand others like it. 1764 01:30:33,300 --> 01:30:34,570 But while hearings are held, 1765 01:30:34,570 --> 01:30:37,170 and anti-trust legislationthreatened, 1766 01:30:37,170 --> 01:30:40,530 the problem is that A.I.has already spread so far 1767 01:30:40,530 --> 01:30:43,630 into our lives and work. 1768 01:30:43,630 --> 01:30:46,100 Well, it's in healthcare,it's in education, 1769 01:30:46,100 --> 01:30:48,670 it's in criminal justice,it's in the experience 1770 01:30:48,670 --> 01:30:51,230 of shopping as you walk downthe street. 1771 01:30:51,230 --> 01:30:54,700 It has pervaded so many elementsof everyday life, 1772 01:30:54,700 --> 01:30:57,370 and in a way that, in manycases, is completely opaque 1773 01:30:57,370 --> 01:30:59,230 to people. 1774 01:30:59,230 --> 01:31:00,830 While we can see a phone andlook at it and we know that 1775 01:31:00,830 --> 01:31:02,970 there's some A.I. technologybehind it, 1776 01:31:02,970 --> 01:31:05,200 many of us don't know that whenwe go for a job interview 1777 01:31:05,200 --> 01:31:07,030 and we sit downand we have a conversation, 1778 01:31:07,030 --> 01:31:09,970 that we're being filmed, andthat our micro expressions 1779 01:31:09,970 --> 01:31:12,670 are being analyzedby hiring companies. 1780 01:31:12,670 --> 01:31:14,700 Or that if you're in thecriminal justice system, 1781 01:31:14,700 --> 01:31:16,670 that there are risk assessmentalgorithms 1782 01:31:16,670 --> 01:31:18,830 that are decidingyour "risk number," 1783 01:31:18,830 --> 01:31:22,530 which could determine whetheror not you receive bail or not. 1784 01:31:22,530 --> 01:31:24,770 These are systems which, in manycases, are hidden 1785 01:31:24,770 --> 01:31:28,400 in the back end of our sortof social institutions. 1786 01:31:28,400 --> 01:31:29,400 And so, one of the bigchallenges we have is, 1787 01:31:29,400 --> 01:31:31,970 how do we make that moreapparent? 1788 01:31:31,970 --> 01:31:32,830 How do we make it transparent? 1789 01:31:32,830 --> 01:31:36,700 And how do we make itaccountable? 1790 01:31:36,700 --> 01:31:39,830 For a very long time,we have felt like as humans, 1791 01:31:39,830 --> 01:31:43,130 as Americans,we have full agency 1792 01:31:43,130 --> 01:31:48,830 in determining our own futures--what we read, what we see, 1793 01:31:48,830 --> 01:31:50,330 we're in charge. 1794 01:31:50,330 --> 01:31:53,070 What Cambridge Analytica taughtus, 1795 01:31:53,070 --> 01:31:55,770 and what Facebook continuesto teach us, 1796 01:31:55,770 --> 01:31:58,830 is that we don't have agency. 1797 01:31:58,830 --> 01:32:00,470 We're not in charge. 1798 01:32:00,470 --> 01:32:05,330 This is machines that areautomating some of our skills, 1799 01:32:05,330 --> 01:32:09,330 but have made decisions aboutwho... 1800 01:32:09,330 --> 01:32:12,630 Who we are. 1801 01:32:12,630 --> 01:32:16,300 And they're using thatinformation to tell others 1802 01:32:16,300 --> 01:32:19,570 the story of us. 1803 01:32:19,570 --> 01:32:22,130 ♪ ♪ 1804 01:32:32,970 --> 01:32:35,470 NARRATOR: In China,in the age of A.I., 1805 01:32:35,470 --> 01:32:38,130 there's no doubtabout who is in charge. 1806 01:32:38,130 --> 01:32:41,370 In an authoritarian state,social stability 1807 01:32:41,370 --> 01:32:43,770 is the watchwordof the government. 1808 01:32:43,770 --> 01:32:47,930 (whistle blowing) 1809 01:32:47,930 --> 01:32:51,070 And artificial intelligence hasincreased its ability to scan 1810 01:32:51,070 --> 01:32:54,430 the country for signs of unrest. 1811 01:32:54,430 --> 01:32:57,470 (whistle blowing) 1812 01:32:57,470 --> 01:33:00,430 It's been projected that over600 million cameras 1813 01:33:00,430 --> 01:33:04,700 will be deployed by 2020. 1814 01:33:04,700 --> 01:33:07,730 Here, they may be used todiscourage jaywalking. 1815 01:33:07,730 --> 01:33:10,400 But they also serve to remindpeople 1816 01:33:10,400 --> 01:33:14,830 that the state is watching. 1817 01:33:14,830 --> 01:33:18,030 And now, there is a projectcalled Sharp Eyes, 1818 01:33:18,030 --> 01:33:22,670 which is putting cameraon every major street 1819 01:33:22,670 --> 01:33:29,730 and the corner of every villagein China-- meaning everywhere. 1820 01:33:29,730 --> 01:33:33,530 Matching with the most advancedartificial intelligence 1821 01:33:33,530 --> 01:33:36,830 algorithm, which they canactually use this data, 1822 01:33:36,830 --> 01:33:39,470 real-time data, to pick upa face or pick up a action. 1823 01:33:39,470 --> 01:33:42,330 ♪ ♪ 1824 01:33:42,330 --> 01:33:44,370 NARRATOR: Frequent securityexpos feature companies 1825 01:33:44,370 --> 01:33:48,530 like Megvii and its facial-recognition technology. 1826 01:33:48,530 --> 01:33:51,970 They show off cameras with A.I.that can track cars, 1827 01:33:51,970 --> 01:33:54,800 and identify individualsby face, 1828 01:33:54,800 --> 01:33:58,270 or just by the way they walk. 1829 01:33:58,270 --> 01:34:02,130 The place is just filled withthese screens where you can see 1830 01:34:02,130 --> 01:34:04,530 the computers are actuallyreading people's faces 1831 01:34:04,530 --> 01:34:07,630 and trying to digest that data,and basically track 1832 01:34:07,630 --> 01:34:09,700 and identify who each person is. 1833 01:34:09,700 --> 01:34:11,600 And it's incredible to see somany, 1834 01:34:11,600 --> 01:34:12,870 because just twoor three years ago, 1835 01:34:12,870 --> 01:34:14,700 we hardly sawthat kind of thing. 1836 01:34:14,700 --> 01:34:16,700 So, a big part of it isgovernment spending. 1837 01:34:16,700 --> 01:34:18,370 And so the technology's reallytaken off, 1838 01:34:18,370 --> 01:34:21,530 and a lot of companies havestarted to sort of glom onto 1839 01:34:21,530 --> 01:34:25,700 this idea that thisis the future. 1840 01:34:25,700 --> 01:34:29,470 China is on its wayto building 1841 01:34:29,470 --> 01:34:32,330 a total surveillance state. 1842 01:34:32,330 --> 01:34:33,830 NARRATOR: And this is thetest lab 1843 01:34:33,830 --> 01:34:36,370 for the surveillance state. 1844 01:34:36,370 --> 01:34:40,170 Here, in the far northwest ofChina, 1845 01:34:40,170 --> 01:34:41,830 is the autonomous regionof Xinjiang. 1846 01:34:41,830 --> 01:34:45,170 Of the 25 million peoplewho live here, 1847 01:34:45,170 --> 01:34:48,170 almost half are a Muslim Turkicspeaking people 1848 01:34:48,170 --> 01:34:52,400 called the Uighurs. 1849 01:34:52,400 --> 01:34:53,900 (people shouting) 1850 01:34:53,900 --> 01:34:57,670 In 2009, tensions with localHan Chinese led to protests 1851 01:34:57,670 --> 01:35:01,300 and then riots in the capital,Urumqi. 1852 01:35:01,300 --> 01:35:04,200 (people shouting, guns firing) 1853 01:35:04,200 --> 01:35:05,670 (people shouting) 1854 01:35:08,200 --> 01:35:11,200 As the conflict has grown,the authorities have brought in 1855 01:35:11,200 --> 01:35:13,670 more police,and deployed extensive 1856 01:35:13,670 --> 01:35:17,530 surveillance technology. 1857 01:35:17,530 --> 01:35:20,700 That data feeds an A.I. systemthat the government claims 1858 01:35:20,700 --> 01:35:24,300 can predict individuals proneto "terrorism" 1859 01:35:24,300 --> 01:35:27,570 and detect those in need of"re-education" 1860 01:35:27,570 --> 01:35:30,730 in scores of recentlybuilt camps. 1861 01:35:30,730 --> 01:35:35,700 It is a campaign that hasalarmed human rights groups. 1862 01:35:35,700 --> 01:35:39,100 Chinese authorities are,without any legal basis, 1863 01:35:39,100 --> 01:35:42,770 arbitrarily detaining upto a million Turkic Muslims 1864 01:35:42,770 --> 01:35:44,800 simply on the basisof their identity. 1865 01:35:44,800 --> 01:35:49,230 But even outside the facilitiesin which these people 1866 01:35:49,230 --> 01:35:51,300 are being held, most of thepopulation there 1867 01:35:51,300 --> 01:35:53,470 is being subjected toextraordinary levels 1868 01:35:53,470 --> 01:35:58,470 of high-tech surveillance suchthat almost no aspect of life 1869 01:35:58,470 --> 01:36:01,100 anymore, you know, takes placeoutside 1870 01:36:01,100 --> 01:36:02,600 the state's line of sight. 1871 01:36:02,600 --> 01:36:06,230 And so the kinds of behaviorthat's now being monitored-- 1872 01:36:06,230 --> 01:36:07,770 you know, which language do youspeak at home, 1873 01:36:07,770 --> 01:36:09,530 whether you're talking to yourrelatives 1874 01:36:09,530 --> 01:36:13,330 in other countries,how often you pray-- 1875 01:36:13,330 --> 01:36:16,230 that information is now beinghoovered up 1876 01:36:16,230 --> 01:36:19,230 and used to decide whetherpeople should be subjected 1877 01:36:19,230 --> 01:36:21,800 to political re-educationin these camps. 1878 01:36:21,800 --> 01:36:24,570 NARRATOR: There have beenreports of torture 1879 01:36:24,570 --> 01:36:27,000 and deaths in the camps. 1880 01:36:27,000 --> 01:36:28,800 And for Uighurs on the outside, 1881 01:36:28,800 --> 01:36:31,670 Xinjiang has already beendescribed 1882 01:36:31,670 --> 01:36:34,770 as an "open-air prison." 1883 01:36:34,770 --> 01:36:36,930 Trying to have a normal lifeas a Uighur 1884 01:36:36,930 --> 01:36:40,430 is impossible both insideand outside of China. 1885 01:36:40,430 --> 01:36:43,530 Just imagine, while you're onyour way to work, 1886 01:36:43,530 --> 01:36:47,600 police subject you to scanyour I.D., 1887 01:36:47,600 --> 01:36:51,770 forcing you to lift your chin,while machines take your photo 1888 01:36:51,770 --> 01:36:54,970 and wait... you wait until youfind out if you can go. 1889 01:36:54,970 --> 01:36:59,100 Imagine police take your phoneand run data scan, 1890 01:36:59,100 --> 01:37:02,500 and force you to installcompulsory software 1891 01:37:02,500 --> 01:37:07,630 allowing your phone calls andmessages to be monitored. 1892 01:37:07,630 --> 01:37:09,730 NARRATOR: Nury Turkel, alawyer and a prominent 1893 01:37:09,730 --> 01:37:14,470 Uighur activist, addresses ademonstration in Washington, DC. 1894 01:37:14,470 --> 01:37:18,700 Many among the Uighur diasporahave lost all contact 1895 01:37:18,700 --> 01:37:21,030 with their families back home. 1896 01:37:21,030 --> 01:37:26,100 Turkel warns that this dystopiandeployment of new technology 1897 01:37:26,100 --> 01:37:29,330 is a demonstration projectfor authoritarian regimes 1898 01:37:29,330 --> 01:37:31,430 around the world. 1899 01:37:31,430 --> 01:37:35,430 They have a bar codes insomebody's home doors 1900 01:37:35,430 --> 01:37:39,730 to identify what kind of citizenthat he is. 1901 01:37:39,730 --> 01:37:42,670 What we're talking about is acollective punishment 1902 01:37:42,670 --> 01:37:45,100 of an ethnic group. 1903 01:37:45,100 --> 01:37:48,200 Not only that, the Chinesegovernment has been promoting 1904 01:37:48,200 --> 01:37:53,100 its methods, its technology,it is... 1905 01:37:53,100 --> 01:37:58,630 to other countries, namelyPakistan, Venezuela, Sudan, 1906 01:37:58,630 --> 01:38:04,400 and others to utilize, tosquelch political resentment 1907 01:38:04,400 --> 01:38:07,970 or prevent a political upheavalin their various societies. 1908 01:38:07,970 --> 01:38:10,430 ♪ ♪ 1909 01:38:10,430 --> 01:38:13,500 NARRATOR: China has a grandscheme to spread its technology 1910 01:38:13,500 --> 01:38:15,470 and influence around the world. 1911 01:38:15,470 --> 01:38:19,770 Launched in 2013, it startedalong the old Silk Road 1912 01:38:19,770 --> 01:38:23,270 out of Xinjiang,and now goes far beyond. 1913 01:38:23,270 --> 01:38:29,570 It's called "the Belt and RoadInitiative." 1914 01:38:29,570 --> 01:38:31,370 So effectivelywhat the Belt and Road 1915 01:38:31,370 --> 01:38:35,630 is is China's attempt to,via spending and investment, 1916 01:38:35,630 --> 01:38:37,700 project its influenceall over the world. 1917 01:38:37,700 --> 01:38:39,830 And we've seen, you know,massive infrastructure projects 1918 01:38:39,830 --> 01:38:43,300 going in in places likePakistan, in, in Venezuela, 1919 01:38:43,300 --> 01:38:45,330 in Ecuador, in Bolivia-- 1920 01:38:45,330 --> 01:38:47,400 you know, all over the world,Argentina, 1921 01:38:47,400 --> 01:38:49,630 in America's backyard,in Africa. 1922 01:38:49,630 --> 01:38:51,530 Africa's been a huge place. 1923 01:38:51,530 --> 01:38:54,230 And what the Belt and Roadultimately does is, it attempts 1924 01:38:54,230 --> 01:38:56,400 to kind of create a politicalleverage 1925 01:38:56,400 --> 01:39:00,070 for the Chinese spendingcampaign all over the globe. 1926 01:39:00,070 --> 01:39:03,600 NARRATOR: Like Xi Jinping's2018 visit to Senegal, 1927 01:39:03,600 --> 01:39:06,700 where Chinese contractors hadjust built a new stadium, 1928 01:39:06,700 --> 01:39:10,970 arranged loans for a newinfrastructure development, 1929 01:39:10,970 --> 01:39:13,270 and, said the Foreign Ministry, 1930 01:39:13,270 --> 01:39:16,370 there would be help"maintaining social stability." 1931 01:39:16,370 --> 01:39:19,470 As China comes into thesecountries and provides 1932 01:39:19,470 --> 01:39:21,770 these loans, what you end upwith is Chinese technology 1933 01:39:21,770 --> 01:39:24,430 being sold and built out by,you know, by Chinese companies 1934 01:39:24,430 --> 01:39:26,370 in these countries. 1935 01:39:26,370 --> 01:39:27,500 We've started to see it alreadyin terms 1936 01:39:27,500 --> 01:39:29,170 of surveillance systems. 1937 01:39:29,170 --> 01:39:31,170 Not the kind of high-level A.I.stuff yet, but, you know, 1938 01:39:31,170 --> 01:39:32,600 lower-level, camera-based,you know, 1939 01:39:32,600 --> 01:39:36,670 manual sort of observation-typethings all over. 1940 01:39:36,670 --> 01:39:38,200 You know, you see it inCambodia, you see it in Ecuador, 1941 01:39:38,200 --> 01:39:39,770 you see it in Venezuela. 1942 01:39:39,770 --> 01:39:42,570 And what they do is, they sella dam, sell some other stuff, 1943 01:39:42,570 --> 01:39:44,100 and they say, "You know,by the way, we can give you 1944 01:39:44,100 --> 01:39:46,730 these camera systems and,for your emergency response. 1945 01:39:46,730 --> 01:39:49,000 And it'll cost you $300 million, 1946 01:39:49,000 --> 01:39:50,500 and we'll build a ton ofcameras, 1947 01:39:50,500 --> 01:39:52,800 and we'll build you a kind of,you know, a main center 1948 01:39:52,800 --> 01:39:55,100 where you have police who canwatch these cameras." 1949 01:39:55,100 --> 01:39:57,870 And that's going in all overthe world already. 1950 01:39:57,870 --> 01:40:03,570 ♪ ♪ 1951 01:40:03,570 --> 01:40:06,600 There are 58 countries thatare starting to plug in 1952 01:40:06,600 --> 01:40:10,230 to China's vision of artificialintelligence. 1953 01:40:10,230 --> 01:40:15,300 Which means effectively thatChina is in the process 1954 01:40:15,300 --> 01:40:17,770 of raising a bamboo curtain. 1955 01:40:17,770 --> 01:40:20,770 One that does not need to... 1956 01:40:20,770 --> 01:40:24,130 One that is sort ofall-encompassing, 1957 01:40:24,130 --> 01:40:26,700 that has shared resources, 1958 01:40:26,700 --> 01:40:28,600 shared telecommunicationssystems, 1959 01:40:28,600 --> 01:40:31,970 shared infrastructure,shared digital systems-- 1960 01:40:31,970 --> 01:40:35,400 even shared mobile-phonetechnologies-- 1961 01:40:35,400 --> 01:40:38,700 that is, that is quickly goingup all around the world 1962 01:40:38,700 --> 01:40:41,700 to the exclusion of usin the West. 1963 01:40:41,700 --> 01:40:43,170 Well, one of the thingsI worry about the most 1964 01:40:43,170 --> 01:40:45,130 is that the worldis gonna split in two, 1965 01:40:45,130 --> 01:40:47,230 and that there will bea Chinese tech sector 1966 01:40:47,230 --> 01:40:48,970 and there will be anAmerican tech sector. 1967 01:40:48,970 --> 01:40:51,830 And countries will effectivelyget to choose 1968 01:40:51,830 --> 01:40:53,170 which one they want. 1969 01:40:53,170 --> 01:40:55,770 It'll be kind of like the ColdWar, where you decide, 1970 01:40:55,770 --> 01:40:57,970 "Oh, are we gonna alignwith the Soviet Union 1971 01:40:57,970 --> 01:40:59,570 or are we gonna alignwith the United States?" 1972 01:40:59,570 --> 01:41:02,200 And the Third World gets tochoose this or that. 1973 01:41:02,200 --> 01:41:06,000 And that's not a world that'sgood for anybody. 1974 01:41:06,000 --> 01:41:09,130 The markets in Asia and theU.S. falling sharply 1975 01:41:09,130 --> 01:41:11,470 on news that a top Chineseexecutive 1976 01:41:11,470 --> 01:41:13,100 has been arrested in Canada. 1977 01:41:13,100 --> 01:41:14,300 Her name is Sabrina Meng. 1978 01:41:14,300 --> 01:41:19,530 She is the CFO of the Chinesetelecom Huawei. 1979 01:41:19,530 --> 01:41:21,270 NARRATOR: News of thedramatic arrest of an important 1980 01:41:21,270 --> 01:41:24,530 Huawei executive was ostensiblyabout the company 1981 01:41:24,530 --> 01:41:26,430 doing business with Iran. 1982 01:41:26,430 --> 01:41:29,600 But it seemed to be more aboutAmerican distrust 1983 01:41:29,600 --> 01:41:32,600 of the company's technology. 1984 01:41:32,600 --> 01:41:33,900 From its headquartersin southern China-- 1985 01:41:33,900 --> 01:41:38,930 designed to look like fancifulEuropean capitals-- 1986 01:41:38,930 --> 01:41:41,570 Huawei is the second-biggestseller of smartphones, 1987 01:41:41,570 --> 01:41:45,630 and the world leaderin building 5G networks, 1988 01:41:45,630 --> 01:41:50,970 the high-speed backbonefor the age of A.I. 1989 01:41:50,970 --> 01:41:53,070 Huawei's C.E.O.,a former officer 1990 01:41:53,070 --> 01:41:54,970 in the People's Liberation Army, 1991 01:41:54,970 --> 01:41:57,830 was defiant aboutthe American actions. 1992 01:41:57,830 --> 01:41:59,470 (speaking Mandarin) 1993 01:41:59,470 --> 01:42:02,600 (translated): There's no waythe U.S. can crush us. 1994 01:42:02,600 --> 01:42:08,500 The world needs Huawei becausewe are more advanced. 1995 01:42:08,500 --> 01:42:12,900 If the lights go out in theWest, the East will still shine. 1996 01:42:12,900 --> 01:42:16,270 And if the North goes dark,then there is still the South. 1997 01:42:16,270 --> 01:42:19,730 America doesn't representthe world. 1998 01:42:19,730 --> 01:42:22,270 NARRATOR: The U.S. governmentfears that as Huawei supplies 1999 01:42:22,270 --> 01:42:26,400 countries around the worldwith 5G, 2000 01:42:26,400 --> 01:42:28,670 the Chinese government couldhave back-door access 2001 01:42:28,670 --> 01:42:30,700 to their equipment. 2002 01:42:30,700 --> 01:42:34,400 Recently, the C.E.O. promisedcomplete transparency 2003 01:42:34,400 --> 01:42:36,700 into the company's software, 2004 01:42:36,700 --> 01:42:39,470 but U.S. authoritiesare not convinced. 2005 01:42:39,470 --> 01:42:44,530 Nothing in China exists freeand clear of the party-state. 2006 01:42:44,530 --> 01:42:48,730 Those companies can only existand prosper 2007 01:42:48,730 --> 01:42:51,030 at the sufferance of the party. 2008 01:42:51,030 --> 01:42:55,030 And it's made very explicit thatwhen the party needs them, 2009 01:42:55,030 --> 01:42:58,900 they either have to respondor they will be dethroned. 2010 01:42:58,900 --> 01:43:03,770 So this is the challenge with acompany like Huawei. 2011 01:43:03,770 --> 01:43:08,900 So Huawei, Ren Zhengfei, thehead of Huawei, he can say, 2012 01:43:08,900 --> 01:43:12,000 "Well, we... we're just aprivate company and we just... 2013 01:43:12,000 --> 01:43:15,470 We don't take ordersfrom the Communist Party." 2014 01:43:15,470 --> 01:43:18,370 Well, maybe they haven't yet. 2015 01:43:18,370 --> 01:43:20,870 But what the Pentagon sees, 2016 01:43:20,870 --> 01:43:23,100 the National IntelligenceCouncil sees, 2017 01:43:23,100 --> 01:43:27,070 and what the FBI sees is,"Well, maybe not yet." 2018 01:43:27,070 --> 01:43:30,200 But when the call comes, 2019 01:43:30,200 --> 01:43:35,430 everybody knows what thecompany's response will be. 2020 01:43:35,430 --> 01:43:37,000 NARRATOR: The U.S. CommerceDepartment 2021 01:43:37,000 --> 01:43:39,400 has recently blacklistedeight companies 2022 01:43:39,400 --> 01:43:42,870 for doing business withgovernment agencies in Xinjiang, 2023 01:43:42,870 --> 01:43:45,370 claiming they are aidingin the "repression" 2024 01:43:45,370 --> 01:43:49,300 of the Muslim minority. 2025 01:43:49,300 --> 01:43:52,270 Among the companies is Megvii. 2026 01:43:52,270 --> 01:43:55,170 They have strongly objectedto the blacklist, 2027 01:43:55,170 --> 01:43:57,630 saying that it's "amisunderstanding of our company 2028 01:43:57,630 --> 01:44:01,500 and our technology." 2029 01:44:01,500 --> 01:44:04,430 ♪ ♪ 2030 01:44:04,430 --> 01:44:07,370 President Xi has increased hisauthoritarian grip 2031 01:44:07,370 --> 01:44:11,070 on the country. 2032 01:44:11,070 --> 01:44:14,530 In 2018, he had the Chineseconstitution changed 2033 01:44:14,530 --> 01:44:20,070 so that he could be presidentfor life. 2034 01:44:20,070 --> 01:44:21,370 If you had asked me20 years ago, 2035 01:44:21,370 --> 01:44:23,230 "What will happen to China?",I would've said, 2036 01:44:23,230 --> 01:44:27,170 "Well, over time, the GreatFirewall will break down. 2037 01:44:27,170 --> 01:44:29,770 Of course, people will getaccess to social media, 2038 01:44:29,770 --> 01:44:31,800 they'll get access to Google... 2039 01:44:31,800 --> 01:44:35,700 Eventually, it'll become a muchmore democratic place, 2040 01:44:35,700 --> 01:44:38,370 with free expressionand lots of Western values." 2041 01:44:38,370 --> 01:44:41,870 And the last time I checked,that has not happened. 2042 01:44:41,870 --> 01:44:46,600 In fact, technology's becomea tool of control. 2043 01:44:46,600 --> 01:44:48,570 And as China has gone throughthis amazing period of growth 2044 01:44:48,570 --> 01:44:51,870 and wealth and openness incertain ways, 2045 01:44:51,870 --> 01:44:53,430 there has not been thedemocratic transformation 2046 01:44:53,430 --> 01:44:55,330 that I thought. 2047 01:44:55,330 --> 01:44:57,570 And it may turn out that,in fact, 2048 01:44:57,570 --> 01:45:00,600 technology is a better tool forauthoritarian governments 2049 01:45:00,600 --> 01:45:02,570 than it is for democraticgovernments. 2050 01:45:02,570 --> 01:45:04,900 NARRATOR: To dominatethe world in A.I., 2051 01:45:04,900 --> 01:45:08,000 President Xi is depending onChinese tech 2052 01:45:08,000 --> 01:45:11,330 to lead the way. 2053 01:45:11,330 --> 01:45:13,030 While companies likeBaidu, Alibaba, 2054 01:45:13,030 --> 01:45:17,870 and Tencent are growing morepowerful and competitive, 2055 01:45:17,870 --> 01:45:20,500 they're also beginning to havedifficulty accessing 2056 01:45:20,500 --> 01:45:24,930 American technology, and areracing to develop their own. 2057 01:45:27,600 --> 01:45:31,200 With a continuing trade warand growing distrust, 2058 01:45:31,200 --> 01:45:33,630 the longtime argument forengagement 2059 01:45:33,630 --> 01:45:38,230 between the two countrieshas been losing ground. 2060 01:45:38,230 --> 01:45:42,100 I've seen more and moreof my colleagues move 2061 01:45:42,100 --> 01:45:44,270 from a position when theythought, 2062 01:45:44,270 --> 01:45:47,330 "Well, if we just keep engagingChina, 2063 01:45:47,330 --> 01:45:50,800 the lines betweenthe two countries 2064 01:45:50,800 --> 01:45:52,800 will slowly converge." 2065 01:45:52,800 --> 01:45:56,570 You know, whether it's ineconomics, technology, politics. 2066 01:45:56,570 --> 01:45:58,370 And the transformation, 2067 01:45:58,370 --> 01:46:01,230 where they now thinkthey're diverging. 2068 01:46:01,230 --> 01:46:05,000 So, in other words, the wholeidea of engagement 2069 01:46:05,000 --> 01:46:07,430 is coming under question. 2070 01:46:07,430 --> 01:46:15,600 And that's cast an entirelydifferent light on technology, 2071 01:46:15,600 --> 01:46:18,800 because if you're diverging andyou're heading into a world 2072 01:46:18,800 --> 01:46:23,900 of antagonism-- you know,conflict, possibly, 2073 01:46:23,900 --> 01:46:25,930 then suddenly, technology issomething 2074 01:46:25,930 --> 01:46:27,930 that you don't want to share. 2075 01:46:27,930 --> 01:46:30,870 You want to sequester, 2076 01:46:30,870 --> 01:46:34,300 to protect your own nationalinterest. 2077 01:46:34,300 --> 01:46:38,130 And I think the tipping-pointmoment we are at now, 2078 01:46:38,130 --> 01:46:41,130 which is what is castingthe whole question of things 2079 01:46:41,130 --> 01:46:45,130 like artificial intelligenceand technological innovation 2080 01:46:45,130 --> 01:46:47,470 into a completely differentframework, 2081 01:46:47,470 --> 01:46:51,700 is that if in fact Chinaand the U.S. are in some way 2082 01:46:51,700 --> 01:46:54,730 fundamentally antagonisticto each other, 2083 01:46:54,730 --> 01:46:59,900 then we're in a completelydifferent world. 2084 01:46:59,900 --> 01:47:05,670 NARRATOR: In the age of A.I.,a new reality is emerging. 2085 01:47:05,670 --> 01:47:07,600 That with so much accumulatedinvestment 2086 01:47:07,600 --> 01:47:11,800 and intellectual power, theworld is already dominated 2087 01:47:11,800 --> 01:47:16,200 by just two A.I. superpowers. 2088 01:47:16,200 --> 01:47:22,130 That's the premise of a new bookwritten by Kai-Fu Lee. 2089 01:47:22,130 --> 01:47:23,500 Hi, I'm Kai-Fu. 2090 01:47:23,500 --> 01:47:25,600 Hi, Dr. Lee, sonice to meet you. 2091 01:47:25,600 --> 01:47:26,670 Really nice to meet you. 2092 01:47:26,670 --> 01:47:28,230 Look at all these dog ears. 2093 01:47:28,230 --> 01:47:30,000 I love, I love that.You like that? 2094 01:47:30,000 --> 01:47:31,970 But I... but I don't like youdidn't buy the book, 2095 01:47:31,970 --> 01:47:33,470 you... you borrowed it. 2096 01:47:33,470 --> 01:47:35,570 I couldn't find it!Oh, really? 2097 01:47:35,570 --> 01:47:36,900 Yeah!And, and you... 2098 01:47:36,900 --> 01:47:39,130 you're coming to my talk?Of course! 2099 01:47:39,130 --> 01:47:40,500 Oh, hi.I did my homework, 2100 01:47:40,500 --> 01:47:41,600 I'm telling you. 2101 01:47:41,600 --> 01:47:42,600 Oh, my goodness, thank you. 2102 01:47:42,600 --> 01:47:44,670 Laurie, can you get thisgentleman a book? 2103 01:47:44,670 --> 01:47:46,430 (people talking in background) 2104 01:47:46,430 --> 01:47:47,730 NARRATOR: In his bookand in life, 2105 01:47:47,730 --> 01:47:50,830 the computerscientist-cum-venture capitalist 2106 01:47:50,830 --> 01:47:52,230 walks a careful path. 2107 01:47:52,230 --> 01:47:56,370 Criticism of the Chinesegovernment is avoided, 2108 01:47:56,370 --> 01:47:58,700 while capitalist successis celebrated. 2109 01:47:58,700 --> 01:48:00,800 I'm studying electricalengineering. 2110 01:48:00,800 --> 01:48:03,230 Sure, send me a resume.Okay, thanks. 2111 01:48:03,230 --> 01:48:06,230 NARRATOR: Now, with the riseof the two superpowers, 2112 01:48:06,230 --> 01:48:09,400 he wants to warn the worldof what's coming. 2113 01:48:09,400 --> 01:48:11,600 Are you the new leaders? 2114 01:48:11,600 --> 01:48:14,230 If we're not the new leaders,we're pretty close. 2115 01:48:14,230 --> 01:48:15,800 (laughs) 2116 01:48:15,800 --> 01:48:18,030 Thank you very much.Thanks. 2117 01:48:18,030 --> 01:48:20,630 NARRATOR: "Never," he writes,"has the potential 2118 01:48:20,630 --> 01:48:22,800 for human flourishing beenhigher 2119 01:48:22,800 --> 01:48:26,230 or the stakes of failuregreater." 2120 01:48:26,230 --> 01:48:27,700 ♪ ♪ 2121 01:48:27,700 --> 01:48:32,070 So if one has to say who'sahead, I would say today, 2122 01:48:32,070 --> 01:48:34,670 China is quickly catching up. 2123 01:48:34,670 --> 01:48:38,970 China actually beganits big push 2124 01:48:38,970 --> 01:48:42,370 in A.I. only two-and-a-halfyears ago, 2125 01:48:42,370 --> 01:48:46,700 when the AlphaGo-Lee Sedol matchbecame the Sputnik moment. 2126 01:48:46,700 --> 01:48:49,870 NARRATOR: He says he believesthat the two A.I. superpowers 2127 01:48:49,870 --> 01:48:52,700 should lead the way and worktogether 2128 01:48:52,700 --> 01:48:55,270 to make A.I. a force for good. 2129 01:48:55,270 --> 01:48:58,400 If we do, we may have a chanceof getting it right. 2130 01:48:58,400 --> 01:49:00,730 If we do a very good jobin the next 20 years, 2131 01:49:00,730 --> 01:49:04,100 A.I. will be viewed as an age ofenlightenment. 2132 01:49:04,100 --> 01:49:08,370 Our children and their childrenwill see A.I. as serendipity. 2133 01:49:08,370 --> 01:49:13,800 That A.I. is here to liberate usfrom having to do routine jobs, 2134 01:49:13,800 --> 01:49:15,830 and push us to do what we love, 2135 01:49:15,830 --> 01:49:19,530 and push us to think what itmeans to be human. 2136 01:49:19,530 --> 01:49:23,600 NARRATOR: But what if humansmishandle this new power? 2137 01:49:23,600 --> 01:49:25,930 Kai-Fu Lee understandsthe stakes. 2138 01:49:25,930 --> 01:49:28,270 After all, he invested earlyin Megvii, 2139 01:49:28,270 --> 01:49:33,030 which is now on the U.S.blacklist. 2140 01:49:33,030 --> 01:49:35,630 He says he's reduced his stakeand doesn't speak 2141 01:49:35,630 --> 01:49:38,070 for the company. 2142 01:49:38,070 --> 01:49:40,370 Asked about the governmentusing A.I. 2143 01:49:40,370 --> 01:49:44,570 for social control,he chose his words carefully. 2144 01:49:44,570 --> 01:49:49,630 Um... A.I. is a technologythat can be used 2145 01:49:49,630 --> 01:49:52,230 for good and for evil. 2146 01:49:52,230 --> 01:50:00,700 So how... how do governmentslimit themselves in, 2147 01:50:00,700 --> 01:50:04,670 on the one hand,using this A.I. technology 2148 01:50:04,670 --> 01:50:07,970 and the database to maintaina safe environment 2149 01:50:07,970 --> 01:50:11,400 for its citizens, but,but not encroach 2150 01:50:11,400 --> 01:50:14,370 on a individual's rightsand privacies? 2151 01:50:14,370 --> 01:50:17,530 That, I think, is also a trickyissue, I think, 2152 01:50:17,530 --> 01:50:19,200 for, for every country. 2153 01:50:19,200 --> 01:50:22,170 I think for... I think everycountry will be tempted 2154 01:50:22,170 --> 01:50:26,030 to use A.I. probablybeyond the limits 2155 01:50:26,030 --> 01:50:29,930 to which that you and I wouldlike the government to use. 2156 01:50:35,370 --> 01:50:40,970 ♪ ♪ 2157 01:50:40,970 --> 01:50:43,030 NARRATOR: Emperor Yao devisedthe game of Go 2158 01:50:43,030 --> 01:50:48,770 to teach his son discipline,concentration, and balance. 2159 01:50:48,770 --> 01:50:52,630 Over 4,000 years later,in the age of A.I., 2160 01:50:52,630 --> 01:50:56,230 those words still resonate withone of its architects. 2161 01:50:56,230 --> 01:50:58,330 ♪ ♪ 2162 01:50:58,330 --> 01:51:02,270 So A.I. can be used in manyways that are very beneficial 2163 01:51:02,270 --> 01:51:03,700 for society. 2164 01:51:03,700 --> 01:51:08,230 But the current use of A.I.isn't necessarily aligned 2165 01:51:08,230 --> 01:51:11,630 with the goals of buildinga better society, 2166 01:51:11,630 --> 01:51:12,900 unfortunately. 2167 01:51:12,900 --> 01:51:16,570 But, but we could change that. 2168 01:51:16,570 --> 01:51:19,570 NARRATOR: In 2016, a game ofGo gave us a glimpse 2169 01:51:19,570 --> 01:51:24,970 of the future of artificialintelligence. 2170 01:51:24,970 --> 01:51:27,500 Since then, it has become clearthat we will need 2171 01:51:27,500 --> 01:51:32,900 a careful strategy to harnessthis new and awesome power. 2172 01:51:35,800 --> 01:51:38,600 I, I do think that democracyis threatened by the progress 2173 01:51:38,600 --> 01:51:42,300 of these tools unless we improveour social norms 2174 01:51:42,300 --> 01:51:46,430 and we increasethe collective wisdom 2175 01:51:46,430 --> 01:51:51,830 at the planet level to, to dealwith that increased power. 2176 01:51:51,830 --> 01:51:57,600 I'm hoping that my concerns arenot founded, 2177 01:51:57,600 --> 01:51:59,900 but the stakes are so high 2178 01:51:59,900 --> 01:52:06,600 that I don't think we shouldtake these concerns lightly. 2179 01:52:06,600 --> 01:52:11,930 I don't think we can play withthose possibilities and just... 2180 01:52:11,930 --> 01:52:17,030 race ahead without thinkingabout the potential outcomes. 2181 01:52:17,030 --> 01:52:20,870 ♪ ♪ 2182 01:52:27,330 --> 01:52:31,100 Go to pbs.org/frontline formore of the impact 2183 01:52:31,100 --> 01:52:32,870 of A.I. on jobs. 2184 01:52:32,870 --> 01:52:37,730 I believe about fifty percentof jobs will be somewhat 2185 01:52:37,730 --> 01:52:41,230 or extremely threatened by A.I.in the next 15 years or so. 2186 01:52:41,230 --> 01:52:43,530 And a look at the potentialfor racial bias 2187 01:52:43,530 --> 01:52:45,200 in this technology. 2188 01:52:45,200 --> 01:52:47,000 We've had issues with bias,with discrimination, 2189 01:52:47,000 --> 01:52:48,670 with poor system design,with errors. 2190 01:52:48,670 --> 01:52:51,500 Connect to the "Frontline"community on Facebook 2191 01:52:51,500 --> 01:52:54,570 and Twitter, and watch anytimeon the PBS Video app 2192 01:52:54,570 --> 01:52:56,500 or pbs.org/frontline. 2193 01:52:58,130 --> 01:53:02,070 ♪ ♪ 2194 01:53:25,000 --> 01:53:26,800 For more on this andother "Frontline" programs, 2195 01:53:26,800 --> 01:53:30,100 visit our websiteat pbs.org/frontline. 2196 01:53:34,930 --> 01:53:37,400 ♪ ♪ 2197 01:53:40,300 --> 01:53:43,530 To order "Frontline's""In the Age of A.I." on DVD, 2198 01:53:43,530 --> 01:53:48,830 visit ShopPBS or call1-800-PLAY-PBS. 2199 01:53:48,830 --> 01:53:52,570 This program is also availableon Amazon Prime Video. 2200 01:53:57,870 --> 01:54:01,170 ♪ ♪ 177822

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.