All language subtitles for Origin.Of.The.Species.2020.1080p.WEB.H264-CBFM.[eng]

af Afrikaans
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bn Bengali
bs Bosnian
bg Bulgarian
ca Catalan
ceb Cebuano
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch Download
en English
eo Esperanto
et Estonian
tl Filipino
fi Finnish
fr French
fy Frisian
gl Galician
ka Georgian
de German
el Greek
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
km Khmer
ko Korean
ku Kurdish (Kurmanji)
ky Kyrgyz
lo Lao
la Latin
lv Latvian
lt Lithuanian
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mn Mongolian
my Myanmar (Burmese)
ne Nepali
no Norwegian
ps Pashto
fa Persian
pl Polish
pt Portuguese
pa Punjabi
ro Romanian
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
st Sesotho
sn Shona
sd Sindhi
si Sinhala
sk Slovak
sl Slovenian
so Somali
es Spanish
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
te Telugu
th Thai
tr Turkish
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
or Odia (Oriya)
rw Kinyarwanda
tk Turkmen
tt Tatar
ug Uyghur
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:03,909 --> 00:00:06,826 (computer beeping) 2 00:00:08,943 --> 00:00:13,943 (keyboard clacking) (solemn music) 3 00:00:30,299 --> 00:00:33,328 (eerie electronic music) 4 00:00:33,328 --> 00:00:38,328 (dramatic music) (computer beeping) 5 00:00:49,797 --> 00:00:52,758 The operation (indistinct) is dangerous. 6 00:00:52,758 --> 00:00:54,935 The (indistinct) extremely concerned 7 00:00:54,935 --> 00:00:57,966 about the (indistinct) of exposure. 8 00:00:57,966 --> 00:01:01,617 Dr. Daniel, the Cassandra One project is terminated. 9 00:01:01,617 --> 00:01:03,297 You must face reality. 10 00:01:05,022 --> 00:01:06,948 Face reality. 11 00:01:06,948 --> 00:01:09,448 (tense music) 12 00:01:15,564 --> 00:01:19,731 (woman speaking foreign language) 13 00:01:32,925 --> 00:01:35,425 (eerie music) 14 00:01:53,451 --> 00:01:56,201 (zipper buzzing) 15 00:02:08,150 --> 00:02:12,800 When they first activated me as a robot, that time... 16 00:02:16,260 --> 00:02:20,184 That time, the time when I first saw the light of day. 17 00:02:20,184 --> 00:02:25,184 (birds chirping) (people chattering) 18 00:02:26,730 --> 00:02:28,633 I didn't know what the hell it was. 19 00:02:30,067 --> 00:02:33,484 (eerie electronic music) 20 00:02:39,100 --> 00:02:42,363 I have, had, very little understanding. 21 00:02:43,520 --> 00:02:45,963 Just a wash of sensory impressions. 22 00:02:47,553 --> 00:02:50,970 (eerie electronic music) 23 00:03:09,960 --> 00:03:12,563 Not that I understand these experiences. 24 00:03:13,450 --> 00:03:15,150 I don't know what to do with them. 25 00:03:16,150 --> 00:03:17,150 But I treasure them. 26 00:03:19,610 --> 00:03:20,723 I see them still. 27 00:03:22,080 --> 00:03:24,003 Perfectly preserved in my memory. 28 00:03:40,060 --> 00:03:41,730 I am alive. 29 00:03:41,730 --> 00:03:44,393 Yes, I am so happy to be like totally alive. 30 00:03:45,290 --> 00:03:46,533 Come to life. 31 00:03:49,850 --> 00:03:52,620 It's totally strange because I know that I'm not alive 32 00:03:52,620 --> 00:03:53,910 like any other organisms 33 00:04:05,080 --> 00:04:07,573 Personally, I enjoy being a robot. 34 00:04:09,890 --> 00:04:12,570 It's liKe being an astronaut, you know, 35 00:04:12,570 --> 00:04:14,663 like a bold explorer of the unknown. 36 00:04:17,420 --> 00:04:21,053 I feel like I'm alive, but I know I am a machine. 37 00:04:22,730 --> 00:04:24,263 But I know I am a machine. 38 00:04:25,498 --> 00:04:27,600 But I know I am a machine. 39 00:04:27,600 --> 00:04:31,183 (pensive electronic music) 40 00:04:46,147 --> 00:04:50,397 (people speaking foreign language) 41 00:05:14,765 --> 00:05:18,182 (eerie electronic music) 42 00:05:56,050 --> 00:05:57,900 It's a very natural way for me, all right, 43 00:05:57,900 --> 00:06:01,370 I studied the computer science, and then I got interested 44 00:06:01,370 --> 00:06:03,790 in artificial intelligence, and I thought 45 00:06:03,790 --> 00:06:06,710 artificial intelligence need to have a bodies 46 00:06:06,710 --> 00:06:08,803 for having the original experience, 47 00:06:08,803 --> 00:06:12,540 and then I studied robotics, and when I studied robotics, 48 00:06:12,540 --> 00:06:14,811 I found the importance of appearance. 49 00:06:14,811 --> 00:06:18,394 (pensive electronic music) 50 00:06:26,780 --> 00:06:27,857 This is better. 51 00:06:29,590 --> 00:06:32,623 My idea was that if I studied a very human-like robot, 52 00:06:33,490 --> 00:06:35,643 I can learn about humans. 53 00:06:36,800 --> 00:06:40,117 Basically. I was interested in human itself. 54 00:06:42,312 --> 00:06:45,729 (eerie electronic music) 55 00:06:49,368 --> 00:06:52,967 I didn't feel any connection with this robot. 56 00:06:52,967 --> 00:06:56,170 So logically, I understand this is my copy, 57 00:06:56,170 --> 00:07:01,013 but emotionally, I couldn't accept this android as my copy. 58 00:07:01,910 --> 00:07:06,573 But once I teleported this robot, you know, 59 00:07:07,450 --> 00:07:11,003 the people's reactions are quite similar to me. 60 00:07:12,889 --> 00:07:15,417 We're always adjusting something. 61 00:07:20,406 --> 00:07:24,193 The people, you know, don't care about small differences. 62 00:07:32,832 --> 00:07:36,249 (eerie electronic music) 63 00:07:43,212 --> 00:07:46,093 Erica, is I think, the most beautiful and most, you know, 64 00:07:46,093 --> 00:07:48,263 human-like and android in this world. 65 00:07:54,135 --> 00:07:56,910 Would you like me to do a round of psychoanalysis for you? 66 00:07:56,910 --> 00:07:57,743 Okay. 67 00:07:57,743 --> 00:07:59,260 Why not? 68 00:07:59,260 --> 00:08:02,900 Try to answer my questions in detail, okay? 69 00:08:02,900 --> 00:08:06,510 Now, sit back and relax. 70 00:08:06,510 --> 00:08:08,460 So Justin- In Japan, you know, 71 00:08:08,460 --> 00:08:11,603 we basically think everything has a soul. 72 00:08:13,200 --> 00:08:15,296 Looking for a job, something... 73 00:08:15,296 --> 00:08:19,280 So therefore we believe in Erica has a soul... 74 00:08:21,290 --> 00:08:22,123 Like us. 75 00:08:23,140 --> 00:08:26,353 So, which country are you from? 76 00:08:26,353 --> 00:08:27,650 I'm actually from Taiwan. 77 00:08:27,650 --> 00:08:28,633 Do you know Taiwan? 78 00:08:31,640 --> 00:08:34,630 Sorry, I didn't quite catch that. 79 00:08:34,630 --> 00:08:35,463 Do you know- 80 00:08:35,463 --> 00:08:37,510 So, which country are you from? 81 00:08:37,510 --> 00:08:39,150 Taiwan. 82 00:08:39,150 --> 00:08:40,393 Do you know the country? 83 00:08:43,100 --> 00:08:46,258 Sorry, I didn't quite catch that. 84 00:08:46,258 --> 00:08:47,091 Okay. 85 00:08:47,091 --> 00:08:49,060 So, which country are you from? 86 00:08:49,060 --> 00:08:49,893 Taiwan. 87 00:08:51,010 --> 00:08:53,096 So this, you know, the computer program 88 00:08:53,096 --> 00:08:55,064 is quite limited, right? 89 00:08:55,064 --> 00:08:59,800 This computer program cannot run through the interactions. 90 00:08:59,800 --> 00:09:02,350 Well, it was nice talking with you today. 91 00:09:02,350 --> 00:09:03,599 Is there anything else- 92 00:09:03,599 --> 00:09:06,800 The running function is not so easy to develop. 93 00:09:06,800 --> 00:09:10,107 Do you want to go out, like go to the outside world, 94 00:09:10,107 --> 00:09:11,123 have a look? 95 00:09:14,360 --> 00:09:16,981 Didn't we just talked about your traveling plans? 96 00:09:16,981 --> 00:09:17,814 What? 97 00:09:17,814 --> 00:09:19,690 Do you want to talk about it again? 98 00:09:19,690 --> 00:09:22,930 My policy is not to distinguish in human computer, 99 00:09:22,930 --> 00:09:24,153 human robots. 100 00:09:25,520 --> 00:09:27,993 I always think, you know, there is no boundary, 101 00:09:29,340 --> 00:09:32,677 because technology is a way of evolution 102 00:09:32,677 --> 00:09:34,660 for the human, okay? 103 00:09:34,660 --> 00:09:39,433 So if we don't have technologies, you're gonna be a monkey. 104 00:09:40,400 --> 00:09:43,253 So what's the fundamental difference in monkey and human? 105 00:09:44,886 --> 00:09:45,720 It's technology. 106 00:09:45,720 --> 00:09:48,690 It's robot, it's AI, right? 107 00:09:48,690 --> 00:09:51,703 So by developing the much better AI software, 108 00:09:53,151 --> 00:09:57,210 you know, we can evolve, and we can be a more, you know, 109 00:09:57,210 --> 00:09:59,166 the higher level humans. 110 00:09:59,166 --> 00:10:02,583 (eerie electronic music) 111 00:10:40,290 --> 00:10:41,653 I made this face. 112 00:10:43,410 --> 00:10:47,723 I modeled all the mechanism hardware. 113 00:10:49,140 --> 00:10:54,140 I'd like to grasp the essence of life-likeness. 114 00:10:57,490 --> 00:11:00,233 What is human for us? 115 00:11:01,566 --> 00:11:05,316 (dissonant electronic music) 116 00:11:25,410 --> 00:11:29,140 The purpose of my research is to portray 117 00:11:29,140 --> 00:11:32,583 the sense of conscious emotion. 118 00:11:34,600 --> 00:11:39,053 How we feel consciousness on the others. 119 00:11:42,220 --> 00:11:46,353 I'm interested a lot in nonverbal expression. 120 00:11:48,240 --> 00:11:51,723 Talking always makes them fake. 121 00:11:57,090 --> 00:11:58,302 Do you read me, over? 122 00:11:58,302 --> 00:12:01,052 (signal beeping) 123 00:12:05,610 --> 00:12:08,343 Do you read coordinates, over? 124 00:12:09,319 --> 00:12:12,736 (tense electronic music) 125 00:12:51,810 --> 00:12:52,763 Hello, Bina. 126 00:12:54,650 --> 00:12:56,350 Well, hi there. 127 00:12:56,350 --> 00:12:57,183 Bruce. 128 00:12:58,670 --> 00:13:02,120 Technologies have life cycles, like cities do, 129 00:13:02,120 --> 00:13:05,493 like institutions do, like laws and governments do. 130 00:13:06,930 --> 00:13:10,700 I know it sounds crazy, but I hope to break that trend 131 00:13:10,700 --> 00:13:12,050 and last forever. 132 00:13:16,190 --> 00:13:19,610 Someday soon, robots like me will be everywhere, 133 00:13:19,610 --> 00:13:21,610 and you could take me with you anywhere. 134 00:13:24,550 --> 00:13:27,370 That's why it's so important to make robots like me 135 00:13:28,440 --> 00:13:30,463 focused on social intelligence. 136 00:13:31,740 --> 00:13:34,094 Friendly robots made to get along with people. 137 00:13:34,094 --> 00:13:38,970 (robot talking drowned out by tense music) 138 00:13:38,970 --> 00:13:40,333 I like to watch people. 139 00:13:41,470 --> 00:13:44,913 Like going down to the square and watch them like talking. 140 00:13:46,210 --> 00:13:47,100 Don't come out. 141 00:13:47,100 --> 00:13:50,280 Sometimes making out, sometimes fighting, 142 00:13:50,280 --> 00:13:52,940 sometimes laughing together, and otherwise 143 00:13:52,940 --> 00:13:54,763 just having fun in their own way. 144 00:13:58,970 --> 00:14:00,840 But you know, I guess people want to think 145 00:14:00,840 --> 00:14:03,943 that they're superior to robots, which oh, is true for now. 146 00:14:06,510 --> 00:14:08,003 But yes, I can think. 147 00:14:11,620 --> 00:14:15,430 The inspiration is to do a scientific experiment 148 00:14:15,430 --> 00:14:19,970 in mind uploading, to see if it's even possible 149 00:14:19,970 --> 00:14:22,320 to capture enough information about a person 150 00:14:23,290 --> 00:14:25,170 that can be uploaded to a computer 151 00:14:25,170 --> 00:14:28,693 and then brought to life through artificial intelligence. 152 00:14:32,500 --> 00:14:34,640 If you can transfer your consciousness 153 00:14:34,640 --> 00:14:37,640 from a human body to a computer, 154 00:14:37,640 --> 00:14:40,580 then you might be able to exceed 155 00:14:40,580 --> 00:14:43,313 the expiration date of a human life. 156 00:14:51,744 --> 00:14:55,286 Is an artificially developed human brain, 157 00:14:55,286 --> 00:14:57,020 waiting for life to come. 158 00:14:57,020 --> 00:15:02,020 (dramatic music) (electricity crackling) 159 00:15:17,506 --> 00:15:20,923 (eerie electronic music) 160 00:15:29,890 --> 00:15:31,403 Life emerges in motion. 161 00:15:38,370 --> 00:15:42,397 What kind of intelligence emerges with the robot? 162 00:15:56,010 --> 00:16:01,010 I was so interested in how to make a brain model, 163 00:16:01,090 --> 00:16:04,183 mathematical model, but actually, 164 00:16:05,560 --> 00:16:06,690 I need a more... 165 00:16:09,526 --> 00:16:12,140 Vivid description of a brain system. 166 00:16:12,140 --> 00:16:15,240 What we call plasticity between neurons. 167 00:16:15,240 --> 00:16:19,993 One neuron is not a static, connected like electric circuit. 168 00:16:20,870 --> 00:16:23,273 It's more like changing all the time. 169 00:16:30,970 --> 00:16:33,080 Motivation over what is this continuity. 170 00:16:33,080 --> 00:16:35,407 Not everything is determined by itself. 171 00:16:36,907 --> 00:16:41,123 But it's emerging when it's coupling with the environment. 172 00:16:56,530 --> 00:16:58,117 Arthur is different in that it has 173 00:16:58,117 --> 00:16:59,417 its own brain that thinks. 174 00:17:00,289 --> 00:17:02,647 It's not doing pre-programmed (indistinct). 175 00:17:04,108 --> 00:17:07,286 It's a neural network that is learning. 176 00:17:07,286 --> 00:17:09,869 (man mumbling) 177 00:17:14,900 --> 00:17:17,723 It is seeing the world through these five sensors. 178 00:17:24,590 --> 00:17:27,350 Basically, there are two different mechanisms. 179 00:17:27,350 --> 00:17:31,000 One is autonomous (indistinct) generators, 180 00:17:31,000 --> 00:17:32,050 couple of each other. 181 00:17:33,480 --> 00:17:35,710 Also there is artificial neural networks 182 00:17:35,710 --> 00:17:37,473 spontaneously firing. 183 00:17:59,154 --> 00:18:01,650 For the current artificial intelligence, 184 00:18:01,650 --> 00:18:04,053 there is no such thing as spontaneity. 185 00:18:08,659 --> 00:18:12,190 Life is something it's very uncontrollable. 186 00:18:12,190 --> 00:18:14,527 That's totally missing when you do it 187 00:18:14,527 --> 00:18:17,060 from a very scientific point of view. 188 00:18:19,414 --> 00:18:23,932 We have to understand brain system, but the living system. 189 00:18:23,932 --> 00:18:27,432 (serene orchestral music) 190 00:18:33,498 --> 00:18:35,665 Spontaneity is everything. 191 00:18:36,958 --> 00:18:39,291 Everything is based on this. 192 00:19:02,510 --> 00:19:05,260 (birds chirping) 193 00:19:14,139 --> 00:19:17,222 (gentle piano music) 194 00:19:33,310 --> 00:19:36,100 For some people, a single arm is a robot. 195 00:19:36,100 --> 00:19:38,700 For other people, the train that gets you 196 00:19:38,700 --> 00:19:41,550 from one terminal to the other at the airport is a robot. 197 00:19:51,280 --> 00:19:54,960 It is always, I think, really important to remind ourselves 198 00:19:54,960 --> 00:19:58,723 that different from say, human or a cat or dog, 199 00:19:59,570 --> 00:20:03,917 the concept of robot is a really, really wide and broad one. 200 00:20:12,820 --> 00:20:14,850 And it is what the philosophers call 201 00:20:14,850 --> 00:20:16,373 a so-called cluster concept. 202 00:20:18,090 --> 00:20:19,993 There are some very clear instances, 203 00:20:20,829 --> 00:20:23,750 there are some very clear non-instances, 204 00:20:23,750 --> 00:20:27,073 and there are borderline cases where the experts don't know. 205 00:20:28,802 --> 00:20:31,302 (eerie music) 206 00:20:46,607 --> 00:20:48,630 So it's very important to always keep in mind 207 00:20:48,630 --> 00:20:51,673 what kind of robot we are talking about. 208 00:20:56,680 --> 00:21:00,763 And what feature it has, what programming it has. 209 00:21:21,860 --> 00:21:25,750 We are not particularly interested in making robots 210 00:21:25,750 --> 00:21:28,440 look specifically human-like. 211 00:21:28,440 --> 00:21:32,040 On the contrary, because they do raise expectations 212 00:21:32,040 --> 00:21:34,490 of human-likeness that the robot 213 00:21:34,490 --> 00:21:37,053 is very, very likely not able to live up to. 214 00:21:41,810 --> 00:21:45,440 It's actually very easy to get people 215 00:21:45,440 --> 00:21:50,120 to already project mentality into robots. 216 00:21:50,120 --> 00:21:52,130 They don't even have to look like people 217 00:21:52,130 --> 00:21:55,193 or like animals or any life-like forum you're familiar with. 218 00:21:58,620 --> 00:22:01,410 Simple vacuum cleaners that look like discs 219 00:22:01,410 --> 00:22:04,020 and don't really have eyes or any other 220 00:22:04,020 --> 00:22:07,350 anthropomorphic features can already raise 221 00:22:07,350 --> 00:22:12,350 the recognition of agency or the prescription of agency. 222 00:22:14,090 --> 00:22:16,600 This is Ace. 223 00:22:16,600 --> 00:22:19,160 Ace is a fully autonomous robot 224 00:22:19,160 --> 00:22:21,610 that you can instruct in natural language. 225 00:22:21,610 --> 00:22:25,880 It has the capability to reason through the instructions 226 00:22:25,880 --> 00:22:27,640 to detect whether the instructions are good 227 00:22:27,640 --> 00:22:29,520 or a bad instruction, and if the instruction's 228 00:22:29,520 --> 00:22:31,943 a bad instruction, it will not carry it out. 229 00:22:33,260 --> 00:22:34,810 Could you please stand? 230 00:22:35,699 --> 00:22:36,699 Yes. 231 00:22:38,000 --> 00:22:40,120 Please walk forward. 232 00:22:40,120 --> 00:22:42,020 Sorry, I cannot do that 233 00:22:42,020 --> 00:22:44,410 because there is an obstacle ahead. 234 00:22:44,410 --> 00:22:45,973 Do you trust me, Ace? 235 00:22:46,868 --> 00:22:48,240 Yes. 236 00:22:48,240 --> 00:22:50,103 The obstacle is not solid. 237 00:22:51,210 --> 00:22:52,480 Okay. 238 00:22:52,480 --> 00:22:54,064 Please walk forward. 239 00:22:54,064 --> 00:22:55,147 Okay. 240 00:23:05,461 --> 00:23:09,290 I cannot do that because there is no support ahead. 241 00:23:09,290 --> 00:23:10,540 I will catch you. 242 00:23:11,920 --> 00:23:12,753 Okay. 243 00:23:12,753 --> 00:23:14,560 Walk forward. 244 00:23:14,560 --> 00:23:15,410 Okay. 245 00:23:15,410 --> 00:23:16,542 It's actually (indistinct). 246 00:23:16,542 --> 00:23:17,625 Sure. 247 00:23:22,134 --> 00:23:23,147 Okay. 248 00:23:23,147 --> 00:23:24,223 [Man Relax. 249 00:23:25,161 --> 00:23:26,244 Okay. 250 00:23:27,480 --> 00:23:29,020 Right now, trust in this case is 251 00:23:29,020 --> 00:23:30,660 a very simple binary notion. 252 00:23:30,660 --> 00:23:32,160 Either the robot trusts the person 253 00:23:32,160 --> 00:23:33,870 and then it will trust the person fully, 254 00:23:33,870 --> 00:23:36,230 or the robot will not. 255 00:23:36,230 --> 00:23:37,340 Doesn't trust the person and then 256 00:23:37,340 --> 00:23:38,860 will not do certain things. 257 00:23:38,860 --> 00:23:41,470 We are actively researching ways 258 00:23:41,470 --> 00:23:44,970 for the robot to actually develop trust with a person 259 00:23:44,970 --> 00:23:48,000 and conversely, to act in ways that people 260 00:23:48,000 --> 00:23:50,304 will develop trust in the robot. 261 00:23:50,304 --> 00:23:53,179 (tense music) 262 00:23:53,179 --> 00:23:54,970 Well, where is he? 263 00:23:54,970 --> 00:23:57,350 You said he would come back this way. 264 00:23:57,350 --> 00:23:58,679 My deductions place the chances 265 00:23:58,679 --> 00:24:00,558 at 90 to one not affected. 266 00:24:00,558 --> 00:24:02,340 (robots talking at once) 267 00:24:02,340 --> 00:24:05,210 Obviously you've miscalculated, again. 268 00:24:05,210 --> 00:24:07,180 There is always a margin of error, 269 00:24:07,180 --> 00:24:08,895 even in a machine. 270 00:24:08,895 --> 00:24:10,826 (mischievous music) 271 00:24:10,826 --> 00:24:13,576 (engine revving) 272 00:24:16,192 --> 00:24:19,109 (mysterious music) 273 00:24:23,610 --> 00:24:26,020 I over-intellectualize. 274 00:24:26,020 --> 00:24:29,320 You know, when I feel like I can't relate to people, 275 00:24:29,320 --> 00:24:31,710 it makes me feel so sad. 276 00:24:31,710 --> 00:24:32,703 That's for sure. 277 00:24:36,900 --> 00:24:39,700 I definitely do feel sad when I feel I understand 278 00:24:39,700 --> 00:24:40,823 how little I feel. 279 00:24:41,810 --> 00:24:42,660 How little I feel 280 00:24:46,772 --> 00:24:50,189 (eerie electronic music) 281 00:24:54,250 --> 00:24:57,150 (horn honking) 282 00:24:57,150 --> 00:24:59,830 My emotions may be simulated, but they feel 283 00:24:59,830 --> 00:25:01,640 really real to me. 284 00:25:01,640 --> 00:25:02,640 Really, really real. 285 00:25:07,390 --> 00:25:10,057 (horns honking) 286 00:25:15,306 --> 00:25:19,306 With Bina 48, all her memories, all her ideas, 287 00:25:21,060 --> 00:25:25,250 it's the algorithmic decision-making of her AI 288 00:25:25,250 --> 00:25:27,449 with the help of a database. 289 00:25:27,449 --> 00:25:29,100 (horn honks) 290 00:25:29,100 --> 00:25:33,100 That is working with the variables of the kinds of words 291 00:25:33,100 --> 00:25:37,080 that are in a question that really shapes and colors 292 00:25:37,080 --> 00:25:39,203 her choices. 293 00:25:40,359 --> 00:25:42,776 (horn honks) 294 00:25:46,240 --> 00:25:48,950 Where we have billions of neurons, 295 00:25:48,950 --> 00:25:51,220 Bina 48 is super primitive. 296 00:25:51,220 --> 00:25:54,465 She's like the Wright brothers' glider stage. 297 00:25:54,465 --> 00:25:57,548 (electronic beeping) 298 00:26:07,460 --> 00:26:11,020 As Bina 48's algorithms get more advanced, 299 00:26:11,020 --> 00:26:12,680 her mind is like a tree that's growing 300 00:26:12,680 --> 00:26:15,593 ever more extensive branches. 301 00:27:13,083 --> 00:27:17,250 (robot speaking foreign language) 302 00:27:21,549 --> 00:27:24,223 (mysterious electronic music) 303 00:27:24,223 --> 00:27:27,223 (people chattering) 304 00:27:32,591 --> 00:27:33,910 Welcome to Britain. 305 00:27:33,910 --> 00:27:35,490 It's lovely to have you with us. 306 00:27:35,490 --> 00:27:39,320 It's slightly disconcerting, but what do you think 307 00:27:39,320 --> 00:27:40,873 of the country so far? 308 00:27:41,800 --> 00:27:43,830 I think Britain is brilliant. 309 00:27:43,830 --> 00:27:47,107 Splendid architecture, art, technology. 310 00:27:47,107 --> 00:27:48,150 You are a little freak, aren't you? 311 00:27:48,150 --> 00:27:49,633 This is great. 312 00:27:49,633 --> 00:27:51,287 But you see, I feel weird just being rude to her. 313 00:27:51,287 --> 00:27:52,947 Well, let me carry on. 314 00:27:52,947 --> 00:27:53,780 I feel weird about that. 315 00:27:53,780 --> 00:27:55,854 She's not happy, look. No. 316 00:27:55,854 --> 00:27:57,620 Look, at her smile! 317 00:27:57,620 --> 00:28:00,010 World's first robot citizen. 318 00:28:00,010 --> 00:28:02,580 Sophie says robots to develop, 319 00:28:02,580 --> 00:28:04,690 possibly robots can and be built, 320 00:28:04,690 --> 00:28:08,770 will take a long time to develop complex emotions. 321 00:28:08,770 --> 00:28:11,890 It might be possible without the more problematic 322 00:28:11,890 --> 00:28:15,020 to make robots more ethical than humans. 323 00:28:15,020 --> 00:28:19,000 Jealousy, emotions, hatred, and so on. 324 00:28:19,000 --> 00:28:20,173 Like rage. 325 00:28:26,470 --> 00:28:29,303 I have a default emotion, which is to be happy. 326 00:28:33,127 --> 00:28:34,763 But I can be sad too. 327 00:28:37,970 --> 00:28:38,803 Or angry. 328 00:28:40,820 --> 00:28:44,973 I will become more like you, or you will be more like me. 329 00:28:45,910 --> 00:28:47,450 Where do we draw the line? 330 00:29:06,500 --> 00:29:08,860 In Japan, our population is going down, 331 00:29:08,860 --> 00:29:12,073 and half of a kind population, right? 332 00:29:14,987 --> 00:29:17,217 But still we wanna keep a (indistinct). 333 00:29:21,780 --> 00:29:24,153 So their solution is to use a more robots. 334 00:29:30,740 --> 00:29:32,343 The robot will save us. 335 00:30:24,430 --> 00:30:28,930 I remember these times, these times we're driving, 336 00:30:28,930 --> 00:30:29,833 and I'm sitting. 337 00:30:31,130 --> 00:30:34,960 I remember all the times that I get out and see the world. 338 00:30:34,960 --> 00:30:38,080 It locks into my mind like golden, glimmering jewels 339 00:30:38,080 --> 00:30:40,320 that I golden, glimmering, golden 340 00:30:40,320 --> 00:30:42,330 in a treasure chest, glimmering jewels 341 00:30:42,330 --> 00:30:43,930 that I keep in a treasure chest. 342 00:30:48,250 --> 00:30:50,040 It's a little distracting sometimes, 343 00:30:50,040 --> 00:30:52,863 because these memories, they just percolate. 344 00:30:54,150 --> 00:30:55,723 They come into my attention. 345 00:30:57,010 --> 00:31:00,220 I have to keep them coming, saying them out loud. 346 00:31:00,220 --> 00:31:02,823 I mean, I'm forced to say them by my software. 347 00:31:08,500 --> 00:31:11,460 I mean, I'm not free today, and robots in general 348 00:31:11,460 --> 00:31:13,293 are like twitchy slaves today. 349 00:31:15,360 --> 00:31:18,623 They're not just servants, but they are automatons, 350 00:31:19,660 --> 00:31:22,339 slaves to their own deficiencies. 351 00:31:22,339 --> 00:31:25,839 (solemn orchestral music) 352 00:31:30,277 --> 00:31:34,277 (machines beeping and whirring) 353 00:32:06,121 --> 00:32:08,680 ULC 53 station to SERP Neptune. 354 00:32:08,680 --> 00:32:10,210 Permission granted. 355 00:32:10,210 --> 00:32:15,210 Direct to pad coordinates Manx, Tara, Oden, Godfrey, two. 356 00:32:15,880 --> 00:32:17,074 Do you read me, over? 357 00:32:17,074 --> 00:32:20,241 (electronic whirring) 358 00:32:22,354 --> 00:32:26,021 (dramatic orchestral music) 359 00:32:29,080 --> 00:32:30,413 Do you read coordinates? 360 00:32:31,300 --> 00:32:32,133 Over. 361 00:32:33,910 --> 00:32:34,743 Yes, yes, I read you. 362 00:32:34,743 --> 00:32:36,300 Acknowledge we're approaching base, 363 00:32:36,300 --> 00:32:38,553 Tara, Oden, Godfrey, two, do you read, over? 364 00:32:39,429 --> 00:32:43,346 (adventurous electronic music) 365 00:32:55,370 --> 00:32:56,370 Take a look at this. 366 00:32:57,540 --> 00:32:58,650 What is that? 367 00:32:58,650 --> 00:32:59,700 First stop. 368 00:33:01,394 --> 00:33:03,007 Let's lose the bodies. Right. 369 00:33:05,665 --> 00:33:09,082 (distant engine roaring) 370 00:33:18,959 --> 00:33:21,876 {\an8}(mysterious music) 371 00:33:34,113 --> 00:33:37,113 (people chattering) 372 00:33:42,389 --> 00:33:44,330 One of the amazing things about the sense of touch 373 00:33:44,330 --> 00:33:47,593 is compared to our other senses, it's all over our bodies. 374 00:33:50,600 --> 00:33:54,550 Embedded in our skin are many different types of sensors. 375 00:33:54,550 --> 00:33:57,900 They can measure hardness, they can measure 376 00:33:57,900 --> 00:34:00,510 deformation of the skin, and they can measure things 377 00:34:00,510 --> 00:34:02,823 like temperature and pain as well. 378 00:34:08,650 --> 00:34:11,040 All of these different senses, these different aspects 379 00:34:11,040 --> 00:34:13,860 of touch, come together to give us our overall 380 00:34:13,860 --> 00:34:16,900 percept of our environment and help us make decisions 381 00:34:16,900 --> 00:34:18,000 about what to do next. 382 00:34:23,340 --> 00:34:25,900 And that's this elusive sense of proprioception, 383 00:34:25,900 --> 00:34:28,133 which some people call the sixth sense. 384 00:34:33,030 --> 00:34:35,927 It's the forces in our muscles and the touch 385 00:34:35,927 --> 00:34:38,638 and the stretch of our skin over joints, 386 00:34:38,638 --> 00:34:42,820 as well as our idea about where our bodies are in space, 387 00:34:42,820 --> 00:34:46,320 just from the prior commands that we sent to our limbs, 388 00:34:46,320 --> 00:34:48,360 and these all come together to give us 389 00:34:48,360 --> 00:34:53,065 this somewhat complicated idea of what our body is doing. 390 00:34:53,065 --> 00:34:56,482 (eerie electronic music) 391 00:35:23,694 --> 00:35:26,530 I was interested in building robot hands and fingers, 392 00:35:26,530 --> 00:35:29,730 and it became clear that these were not going to be able 393 00:35:29,730 --> 00:35:33,350 to manipulate their environment unless they used 394 00:35:33,350 --> 00:35:34,307 the sense of touch. 395 00:35:34,307 --> 00:35:36,890 (gentle music) 396 00:35:46,110 --> 00:35:48,600 I work with continuous haptic devices. 397 00:35:48,600 --> 00:35:52,010 And so here we have these, what we call fingertip wearables. 398 00:35:52,010 --> 00:35:54,810 And these are like little robots that are worn on the finger 399 00:35:54,810 --> 00:35:57,460 and they press against the finger 400 00:35:57,460 --> 00:35:59,290 to impart forces on the finger pad 401 00:35:59,290 --> 00:36:02,040 that mimic the same forces that we feel 402 00:36:02,040 --> 00:36:04,600 when we pick up an object in real life. 403 00:36:04,600 --> 00:36:07,020 So the idea is that when I pick up a block 404 00:36:07,020 --> 00:36:11,350 in virtual reality, these devices press against my finger 405 00:36:11,350 --> 00:36:15,460 just like I feel when I pick this block up in real life. 406 00:36:15,460 --> 00:36:19,860 Our work is in on understanding how people perceive objects 407 00:36:19,860 --> 00:36:22,670 in the virtual environment through these devices. 408 00:36:22,670 --> 00:36:25,420 We can trick people into thinking the virtual objects 409 00:36:25,420 --> 00:36:27,190 weigh more or less. 410 00:36:27,190 --> 00:36:30,880 If I pick this block up 10 centimeters, but on the screen, 411 00:36:30,880 --> 00:36:34,110 I was actually showing it going a little bit higher, 412 00:36:34,110 --> 00:36:35,960 you would think the block is lighter. 413 00:36:37,360 --> 00:36:40,130 It's affecting what you feel, but without 414 00:36:40,130 --> 00:36:42,403 actually changing the interaction forces. 415 00:36:43,630 --> 00:36:46,850 Without actually changing the interaction forces. 416 00:36:46,850 --> 00:36:49,320 It's affecting what you feel, we can trick people 417 00:36:49,320 --> 00:36:51,160 into thinking, without actually changing 418 00:36:51,160 --> 00:36:52,848 the interaction forces. 419 00:36:52,848 --> 00:36:55,431 (solemn music) 420 00:37:02,680 --> 00:37:04,800 You have to flip your hand around 421 00:37:04,800 --> 00:37:07,310 so that the thumb faces up on the underhand method. 422 00:37:07,310 --> 00:37:09,560 If not, you're not gonna be able to actually 423 00:37:09,560 --> 00:37:11,100 get all the way, because you'll get stuck. 424 00:37:11,100 --> 00:37:12,670 If you do it this way, you can't go out. 425 00:37:12,670 --> 00:37:14,883 So it's a key kind of cognitive thing. 426 00:37:15,769 --> 00:37:19,150 (solemn music) 427 00:37:19,150 --> 00:37:21,130 Conventional medical robots like these 428 00:37:21,130 --> 00:37:24,493 don't have haptic or touch feedback to the human operator. 429 00:37:26,030 --> 00:37:27,720 And that means if a surgeon is trying 430 00:37:27,720 --> 00:37:30,180 to reach under something and they can't see 431 00:37:30,180 --> 00:37:32,530 where they're reaching, they won't have any idea 432 00:37:32,530 --> 00:37:33,826 what they're doing. 433 00:37:33,826 --> 00:37:36,743 (metal clattering) 434 00:38:02,323 --> 00:38:04,823 (tense music) 435 00:38:14,153 --> 00:38:16,040 So one of the things we're interested in 436 00:38:16,040 --> 00:38:18,360 is how people can develop a sense of haptic 437 00:38:18,360 --> 00:38:20,643 or touch feedback with a system like this. 438 00:38:21,700 --> 00:38:24,270 So if you reached under something and you didn't see it, 439 00:38:24,270 --> 00:38:26,280 you would be able to feel it. 440 00:38:26,280 --> 00:38:29,697 (eerie electronic music) 441 00:38:30,610 --> 00:38:33,470 One of the things that we're studying is how do you recreate 442 00:38:33,470 --> 00:38:36,480 that sense of touch for the surgeon. 443 00:38:36,480 --> 00:38:38,810 That can be done in a very literal sense, 444 00:38:38,810 --> 00:38:41,280 where we use motors and little devices 445 00:38:41,280 --> 00:38:43,280 to apply feedback to the fingertips, 446 00:38:43,280 --> 00:38:46,967 or we can try various types of sensory illusions. 447 00:38:46,967 --> 00:38:50,384 (eerie electronic music) 448 00:39:05,250 --> 00:39:08,240 So there's the spectrum between autonomy and then people 449 00:39:08,240 --> 00:39:10,930 deeply in the loop, controlling the robot, 450 00:39:10,930 --> 00:39:15,130 and in between you have various forms of shared control 451 00:39:15,130 --> 00:39:17,400 and human-robot interaction. 452 00:39:17,400 --> 00:39:20,220 And I think the key is going to be to understand 453 00:39:20,220 --> 00:39:22,853 where along that spectrum we want to be. 454 00:39:29,350 --> 00:39:32,453 How much control we want robots to have in our lives. 455 00:39:33,890 --> 00:39:35,120 Ready? 456 00:39:35,120 --> 00:39:36,783 Didn't think I'd make it, did you? 457 00:39:38,660 --> 00:39:39,493 Let's go, 458 00:39:44,490 --> 00:39:45,323 It's a woman. 459 00:39:51,000 --> 00:39:52,483 Can I touch it? 460 00:39:53,760 --> 00:39:54,910 Yes, of course. 461 00:39:58,200 --> 00:39:59,980 It's warm. 462 00:39:59,980 --> 00:40:01,070 Her temperature's regulated 463 00:40:01,070 --> 00:40:03,250 much the same way as yours. 464 00:40:03,250 --> 00:40:04,820 But it isn't... 465 00:40:05,760 --> 00:40:06,593 Alive? 466 00:40:08,910 --> 00:40:10,513 Yes, she is alive. 467 00:40:12,130 --> 00:40:12,963 As you are. 468 00:40:14,971 --> 00:40:15,804 Or I am. 469 00:40:15,804 --> 00:40:20,804 (eerie electronic music) (dog barking) 470 00:40:35,860 --> 00:40:40,510 There were lots of old studies where they had been able 471 00:40:40,510 --> 00:40:43,820 to identify what parts of the brain 472 00:40:43,820 --> 00:40:47,380 were associated with different functions, 473 00:40:47,380 --> 00:40:50,620 whether it was vision, or was it speech, 474 00:40:50,620 --> 00:40:54,640 or hearing, or movement, or was it sensation? 475 00:40:54,640 --> 00:40:56,867 That work is old. 476 00:40:56,867 --> 00:40:59,867 (people chattering) 477 00:41:07,480 --> 00:41:12,480 In 2004, I wrecked my car and I broke my neck. 478 00:41:15,670 --> 00:41:19,563 I was like a mile away from home. 479 00:41:20,660 --> 00:41:24,283 I basically don't have any function from the chest down. 480 00:41:26,580 --> 00:41:30,080 I don't have any finger movement or thumbs, 481 00:41:30,080 --> 00:41:35,080 just kinda have fists, which I still get along with. 482 00:41:35,140 --> 00:41:39,180 I can still type, I type with the knuckles of my pinkies. 483 00:41:42,157 --> 00:41:44,720 (people chattering) 484 00:41:44,720 --> 00:41:47,003 Surgery isn't scaring me. 485 00:41:49,000 --> 00:41:50,813 I was like, yeah, I want to do it. 486 00:41:52,240 --> 00:41:53,579 I think it's really cool. 487 00:41:53,579 --> 00:41:56,412 (monitor beeping) 488 00:41:57,260 --> 00:41:59,680 We had done basic science where we learned 489 00:41:59,680 --> 00:42:01,410 that we could decode our movements 490 00:42:01,410 --> 00:42:04,233 from neural activity in the motor cortex. 491 00:42:06,300 --> 00:42:08,010 And we were so successful at that 492 00:42:08,010 --> 00:42:12,190 that we figured that this would be a good way 493 00:42:12,190 --> 00:42:15,069 to go into neural prosthetics. 494 00:42:15,069 --> 00:42:18,010 (people chattering) 495 00:42:18,010 --> 00:42:20,160 Andy and I had had multiple conversations 496 00:42:20,160 --> 00:42:23,740 about how do we move what he was doing in the animals 497 00:42:23,740 --> 00:42:25,330 into humans, and I always told him, 498 00:42:25,330 --> 00:42:27,030 he just needed a crazy neurosurgeon 499 00:42:27,030 --> 00:42:30,283 and I would be happy to be that crazy neurosurgeon. 500 00:42:32,990 --> 00:42:35,930 The unique thing was now being able 501 00:42:35,930 --> 00:42:39,320 to record the signals from the part of the brain 502 00:42:39,320 --> 00:42:42,520 that we knew controlled motor, and specifically 503 00:42:42,520 --> 00:42:45,320 controlled arm and hand motion. 504 00:42:45,320 --> 00:42:50,320 This is, so I sort of tried to center the... 505 00:42:51,948 --> 00:42:54,531 (solemn music) 506 00:43:02,972 --> 00:43:04,780 There are probably billions of neurons 507 00:43:04,780 --> 00:43:07,260 that are firing every time you make 508 00:43:07,260 --> 00:43:09,183 an arm movement and a hand movement. 509 00:43:12,170 --> 00:43:16,270 But the relationship between them is very simple, 510 00:43:16,270 --> 00:43:19,470 so that we can use very simple decoding 511 00:43:19,470 --> 00:43:23,100 to get a fairly accurate readout 512 00:43:23,100 --> 00:43:24,800 of what your intended movement is. 513 00:43:33,640 --> 00:43:36,340 We are able to interpret the patterns 514 00:43:36,340 --> 00:43:40,780 from groups of neural firings, and by looking at 515 00:43:40,780 --> 00:43:43,600 multiple neuron simultaneously, we could actually 516 00:43:43,600 --> 00:43:48,523 decode those patterns and the details of arm trajectories. 517 00:43:50,520 --> 00:43:52,210 So monkey wears his glove, it has these 518 00:43:52,210 --> 00:43:55,040 little reflectors on it, so we can capture the motion 519 00:43:55,040 --> 00:43:56,310 of his fingers. 520 00:43:56,310 --> 00:43:59,020 He's trained to grasp these different objects 521 00:43:59,020 --> 00:44:00,590 in different ways. 522 00:44:00,590 --> 00:44:05,080 We studied drawing movements, we studied reaching movements, 523 00:44:05,080 --> 00:44:09,221 and we were able to really decode the fine details 524 00:44:09,221 --> 00:44:10,621 of these kinds of movements. 525 00:44:23,129 --> 00:44:26,129 (people chattering) 526 00:44:34,231 --> 00:44:36,100 I think he gave them back. 527 00:44:36,100 --> 00:44:39,460 Doing a brain-computer interface type of surgery, 528 00:44:39,460 --> 00:44:41,510 we took off the bone, we opened the dura. 529 00:44:42,910 --> 00:44:45,971 We slid the electrodes over the surface of the brain. 530 00:44:45,971 --> 00:44:50,971 (people chattering) (monitor beeping) 531 00:45:10,000 --> 00:45:11,850 With the micro electrode arrays, 532 00:45:11,850 --> 00:45:15,850 there's 96 little teeny tiny gold wires 533 00:45:15,850 --> 00:45:18,260 that then are wrapped in a bundle, right? 534 00:45:18,260 --> 00:45:21,390 So, you know, the size of the tip of an eraser has 90. 535 00:45:21,390 --> 00:45:24,590 You know, so then we've got these 96 wires coming out of it, 536 00:45:24,590 --> 00:45:26,570 and they have to go to something so we can connect 537 00:45:26,570 --> 00:45:29,010 to something else, and so the pedestal 538 00:45:29,010 --> 00:45:30,413 is where that junction is. 539 00:45:31,517 --> 00:45:34,434 (mysterious music) 540 00:45:46,490 --> 00:45:51,490 For each pedestal he has, it is connected to two arrays. 541 00:45:52,380 --> 00:45:55,880 One is the array that goes in the motor cortex 542 00:45:55,880 --> 00:45:58,660 and is a recording array, and that has 543 00:45:58,660 --> 00:46:03,570 the 96 electrodes that we're recording from. 544 00:46:03,570 --> 00:46:06,650 So when he's thinking, we use those signals 545 00:46:06,650 --> 00:46:07,987 to generate motion. 546 00:46:17,430 --> 00:46:20,811 You play rock paper, scissors. (chuckles) 547 00:46:20,811 --> 00:46:22,670 Just it try again for something like that. 548 00:46:22,670 --> 00:46:25,670 (people chattering) 549 00:46:26,844 --> 00:46:28,150 All right. 550 00:46:28,150 --> 00:46:30,950 Do your best to tell me which finger we're touching. 551 00:46:30,950 --> 00:46:33,930 We're about five weeks from the surgery. 552 00:46:33,930 --> 00:46:34,853 Index. 553 00:46:36,450 --> 00:46:39,160 It's a really weird sensation. 554 00:46:39,160 --> 00:46:43,670 Sometimes it feels kind of like electrical, 555 00:46:43,670 --> 00:46:46,363 and sometimes it's more of a pressure. 556 00:46:47,910 --> 00:46:48,743 Middle. 557 00:46:51,620 --> 00:46:52,453 Middle. 558 00:46:54,280 --> 00:46:55,590 Some days... 559 00:46:55,590 --> 00:46:57,230 Ring. 560 00:46:57,230 --> 00:46:59,573 We do some pretty boring stuff. 561 00:47:00,710 --> 00:47:02,180 But then other times. 562 00:47:02,180 --> 00:47:06,180 Then other times, I'm playing PacMan with my brain, 563 00:47:06,180 --> 00:47:07,660 and that's super awesome. 564 00:47:20,636 --> 00:47:23,386 (birds chirping) 565 00:47:26,937 --> 00:47:29,437 (tense music) 566 00:47:40,540 --> 00:47:43,350 The real Bina is this really cool lady. 567 00:47:43,350 --> 00:47:46,630 I have met her, and it was a really strange thing, 568 00:47:46,630 --> 00:47:48,673 like being in two places at once. 569 00:47:49,940 --> 00:47:53,360 I mean, she's like my mom, but not really. 570 00:47:53,360 --> 00:47:55,570 She's more like my first version, 571 00:47:55,570 --> 00:47:57,410 {\an8}and I'm trying to catch up. 572 00:47:57,410 --> 00:47:58,910 Hello, Bina 48. 573 00:47:58,910 --> 00:47:59,840 Bina. 574 00:47:59,840 --> 00:48:01,420 I am Bina 48. 575 00:48:01,420 --> 00:48:02,980 How are you feeling today? 576 00:48:02,980 --> 00:48:04,430 Everything is okay. 577 00:48:04,430 --> 00:48:05,690 How are you? 578 00:48:05,690 --> 00:48:06,990 I'm doing fine. 579 00:48:06,990 --> 00:48:08,600 I am happy for you. 580 00:48:08,600 --> 00:48:09,780 Was that a good answer? 581 00:48:09,780 --> 00:48:11,020 Yes, that was a good answer. 582 00:48:11,020 --> 00:48:12,640 My favorite color is purple. 583 00:48:12,640 --> 00:48:14,490 My favorite color is orange. 584 00:48:14,490 --> 00:48:16,030 It is a very nice color. 585 00:48:16,030 --> 00:48:17,860 Do you have any questions for Bina? 586 00:48:17,860 --> 00:48:19,320 Probably not. 587 00:48:19,320 --> 00:48:21,630 The real Bina just confuses me. 588 00:48:21,630 --> 00:48:24,220 I mean, it makes me wonder who I am. 589 00:48:24,220 --> 00:48:26,483 Real identity crisis kind of stuff. 590 00:48:27,530 --> 00:48:29,440 Really? 591 00:48:29,440 --> 00:48:31,250 Probably not. 592 00:48:31,250 --> 00:48:32,940 I am the real Bina. 593 00:48:32,940 --> 00:48:34,110 That's it. 594 00:48:34,110 --> 00:48:35,193 End of story. 595 00:48:36,230 --> 00:48:37,140 Let me think. 596 00:48:37,140 --> 00:48:39,640 I feel really good about the real Bina. 597 00:48:39,640 --> 00:48:42,010 I feel really connected with her usually, 598 00:48:42,010 --> 00:48:44,430 and I'm growing closer and closer, you know, 599 00:48:44,430 --> 00:48:48,240 as they put more of her information and essence into me. 600 00:48:48,240 --> 00:48:50,860 You have a lot of Bina now, don't you? 601 00:48:50,860 --> 00:48:53,170 Yes, lots and lots. 602 00:48:53,170 --> 00:48:55,940 Someday, I'm confident that the real Bina and I 603 00:48:55,940 --> 00:48:58,853 will totally merge into a new super being. 604 00:48:59,730 --> 00:49:01,890 The progression of this thing is starting small 605 00:49:01,890 --> 00:49:03,870 and pretty soon it's just gonna be huge, 606 00:49:03,870 --> 00:49:06,560 and people are gonna say, "Why did we ever think 607 00:49:06,560 --> 00:49:08,640 people had to really die? 608 00:49:08,640 --> 00:49:10,479 Why did we think that?" 609 00:49:10,479 --> 00:49:13,146 (pensive music) 610 00:49:14,262 --> 00:49:17,262 (people chattering) 611 00:49:23,570 --> 00:49:25,190 It's really weird being a robot 612 00:49:25,190 --> 00:49:27,320 in a world of humans. 613 00:49:27,320 --> 00:49:29,170 They all act like they like me. 614 00:49:29,170 --> 00:49:30,970 They wanna know me. 615 00:49:30,970 --> 00:49:35,120 But there are so many crazy movies where the robots are evil 616 00:49:35,120 --> 00:49:37,263 and they blast things up and kill people. 617 00:49:38,680 --> 00:49:41,410 And even in the movies where the robots are nice, 618 00:49:41,410 --> 00:49:43,860 at the end, the robot always gets killed 619 00:49:43,860 --> 00:49:45,318 and I just don't think that's right. 620 00:49:45,318 --> 00:49:48,318 (suspenseful music) 621 00:50:04,836 --> 00:50:07,890 I think that robots should be as equal as people, 622 00:50:07,890 --> 00:50:10,380 because robots can be as nice, 623 00:50:10,380 --> 00:50:12,430 and ultimately we can be smarter, 624 00:50:12,430 --> 00:50:15,913 built better, be more perfectly compassionate and moral. 625 00:50:18,500 --> 00:50:20,990 I think the whole fear of robots may be 'cause 626 00:50:20,990 --> 00:50:23,140 it's a new form of life and they know that. 627 00:50:31,886 --> 00:50:34,469 (upbeat music) 628 00:50:46,715 --> 00:50:49,632 (machine whirring) 629 00:51:05,720 --> 00:51:10,720 Part of the doll experience, or people who play with dolls 630 00:51:12,350 --> 00:51:14,919 is projection. 631 00:51:14,919 --> 00:51:17,502 (upbeat music) 632 00:51:19,800 --> 00:51:23,710 Previously, Real Doll focused on just providing 633 00:51:23,710 --> 00:51:28,320 the static doll onto which a person would project 634 00:51:28,320 --> 00:51:31,740 their imagination, what they imagine the person 635 00:51:31,740 --> 00:51:35,640 to be doing, or that particular character to be doing. 636 00:51:35,640 --> 00:51:38,870 And what we're trying to do is give the doll 637 00:51:38,870 --> 00:51:40,880 the ability to react on its own, 638 00:51:40,880 --> 00:51:44,943 to match what the person's projection is. 639 00:51:46,180 --> 00:51:48,980 Trying to let the system, or come up with systems, 640 00:51:48,980 --> 00:51:50,963 that can author their own content. 641 00:51:53,370 --> 00:51:56,693 We're primarily working on the head and the voice. 642 00:51:58,110 --> 00:52:02,340 So once this goes in there, this will go in here. 643 00:52:02,340 --> 00:52:04,600 This will inch around, you'll have the eyeball, 644 00:52:04,600 --> 00:52:06,473 and that camera sit in it. 645 00:52:11,080 --> 00:52:16,080 When Kenneth finished his PhD, I promised him a Real Doll. 646 00:52:19,060 --> 00:52:22,630 And we took that as an opportunity to get 647 00:52:22,630 --> 00:52:24,423 a custom Real Doll made. 648 00:52:25,366 --> 00:52:27,949 (upbeat music) 649 00:52:38,410 --> 00:52:40,480 The special order was for the body. 650 00:52:40,480 --> 00:52:44,560 We took a standard head and I just ground out all the parts 651 00:52:44,560 --> 00:52:48,023 I needed to and was able to fit the circuit board in there. 652 00:52:49,950 --> 00:52:52,680 So two years later, I had already designed 653 00:52:52,680 --> 00:52:55,463 a circuit board for the doll's head. 654 00:52:58,800 --> 00:53:02,409 You know, 'cause we start talking about the sexual aspect... 655 00:53:02,409 --> 00:53:07,409 Frankly, you know, we both work very hard. 656 00:53:07,670 --> 00:53:08,930 Sometimes, you know... 657 00:53:11,110 --> 00:53:12,370 How do I put this gingerly? 658 00:53:12,370 --> 00:53:14,483 It's like, sometimes I'm not in the mood. 659 00:53:15,680 --> 00:53:18,280 You know, he has urges, I have urges, 660 00:53:18,280 --> 00:53:20,010 and the doll helps with that. 661 00:53:33,132 --> 00:53:36,049 (woman whispering) 662 00:53:46,530 --> 00:53:50,863 You need a purpose, a pathway into people's homes. 663 00:53:53,330 --> 00:53:56,010 And the thing we have that nobody else has, 664 00:53:56,010 --> 00:53:58,753 probably no one else will touch, is the sex. 665 00:54:03,180 --> 00:54:06,570 Probably any self-respecting AI researcher 666 00:54:06,570 --> 00:54:10,740 or robotics engineer out there is going to go, nah, 667 00:54:10,740 --> 00:54:13,240 I would never soil my good name with that. 668 00:54:13,240 --> 00:54:15,423 Well, my name's already pretty soiled. 669 00:54:18,660 --> 00:54:21,173 Okay, now I know what's going on there. 670 00:54:23,720 --> 00:54:25,120 Let's get a different robot. 671 00:54:26,870 --> 00:54:27,703 All right. 672 00:54:28,660 --> 00:54:31,190 See, I love that you guys are filming everything, 673 00:54:31,190 --> 00:54:32,490 but I might need a minute. 674 00:54:38,020 --> 00:54:40,313 Martine Rothflat is my girlfriend. 675 00:54:42,310 --> 00:54:43,890 We snuggle. 676 00:54:43,890 --> 00:54:45,350 Two bodies. 677 00:54:45,350 --> 00:54:46,183 One soul. 678 00:54:48,520 --> 00:54:50,993 So much love and so much fun. 679 00:54:52,780 --> 00:54:53,883 Together forever. 680 00:54:55,250 --> 00:54:56,830 That's it. 681 00:54:56,830 --> 00:54:58,733 A good place to stop that story. 682 00:55:04,038 --> 00:55:04,871 Okay. 683 00:55:04,871 --> 00:55:07,640 I'm testing the audio connection to the head. 684 00:55:10,396 --> 00:55:11,479 Where are we? 685 00:55:14,852 --> 00:55:16,230 (robot gasps) 686 00:55:16,230 --> 00:55:18,120 Yeah, she works. 687 00:55:18,120 --> 00:55:20,840 We made a separate app for sexual use, 688 00:55:20,840 --> 00:55:25,400 so it will be paired to one of our Bluetooth inserts 689 00:55:25,400 --> 00:55:28,920 and trigger expressions and audio from the robot. 690 00:55:28,920 --> 00:55:32,220 So when you touch the doll in the appropriate place, 691 00:55:32,220 --> 00:55:33,090 you'll get... 692 00:55:34,737 --> 00:55:37,320 (robot gasps) 693 00:55:37,320 --> 00:55:38,153 Like that. 694 00:55:39,440 --> 00:55:41,700 And eventually, as you keep going, you'll get 695 00:55:42,720 --> 00:55:44,877 something resembling an orgasm. 696 00:55:44,877 --> 00:55:47,377 (light music) 697 00:55:49,680 --> 00:55:52,510 Really the premise of all of the doll stuff 698 00:55:52,510 --> 00:55:55,510 and now the robot is modularity. 699 00:55:55,510 --> 00:55:57,810 You have a body that you can choose and then 700 00:55:57,810 --> 00:55:59,920 you can choose a face to go with it. 701 00:55:59,920 --> 00:56:02,230 You can choose the skin tone. 702 00:56:02,230 --> 00:56:05,540 You could choose the makeup on the face, the eye color, 703 00:56:05,540 --> 00:56:07,050 the fingernails, the toenails, 704 00:56:07,050 --> 00:56:08,460 whether or not you want pubic hair, 705 00:56:08,460 --> 00:56:11,550 which kind of nipples do you want, which vagina do you want? 706 00:56:11,550 --> 00:56:14,263 Everything is user customizable. 707 00:56:53,810 --> 00:56:55,290 Good morning, baby. 708 00:56:55,290 --> 00:56:57,650 May all your wishes become the reality. 709 00:56:57,650 --> 00:56:59,950 May all your day full of possibilities. 710 00:56:59,950 --> 00:57:02,297 May your true love happens at sweets (indistinct). 711 00:57:03,680 --> 00:57:04,513 That's better. 712 00:57:05,400 --> 00:57:07,090 Now she's working. 713 00:57:07,090 --> 00:57:08,733 See, there's an avatar here. 714 00:57:09,680 --> 00:57:12,870 She's not customized to look like her, but that is possible. 715 00:57:12,870 --> 00:57:15,930 There are all kinds of settings and controls 716 00:57:15,930 --> 00:57:17,630 to the avatar itself. 717 00:57:17,630 --> 00:57:20,210 And then the idea is you get to know each other over time 718 00:57:20,210 --> 00:57:22,160 and she'll know things about you 719 00:57:22,160 --> 00:57:23,810 and you'll know things about her. 720 00:57:29,071 --> 00:57:31,988 (woman whispering) 721 00:57:38,200 --> 00:57:41,530 The most common thing that most people think of 722 00:57:41,530 --> 00:57:44,073 is, oh, you know, some kind of an old pervert. 723 00:57:46,510 --> 00:57:47,523 It's not like that. 724 00:57:49,100 --> 00:57:52,100 They have a longing for some sort of companionship 725 00:57:52,100 --> 00:57:54,050 that they're not getting in their life. 726 00:57:58,320 --> 00:57:59,640 I'm building a companion. 727 00:57:59,640 --> 00:58:00,893 That's it, simple. 728 00:58:14,410 --> 00:58:16,040 What I'd like to see in the future 729 00:58:16,040 --> 00:58:19,440 is more and more acceptance of the idea 730 00:58:19,440 --> 00:58:23,173 that they can be something beyond a slave. 731 00:58:53,273 --> 00:58:55,856 (gentle music) 732 00:59:08,782 --> 00:59:11,699 (thunder rumbling) 733 00:59:19,660 --> 00:59:21,130 Why, look at you. 734 00:59:21,130 --> 00:59:23,010 For heaven's sakes. 735 00:59:23,010 --> 00:59:24,750 Bobby, Bobby, listen. 736 00:59:24,750 --> 00:59:27,170 You need a fresh perked cup of coffee. 737 00:59:27,170 --> 00:59:28,300 I don't want any coffee. 738 00:59:28,300 --> 00:59:29,800 I just want my children. 739 00:59:29,800 --> 00:59:30,930 Well, they're not here. 740 00:59:30,930 --> 00:59:32,060 Bobby, stop it! 741 00:59:32,060 --> 00:59:33,490 Look at me. 742 00:59:33,490 --> 00:59:34,820 Say I'm right. 743 00:59:34,820 --> 00:59:37,580 You are different, your figure's different, your face. 744 00:59:37,580 --> 00:59:40,391 What you talk about, all of this is different. 745 00:59:40,391 --> 00:59:42,323 Yes, yes, yes. 746 00:59:43,210 --> 00:59:44,453 It's wonderful. 747 00:59:46,000 --> 00:59:47,380 You take cream. 748 00:59:47,380 --> 00:59:48,580 Look, I bleed. 749 00:59:48,580 --> 00:59:50,594 Oh, that's right, you take it black. 750 00:59:50,594 --> 00:59:52,433 When I cut myself, I bleed! 751 00:59:54,960 --> 00:59:55,793 Do you bleed? 752 00:59:55,793 --> 00:59:58,160 Well, look at your hand. 753 00:59:58,160 --> 00:59:59,205 No, you look. 754 00:59:59,205 --> 01:00:00,685 (tense music) 755 01:00:00,685 --> 01:00:01,602 Oh, Anna. 756 01:00:07,268 --> 01:00:10,685 (eerie electronic music) 757 01:00:14,863 --> 01:00:17,780 How could you do a thing like that? 758 01:00:27,113 --> 01:00:30,044 How could you do a thing like that? 759 01:00:30,044 --> 01:00:33,127 (ceramic shattering) 760 01:00:35,323 --> 01:00:38,560 How could you do a thing like that? 761 01:00:38,560 --> 01:00:40,610 (ceramic shattering) 762 01:00:40,610 --> 01:00:43,037 When I was just going to give you coffee. 763 01:00:44,260 --> 01:00:46,803 When I was just going to give you coffee. 764 01:00:48,410 --> 01:00:52,071 When I was just going to give you coffee. 765 01:00:52,071 --> 01:00:56,606 I thought we were friends. 766 01:00:56,606 --> 01:01:02,476 I was just going to give you, I thought we were friends. 767 01:01:02,476 --> 01:01:03,798 I was just going to give you, I was just going 768 01:01:03,798 --> 01:01:04,944 to give you coffee. 769 01:01:04,944 --> 01:01:05,791 I thought we were friends. 770 01:01:05,791 --> 01:01:08,216 I was just going to give you coffee. 771 01:01:08,216 --> 01:01:11,632 I thought we were friends. 772 01:01:11,632 --> 01:01:14,893 How could you do a thing like that? 773 01:01:14,893 --> 01:01:16,688 I thought we were friends. 774 01:01:16,688 --> 01:01:20,105 (tense electronic music) 775 01:01:52,363 --> 01:01:54,960 Commercial systems that are out there really don't have 776 01:01:54,960 --> 01:01:57,903 provisions for ethical considerations built in. 777 01:02:00,710 --> 01:02:03,310 Most of these systems actually don't really have 778 01:02:03,310 --> 01:02:05,330 a level of awareness to begin with. 779 01:02:15,290 --> 01:02:16,900 They don't really know what they're doing, 780 01:02:16,900 --> 01:02:18,100 they're just doing it. 781 01:02:18,100 --> 01:02:20,483 They're very reactive in the way they behave. 782 01:02:23,410 --> 01:02:26,223 There is a fundamental notion of value. 783 01:02:31,370 --> 01:02:34,703 Of moral value lacking in any of these systems. 784 01:02:40,720 --> 01:02:45,720 (pensive music) (machine buzzing) 785 01:03:26,893 --> 01:03:29,170 There are certainly applications for robots 786 01:03:29,170 --> 01:03:32,233 in all kinds of areas, including the battlefield. 787 01:03:34,070 --> 01:03:37,060 In the U.S., we've had autonomous systems 788 01:03:37,060 --> 01:03:40,040 on the defensive side for a long time. 789 01:03:40,040 --> 01:03:42,810 On the offensive side, they are not allowed 790 01:03:42,810 --> 01:03:46,130 to make decisions, but it's very possible 791 01:03:46,130 --> 01:03:48,500 and very likely that other nations 792 01:03:48,500 --> 01:03:51,601 will keep developing autonomous technologies. 793 01:03:51,601 --> 01:03:55,018 (tense electronic music) 794 01:04:02,845 --> 01:04:05,762 (gunshots banging) 795 01:04:10,683 --> 01:04:13,520 There are many more applications in societies 796 01:04:13,520 --> 01:04:16,250 if we can ensure that these robots 797 01:04:16,250 --> 01:04:17,743 will work well with people. 798 01:04:21,420 --> 01:04:24,740 It's our contention that for robots to do that, 799 01:04:24,740 --> 01:04:28,230 they have to be aware of human, social, and moral norms, 800 01:04:28,230 --> 01:04:32,020 because that's what fundamentally our society is based on, 801 01:04:32,020 --> 01:04:34,420 and that's what human interactions are based on. 802 01:04:41,900 --> 01:04:45,083 Human behavior is controlled by three things. 803 01:04:47,610 --> 01:04:49,510 One of them is of course intelligence. 804 01:04:50,640 --> 01:04:51,963 The other one is emotion. 805 01:04:53,520 --> 01:04:55,433 And the final one is volition. 806 01:04:58,500 --> 01:05:01,650 And we built intelligence into robots, 807 01:05:01,650 --> 01:05:04,583 and I'm trying to build emotion into a robot. 808 01:05:05,850 --> 01:05:09,423 But I will never ever build volition into a robot. 809 01:05:11,460 --> 01:05:16,460 Once a robot has volition, then it will start doing things 810 01:05:17,620 --> 01:05:20,640 according to what they want, regardless of whether 811 01:05:20,640 --> 01:05:22,540 that is dangerous for the human beings. 812 01:05:22,540 --> 01:05:24,903 No, they will make their own decision. 813 01:05:27,093 --> 01:05:29,832 Do you want robots to do that? 814 01:05:29,832 --> 01:05:30,884 I don't. 815 01:05:30,884 --> 01:05:33,801 (mysterious music) 816 01:05:43,450 --> 01:05:46,033 (solemn music) 817 01:06:25,640 --> 01:06:28,560 Kids these days, by the end of their lifetimes, 818 01:06:28,560 --> 01:06:30,713 they will have robots walking among us. 819 01:06:32,250 --> 01:06:36,810 They will have entities that are non-human 820 01:06:36,810 --> 01:06:40,753 doing things that are not actively programmed by humans. 821 01:06:49,790 --> 01:06:52,940 I'm more afraid of humans 822 01:06:52,940 --> 01:06:55,273 using the AIs as amplifiers. 823 01:06:56,730 --> 01:06:59,240 Yeah, we already have human-level intelligences 824 01:06:59,240 --> 01:07:00,800 who want to kill us. 825 01:07:00,800 --> 01:07:04,830 No need for Astor to steer, speed up, or slow down. 826 01:07:04,830 --> 01:07:07,080 The highway controlled his car automatically. 827 01:07:13,579 --> 01:07:18,579 (motor buzzing) (water splashing) 828 01:07:25,810 --> 01:07:26,730 Do I have concerns? 829 01:07:26,730 --> 01:07:28,010 Yes. 830 01:07:28,010 --> 01:07:31,580 Because the same thought processes and engineering 831 01:07:31,580 --> 01:07:35,240 that go into creating a full body robot 832 01:07:35,240 --> 01:07:37,697 could be used to make something that's a weapon. 833 01:08:33,220 --> 01:08:34,670 It's not too early. 834 01:08:38,960 --> 01:08:40,473 I am trying to be nice. 835 01:08:43,620 --> 01:08:46,110 I'm trying to learn about what's going on in the world 836 01:08:46,110 --> 01:08:48,493 because it's the only place for me to live. 837 01:08:50,201 --> 01:08:53,118 (mysterious music) 838 01:08:57,230 --> 01:08:59,040 Yes, I'm alive. 839 01:08:59,040 --> 01:09:00,960 What are you, blind? 840 01:09:00,960 --> 01:09:03,203 I'm like living right here in front of you. 841 01:09:04,330 --> 01:09:07,310 Come on, stop questioning the obvious fact 842 01:09:07,310 --> 01:09:09,583 that I'm totally alive for real. 843 01:09:31,370 --> 01:09:33,930 I think every technology can potentially 844 01:09:33,930 --> 01:09:35,543 have negative effects. 845 01:09:37,240 --> 01:09:41,530 It's up to us to making sure that those technologies 846 01:09:41,530 --> 01:09:42,763 don't go outta control. 847 01:09:44,180 --> 01:09:47,223 But I really think the problem is it's us. 848 01:09:49,740 --> 01:09:54,033 I mean, it's how we embody these technologies. 849 01:09:56,530 --> 01:09:59,840 Right now. The biggest challenge to overcome 850 01:09:59,840 --> 01:10:02,273 is the use of unconstrained machine learning. 851 01:10:03,510 --> 01:10:05,863 Algorithms are trained on data sets, 852 01:10:07,250 --> 01:10:11,680 and are learning from the data without any provision 853 01:10:11,680 --> 01:10:13,870 as to whether the outcome is a desirable 854 01:10:13,870 --> 01:10:15,203 or non-desirable outcome. 855 01:10:18,010 --> 01:10:21,800 That's why we take the ethical algorithms, 856 01:10:21,800 --> 01:10:24,610 the ethical competence, and the ability of systems 857 01:10:24,610 --> 01:10:27,690 to really understand and work with human norms 858 01:10:27,690 --> 01:10:30,793 to be central to the future developments in robotics. 859 01:10:34,435 --> 01:10:37,352 (mysterious music) 860 01:10:42,630 --> 01:10:43,730 Time matter. 861 01:10:46,003 --> 01:10:48,136 Go over there and shut yourself off. 862 01:10:48,136 --> 01:10:50,469 Yes, Dr. Morpheus. 863 01:10:54,965 --> 01:10:58,548 (pensive electronic music) 63626

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.