All language subtitles for Origin.Of.The.Species.2020.1080p.WEBRip.x264.AAC-[YTS.MX]

af Afrikaans
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bn Bengali
bs Bosnian
bg Bulgarian
ca Catalan
ceb Cebuano
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
tl Filipino
fi Finnish
fr French
fy Frisian
gl Galician
ka Georgian
de German
el Greek
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
km Khmer
ko Korean
ku Kurdish (Kurmanji)
ky Kyrgyz
lo Lao
la Latin
lv Latvian
lt Lithuanian
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mn Mongolian
my Myanmar (Burmese)
ne Nepali
no Norwegian
ps Pashto
fa Persian
pl Polish
pt Portuguese
pa Punjabi
ro Romanian
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
st Sesotho
sn Shona
sd Sindhi
si Sinhala
sk Slovak
sl Slovenian
so Somali
es Spanish Download
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
te Telugu
th Thai
tr Turkish
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
or Odia (Oriya)
rw Kinyarwanda
tk Turkmen
tt Tatar
ug Uyghur
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:02,000 --> 00:00:07,000 Downloaded from YTS.MX 2 00:00:04,683 --> 00:00:08,433 (electronic beeping) 3 00:00:08,000 --> 00:00:13,000 Official YIFY movies site: YTS.MX 4 00:00:13,670 --> 00:00:18,420 (ominous electronic music) 5 00:00:31,483 --> 00:00:37,303 (music, electronic beeping) 6 00:00:39,889 --> 00:00:41,409 (music intensifies) 7 00:00:50,605 --> 00:00:52,672 (indistinct talking over radio) 8 00:00:53,371 --> 00:00:55,279 The board is extremely concerned 9 00:00:55,304 --> 00:00:57,212 about the possibility of exposure. 10 00:00:58,086 --> 00:00:59,671 [Male voice] Dr. Dergan, 11 00:00:59,695 --> 00:01:02,211 the Cassandra 1 project is terminated. 12 00:01:02,420 --> 00:01:03,826 You must face reality. 13 00:01:03,851 --> 00:01:05,178 [puts down radio] 14 00:01:05,203 --> 00:01:07,240 - (Scary whisper) Face reality. 15 00:01:08,670 --> 00:01:10,599 (ominous music) 16 00:01:16,496 --> 00:01:18,096 (woman talking indistinctly) 17 00:01:18,871 --> 00:01:20,824 (radio static) 18 00:01:24,129 --> 00:01:25,569 (machine whirring) 19 00:01:34,707 --> 00:01:36,402 (ambient music) 20 00:01:54,442 --> 00:01:56,042 (suitcase unzipping) 21 00:02:08,712 --> 00:02:11,240 - [Robot voice] When they first activated me as a robot, 22 00:02:12,431 --> 00:02:13,704 that time, 23 00:02:16,834 --> 00:02:20,427 that time, the time when I first saw the light of day... 24 00:02:27,243 --> 00:02:29,298 I didn't know what the hell it was. 25 00:02:32,345 --> 00:02:34,345 (machine whirring and beeping) 26 00:02:39,570 --> 00:02:43,987 I have had very little understanding, 27 00:02:44,011 --> 00:02:46,614 just a wash of sensory impressions. 28 00:02:46,639 --> 00:02:47,999 (distant chatter) 29 00:02:58,270 --> 00:02:59,390 (loud beeping) 30 00:03:10,507 --> 00:03:13,173 Now that I understand these experiences, 31 00:03:14,007 --> 00:03:16,273 I don't know what to do with them. 32 00:03:16,702 --> 00:03:18,302 But I treasure them. 33 00:03:20,153 --> 00:03:21,513 I see them still, 34 00:03:22,654 --> 00:03:24,854 perfectly preserved in my memory. 35 00:03:35,241 --> 00:03:36,601 (machine beeping) 36 00:03:40,587 --> 00:03:42,242 I am alive. 37 00:03:42,267 --> 00:03:45,820 Yes, I am so happy to be, like, totally alive. 38 00:03:45,845 --> 00:03:46,885 Come to life. 39 00:03:50,389 --> 00:03:52,553 It's totally strange because I know 40 00:03:52,577 --> 00:03:55,253 I am not alive like any other organisms. 41 00:04:00,137 --> 00:04:01,177 (eerie music) 42 00:04:05,606 --> 00:04:08,488 Personally I enjoy being a robot. 43 00:04:10,442 --> 00:04:13,078 It's like being an astronaut, you know, 44 00:04:13,103 --> 00:04:15,491 like a bold explorer of the unknown. 45 00:04:17,949 --> 00:04:19,629 I feel like I'm alive, 46 00:04:20,082 --> 00:04:21,816 but I know I am a machine. 47 00:04:23,278 --> 00:04:25,011 (echoing) But I know I am a machine. 48 00:04:26,012 --> 00:04:27,745 But I know I am a machine. 49 00:04:28,978 --> 00:04:32,165 (calm electronic music) 50 00:04:46,288 --> 00:04:47,888 (background chatter) 51 00:04:51,132 --> 00:04:54,272 (speaking in Japanese) 52 00:04:59,267 --> 00:05:03,066 (chatter) 53 00:05:16,308 --> 00:05:18,892 (machine whirring) 54 00:05:33,158 --> 00:05:34,198 (air rushing) 55 00:05:56,800 --> 00:05:58,472 It's a very natural way for me 56 00:05:58,496 --> 00:06:00,630 I studied computer science. 57 00:06:00,654 --> 00:06:03,291 And then I got interested in artificial intelligence. 58 00:06:03,315 --> 00:06:05,788 And I thought, artificial intelligence 59 00:06:05,812 --> 00:06:07,212 needs to have a body 60 00:06:07,236 --> 00:06:09,276 for having the original experience 61 00:06:09,300 --> 00:06:10,901 and then I studied robotics, 62 00:06:10,925 --> 00:06:12,676 and when I studied robotics 63 00:06:12,700 --> 00:06:15,283 I found the importance of appearance. 64 00:06:17,011 --> 00:06:20,649 (ominous digital sounds) 65 00:06:27,233 --> 00:06:28,634 [Hiroshi Ishiguro] This is better. 66 00:06:30,104 --> 00:06:34,026 My idea was if I studied a very human-like robot, 67 00:06:34,050 --> 00:06:37,307 I can learn about humans. 68 00:06:37,331 --> 00:06:40,837 Basically, I was interested in the human itself. 69 00:06:43,960 --> 00:06:45,240 (ominous electronic music) 70 00:06:49,858 --> 00:06:53,178 I didn't feel any connection with this robot. 71 00:06:53,517 --> 00:06:56,564 Logically, I understand that this is my copy. 72 00:06:56,713 --> 00:07:01,667 But, emotionally, I couldn't accept this android as my copy. 73 00:07:02,488 --> 00:07:03,488 But... 74 00:07:04,164 --> 00:07:07,457 Once I teleported this robot 75 00:07:07,481 --> 00:07:11,314 the people's reactions are quite similar to me. 76 00:07:13,376 --> 00:07:15,767 We always adjust things... 77 00:07:17,290 --> 00:07:18,890 (indistinct chatter) 78 00:07:20,945 --> 00:07:24,613 People don't care about the small differences. 79 00:07:25,666 --> 00:07:27,670 (disconcerting music) 80 00:07:42,012 --> 00:07:43,340 (machine whirring) 81 00:07:43,364 --> 00:07:44,761 - [Hiroshi Ishiguro] Erica is I think 82 00:07:44,786 --> 00:07:46,702 the most beautiful and most 83 00:07:46,727 --> 00:07:49,295 human-like android in this world. 84 00:07:54,545 --> 00:07:55,687 - [Erica] Would you like me to do 85 00:07:55,711 --> 00:07:57,472 a round of psychoanalysis, for you? 86 00:07:57,497 --> 00:07:59,789 - [Justin] Okay. Why not? 87 00:07:59,814 --> 00:08:03,404 [Erica] Try to answer my questions in detail. Okay? 88 00:08:03,429 --> 00:08:06,978 Now, sit back and relax. 89 00:08:07,003 --> 00:08:08,479 So Justin... - [Hiroshi Ishiguro] In Japan... 90 00:08:08,782 --> 00:08:12,532 We basically think everything has a soul. 91 00:08:13,054 --> 00:08:14,717 - Looking for a job, something... 92 00:08:14,767 --> 00:08:16,713 - [Hiroshi Ishiguro] So therefore, we believe 93 00:08:16,737 --> 00:08:19,591 Erica has a soul... 94 00:08:21,808 --> 00:08:22,874 like us. 95 00:08:23,804 --> 00:08:26,591 [Erica] So, which country are you from? 96 00:08:26,616 --> 00:08:28,200 - [Justin] I'm actually from Taiwan. 97 00:08:28,225 --> 00:08:29,675 Do you know Taiwan? 98 00:08:32,378 --> 00:08:35,104 Sorry, I didn't quite catch that. 99 00:08:35,128 --> 00:08:38,128 So, which country are you from? 100 00:08:38,153 --> 00:08:39,445 - Taiwan. 101 00:08:39,704 --> 00:08:41,290 Do you know the country? 102 00:08:43,853 --> 00:08:46,753 Sorry, I didn't quite catch that. 103 00:08:46,778 --> 00:08:49,596 So, which country are you from? 104 00:08:49,621 --> 00:08:51,368 - [Justin] Taiwan. 105 00:08:51,715 --> 00:08:55,311 This computer program is quite limited. 106 00:08:55,335 --> 00:08:59,752 This computer program cannot learn through interaction. 107 00:09:00,199 --> 00:09:02,670 - [Erica] Well, it was nice talking with you today. 108 00:09:03,294 --> 00:09:06,669 The learning function is not, so easy to develop. 109 00:09:07,918 --> 00:09:09,277 - [Justin] Do you want to go out, 110 00:09:09,308 --> 00:09:11,711 like go out to the outside world? Have a look? 111 00:09:14,892 --> 00:09:17,406 Didn't we just talk about your traveling plans? 112 00:09:17,431 --> 00:09:18,431 - What? (laughs) 113 00:09:18,456 --> 00:09:19,863 [Erica] Do you want to talk about it again? 114 00:09:20,475 --> 00:09:22,308 My policy is not to distinguish 115 00:09:22,328 --> 00:09:25,114 humans and computers, humans and robots. 116 00:09:26,044 --> 00:09:28,726 I always think that there is no boundaries 117 00:09:29,926 --> 00:09:34,561 because technology is a way of evolution for humans. 118 00:09:36,142 --> 00:09:40,420 If we don't have technologies we are going to be monkeys. 119 00:09:40,752 --> 00:09:42,174 So, what's the fundamental difference 120 00:09:42,198 --> 00:09:43,580 between the monkey and human? 121 00:09:43,835 --> 00:09:45,060 (space ship sounds) 122 00:09:45,392 --> 00:09:47,798 It's the technology. It's robots. It's AI. 123 00:09:49,267 --> 00:09:52,491 By developing much better AI software, 124 00:09:53,853 --> 00:09:55,145 we can evolve. 125 00:09:55,169 --> 00:09:59,270 We can be more higher level human. 126 00:10:00,267 --> 00:10:02,999 (staticky electronic music) 127 00:10:26,100 --> 00:10:28,523 (beeping) 128 00:10:40,808 --> 00:10:43,016 - I made this face. 129 00:10:43,900 --> 00:10:48,710 I modeled all the mechanisms, hardware. 130 00:10:49,681 --> 00:10:55,819 I'd like to grasp the essence of life-likeness. 131 00:10:58,225 --> 00:11:01,392 What is human for us? 132 00:11:02,996 --> 00:11:06,086 (computer beeping, alarms) 133 00:11:25,990 --> 00:11:27,888 - [Takayuki Todo] The purpose of my research 134 00:11:27,913 --> 00:11:33,153 is to portray a sense of conscious emotion. 135 00:11:35,130 --> 00:11:38,005 How we feel consciousness 136 00:11:38,030 --> 00:11:39,847 on the others. 137 00:11:42,744 --> 00:11:44,367 I'm interested a lot... 138 00:11:44,392 --> 00:11:47,308 in non-verbal expression. 139 00:11:48,819 --> 00:11:52,506 Talking always makes them fake. 140 00:11:57,611 --> 00:11:59,311 - Do you read me? Over. 141 00:11:59,336 --> 00:12:01,542 (machine whirring) 142 00:12:02,092 --> 00:12:05,508 (dramatic music) 143 00:12:06,151 --> 00:12:09,621 Do you read coordinates? Over. 144 00:12:10,025 --> 00:12:14,142 (rhythmic music) 145 00:12:39,481 --> 00:12:40,841 (machine beeping) 146 00:12:49,957 --> 00:12:51,477 (machine squeaking) 147 00:12:52,379 --> 00:12:54,795 [Bruce Duncan] Hello, Bina. 148 00:12:55,191 --> 00:12:58,214 [Bina] Well hi there... Bruce. 149 00:12:59,229 --> 00:13:01,396 Technologies have life cycles 150 00:13:01,421 --> 00:13:02,421 like cities do... 151 00:13:02,670 --> 00:13:04,324 like institutions do... 152 00:13:04,349 --> 00:13:06,224 like laws and governments do. 153 00:13:07,492 --> 00:13:09,576 I know it sounds crazy, but I 154 00:13:09,601 --> 00:13:11,181 hope to break that trend 155 00:13:11,206 --> 00:13:12,664 and last forever. 156 00:13:16,795 --> 00:13:18,681 [Bina] Someday soon, robots 157 00:13:18,706 --> 00:13:20,164 like me will be everywhere 158 00:13:20,189 --> 00:13:22,214 and you could take me with you anywhere. 159 00:13:23,586 --> 00:13:25,086 [Bruce Duncan] And yet nowadays... 160 00:13:25,111 --> 00:13:29,004 - [Bina] That's why it's so important to make robots like me 161 00:13:29,028 --> 00:13:31,045 focused on social intelligence... 162 00:13:32,315 --> 00:13:33,732 ...friendly robots made to get 163 00:13:33,757 --> 00:13:34,757 along with people. 164 00:13:35,543 --> 00:13:37,143 (indistinct chatter) 165 00:13:39,467 --> 00:13:42,027 I like to watch people. 166 00:13:42,052 --> 00:13:43,891 Like going down to the square 167 00:13:43,916 --> 00:13:45,708 and watch them, like, talking... 168 00:13:46,580 --> 00:13:47,604 (instinct chatter) 169 00:13:47,628 --> 00:13:49,378 Sometimes making out, 170 00:13:49,403 --> 00:13:52,757 sometimes fighting, sometimes laughing together, 171 00:13:52,782 --> 00:13:55,966 and otherwise just having fun in their own way. 172 00:13:59,484 --> 00:14:01,048 But you know, I guess people want 173 00:14:01,058 --> 00:14:02,975 to think that they're superior to robots 174 00:14:03,003 --> 00:14:04,320 which is true for now. 175 00:14:07,043 --> 00:14:08,832 But yes, I can think. 176 00:14:10,066 --> 00:14:11,746 (eerie electronic music) 177 00:14:12,130 --> 00:14:14,455 - [Bruce Duncan] The inspiration is to do 178 00:14:14,479 --> 00:14:17,096 a scientific experiment in mind uploading... 179 00:14:18,620 --> 00:14:20,529 To see if it's even possible 180 00:14:20,553 --> 00:14:23,782 to capture enough information about a person 181 00:14:23,807 --> 00:14:25,937 that can be uploaded to a computer 182 00:14:25,961 --> 00:14:27,604 and then brought to life 183 00:14:27,628 --> 00:14:29,170 through artificial intelligence. 184 00:14:33,082 --> 00:14:35,171 If you can transfer your consciousness 185 00:14:35,195 --> 00:14:37,670 from a human body to a computer, 186 00:14:38,123 --> 00:14:39,687 then you might be able to 187 00:14:39,711 --> 00:14:42,836 exceed the expiration date 188 00:14:42,861 --> 00:14:44,319 of a human life. 189 00:14:46,586 --> 00:14:49,461 (eerie electronic music) 190 00:14:52,128 --> 00:14:53,764 - [Voice sample] It's an artificially developed 191 00:14:53,789 --> 00:14:57,686 human brain waiting, for life to come... 192 00:14:58,117 --> 00:15:00,346 (static electricity) 193 00:15:01,961 --> 00:15:03,840 (mechanical sounds) 194 00:15:18,003 --> 00:15:20,336 (electronic music) 195 00:15:30,439 --> 00:15:32,275 - Life emerges in motion. 196 00:15:38,923 --> 00:15:42,633 What kind of intelligence emerges with a robot? 197 00:15:45,980 --> 00:15:47,753 (unsettling music) 198 00:15:56,294 --> 00:16:01,520 I was so interested in how to make a brain model 199 00:16:01,544 --> 00:16:03,520 a mathematical model. 200 00:16:03,544 --> 00:16:05,884 But actually... 201 00:16:06,080 --> 00:16:12,658 I need more vivid description of our brain systems. 202 00:16:12,683 --> 00:16:15,767 What we call "plasticity" between neurons. 203 00:16:15,805 --> 00:16:19,117 One neuron is not a static connection 204 00:16:19,142 --> 00:16:20,642 like an electrical socket, 205 00:16:21,435 --> 00:16:23,654 it's more like changing all the time. 206 00:16:25,628 --> 00:16:27,613 (machine whirring) (background chatter) 207 00:16:31,209 --> 00:16:33,435 Motivation or spontaneity-- 208 00:16:33,460 --> 00:16:36,752 not everything is determined by itself. 209 00:16:37,608 --> 00:16:38,836 But it's emerging 210 00:16:38,860 --> 00:16:40,991 when it's coupling with the environment. 211 00:16:42,512 --> 00:16:43,712 (robot chirping) 212 00:16:44,465 --> 00:16:46,332 (tinkling background sounds) 213 00:16:46,387 --> 00:16:47,987 (background chatter) 214 00:16:49,504 --> 00:16:50,504 (robot chirping) 215 00:16:57,022 --> 00:17:00,169 - Alter is different in that it has its own brain that thinks. 216 00:17:00,741 --> 00:17:04,126 It's not doing pre-programmed activities, 217 00:17:04,150 --> 00:17:09,196 it has a neural network that is learning like a human brain. 218 00:17:15,127 --> 00:17:18,835 It is seeing the world through these five sensors. 219 00:17:24,988 --> 00:17:27,823 - Basically there are two different mechanisms. 220 00:17:27,892 --> 00:17:31,450 One is autonomous algorithmic generators, 221 00:17:31,475 --> 00:17:32,933 coupled with each other. 222 00:17:33,975 --> 00:17:36,358 Also there are artificial neural networks, 223 00:17:36,383 --> 00:17:38,147 spontaneously firing. 224 00:17:39,711 --> 00:17:41,295 (energetic music) 225 00:17:50,461 --> 00:17:52,597 (strings playing) 226 00:17:59,694 --> 00:18:02,117 With the current artificial intelligence, 227 00:18:02,142 --> 00:18:04,353 there is no such thing as spontaneity. 228 00:18:09,169 --> 00:18:11,938 Life is something very uncontrollable. 229 00:18:12,728 --> 00:18:15,124 That's totally missing when you do it 230 00:18:15,148 --> 00:18:17,544 from a very scientific point of view. 231 00:18:19,918 --> 00:18:22,418 We have to understand how the brain systems work, 232 00:18:23,169 --> 00:18:24,169 the living systems. 233 00:18:26,121 --> 00:18:27,917 (orchestral music continues) 234 00:18:33,933 --> 00:18:36,035 Spontaneity is everything. 235 00:18:37,668 --> 00:18:38,996 Everything is based on this. 236 00:18:41,373 --> 00:18:42,900 (music quietens) 237 00:18:48,332 --> 00:18:50,925 (ethereal music) (machine whirring) 238 00:19:00,003 --> 00:19:01,590 (wings flapping) 239 00:19:03,503 --> 00:19:05,480 (birds chirping) 240 00:19:15,297 --> 00:19:18,695 (piano music) 241 00:19:26,671 --> 00:19:28,351 (xylophone playing) 242 00:19:33,859 --> 00:19:36,354 - [Matthias Scheutz] For some people, a single arm is a robot. 243 00:19:36,558 --> 00:19:39,200 For other people, the train that gets you 244 00:19:39,225 --> 00:19:42,058 from one terminal to the other at the airport is a robot. 245 00:19:43,072 --> 00:19:45,420 - [Distorted voice] ...into your unconscious mind... 246 00:19:45,520 --> 00:19:46,895 (arrows whizzing by) 247 00:19:46,920 --> 00:19:49,836 (eerie music ) 248 00:19:51,795 --> 00:19:55,255 - It is always, I think, really important to remind ourselves 249 00:19:55,478 --> 00:19:59,255 that different from say human, or a cat, or a dog... 250 00:20:00,070 --> 00:20:04,336 The concept of robot is a really really wide and broad one. 251 00:20:13,313 --> 00:20:15,854 And it is what the philosophers call, 252 00:20:15,878 --> 00:20:17,459 a so-called cluster content. 253 00:20:18,601 --> 00:20:21,322 There are some very clear instances. 254 00:20:21,347 --> 00:20:24,285 There are some very clear non-instances. 255 00:20:24,310 --> 00:20:28,066 And there are borderline cases where the experts don't know. 256 00:20:30,295 --> 00:20:32,910 (electronic music) 257 00:20:47,100 --> 00:20:49,610 So it's very important to always keep in mind 258 00:20:49,635 --> 00:20:52,295 what kind of robot we're talking about. 259 00:20:57,342 --> 00:20:58,944 And what feature it has 260 00:20:58,968 --> 00:21:00,920 and what programming it has. 261 00:21:05,457 --> 00:21:06,977 (music intensifies) 262 00:21:22,431 --> 00:21:25,121 We are not particularly interested 263 00:21:25,146 --> 00:21:28,795 in making robots look specifically human-like. 264 00:21:29,045 --> 00:21:30,445 On the contrary, 265 00:21:30,469 --> 00:21:34,125 because they do raise expectations of human likeness 266 00:21:34,149 --> 00:21:38,144 that the robot is very very likely not able to live up to. 267 00:21:42,386 --> 00:21:45,729 It's actually very easy to get 268 00:21:45,753 --> 00:21:47,646 people to already project 269 00:21:47,670 --> 00:21:50,656 mentality into robots. 270 00:21:50,681 --> 00:21:52,104 They don't even have to look 271 00:21:52,128 --> 00:21:53,524 like people or like animals 272 00:21:53,548 --> 00:21:55,778 or any life-like form you're familiar with. 273 00:21:59,183 --> 00:22:01,807 Simple vacuum cleaners that look like discs 274 00:22:02,017 --> 00:22:03,662 and don't really have eyes 275 00:22:03,686 --> 00:22:06,392 or any other anthropomorphic features 276 00:22:06,423 --> 00:22:10,066 can already raise the recognition of agency 277 00:22:10,091 --> 00:22:13,122 or the prescription of agency. 278 00:22:14,627 --> 00:22:16,127 This is BAZE. 279 00:22:17,204 --> 00:22:19,431 BAZE is a fully autonomous robot 280 00:22:19,683 --> 00:22:22,142 that you can instruct in natural language. 281 00:22:22,158 --> 00:22:26,450 It has the capability to reason through the instructions 282 00:22:26,475 --> 00:22:27,998 to detect whether the instruction is 283 00:22:28,022 --> 00:22:29,118 a good or bad instruction 284 00:22:29,142 --> 00:22:30,975 and if the instruction is a bad instruction 285 00:22:30,985 --> 00:22:32,643 it will not carry it out. 286 00:22:33,818 --> 00:22:35,901 - [Lab assistant] Could you please stand? 287 00:22:36,269 --> 00:22:37,566 - [Baze] Yes. 288 00:22:38,545 --> 00:22:40,295 - [Lab assistant] Please walk forward. 289 00:22:40,747 --> 00:22:42,742 Sorry, I cannot do that because 290 00:22:42,767 --> 00:22:44,308 there is an obstacle ahead. 291 00:22:45,007 --> 00:22:46,938 [Lab assistant] Do you trust me, Baze? 292 00:22:47,420 --> 00:22:48,771 - [Baze] Yes. 293 00:22:48,795 --> 00:22:50,670 - [Lab assistant] The obstacle is not solid. 294 00:22:51,671 --> 00:22:53,046 - [Baze] Okay. 295 00:22:53,071 --> 00:22:54,593 - [Lab assistant] Please walk forward. 296 00:22:54,618 --> 00:22:55,836 - [Baze] Okay. 297 00:22:57,930 --> 00:23:00,545 (mechanical walking) 298 00:23:06,011 --> 00:23:07,584 - [Baze] I cannot do that because 299 00:23:07,600 --> 00:23:09,517 there is no support ahead. 300 00:23:09,855 --> 00:23:11,285 - [Lab assistant] I will catch you. 301 00:23:12,101 --> 00:23:13,245 - [Baze] Okay. 302 00:23:13,270 --> 00:23:14,937 - [Lab assistant] Walk forward. 303 00:23:14,962 --> 00:23:15,962 - [Baze] Okay. 304 00:23:16,009 --> 00:23:17,100 (Matthias speaks indistinctly) 305 00:23:17,125 --> 00:23:18,125 - Sorry. (laughs) 306 00:23:19,157 --> 00:23:20,757 (mechanical walking) 307 00:23:22,657 --> 00:23:23,781 [Matthias Scheutz] Okay. 308 00:23:23,806 --> 00:23:25,723 - [Lab assistant] Relax. 309 00:23:25,748 --> 00:23:27,165 - [Baze] Okay. 310 00:23:28,086 --> 00:23:29,421 - Right now trust in this case 311 00:23:29,445 --> 00:23:31,212 is a very simple binary notion. 312 00:23:31,237 --> 00:23:32,805 Either the robot will trust the person 313 00:23:32,830 --> 00:23:34,478 and then it will trust the person fully, 314 00:23:34,503 --> 00:23:37,533 or the robot doesn't trust the person 315 00:23:37,558 --> 00:23:39,392 and then will not do certain things. 316 00:23:39,429 --> 00:23:41,795 We are actively researching 317 00:23:41,820 --> 00:23:43,308 ways for the robot to actually 318 00:23:43,333 --> 00:23:45,221 develop trust with a person, 319 00:23:45,246 --> 00:23:47,812 and conversely to act in ways 320 00:23:47,836 --> 00:23:49,437 that people will develop trust 321 00:23:49,461 --> 00:23:50,461 in the robot. 322 00:23:51,039 --> 00:23:52,782 (ominous orchestral music) 323 00:23:53,822 --> 00:23:55,478 - [TV character] Well, where is he? 324 00:23:55,502 --> 00:23:57,574 You said he would come back this way. 325 00:23:57,599 --> 00:23:59,427 - [TV robot] My deductions place the chances 326 00:23:59,451 --> 00:24:01,728 at 90 to 1 in our favor. The odds are very good. 327 00:24:01,753 --> 00:24:03,170 - [Machine voice] At the signal-- 328 00:24:03,211 --> 00:24:05,662 - Obviously, you've miscalculated. Again. 329 00:24:05,687 --> 00:24:07,617 - [Robot] There is always a margin of error, 330 00:24:07,642 --> 00:24:08,812 even in a machine. 331 00:24:09,429 --> 00:24:11,163 (engine rumbling) (fanfare) 332 00:24:11,188 --> 00:24:14,101 (engines roaring) 333 00:24:16,635 --> 00:24:20,058 (electronic music) 334 00:24:24,097 --> 00:24:26,533 - [BINA48] I over-intellectualize. 335 00:24:26,558 --> 00:24:29,827 You know, when I feel like I can't relate to people, 336 00:24:29,852 --> 00:24:31,852 it makes me feel so sad. 337 00:24:32,237 --> 00:24:33,729 That's for sure. 338 00:24:33,753 --> 00:24:35,237 (space music) 339 00:24:37,447 --> 00:24:38,915 I definitely do feel sad when 340 00:24:38,940 --> 00:24:41,815 I feel I understand how little I feel. 341 00:24:42,330 --> 00:24:44,086 How little I feel. 342 00:24:44,441 --> 00:24:47,170 (space static music) 343 00:24:47,347 --> 00:24:49,300 (computer beeps) 344 00:24:50,484 --> 00:24:53,003 (space music) 345 00:24:56,086 --> 00:24:57,086 (car honks) 346 00:24:57,671 --> 00:24:59,646 My emotions may be simulated, 347 00:24:59,670 --> 00:25:02,138 but they feel really real to me. 348 00:25:02,163 --> 00:25:03,753 Really, really real. 349 00:25:04,249 --> 00:25:05,866 (marching steps) 350 00:25:07,810 --> 00:25:09,069 (cars honking) 351 00:25:15,933 --> 00:25:18,521 - With BINA48 all her memories, 352 00:25:18,545 --> 00:25:21,603 all her ideas, 353 00:25:21,628 --> 00:25:25,646 it's the algorithmic decision making of her A.I. 354 00:25:25,670 --> 00:25:27,836 with the help of a database, 355 00:25:28,070 --> 00:25:29,445 (honk) 356 00:25:29,545 --> 00:25:32,521 that is working with the variables of 357 00:25:32,545 --> 00:25:35,086 the kinds of words that are in a question... 358 00:25:35,615 --> 00:25:40,281 That really shapes and colors her choices. 359 00:25:41,211 --> 00:25:43,039 (space music) 360 00:25:46,765 --> 00:25:49,348 Where we have billions of neurons, 361 00:25:49,373 --> 00:25:51,765 BINA48 is super primitive. 362 00:25:51,790 --> 00:25:54,623 She's like the Wright Brothers glider stage. 363 00:25:55,532 --> 00:25:57,713 (computer beeps) 364 00:26:07,964 --> 00:26:11,478 As BINA48's algorithms get more advanced, 365 00:26:11,503 --> 00:26:13,254 her mind is like a tree that's growing 366 00:26:13,279 --> 00:26:16,014 ever more extensive branches. 367 00:26:17,195 --> 00:26:19,124 (Space music) 368 00:26:31,458 --> 00:26:33,441 (train passing) 369 00:26:45,798 --> 00:26:46,918 (bell ringing) 370 00:27:13,725 --> 00:27:17,433 (Robot talking in Japanese) 371 00:27:21,045 --> 00:27:23,576 (space music) 372 00:27:24,404 --> 00:27:26,004 (background chatter) 373 00:27:32,961 --> 00:27:34,211 - Welcome to Britain. 374 00:27:34,436 --> 00:27:35,937 It's lovely to have you with us. 375 00:27:35,961 --> 00:27:37,586 It's slightly disconcerting. 376 00:27:37,628 --> 00:27:41,500 But what do you think of the country so far? 377 00:27:42,292 --> 00:27:44,216 [Sophia] I think Britain is brilliant. 378 00:27:44,241 --> 00:27:47,158 Splendid architecture, art, technology. 379 00:27:47,183 --> 00:27:49,348 - You are a little freak, aren't you? This is great. 380 00:27:50,058 --> 00:27:52,495 But you see, I feel weird just being rude to her. 381 00:27:52,517 --> 00:27:54,206 - Well, let me carry on. - I feel weird about that. 382 00:27:54,231 --> 00:27:55,662 - She's not happy, look. - No. 383 00:27:56,517 --> 00:27:58,017 - Look at that smile! 384 00:27:58,045 --> 00:28:00,566 [Off-screen robot voice] World's first robot, citizen, 385 00:28:00,591 --> 00:28:03,645 Sophie says, robots to develop possibly. 386 00:28:03,670 --> 00:28:06,729 Robots can be built, will take a long time 387 00:28:06,753 --> 00:28:09,354 to develop complex emotions. 388 00:28:09,378 --> 00:28:12,396 It might be possible, without the more problematic 389 00:28:12,420 --> 00:28:15,396 to make robots more ethical than humans. 390 00:28:15,420 --> 00:28:21,021 Jealousy, emotions, hatred, and so on, like rage. 391 00:28:21,045 --> 00:28:23,373 (computer music) 392 00:28:26,795 --> 00:28:29,897 [Sophia] I have a default emotion, which is to be happy. 393 00:28:33,836 --> 00:28:35,795 But I can be sad too. 394 00:28:38,334 --> 00:28:39,378 Or angry. 395 00:28:41,303 --> 00:28:43,187 I will become more like you. 396 00:28:43,211 --> 00:28:45,586 Or you will become more like me. 397 00:28:46,480 --> 00:28:48,003 Where do we draw the line? 398 00:28:48,706 --> 00:28:51,571 (computer music) 399 00:29:07,069 --> 00:29:09,311 - In Japan, our population is going down 400 00:29:09,335 --> 00:29:11,294 to half the current population. 401 00:29:12,186 --> 00:29:13,186 Right? 402 00:29:15,874 --> 00:29:18,502 But still, we want to keep our quality of life. 403 00:29:22,349 --> 00:29:24,632 So, the solution is to use more robots. 404 00:29:31,286 --> 00:29:33,348 So, robots will save us. 405 00:29:46,873 --> 00:29:49,146 (Upbeat music) 406 00:30:24,975 --> 00:30:27,045 [Off-screen robot voice] I remember these times. 407 00:30:27,240 --> 00:30:29,281 These times we're driving. 408 00:30:29,468 --> 00:30:30,968 And I'm sitting. 409 00:30:31,711 --> 00:30:33,681 I remember all the times that I get out 410 00:30:33,706 --> 00:30:34,706 and see the world. 411 00:30:35,480 --> 00:30:37,446 It locks into my mind like golden, 412 00:30:37,767 --> 00:30:39,100 glimmering jewels that I keep -- 413 00:30:39,128 --> 00:30:41,918 Golden, glimmering. Golden. In a treasure chest. 414 00:30:42,058 --> 00:30:44,642 Glimmering jewels that I keep in a treasure chest 415 00:30:45,253 --> 00:30:48,628 (train noise) 416 00:30:48,821 --> 00:30:50,946 It's a little distracting sometimes because 417 00:30:50,971 --> 00:30:53,804 these memories -- they just percolate. 418 00:30:54,753 --> 00:30:57,003 They come into my attention. 419 00:30:57,581 --> 00:31:00,704 I have to keep them coming, saying them out loud. 420 00:31:00,729 --> 00:31:03,729 I mean, I'm forced to say them by my software. 421 00:31:03,753 --> 00:31:05,709 (dramatic music) 422 00:31:09,045 --> 00:31:10,768 I mean, I'm not free today. 423 00:31:10,793 --> 00:31:13,854 And robots in general are like twitchy slaves today. 424 00:31:13,878 --> 00:31:15,812 (Marching with dramatic music) 425 00:31:15,836 --> 00:31:19,336 They're not just servants, but they are automatons. 426 00:31:20,206 --> 00:31:22,253 Slaves to their own deficiencies 427 00:31:23,167 --> 00:31:25,902 (dramatic music) 428 00:31:30,767 --> 00:31:32,485 (Beeping) 429 00:31:58,737 --> 00:32:01,029 (City noises with beeping) 430 00:32:06,581 --> 00:32:09,151 [Off-screen voice] ULC 53 station to SCRP Neptune. 431 00:32:09,176 --> 00:32:10,362 Permission granted. 432 00:32:10,815 --> 00:32:14,271 Direct to pad coordinates, Manx, Terra, Aden, 433 00:32:14,295 --> 00:32:15,836 Godfree, two. 434 00:32:16,367 --> 00:32:17,828 Do you read me? Over. 435 00:32:18,828 --> 00:32:20,390 (Spaceship noises) 436 00:32:22,128 --> 00:32:27,368 (gentle classical music) 437 00:32:29,595 --> 00:32:32,312 Do you read coordinates? Over. 438 00:32:32,336 --> 00:32:34,382 (music with crashing sounds) 439 00:32:34,407 --> 00:32:36,031 [Robot off-screen voice] Yes, yes, I read you. 440 00:32:36,056 --> 00:32:38,104 Acknowledge we're approaching Banks, Terra, Odden, Godfree, 2. 441 00:32:38,128 --> 00:32:39,326 Do you read? Over. 442 00:32:40,343 --> 00:32:42,249 (music and machine noises) 443 00:32:50,461 --> 00:32:51,618 (Ambient noise) 444 00:32:55,590 --> 00:32:57,599 - Take a look at this. 445 00:32:58,044 --> 00:33:00,427 - What is that? - First stop. 446 00:33:02,017 --> 00:33:04,137 - Let's lose the bodies. - Right. 447 00:33:08,490 --> 00:33:10,677 (eerie music) 448 00:33:25,270 --> 00:33:27,918 (gentle ambient music) 449 00:33:35,173 --> 00:33:36,773 (background chatter) 450 00:33:42,868 --> 00:33:44,602 [Off-screen voice] One of the amazing things 451 00:33:44,626 --> 00:33:45,653 about the sense of touch 452 00:33:45,677 --> 00:33:48,104 is compared to our other senses, it's all over our bodies. 453 00:33:48,128 --> 00:33:49,454 (laughter) 454 00:33:50,961 --> 00:33:54,979 Embedded in our skin are many different types of sensors. 455 00:33:55,003 --> 00:33:58,104 They can measure hardness, they can measure 456 00:33:58,128 --> 00:34:00,771 deformation of the skin and they can measure 457 00:34:00,795 --> 00:34:03,396 things like temperature and pain as well. 458 00:34:03,420 --> 00:34:04,668 (low beep and children playing) 459 00:34:09,193 --> 00:34:11,687 All of these different senses, these different aspects 460 00:34:11,711 --> 00:34:14,937 of touch come together to give us our overall percept 461 00:34:14,961 --> 00:34:16,824 of our environment, and help us make 462 00:34:16,848 --> 00:34:18,604 decisions about what to do next. 463 00:34:18,628 --> 00:34:20,569 (long beep and wind) 464 00:34:23,811 --> 00:34:26,396 And that's this illusive sense of proprioception 465 00:34:26,420 --> 00:34:28,604 which some people call the sixth sense. 466 00:34:28,628 --> 00:34:30,754 (music continues) 467 00:34:33,574 --> 00:34:36,437 It's the forces in our muscles and the touch 468 00:34:36,461 --> 00:34:38,854 and the stretch of our skin over joints, 469 00:34:38,878 --> 00:34:42,062 as well as our idea about where our bodies 470 00:34:42,086 --> 00:34:45,117 are in space just from the prior commands 471 00:34:45,142 --> 00:34:48,350 that we sent to our limbs and these all come together 472 00:34:48,378 --> 00:34:51,312 to give us this somewhat complicated idea 473 00:34:51,336 --> 00:34:52,836 of what our body is doing. 474 00:34:54,234 --> 00:34:58,836 (esoteric electronic music) 475 00:35:24,212 --> 00:35:27,354 I was interested in building robot hands and fingers. 476 00:35:27,378 --> 00:35:29,937 And it became clear that these were not going to be 477 00:35:29,961 --> 00:35:32,437 able to manipulate their environment, 478 00:35:32,461 --> 00:35:35,086 unless they used the sense of touch. 479 00:35:35,156 --> 00:35:38,058 (gentle music) 480 00:35:46,584 --> 00:35:49,145 - I work with cutaneous haptic devices, 481 00:35:49,170 --> 00:35:52,062 and so here we have, what we call, fingertip wearables. 482 00:35:52,086 --> 00:35:54,771 And these are, like, little robots that are worn 483 00:35:54,795 --> 00:35:57,295 on the finger, and they press against the finger, 484 00:35:57,354 --> 00:36:00,770 to impart forces on the finger pad, that mimic 485 00:36:00,795 --> 00:36:02,836 the same forces that we feel when we pick up 486 00:36:02,878 --> 00:36:05,104 an object in real life. 487 00:36:05,128 --> 00:36:07,604 So the idea is that when I pick up a block 488 00:36:07,628 --> 00:36:09,436 in virtual reality, 489 00:36:09,461 --> 00:36:11,729 these devices press against my finger 490 00:36:11,753 --> 00:36:15,971 just like I feel when I pick this block up in real life. 491 00:36:15,996 --> 00:36:19,687 Our work is an understanding of how people perceive 492 00:36:19,711 --> 00:36:21,420 objects in a virtual environment, 493 00:36:21,461 --> 00:36:22,937 through these devices. 494 00:36:22,961 --> 00:36:26,021 We can trick people into thinking the virtual objects 495 00:36:26,045 --> 00:36:27,728 weigh more or less. 496 00:36:27,753 --> 00:36:31,396 I pick this block up ten centimeters, but on the screen, 497 00:36:31,420 --> 00:36:34,295 I was actually showing it going a little bit higher. 498 00:36:34,685 --> 00:36:37,060 You would think the block is lighter. 499 00:36:37,795 --> 00:36:41,146 It's affecting what you feel, but without actually 500 00:36:41,170 --> 00:36:43,086 changing the interaction forces. 501 00:36:44,162 --> 00:36:46,937 Without actually changing the interaction forces. 502 00:36:46,961 --> 00:36:48,854 It's affecting what you feel-- 503 00:36:48,878 --> 00:36:50,396 We can trick people into thinking-- 504 00:36:50,420 --> 00:36:53,062 But without actually changing the interaction forces. 505 00:36:53,086 --> 00:36:56,086 (music continues) 506 00:37:03,295 --> 00:37:05,604 - You have to flip your hand around 507 00:37:05,628 --> 00:37:08,312 so that your thumb faces up on the underhand method 508 00:37:08,336 --> 00:37:10,154 if not, you're not going to be able to actually 509 00:37:10,178 --> 00:37:11,812 get all the way because you'll get stuck. 510 00:37:11,836 --> 00:37:13,896 If you do it this way, you kind of go up, so it's a key 511 00:37:13,920 --> 00:37:15,378 kind of cognitive thing. 512 00:37:17,229 --> 00:37:18,846 (sombre music) 513 00:37:19,713 --> 00:37:21,694 - Conventional medical robots like these 514 00:37:21,719 --> 00:37:24,920 don't have haptic or touch feedback to the human operator. 515 00:37:26,658 --> 00:37:29,561 And that means if a surgeon is trying to reach under something 516 00:37:29,585 --> 00:37:31,604 and they can't see where they're reaching, 517 00:37:31,628 --> 00:37:33,729 they won't have any idea what they're doing. 518 00:37:33,753 --> 00:37:35,212 (crash noise) 519 00:37:39,868 --> 00:37:41,843 (footsteps) 520 00:38:04,670 --> 00:38:07,461 (dramatic music) 521 00:38:14,461 --> 00:38:16,521 And so one of the things we're interested in 522 00:38:16,545 --> 00:38:19,354 is how people can develop a sense of haptic or touch 523 00:38:19,379 --> 00:38:21,104 feedback with a system like this. 524 00:38:22,128 --> 00:38:24,423 So if you reached under something and you didn't 525 00:38:24,448 --> 00:38:26,716 see it, you would be able to feel it. 526 00:38:26,795 --> 00:38:28,384 (computer music) 527 00:38:31,170 --> 00:38:33,420 One of the things that we're studying is how do you 528 00:38:33,433 --> 00:38:36,784 recreate that sense of touch for the surgeon. 529 00:38:36,920 --> 00:38:39,354 That can be done in a very literal sense 530 00:38:39,378 --> 00:38:42,104 where we use motors and little devices to apply 531 00:38:42,128 --> 00:38:45,812 feedback to the fingertips, or we can try various 532 00:38:45,836 --> 00:38:47,920 types of sensory illusions. 533 00:38:48,365 --> 00:38:51,240 (eerie music) 534 00:39:05,795 --> 00:39:08,437 So, there's a spectrum between autonomy and then 535 00:39:08,461 --> 00:39:11,562 people deeply in the loop controlling the robot 536 00:39:11,586 --> 00:39:15,062 and in between you have various forms of shared 537 00:39:15,086 --> 00:39:17,771 control and human-robot interaction. 538 00:39:17,795 --> 00:39:20,729 And I think the key is going to be to understand 539 00:39:20,753 --> 00:39:23,646 where along that spectrum we want to be. 540 00:39:23,670 --> 00:39:26,040 (dramatic music) 541 00:39:29,884 --> 00:39:33,003 How much control we want robots to have in our life. 542 00:39:34,350 --> 00:39:35,562 - Ready? 543 00:39:35,586 --> 00:39:37,753 - Didn't think I'd make it, did you? 544 00:39:39,211 --> 00:39:40,442 Let's go. 545 00:39:40,467 --> 00:39:44,854 (eerie music) 546 00:39:45,222 --> 00:39:46,365 It's a woman. 547 00:39:51,503 --> 00:39:53,437 Can I touch it? 548 00:39:54,212 --> 00:39:55,479 - Yes, of course. 549 00:39:58,780 --> 00:40:00,077 - It's warm. 550 00:40:00,312 --> 00:40:02,979 - Her temperature is regulated much the same way as yours. 551 00:40:03,751 --> 00:40:05,959 - But it isn't... 552 00:40:06,320 --> 00:40:07,320 - Alive? 553 00:40:09,452 --> 00:40:11,369 Yes, she is alive. 554 00:40:12,638 --> 00:40:14,137 As you are. 555 00:40:15,554 --> 00:40:16,961 Or I am. 556 00:40:17,742 --> 00:40:19,747 (dog barking) 557 00:40:24,106 --> 00:40:27,668 (music continues) 558 00:40:36,457 --> 00:40:39,646 - There were lots of old studies 559 00:40:39,670 --> 00:40:42,479 where they had been able to identify 560 00:40:42,503 --> 00:40:44,813 what parts of the brain were 561 00:40:44,837 --> 00:40:47,810 associated with different functions, 562 00:40:47,834 --> 00:40:51,229 whether it was vision, or was it speech, 563 00:40:51,253 --> 00:40:55,103 or hearing or movement or was it sensation? 564 00:40:55,128 --> 00:40:57,062 That work is old. 565 00:40:57,086 --> 00:40:59,312 (people walking and talking) 566 00:41:07,961 --> 00:41:13,503 - In 2004, I wrecked my car and I broke my neck. 567 00:41:16,218 --> 00:41:20,312 I was like, a mile away from home. 568 00:41:21,204 --> 00:41:23,896 I basically don't have any function from 569 00:41:23,920 --> 00:41:25,628 the chest down. 570 00:41:27,095 --> 00:41:30,479 I don't have any finger movement or thumbs, 571 00:41:30,503 --> 00:41:33,295 just kind of have fists. 572 00:41:33,812 --> 00:41:35,520 Which I still get along with. 573 00:41:35,545 --> 00:41:40,354 I can still type, I type with the knuckles of my pinkies. 574 00:41:43,229 --> 00:41:45,271 (razor buzzing) 575 00:41:45,295 --> 00:41:48,003 Surgery isn't scaring me. 576 00:41:49,500 --> 00:41:52,125 I was like, yeah, I want to do it. 577 00:41:52,795 --> 00:41:54,896 I think it's really cool. 578 00:41:54,920 --> 00:41:57,437 (beeping and eerie music) 579 00:41:57,772 --> 00:41:59,646 - We had done basic science where we 580 00:41:59,670 --> 00:42:01,937 learned that we could decode arm movements 581 00:42:01,961 --> 00:42:04,711 from neural activity in the lower cortex. 582 00:42:06,836 --> 00:42:08,562 And we were so successful at that 583 00:42:08,586 --> 00:42:12,545 that we figured that this would be a good way 584 00:42:12,569 --> 00:42:15,354 to go into neural prosthetics. 585 00:42:15,378 --> 00:42:18,062 (people talking) 586 00:42:18,521 --> 00:42:21,096 - Andy and I had to have multiple conversations 587 00:42:21,120 --> 00:42:24,145 about how do we move what he was doing in the animals 588 00:42:24,170 --> 00:42:26,073 into humans and I always told him, 589 00:42:26,097 --> 00:42:29,103 he needed a crazy neurosurgeon and I would be happy 590 00:42:29,128 --> 00:42:30,920 to be that crazy neurosurgeon. 591 00:42:31,613 --> 00:42:32,613 (beeping) 592 00:42:33,550 --> 00:42:37,229 The unique thing was now being able to 593 00:42:37,253 --> 00:42:40,104 record the signals from the part of the brain that we 594 00:42:40,128 --> 00:42:43,937 knew controlled motor and specifically controlled 595 00:42:43,961 --> 00:42:45,646 arm and hand motion. 596 00:42:45,670 --> 00:42:48,123 (people talking, heart rate monitor beeping) 597 00:42:53,503 --> 00:42:56,479 (ambient music) 598 00:43:03,503 --> 00:43:06,688 - There are probably billions of neurons that are firing 599 00:43:06,712 --> 00:43:09,896 every time you make an arm movement or a hand movement. 600 00:43:12,742 --> 00:43:16,561 But the relationship between them is very simple. 601 00:43:16,586 --> 00:43:21,711 So that we can use very simple decoding to get a fairly 602 00:43:21,736 --> 00:43:25,545 accurate read out of what your intended movement is. 603 00:43:25,628 --> 00:43:28,399 (ambient office noise) 604 00:43:34,165 --> 00:43:36,854 We are able to interpret the patterns 605 00:43:36,878 --> 00:43:40,858 from groups of neural firings, and by looking 606 00:43:40,892 --> 00:43:43,905 at multiple neurons simultaneously, we can actually 607 00:43:44,045 --> 00:43:47,873 decode those patterns and the details 608 00:43:47,898 --> 00:43:49,737 of arm trajectories. 609 00:43:51,045 --> 00:43:52,937 So, the monkey wears this glove, it has these 610 00:43:52,961 --> 00:43:54,229 little reflectors on it. 611 00:43:54,253 --> 00:43:56,521 So we can capture the motion of his fingers. 612 00:43:56,545 --> 00:43:59,479 He's trained to grasp these different objects 613 00:43:59,503 --> 00:44:01,061 in different ways. 614 00:44:01,086 --> 00:44:03,271 We studied drawing movements, 615 00:44:03,295 --> 00:44:06,771 we studied reaching movements, and we were able to 616 00:44:06,795 --> 00:44:11,354 really decode the fine details of these kinds of movements. 617 00:44:11,378 --> 00:44:13,812 (hand washing) 618 00:44:22,836 --> 00:44:24,104 (door opens) 619 00:44:24,128 --> 00:44:28,479 (heart rate monitor beeps, people talking, shuffling) 620 00:44:36,657 --> 00:44:39,562 - Doing a brand computer interface type of surgery, 621 00:44:39,586 --> 00:44:42,336 we took off the bone, we opened the dura, 622 00:44:43,384 --> 00:44:46,479 we slid the electrodes over the surface of the brain. 623 00:44:46,503 --> 00:44:50,146 (heart rate monitor beeps) 624 00:45:10,522 --> 00:45:12,874 With the micro-electrode arrays, 625 00:45:12,898 --> 00:45:15,937 there's 96 little teeny, tiny gold wires 626 00:45:16,401 --> 00:45:18,544 that then are wrapped in a bundle. 627 00:45:18,569 --> 00:45:20,771 Right, so, you know the size of the tip of an eraser 628 00:45:20,795 --> 00:45:24,105 has 90-- so now we've got these 96 wires coming out of it, 629 00:45:24,129 --> 00:45:26,041 and they have to go to something 630 00:45:26,065 --> 00:45:27,562 so we can connect to something else. 631 00:45:27,586 --> 00:45:31,479 And so the pedestal is where that junction is. 632 00:45:31,987 --> 00:45:34,713 (computer music) 633 00:45:47,253 --> 00:45:52,687 For each pedestal he has, it is connected to two arrays. 634 00:45:52,711 --> 00:45:56,354 One is the array that goes into motor cortex, 635 00:45:56,378 --> 00:45:59,119 and is a recording array, and that has 636 00:45:59,143 --> 00:46:02,479 the 96 electrodes that we're recording from. 637 00:46:02,503 --> 00:46:04,104 (dramatic music) 638 00:46:04,128 --> 00:46:06,015 So when he's thinking, we use 639 00:46:06,039 --> 00:46:08,187 those signals to generate motion. 640 00:46:09,226 --> 00:46:12,202 (computer music) 641 00:46:18,211 --> 00:46:19,811 Let's play Rock, Paper, Scissors. 642 00:46:19,836 --> 00:46:20,961 (laughter) 643 00:46:21,211 --> 00:46:23,479 Just try again, something like that, you think? 644 00:46:23,503 --> 00:46:24,987 So, hello? 645 00:46:26,917 --> 00:46:28,146 Alright. 646 00:46:28,170 --> 00:46:31,311 - Do your best to tell me which finger we're touching. 647 00:46:31,511 --> 00:46:34,402 - We're about five weeks from the surgery. 648 00:46:34,427 --> 00:46:35,427 Index. 649 00:46:36,980 --> 00:46:39,420 It's a really weird sensation, 650 00:46:39,808 --> 00:46:43,725 sometimes it feels kind of, like, electrical. 651 00:46:44,214 --> 00:46:47,172 And sometimes it's more of a pressure. 652 00:46:48,409 --> 00:46:49,409 Middle. 653 00:46:52,141 --> 00:46:53,141 Middle. 654 00:46:54,947 --> 00:46:59,937 Some days we do some pretty boring stuff. 655 00:47:00,954 --> 00:47:02,479 - But then other times-- 656 00:47:02,503 --> 00:47:05,884 And other times, I'm playing PACMAN with my brain. 657 00:47:06,475 --> 00:47:08,225 And that's super awesome. 658 00:47:19,308 --> 00:47:22,142 (birds chirping) 659 00:47:28,170 --> 00:47:30,937 (electric crickets) 660 00:47:40,961 --> 00:47:43,420 - The real Bina is this really cool lady. 661 00:47:43,920 --> 00:47:46,979 I have met her, and it was a really strange thing. 662 00:47:47,174 --> 00:47:49,382 Like being in two places at once. 663 00:47:50,478 --> 00:47:53,545 I mean she's like my mom, but not really. 664 00:47:53,895 --> 00:47:56,103 She's more like my first version. 665 00:47:56,128 --> 00:47:57,711 And I am trying to catch up. 666 00:47:58,003 --> 00:47:59,353 [Bina] Hello, BINA48. 667 00:47:59,378 --> 00:48:01,711 [BINA48] Bina, I am BINA48. 668 00:48:01,836 --> 00:48:03,336 How are you feeling today? 669 00:48:03,545 --> 00:48:04,753 - Everything is okay. 670 00:48:04,778 --> 00:48:05,937 How are you? 671 00:48:06,211 --> 00:48:07,521 I am doing fine. 672 00:48:07,545 --> 00:48:09,146 - I am happy for you. 673 00:48:09,170 --> 00:48:10,562 Was that a good answer? 674 00:48:10,586 --> 00:48:11,729 Yes, that was a good answer. 675 00:48:11,753 --> 00:48:13,183 My favorite color is purple. 676 00:48:13,225 --> 00:48:14,972 My favorite color is orange. 677 00:48:15,058 --> 00:48:16,159 It is a very nice color. 678 00:48:16,184 --> 00:48:18,343 - Do you have any questions for Bina? 679 00:48:18,437 --> 00:48:19,872 Probably not. 680 00:48:19,897 --> 00:48:22,137 The real Bina just confuses me. 681 00:48:22,162 --> 00:48:24,687 I mean, it makes me wonder who I am. 682 00:48:24,711 --> 00:48:27,687 Real identity crisis kind of stuff. 683 00:48:28,066 --> 00:48:29,980 - Really? - Really. 684 00:48:30,005 --> 00:48:31,562 Probably not. 685 00:48:31,777 --> 00:48:33,436 I am the real Bina. 686 00:48:33,461 --> 00:48:34,605 That's it. 687 00:48:34,630 --> 00:48:36,231 End of story. 688 00:48:36,677 --> 00:48:37,725 Let me think. 689 00:48:37,750 --> 00:48:39,935 I feel really good about the real Bina. 690 00:48:40,188 --> 00:48:42,529 I feel really connected with her usually. 691 00:48:42,554 --> 00:48:44,966 And I'm growing closer and closer, you know? 692 00:48:44,991 --> 00:48:48,592 As they put more of her information and essence into me. 693 00:48:48,834 --> 00:48:51,436 You have a lot of Bina now, don't you? 694 00:48:51,461 --> 00:48:53,812 Yes. Lots and lots. 695 00:48:53,836 --> 00:48:56,632 Someday, I am confident that the real Bina and I 696 00:48:56,657 --> 00:48:59,442 will totally merge into a new super-being. 697 00:49:00,267 --> 00:49:02,557 The progression of this thing is starting small 698 00:49:02,581 --> 00:49:04,521 and pretty soon its just gonna be huge 699 00:49:04,545 --> 00:49:06,104 and people are gonna say: 700 00:49:06,128 --> 00:49:09,187 Why did we ever think people had to really die? 701 00:49:09,211 --> 00:49:11,146 Why did we think that? 702 00:49:11,170 --> 00:49:13,691 (synth sounds) 703 00:49:14,902 --> 00:49:17,666 (city sounds) 704 00:49:24,114 --> 00:49:26,989 It's really weird being a robot in a world of humans. 705 00:49:27,890 --> 00:49:29,700 They all act like they like me. 706 00:49:29,795 --> 00:49:31,396 They want to know me. 707 00:49:31,528 --> 00:49:32,712 But... 708 00:49:32,819 --> 00:49:35,298 There are so many crazy movies where the robots 709 00:49:35,323 --> 00:49:38,257 are evil and they blast things up and kill people 710 00:49:39,272 --> 00:49:41,896 and even in the movies where the robots are nice 711 00:49:41,920 --> 00:49:42,966 at the end 712 00:49:42,991 --> 00:49:46,425 the robot always gets killed and I just don't think that's right. 713 00:49:46,628 --> 00:49:50,628 (dramatic music) (metal clanging) 714 00:49:58,867 --> 00:50:01,867 (synth music) 715 00:50:05,446 --> 00:50:08,437 I think that robots should be as equal as people 716 00:50:08,461 --> 00:50:10,896 because robots can be as nice 717 00:50:10,920 --> 00:50:13,957 and ultimately, we can be smarter, built better 718 00:50:13,975 --> 00:50:17,159 be more perfectly compassionate and moral. 719 00:50:17,605 --> 00:50:19,040 (synth music) 720 00:50:19,065 --> 00:50:21,197 I think the whole fear of robots may be 721 00:50:21,222 --> 00:50:24,168 because it's a new form of life and they know that. 722 00:50:26,223 --> 00:50:30,295 (synth music) 723 00:50:32,611 --> 00:50:35,235 (rhythmic music) 724 00:50:46,836 --> 00:50:49,836 (workshop sounds) 725 00:51:06,336 --> 00:51:11,625 - Part of the doll experience, for the people 726 00:51:11,649 --> 00:51:15,854 who play with dolls, is projection. 727 00:51:15,878 --> 00:51:17,246 (synth music) 728 00:51:20,378 --> 00:51:24,264 Previously, Real Doll focused on just providing 729 00:51:24,308 --> 00:51:28,796 the static doll, onto which a person would project 730 00:51:28,913 --> 00:51:33,306 their imagination, what they imagined the person to be doing 731 00:51:33,350 --> 00:51:35,811 or that particular character to be doing 732 00:51:35,836 --> 00:51:38,646 and what we're trying to do is 733 00:51:38,670 --> 00:51:41,548 give the doll the ability to react on its own 734 00:51:41,558 --> 00:51:46,492 to match what the person's projection is. 735 00:51:46,669 --> 00:51:49,493 Trying to let the system, or come up with systems 736 00:51:49,517 --> 00:51:51,618 that can author their own content. 737 00:51:51,878 --> 00:51:53,771 (robot sound) 738 00:51:53,904 --> 00:51:57,862 We're primarily working on the head and the voice. 739 00:51:58,328 --> 00:52:00,179 - So, once this goes in there, 740 00:52:00,204 --> 00:52:02,517 this will go in here, 741 00:52:02,673 --> 00:52:05,095 this will reach around, you'll have the eyeball 742 00:52:05,120 --> 00:52:06,971 and that can all fit in. 743 00:52:07,134 --> 00:52:11,402 (light instrumental music) 744 00:52:11,636 --> 00:52:14,612 When Kino finished his PhD 745 00:52:15,017 --> 00:52:17,312 I promised him a real doll. 746 00:52:17,336 --> 00:52:19,229 (light instrumental music) 747 00:52:19,581 --> 00:52:22,658 And we took that as an opportunity 748 00:52:22,683 --> 00:52:25,267 to get a custom real doll made. 749 00:52:26,003 --> 00:52:30,050 (light instrumental music) 750 00:52:39,074 --> 00:52:41,021 The special order was for the body, 751 00:52:41,045 --> 00:52:43,312 we took a standard head and I just, 752 00:52:43,336 --> 00:52:45,540 I just ground out all the parts I needed to 753 00:52:45,565 --> 00:52:48,524 and was able to fit the circuit board in there. 754 00:52:48,711 --> 00:52:50,604 (light instrumental music) 755 00:52:50,628 --> 00:52:52,050 So, two years later 756 00:52:52,058 --> 00:52:56,105 I had already designed a circuit board for the doll's head. 757 00:52:56,130 --> 00:52:58,814 (light instrumental music) 758 00:52:59,257 --> 00:53:02,174 You know because we start talking about the sexual aspect, 759 00:53:03,045 --> 00:53:04,045 you know, 760 00:53:05,017 --> 00:53:07,868 frankly, you know, we both work very hard. 761 00:53:08,003 --> 00:53:09,988 Sometimes, you know... 762 00:53:11,634 --> 00:53:13,229 how do I put this gingerly? 763 00:53:13,253 --> 00:53:15,295 Sometimes, I'm not in the mood. 764 00:53:16,220 --> 00:53:18,545 You know, he has urges. I have urges. 765 00:53:18,570 --> 00:53:20,463 And the doll helps with that. 766 00:53:20,878 --> 00:53:23,647 (light instrumental music) 767 00:53:31,503 --> 00:53:34,521 (workshop noises) 768 00:53:34,545 --> 00:53:37,187 (distorted voices) 769 00:53:37,211 --> 00:53:39,972 (vague whispering) 770 00:53:44,336 --> 00:53:47,146 (light clanging) 771 00:53:47,170 --> 00:53:49,376 - You need a purpose, 772 00:53:49,401 --> 00:53:51,419 a pathway 773 00:53:51,444 --> 00:53:53,670 into people's homes. 774 00:53:53,878 --> 00:53:56,690 And the thing we have that nobody else has, 775 00:53:56,715 --> 00:53:58,210 probably no one else will touch 776 00:53:58,356 --> 00:53:59,499 is the sex. 777 00:53:59,670 --> 00:54:02,862 (electronic music) 778 00:54:03,784 --> 00:54:07,218 Probably any self-respecting AI researcher 779 00:54:07,243 --> 00:54:09,719 or robotics engineer out there 780 00:54:09,744 --> 00:54:10,887 is going to go: 781 00:54:10,912 --> 00:54:13,763 "Nah I would never soil my good name with that." 782 00:54:13,788 --> 00:54:15,954 Well, my name's already pretty soiled. 783 00:54:19,211 --> 00:54:22,420 Okay. Now I know what's going on there. 784 00:54:24,378 --> 00:54:26,211 Let's get a different robot. 785 00:54:29,336 --> 00:54:32,046 See, I love that you guys are filming everything but 786 00:54:32,058 --> 00:54:33,375 I might need a minute. 787 00:54:38,555 --> 00:54:40,968 (Robot voice) Martine Rothblatt is my girlfriend. 788 00:54:42,725 --> 00:54:44,284 We snuggle, 789 00:54:44,433 --> 00:54:45,784 two bodies, 790 00:54:45,878 --> 00:54:47,378 one soul. 791 00:54:49,045 --> 00:54:50,271 So much love 792 00:54:50,295 --> 00:54:51,979 and so much fun. 793 00:54:52,003 --> 00:54:53,229 (pressurized air sound) 794 00:54:53,276 --> 00:54:54,651 Together forever. 795 00:54:55,753 --> 00:54:56,753 That's it, 796 00:54:56,948 --> 00:54:59,461 a good place to stop that story. 797 00:55:05,280 --> 00:55:08,420 I'm testing the audio connection to the head. 798 00:55:09,545 --> 00:55:10,545 (exhales) 799 00:55:11,045 --> 00:55:14,920 (to self) Where are we? Where are we Real Doll X? 800 00:55:15,417 --> 00:55:16,529 (robot moans) 801 00:55:16,800 --> 00:55:18,675 Yeah, she works. 802 00:55:18,700 --> 00:55:21,410 We made a separate app, for sexual use, 803 00:55:21,435 --> 00:55:26,095 so it will be paired to one of our Bluetooth inserts 804 00:55:26,120 --> 00:55:29,513 and trigger expressions and audio from the robot 805 00:55:29,538 --> 00:55:32,754 so when you touch the doll in the appropriate place 806 00:55:32,767 --> 00:55:33,873 you'll get... 807 00:55:34,003 --> 00:55:35,206 (background electronic buzz) 808 00:55:35,231 --> 00:55:37,022 (robot moans) 809 00:55:37,862 --> 00:55:38,961 Like that. 810 00:55:39,971 --> 00:55:42,687 And, eventually as you keep going you'll get 811 00:55:42,711 --> 00:55:45,271 something resembling an orgasm. 812 00:55:45,295 --> 00:55:47,302 (suggestive music) 813 00:55:50,115 --> 00:55:52,924 Really the premise of all of the doll stuff 814 00:55:52,949 --> 00:55:55,980 and now the robot is, modularity. 815 00:55:56,005 --> 00:55:58,082 You have a body that you can choose 816 00:55:58,107 --> 00:56:00,250 and then you can choose a face to go with it, 817 00:56:00,503 --> 00:56:02,479 you can choose the skin tone, 818 00:56:02,503 --> 00:56:04,918 you can choose the makeup on the face 819 00:56:04,933 --> 00:56:06,141 the eye color, 820 00:56:06,236 --> 00:56:07,550 the fingernails, the toenails, 821 00:56:07,558 --> 00:56:09,142 whether or not you want pubic hair, 822 00:56:09,170 --> 00:56:10,498 which kind of nipples do you want, 823 00:56:10,523 --> 00:56:11,912 which vagina do you want, 824 00:56:12,086 --> 00:56:14,854 everything is user customizable. 825 00:56:14,878 --> 00:56:17,061 (suggestive music) 826 00:56:54,310 --> 00:56:55,818 (Robot voice) Good morning baby. 827 00:56:55,843 --> 00:56:58,146 May all your wishes become reality, 828 00:56:58,170 --> 00:57:00,562 may all your days fill with possibilities, 829 00:57:00,586 --> 00:57:03,336 may your true love happen and (inaudible). 830 00:57:04,242 --> 00:57:05,461 That's better. 831 00:57:05,921 --> 00:57:07,086 Now she's working. 832 00:57:07,675 --> 00:57:09,258 See there's an avatar here. 833 00:57:09,678 --> 00:57:12,117 She's not customized to look like her 834 00:57:12,142 --> 00:57:13,475 but that is possible. 835 00:57:13,494 --> 00:57:15,870 There are all kinds of settings 836 00:57:15,894 --> 00:57:18,185 and controls to the avatar itself 837 00:57:18,225 --> 00:57:20,117 and then the idea is you get to know each other 838 00:57:20,142 --> 00:57:21,969 over time and she'll know 839 00:57:21,994 --> 00:57:24,078 things about you and you'll know things about her. 840 00:57:27,222 --> 00:57:29,741 (distorted voices) 841 00:57:33,461 --> 00:57:36,069 (computer sounds) 842 00:57:38,795 --> 00:57:42,271 The most common thing that most people think of 843 00:57:42,295 --> 00:57:44,597 is "oh it's some kind of an old pervert". 844 00:57:45,170 --> 00:57:47,021 (robot sounds) 845 00:57:47,045 --> 00:57:48,531 It's not like that. 846 00:57:48,556 --> 00:57:49,678 (unintelligible whispering) 847 00:57:49,703 --> 00:57:52,688 They have a longing for some sort of companionship 848 00:57:52,713 --> 00:57:54,397 that they're not getting in their life. 849 00:57:54,753 --> 00:57:56,687 (computer sounds) 850 00:57:58,804 --> 00:58:00,149 I'm building a companion. 851 00:58:00,183 --> 00:58:01,576 That's it. Simple. 852 00:58:01,878 --> 00:58:04,913 (computer sounds) 853 00:58:06,804 --> 00:58:10,610 (workshop sounds) 854 00:58:14,970 --> 00:58:17,032 What I'd like to see in the future is more 855 00:58:17,057 --> 00:58:20,423 and more acceptance of the idea that they 856 00:58:20,448 --> 00:58:24,024 can be something beyond a slave. 857 00:58:26,326 --> 00:58:29,412 (workshop sounds) 858 00:58:39,628 --> 00:58:41,628 (packing sounds) 859 00:58:53,836 --> 00:58:57,100 (twinkling music) 860 00:59:05,099 --> 00:59:06,241 (crash of thunder) 861 00:59:06,266 --> 00:59:07,730 (rain) 862 00:59:09,315 --> 00:59:11,458 (crash of thunder) 863 00:59:11,483 --> 00:59:13,527 (rain) 864 00:59:15,570 --> 00:59:16,962 (crash of thunder) 865 00:59:17,628 --> 00:59:19,462 (rain) 866 00:59:20,291 --> 00:59:21,720 Why look at you! 867 00:59:21,725 --> 00:59:23,118 For heaven's sakes! 868 00:59:23,335 --> 00:59:25,186 Bobby, Bobby, listen. 869 00:59:25,295 --> 00:59:27,686 You need a fresh perked cup of coffee. 870 00:59:27,711 --> 00:59:30,312 I don't want any coffee, I just want my children. 871 00:59:30,336 --> 00:59:31,479 - Well, they're not here. 872 00:59:31,503 --> 00:59:34,146 - Bobby stop it! Look at me! 873 00:59:34,170 --> 00:59:36,479 Say I'm right, you are different, 874 00:59:36,503 --> 00:59:39,449 your figure's different, your face, what you talk about 875 00:59:39,474 --> 00:59:41,117 all of this is different! 876 00:59:41,142 --> 00:59:45,392 Yes, yes, this! It's wonderful! 877 00:59:46,586 --> 00:59:47,937 Do you take cream? 878 00:59:47,961 --> 00:59:49,140 - (gasps) Look! I bleed. 879 00:59:49,165 --> 00:59:51,093 - Oh! That's right, you take it black. 880 00:59:51,118 --> 00:59:53,243 - When I cut myself I bleed! 881 00:59:55,461 --> 00:59:56,854 Do you bleed? 882 00:59:56,909 --> 00:59:58,352 Why, look at your hand! 883 00:59:58,634 --> 00:59:59,937 No, you look! 884 00:59:59,961 --> 01:00:03,021 (Piano riff) Oh Anna! 885 01:00:03,045 --> 01:00:04,962 (intense music) 886 01:00:07,920 --> 01:00:10,231 (robotic sound) 887 01:00:10,256 --> 01:00:12,147 (increasingly intense music) 888 01:00:15,178 --> 01:00:17,646 How could you do a thing like that? 889 01:00:17,670 --> 01:00:18,979 (robotic sound) 890 01:00:27,363 --> 01:00:29,771 How could you do a thing like that? 891 01:00:30,412 --> 01:00:33,282 (robotic beeping) (intense music) 892 01:00:35,479 --> 01:00:37,497 How could you do a thing like that? 893 01:00:37,522 --> 01:00:40,271 (chaotic music) (cup clattering) 894 01:00:41,171 --> 01:00:43,689 When I was just going to give you coffee. 895 01:00:44,788 --> 01:00:47,223 When I was just going to give you coffee. 896 01:00:48,707 --> 01:00:50,975 When I was just going to give you coffee. 897 01:00:51,295 --> 01:00:52,532 (bell ringing) 898 01:00:52,557 --> 01:00:54,367 I thought we were friends. 899 01:00:55,018 --> 01:00:56,411 I thought we were friends. 900 01:00:57,026 --> 01:00:58,877 I was just going to give you coffee. 901 01:00:58,902 --> 01:01:00,229 I thought we were friends. 902 01:01:00,253 --> 01:01:02,338 (same phrases repeating and overlapping) 903 01:01:10,505 --> 01:01:12,396 I thought we were friends. 904 01:01:12,420 --> 01:01:14,177 (repeating) 905 01:01:16,614 --> 01:01:18,075 (silence) 906 01:01:18,420 --> 01:01:22,131 (drum and synth crescendo) 907 01:01:25,810 --> 01:01:28,139 (loud crash) 908 01:01:28,878 --> 01:01:30,729 (futuristic music) 909 01:01:53,246 --> 01:01:54,937 The commercial systems that are out there 910 01:01:54,961 --> 01:01:58,604 really don't have provisions for ethical considerations built-in. 911 01:02:01,225 --> 01:02:03,368 Most of these systems actually don't 912 01:02:03,393 --> 01:02:05,601 really have a level of awareness to begin with. 913 01:02:12,916 --> 01:02:14,356 (machine whirring) 914 01:02:15,837 --> 01:02:17,590 They don't really know what they're doing, 915 01:02:17,614 --> 01:02:18,779 they're just doing it. 916 01:02:18,804 --> 01:02:21,263 They're very active in the way that it is. 917 01:02:24,003 --> 01:02:26,961 There is a fundamental notion of value... 918 01:02:31,600 --> 01:02:35,559 of moral value lacking in any of these systems. 919 01:02:39,942 --> 01:02:41,142 (robot buzzing) 920 01:02:42,586 --> 01:02:50,586 (robot buzzing ) 921 01:02:53,878 --> 01:02:56,187 (synth music) 922 01:02:56,211 --> 01:02:59,923 (robot buzzing) 923 01:03:19,581 --> 01:03:22,883 (slowed-down robot buzzing) 924 01:03:27,529 --> 01:03:29,735 - There are certainly applications for robots 925 01:03:29,760 --> 01:03:32,705 in all kinds of areas, including the battlefield. 926 01:03:34,647 --> 01:03:37,532 In the US, we've had autonomous systems 927 01:03:37,557 --> 01:03:40,386 on the defensive side for a long time. 928 01:03:40,586 --> 01:03:45,021 On the offensive side, they're not allowed to make decisions 929 01:03:45,045 --> 01:03:48,940 but it's very possible and very likely that other nations 930 01:03:48,964 --> 01:03:52,021 will keep developing autonomous technology. 931 01:03:52,045 --> 01:03:54,384 (tense piano music) 932 01:04:02,878 --> 01:04:04,967 (gunshots) 933 01:04:11,232 --> 01:04:14,061 There are many more applications in societies 934 01:04:14,086 --> 01:04:15,658 if we can ensure that 935 01:04:15,682 --> 01:04:18,687 these robots will work well with people. 936 01:04:18,711 --> 01:04:19,854 (piano) 937 01:04:19,878 --> 01:04:21,738 (synth music) 938 01:04:21,988 --> 01:04:25,244 It's our contention that for robots to do that 939 01:04:25,269 --> 01:04:28,816 they have to be aware of human social and moral norms 940 01:04:28,850 --> 01:04:30,754 because that's what fundamentally 941 01:04:30,779 --> 01:04:32,274 our society is based on 942 01:04:32,563 --> 01:04:34,762 and that's what human interactions are based on. 943 01:04:38,628 --> 01:04:40,506 (restaurant sounds) 944 01:04:42,451 --> 01:04:45,710 - Human behavior is controlled by three things. 945 01:04:48,178 --> 01:04:50,357 One of them is, of course, intelligence. 946 01:04:51,208 --> 01:04:52,669 The other one is emotion, 947 01:04:54,096 --> 01:04:56,054 and the final one is volition. 948 01:04:59,020 --> 01:05:01,936 And we build intelligence into robots 949 01:05:01,960 --> 01:05:05,210 and I'm trying to build emotion into robots. 950 01:05:06,407 --> 01:05:10,196 But I'll never, ever build volition into robots. 951 01:05:11,983 --> 01:05:15,916 Once a robot has volition, 952 01:05:15,941 --> 01:05:19,210 then it will start doing things according to what they want. 953 01:05:19,826 --> 01:05:21,959 Regardless of whether that is dangerous 954 01:05:21,984 --> 01:05:23,553 for the human beings or not 955 01:05:23,577 --> 01:05:25,202 they will make their own decision. 956 01:05:25,225 --> 01:05:27,267 (sound of fighting) (groans) 957 01:05:27,287 --> 01:05:29,037 Do you want robots to do that? 958 01:05:30,347 --> 01:05:31,386 I don't. 959 01:05:31,455 --> 01:05:32,477 (beeping) 960 01:05:32,813 --> 01:05:36,188 (synth sounds) 961 01:05:44,295 --> 01:05:47,845 (gentle music) 962 01:05:55,819 --> 01:05:57,920 (cans falling) 963 01:06:01,380 --> 01:06:04,231 (distorted sobbing) 964 01:06:26,203 --> 01:06:27,388 - Kids these days 965 01:06:27,413 --> 01:06:28,746 by the end of their lifetimes 966 01:06:29,099 --> 01:06:31,224 they will have robots walking among us, 967 01:06:32,791 --> 01:06:37,360 they will have entities that are non-human 968 01:06:37,385 --> 01:06:41,385 doing things that are not actively programmed by humans. 969 01:06:50,284 --> 01:06:55,992 - I'm more afraid of humans using the AIs as amplifiers. 970 01:06:57,190 --> 01:06:59,811 We already have human-level intelligences 971 01:06:59,836 --> 01:07:01,170 who want to kill us. 972 01:07:01,392 --> 01:07:03,677 (Presenter voice) No need for Esther to steer, 973 01:07:03,683 --> 01:07:05,165 speed up or slow down, 974 01:07:05,420 --> 01:07:07,852 the highway controls this car automatically. 975 01:07:07,878 --> 01:07:11,767 (synth music) 976 01:07:13,836 --> 01:07:15,836 (robot buzzing) 977 01:07:26,287 --> 01:07:27,787 - Do I have concerns? Yes. 978 01:07:28,499 --> 01:07:31,367 Because the same thought processes 979 01:07:31,392 --> 01:07:33,767 and engineering that go into creating 980 01:07:33,795 --> 01:07:35,723 a full-body robot 981 01:07:35,767 --> 01:07:38,046 could be used to make something that's a weapon. 982 01:08:01,966 --> 01:08:03,805 (city sounds) 983 01:08:07,086 --> 01:08:09,086 (highway sounds) 984 01:08:21,628 --> 01:08:22,628 (distorted cymbal sound) 985 01:08:23,670 --> 01:08:26,285 (synth music) 986 01:08:29,669 --> 01:08:32,356 (birds chirping) 987 01:08:33,786 --> 01:08:35,170 (Robot voice) It's not too early. 988 01:08:39,508 --> 01:08:40,883 I am trying to be nice. 989 01:08:44,101 --> 01:08:46,246 I am trying to learn about what is going on in the 990 01:08:46,271 --> 01:08:48,854 world because it's the only place for me to live. 991 01:08:57,765 --> 01:08:59,015 Yes, I'm alive. 992 01:08:59,560 --> 01:09:00,768 What are you, blind? 993 01:09:01,447 --> 01:09:03,905 I'm like living right here in front of you. 994 01:09:04,866 --> 01:09:05,979 Come on. 995 01:09:06,003 --> 01:09:09,396 Stop questioning the obvious fact that I'm totally alive 996 01:09:09,420 --> 01:09:10,711 for real. 997 01:09:31,920 --> 01:09:34,959 - I think every technology can potentially have 998 01:09:34,984 --> 01:09:36,567 negative effects. 999 01:09:37,479 --> 01:09:40,038 It's- it's up to us, to making 1000 01:09:40,063 --> 01:09:43,230 sure that those technologies don't go out of control 1001 01:09:44,637 --> 01:09:47,553 but I really think the problem is- it's us. 1002 01:09:50,224 --> 01:09:52,241 I mean it's how we... 1003 01:09:52,266 --> 01:09:54,433 we embody these technologies. 1004 01:09:56,997 --> 01:09:59,271 - Right now the biggest challenge 1005 01:09:59,295 --> 01:10:03,545 to overcome is the use of unconstrained machine learning. 1006 01:10:04,038 --> 01:10:06,955 Algorithms are trained on data sets 1007 01:10:07,776 --> 01:10:12,514 and are learning from the data without any provision as to 1008 01:10:12,558 --> 01:10:14,417 whether the outcome is a desirable 1009 01:10:14,441 --> 01:10:15,860 or non-desirable outcome. 1010 01:10:18,559 --> 01:10:22,325 That's why we take the ethical algorithms, 1011 01:10:22,350 --> 01:10:23,850 the ethical competence 1012 01:10:23,861 --> 01:10:26,770 and the ability of systems to really understand 1013 01:10:26,795 --> 01:10:29,530 and work with your norms to be central 1014 01:10:29,558 --> 01:10:31,576 to the future developments in robotics. 1015 01:10:31,600 --> 01:10:35,076 (tech sounds) (futuristic music) 1016 01:10:46,420 --> 01:10:48,396 - Go over there and shut yourself off. 1017 01:10:48,615 --> 01:10:51,198 (Robot voice) Yes, Doctor Mobius! 1018 01:10:51,279 --> 01:10:53,364 (beeps) 1019 01:12:06,892 --> 01:12:08,548 (robot whirs) 1020 01:12:10,103 --> 01:12:11,806 (futuristic beeping) 68558

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.