All language subtitles for 04-Lecture 1 Segment 4 What about the brain.en

af Afrikaans
ak Akan
sq Albanian
am Amharic
ar Arabic Download
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
el Greek
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranî)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian
pl Polish
pt-BR Portuguese (Brazil)
pt Portuguese (Portugal)
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese
sk Slovak
sl Slovenian
so Somali
es Spanish
es-419 Spanish (Latin American)
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:01,879 --> 00:00:05,000 And so you say, what about the brain? I mean we're engineers right. We want to build something. 2 00:00:05,000 --> 00:00:09,050 we want to build something that's intelligent. It'd be a little silly to ignore the fact that we 3 00:00:09,050 --> 00:00:12,270 actually have a working prototype, and we're trying to do this really hard thing. We 4 00:00:12,270 --> 00:00:13,719 should look at the prototype. 5 00:00:13,719 --> 00:00:17,239 And you know human minds are not perfect decision-makers, but by and large they are 6 00:00:17,239 --> 00:00:18,699 actually very good. 7 00:00:18,699 --> 00:00:22,449 And even the cases where humans are famously not rational, there are many cases 8 00:00:22,449 --> 00:00:24,899 there where you can actually say, well they are rational if you think of their 9 00:00:24,899 --> 00:00:27,739 goals in a different way. That's a whole sticky subject, we'lll talk about that 10 00:00:27,739 --> 00:00:30,979 later in the course. But basically human brains are in many ways 11 00:00:30,979 --> 00:00:31,960 better than what we can build. 12 00:00:31,960 --> 00:00:36,120 So let's just reverse engineer them. So it turns out, brains aren't as modular as software, and 13 00:00:36,120 --> 00:00:39,970 they're a whole lot more squishy. And so they're hard to reverse engineer. 14 00:00:39,970 --> 00:00:42,690 And because we can't reverse engineer them we have a very limited understanding of 15 00:00:42,690 --> 00:00:45,660 what brains actually do, we really don't 16 00:00:45,660 --> 00:00:47,540 kind of reverse engineer them and build a kind of 17 00:00:47,540 --> 00:00:49,230 biomimetic brain. 18 00:00:49,230 --> 00:00:53,750 And in particular there is an idea, which, you know, this is one point of view, 19 00:00:53,750 --> 00:00:56,620 but I think it's got a lot of merit to it. People say that brains are to 20 00:00:56,620 --> 00:00:59,480 intelligence as wings are to flight. 21 00:00:59,480 --> 00:01:01,080 That is to say, that 22 00:01:01,080 --> 00:01:04,230 the biggest breakthrough in manned flight--does anybody know what was the 23 00:01:04,230 --> 00:01:06,240 biggest breakthrough in manned flight? 24 00:01:06,240 --> 00:01:10,520 Stop making the wings flap, right. Stop trying to make a really big bird 25 00:01:10,520 --> 00:01:14,500 made out of metal or something, or fabric right. And so this idea of, okay, maybe the 26 00:01:14,500 --> 00:01:17,180 wings can be fixed 'cause we're going to be going faster or we're going to be going higher, 27 00:01:17,180 --> 00:01:20,210 it's going to be rigid. And so when people started thinking, okay, we're going to learn something about 28 00:01:20,210 --> 00:01:25,010 aerodynamics from the proofs of concept, say the birds, but we're not gonna just 29 00:01:25,010 --> 00:01:28,220 slavishly mimic what's in the biology. Same thing with the brains. We're gonna 30 00:01:28,220 --> 00:01:31,420 learn what we can, but we're not gonna just mimic that because we got different constraints 31 00:01:31,420 --> 00:01:36,310 and a limited understanding of what the brain does to begin with. 32 00:01:36,310 --> 00:01:39,410 So, what do we actually get. Well, we do know something from the brain and it boils down 33 00:01:39,410 --> 00:01:41,900 to this, kind of at the very highest level. 34 00:01:41,900 --> 00:01:44,490 We know that in order to make good decisions, 35 00:01:44,490 --> 00:01:46,210 there's really two parts to that. 36 00:01:46,210 --> 00:01:48,230 One way you can make a good decision 37 00:01:48,230 --> 00:01:52,490 is by remembering that in the past you did this thing and it was bad, 38 00:01:52,490 --> 00:01:55,680 so you know not to do that thing again. You know that boils down to memory. That leads into 39 00:01:55,680 --> 00:01:58,659 learning, machine learning, and that's going to be a big part of this course. 40 00:01:58,659 --> 00:02:01,760 The other way you can make a good prediction, is through simulation--having a 41 00:02:01,760 --> 00:02:04,990 model of the world. We think, alright, what would happen if I do this, well then this 42 00:02:04,990 --> 00:02:07,799 would happen, then this would happen--oh, and then that would happen, 43 00:02:07,799 --> 00:02:10,700 that would be bad. And so you can realize that something is a bad decision by thinking 44 00:02:10,700 --> 00:02:14,400 through a chain of consequences. Not actually doing that chain, 45 00:02:14,400 --> 00:02:16,529 but thinking it through in a simulated model of the world 46 00:02:16,529 --> 00:02:19,619 and that's going to occupy the first half of the class, when we think about 47 00:02:19,619 --> 00:02:21,760 what it means to have a model of the world, that's an abstraction of the world, 48 00:02:21,760 --> 00:02:24,629 and what it means to think ahead along a kind of tree of outcomes. 4886

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.