All language subtitles for 045 How Do GANs Work (Step 3)-en

af Afrikaans
ak Akan
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
el Greek
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranî)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian
pl Polish
pt-BR Portuguese (Brazil)
pt Portuguese (Portugal)
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese
sk Slovak
sl Slovenian
so Somali
es Spanish
es-419 Spanish (Latin American)
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese Download
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:00,390 --> 00:00:06,470 Crais final step that we're going to look at noise goes into the generator generator generates dogs 2 00:00:06,540 --> 00:00:13,090 with eyes this time and we can look more three dimensional I have ears for different. 3 00:00:13,100 --> 00:00:15,680 Too bad they have too many eyes. 4 00:00:15,680 --> 00:00:22,990 But some of them but already you know it's learned from the mistakes it made previously. 5 00:00:23,160 --> 00:00:28,560 And so now we want to train the discriminator and we're going to need another batch of dog images we're 6 00:00:28,560 --> 00:00:30,000 going to put them in. 7 00:00:30,000 --> 00:00:33,360 And there you go see what outputs you've got. 8 00:00:33,360 --> 00:00:44,180 We've got some valuable top to bottom and so what we can see now is that these values are slowly converging 9 00:00:44,180 --> 00:00:51,470 so the value of getting closer the values for the dogs or the dogs are getting bigger with time and 10 00:00:51,470 --> 00:00:57,550 it's harder it is becoming harder for the discriminator to discriminate between dogs and dogs. 11 00:00:57,550 --> 00:01:03,230 I also noticed that in the Paduan retraining the general and that's good that means these images are 12 00:01:03,230 --> 00:01:05,170 getting closer to what we want to. 13 00:01:05,210 --> 00:01:07,130 More realistic dogs. 14 00:01:07,130 --> 00:01:12,590 And again we're going to take these values Kellow can calculate the errors based on what we know they 15 00:01:12,590 --> 00:01:15,410 should be they should be zeros at the top ones at the bottom. 16 00:01:15,740 --> 00:01:20,300 And that error is going to be back propagated through the network of the discriminator. 17 00:01:20,320 --> 00:01:26,350 Weights are going to be updated and then then it's time to train the generators. 18 00:01:26,360 --> 00:01:28,930 Now we're going to train the generator. 19 00:01:28,940 --> 00:01:32,800 We're going to have to put these images into the network. 20 00:01:32,990 --> 00:01:36,500 And what happens next is we get an output. 21 00:01:36,500 --> 00:01:43,700 You can see this output is lower than it was just now but also as you can see it slowly. 22 00:01:43,760 --> 00:01:46,470 It's not as low as at the very start. 23 00:01:46,560 --> 00:01:55,100 Again this is a sign it's showing us that these images are actually getting closer to realistic dog 24 00:01:55,130 --> 00:01:56,050 images. 25 00:01:56,140 --> 00:02:01,500 Again that these valves are going to be we're going to get the values collate the error and get back 26 00:02:01,510 --> 00:02:08,260 Burkitt the error through the network of the generator and the weights there. 27 00:02:08,750 --> 00:02:10,970 So there we go that's how it works. 28 00:02:10,970 --> 00:02:16,550 These are just three steps in training in reality these like hundreds and thousands of these steps and 29 00:02:16,550 --> 00:02:19,670 then multiple airports as well. 30 00:02:19,670 --> 00:02:24,410 And as you can imagine through many many many many iterations these images are going to get better and 31 00:02:24,410 --> 00:02:31,790 better and better and better through this struggle through this confrontation between the generator 32 00:02:31,850 --> 00:02:34,100 and the discriminant. 33 00:02:34,340 --> 00:02:37,740 Additional information is definitely available. 34 00:02:37,760 --> 00:02:43,400 And one of the best place to go to is the original paper by in good fellow it's called generative adversarial 35 00:02:43,490 --> 00:02:48,420 Annetts 2014 paper which you can find archive. 36 00:02:48,530 --> 00:02:48,880 Yeah. 37 00:02:48,890 --> 00:02:51,110 And he explains everything there. 38 00:02:51,500 --> 00:02:54,570 And one more thing I wanted to mention here. 39 00:02:54,620 --> 00:02:58,850 When we were talking about the generator we just said it's a neural network. 40 00:02:58,850 --> 00:03:03,850 But this type of neural network we haven't discussed it is not even discussed in the annex. 41 00:03:03,950 --> 00:03:10,300 It's called a deconvolution or neural network in that case you'll find artificial neural networks and 42 00:03:10,300 --> 00:03:14,070 can we do illusional neural networks but not deconvolution all neural networks. 43 00:03:14,240 --> 00:03:20,090 So I just wanted to make a quick note here so there is this illusional neural network what a deconvolution 44 00:03:20,090 --> 00:03:27,350 all neural network is if you flip this upside down or run it back to front and then run it this way 45 00:03:28,460 --> 00:03:30,660 instead of the normal way you run the other way. 46 00:03:30,680 --> 00:03:35,870 Basically you start with the vector of values which is our Random signal. 47 00:03:35,880 --> 00:03:38,180 Then at the end you'll get the image. 48 00:03:38,420 --> 00:03:43,940 And if you'd like to learn more about decompositional neural networks there is some additional reading 49 00:03:43,940 --> 00:03:44,700 over here. 50 00:03:44,750 --> 00:03:52,730 We're not discussing it here because it's not the main concept of for us to focus on one focus on Ganns 51 00:03:52,760 --> 00:03:59,450 But if you'd like to learn more about D-Conn. finance then this is a paper which talks about them is 52 00:03:59,450 --> 00:04:06,260 called adaptive deconvolution networks for mid and high level feature learning by Mathieu's and others. 53 00:04:06,500 --> 00:04:14,780 So yeah that's a go to people are forged continents hopefully you enjoyed this tutorial and you now 54 00:04:14,780 --> 00:04:20,840 know how Ganns work in the background and then report seeing you back here next time. 55 00:04:20,850 --> 00:04:22,710 And until then enjoy computer vision. 6181

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.