All language subtitles for 045 How Do GANs Work (Step 3)-en

af Afrikaans
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bn Bengali
bs Bosnian
bg Bulgarian
ca Catalan
ceb Cebuano
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
tl Filipino
fi Finnish
fr French
fy Frisian
gl Galician
ka Georgian
de German
el Greek
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
km Khmer
ko Korean
ku Kurdish (Kurmanji)
ky Kyrgyz
lo Lao
la Latin
lv Latvian
lt Lithuanian
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mn Mongolian
my Myanmar (Burmese)
ne Nepali
no Norwegian
ps Pashto
fa Persian
pl Polish
pt Portuguese
pa Punjabi
ro Romanian
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
st Sesotho
sn Shona
sd Sindhi
si Sinhala
sk Slovak
sl Slovenian
so Somali
es Spanish
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
te Telugu
th Thai
tr Turkish
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese Download
cy Welsh
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
or Odia (Oriya)
rw Kinyarwanda
tk Turkmen
tt Tatar
ug Uyghur
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:00,390 --> 00:00:06,470 Crais final step that we're going to look at noise goes into the generator generator generates dogs 2 00:00:06,540 --> 00:00:13,090 with eyes this time and we can look more three dimensional I have ears for different. 3 00:00:13,100 --> 00:00:15,680 Too bad they have too many eyes. 4 00:00:15,680 --> 00:00:22,990 But some of them but already you know it's learned from the mistakes it made previously. 5 00:00:23,160 --> 00:00:28,560 And so now we want to train the discriminator and we're going to need another batch of dog images we're 6 00:00:28,560 --> 00:00:30,000 going to put them in. 7 00:00:30,000 --> 00:00:33,360 And there you go see what outputs you've got. 8 00:00:33,360 --> 00:00:44,180 We've got some valuable top to bottom and so what we can see now is that these values are slowly converging 9 00:00:44,180 --> 00:00:51,470 so the value of getting closer the values for the dogs or the dogs are getting bigger with time and 10 00:00:51,470 --> 00:00:57,550 it's harder it is becoming harder for the discriminator to discriminate between dogs and dogs. 11 00:00:57,550 --> 00:01:03,230 I also noticed that in the Paduan retraining the general and that's good that means these images are 12 00:01:03,230 --> 00:01:05,170 getting closer to what we want to. 13 00:01:05,210 --> 00:01:07,130 More realistic dogs. 14 00:01:07,130 --> 00:01:12,590 And again we're going to take these values Kellow can calculate the errors based on what we know they 15 00:01:12,590 --> 00:01:15,410 should be they should be zeros at the top ones at the bottom. 16 00:01:15,740 --> 00:01:20,300 And that error is going to be back propagated through the network of the discriminator. 17 00:01:20,320 --> 00:01:26,350 Weights are going to be updated and then then it's time to train the generators. 18 00:01:26,360 --> 00:01:28,930 Now we're going to train the generator. 19 00:01:28,940 --> 00:01:32,800 We're going to have to put these images into the network. 20 00:01:32,990 --> 00:01:36,500 And what happens next is we get an output. 21 00:01:36,500 --> 00:01:43,700 You can see this output is lower than it was just now but also as you can see it slowly. 22 00:01:43,760 --> 00:01:46,470 It's not as low as at the very start. 23 00:01:46,560 --> 00:01:55,100 Again this is a sign it's showing us that these images are actually getting closer to realistic dog 24 00:01:55,130 --> 00:01:56,050 images. 25 00:01:56,140 --> 00:02:01,500 Again that these valves are going to be we're going to get the values collate the error and get back 26 00:02:01,510 --> 00:02:08,260 Burkitt the error through the network of the generator and the weights there. 27 00:02:08,750 --> 00:02:10,970 So there we go that's how it works. 28 00:02:10,970 --> 00:02:16,550 These are just three steps in training in reality these like hundreds and thousands of these steps and 29 00:02:16,550 --> 00:02:19,670 then multiple airports as well. 30 00:02:19,670 --> 00:02:24,410 And as you can imagine through many many many many iterations these images are going to get better and 31 00:02:24,410 --> 00:02:31,790 better and better and better through this struggle through this confrontation between the generator 32 00:02:31,850 --> 00:02:34,100 and the discriminant. 33 00:02:34,340 --> 00:02:37,740 Additional information is definitely available. 34 00:02:37,760 --> 00:02:43,400 And one of the best place to go to is the original paper by in good fellow it's called generative adversarial 35 00:02:43,490 --> 00:02:48,420 Annetts 2014 paper which you can find archive. 36 00:02:48,530 --> 00:02:48,880 Yeah. 37 00:02:48,890 --> 00:02:51,110 And he explains everything there. 38 00:02:51,500 --> 00:02:54,570 And one more thing I wanted to mention here. 39 00:02:54,620 --> 00:02:58,850 When we were talking about the generator we just said it's a neural network. 40 00:02:58,850 --> 00:03:03,850 But this type of neural network we haven't discussed it is not even discussed in the annex. 41 00:03:03,950 --> 00:03:10,300 It's called a deconvolution or neural network in that case you'll find artificial neural networks and 42 00:03:10,300 --> 00:03:14,070 can we do illusional neural networks but not deconvolution all neural networks. 43 00:03:14,240 --> 00:03:20,090 So I just wanted to make a quick note here so there is this illusional neural network what a deconvolution 44 00:03:20,090 --> 00:03:27,350 all neural network is if you flip this upside down or run it back to front and then run it this way 45 00:03:28,460 --> 00:03:30,660 instead of the normal way you run the other way. 46 00:03:30,680 --> 00:03:35,870 Basically you start with the vector of values which is our Random signal. 47 00:03:35,880 --> 00:03:38,180 Then at the end you'll get the image. 48 00:03:38,420 --> 00:03:43,940 And if you'd like to learn more about decompositional neural networks there is some additional reading 49 00:03:43,940 --> 00:03:44,700 over here. 50 00:03:44,750 --> 00:03:52,730 We're not discussing it here because it's not the main concept of for us to focus on one focus on Ganns 51 00:03:52,760 --> 00:03:59,450 But if you'd like to learn more about D-Conn. finance then this is a paper which talks about them is 52 00:03:59,450 --> 00:04:06,260 called adaptive deconvolution networks for mid and high level feature learning by Mathieu's and others. 53 00:04:06,500 --> 00:04:14,780 So yeah that's a go to people are forged continents hopefully you enjoyed this tutorial and you now 54 00:04:14,780 --> 00:04:20,840 know how Ganns work in the background and then report seeing you back here next time. 55 00:04:20,850 --> 00:04:22,710 And until then enjoy computer vision. 6181

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.