All language subtitles for 006 Discrete Distributions The Binomial Distribution_en

af Afrikaans
ak Akan
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
el Greek
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranî)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian Download
pl Polish
pt-BR Portuguese (Brazil)
pt Portuguese (Portugal)
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese
sk Slovak
sl Slovenian
so Somali
es Spanish
es-419 Spanish (Latin American)
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:03,180 --> 00:00:04,680 Instructor: Welcome back. 2 00:00:04,680 --> 00:00:08,340 In the last video we mentioned binomial distributions. 3 00:00:08,340 --> 00:00:10,680 In essence, binomial events are a sequence 4 00:00:10,680 --> 00:00:12,573 of identical Bernoulli events. 5 00:00:13,560 --> 00:00:14,940 Before we get into the differences 6 00:00:14,940 --> 00:00:17,700 and similarities between these two distributions 7 00:00:17,700 --> 00:00:19,980 let us examine the proper notation 8 00:00:19,980 --> 00:00:21,603 for a binomial distribution. 9 00:00:22,920 --> 00:00:27,000 We use the letter B to express a binomial distribution 10 00:00:27,000 --> 00:00:28,680 followed by the number of trials 11 00:00:28,680 --> 00:00:31,113 and the probability of success in each one. 12 00:00:32,100 --> 00:00:34,800 Therefore, we read the following statement as 13 00:00:34,800 --> 00:00:39,600 variable X follows a binomial distribution with 10 trials 14 00:00:39,600 --> 00:00:44,193 and a likelihood of 0.6 to succeed on each individual trial. 15 00:00:45,870 --> 00:00:49,200 Additionally, we can express a Bernoulli distribution 16 00:00:49,200 --> 00:00:51,783 as a binomial distribution with a single trial. 17 00:00:53,970 --> 00:00:56,520 All right, to better understand the differences 18 00:00:56,520 --> 00:00:58,380 between the two types of events 19 00:00:58,380 --> 00:01:00,033 suppose the following scenario. 20 00:01:01,140 --> 00:01:02,760 You go to class and your professor 21 00:01:02,760 --> 00:01:05,190 gives the class a surprise pop quiz 22 00:01:05,190 --> 00:01:07,590 which you have not prepared for. 23 00:01:07,590 --> 00:01:09,270 Luckily for you, the quiz consists 24 00:01:09,270 --> 00:01:12,240 of 10 true or false questions. 25 00:01:12,240 --> 00:01:14,460 In this case, guessing a single true 26 00:01:14,460 --> 00:01:17,670 or false question is a Bernoulli event, 27 00:01:17,670 --> 00:01:21,183 but guessing the entire quiz is a binomial event. 28 00:01:23,220 --> 00:01:24,210 All right, let's go back 29 00:01:24,210 --> 00:01:26,110 to the quiz example we just mentioned. 30 00:01:27,000 --> 00:01:30,390 In it, the expected value of the Bernoulli distribution 31 00:01:30,390 --> 00:01:33,483 suggests which outcome we expect for a single trial. 32 00:01:34,470 --> 00:01:35,820 Now, the expected value 33 00:01:35,820 --> 00:01:38,730 of the binomial distribution would suggest the number 34 00:01:38,730 --> 00:01:42,093 of times we expect to get a specific outcome. 35 00:01:43,800 --> 00:01:45,900 Great. Now, the graph 36 00:01:45,900 --> 00:01:48,870 of the binomial distribution represents the likelihood 37 00:01:48,870 --> 00:01:52,113 of attaining our desired outcome a specific number of times. 38 00:01:52,980 --> 00:01:55,620 If we run n trials, our graph 39 00:01:55,620 --> 00:01:59,220 would consist n plus one many bars 40 00:01:59,220 --> 00:02:02,910 one for each unique value from zero to n. 41 00:02:02,910 --> 00:02:05,790 For instance, we could be flipping the same unfair coin 42 00:02:05,790 --> 00:02:07,830 we had from the last lecture. 43 00:02:07,830 --> 00:02:11,190 If we toss it twice, we need bars for the three 44 00:02:11,190 --> 00:02:15,573 different outcomes; zero, one, or two tails. 45 00:02:17,040 --> 00:02:20,760 Fantastic. If we wish to find the associated likelihood 46 00:02:20,760 --> 00:02:23,580 of getting a given outcome a precise number of times 47 00:02:23,580 --> 00:02:26,910 over the course of n trials, we need to introduce 48 00:02:26,910 --> 00:02:29,943 the probability function of the binomial distribution. 49 00:02:30,796 --> 00:02:35,610 For starters, each individual trial is a Bernoulli trial 50 00:02:35,610 --> 00:02:38,730 so we express the probability of getting our desired outcome 51 00:02:38,730 --> 00:02:43,730 as p and the likelihood of the other one as one minus p 52 00:02:45,240 --> 00:02:49,080 in order to get our favorite outcome exactly y many times 53 00:02:49,080 --> 00:02:52,470 over the n trials, we also need to get the alternative 54 00:02:52,470 --> 00:02:55,923 outcome n minus y many times. 55 00:02:56,970 --> 00:02:58,710 If we don't account for this, we would 56 00:02:58,710 --> 00:03:01,380 be estimating the likelihood of getting our desired outcome 57 00:03:01,380 --> 00:03:03,513 at least y many times. 58 00:03:05,070 --> 00:03:06,990 Additionally, more than one way to 59 00:03:06,990 --> 00:03:09,990 reach our desired outcome could exist. 60 00:03:09,990 --> 00:03:12,180 To account for this, we need to find the number 61 00:03:12,180 --> 00:03:15,090 of scenarios in which y out of the n 62 00:03:15,090 --> 00:03:17,223 many outcomes would be favorable. 63 00:03:18,510 --> 00:03:21,870 But these are actually the combinations we already know. 64 00:03:21,870 --> 00:03:23,700 For instance, if we wish to find 65 00:03:23,700 --> 00:03:25,800 out the number of ways in which four 66 00:03:25,800 --> 00:03:28,290 out of the six trials could be successful, 67 00:03:28,290 --> 00:03:30,330 it is the same as picking four elements 68 00:03:30,330 --> 00:03:32,640 out of a sample space of six. 69 00:03:32,640 --> 00:03:35,760 Now you see why combinatorics are a fundamental part 70 00:03:35,760 --> 00:03:38,700 of probability; thus, we need 71 00:03:38,700 --> 00:03:40,740 to find the number of combinations 72 00:03:40,740 --> 00:03:44,673 in which y out of the n outcomes would be favorable. 73 00:03:45,570 --> 00:03:47,280 For instance, there are three different ways to 74 00:03:47,280 --> 00:03:50,793 get tails exactly twice in three coin flips. 75 00:03:51,630 --> 00:03:53,790 Therefore, the probability function 76 00:03:53,790 --> 00:03:56,670 for a binomial distribution is the product 77 00:03:56,670 --> 00:03:58,680 of the number of combinations of picking 78 00:03:58,680 --> 00:04:01,170 y many elements out of n times 79 00:04:01,170 --> 00:04:02,970 p to the power of y 80 00:04:02,970 --> 00:04:05,250 times one minus p 81 00:04:05,250 --> 00:04:08,313 to the power of n minus y. 82 00:04:10,740 --> 00:04:15,003 Great. To see this in action, let us look at an example. 83 00:04:15,960 --> 00:04:19,350 Imagine you bought a single stock of General Motors. 84 00:04:19,350 --> 00:04:22,200 Historically, you know there is a 60% chance the price 85 00:04:22,200 --> 00:04:25,260 of your stock will go up on any given day 86 00:04:25,260 --> 00:04:27,960 and a 40% chance it will drop. 87 00:04:27,960 --> 00:04:29,850 By the price going up, we mean that 88 00:04:29,850 --> 00:04:32,400 the closing price is higher than the opening price. 89 00:04:33,360 --> 00:04:35,520 With the probability distribution function 90 00:04:35,520 --> 00:04:37,980 you can calculate the likelihood of the stock price 91 00:04:37,980 --> 00:04:41,463 increasing three times during the five workday week. 92 00:04:42,540 --> 00:04:45,240 If we wish to use the probability distribution formula 93 00:04:45,240 --> 00:04:50,240 we need to plug in three for y, five for n and 0.6 for p. 94 00:04:53,550 --> 00:04:57,360 After plugging in, we get number of different possible 95 00:04:57,360 --> 00:05:01,410 combinations of picking three elements out of five times 0.6 96 00:05:01,410 --> 00:05:05,283 to the power of three times 0.4 to the power of two. 97 00:05:06,300 --> 00:05:11,000 This is equivalent to 10 times 0.216 times 0.16 98 00:05:13,185 --> 00:05:14,890 or 0.3456 99 00:05:16,050 --> 00:05:20,790 Thus, we have a 34.56% chance of getting exactly 100 00:05:20,790 --> 00:05:23,643 three increases over the course of a work week. 101 00:05:24,810 --> 00:05:27,540 The big advantage of recognizing the distribution is 102 00:05:27,540 --> 00:05:30,030 that you can simply use these formulas and plug 103 00:05:30,030 --> 00:05:31,863 in the information you already have. 104 00:05:34,620 --> 00:05:37,830 All right, now that we know the probability function 105 00:05:37,830 --> 00:05:40,173 we can move on to the expected value. 106 00:05:41,246 --> 00:05:44,310 By definition, the expected value equals the sum 107 00:05:44,310 --> 00:05:47,310 of all values in the sample space multiplied 108 00:05:47,310 --> 00:05:49,083 by their respective probabilities. 109 00:05:50,280 --> 00:05:53,040 The expected value formula for a binomial event 110 00:05:53,040 --> 00:05:54,180 equals the probability 111 00:05:54,180 --> 00:05:56,280 of success for a given value 112 00:05:56,280 --> 00:05:59,310 multiplied by the number of trials we carry out. 113 00:05:59,310 --> 00:06:00,510 This seems familiar 114 00:06:00,510 --> 00:06:02,910 because this is the exact formula we used when 115 00:06:02,910 --> 00:06:04,500 computing the expected values 116 00:06:04,500 --> 00:06:07,593 for categorical variables in the beginning of the course. 117 00:06:09,120 --> 00:06:10,920 After computing the expected value 118 00:06:10,920 --> 00:06:13,560 we can finally calculate the variance. 119 00:06:13,560 --> 00:06:15,870 We do so by applying the short formula 120 00:06:15,870 --> 00:06:17,550 we learned earlier; 121 00:06:17,550 --> 00:06:20,370 Variance of y equals the expected value 122 00:06:20,370 --> 00:06:25,370 of y squared minus the expected value of y squared. 123 00:06:27,450 --> 00:06:30,090 After some simplifications, this results in 124 00:06:30,090 --> 00:06:33,273 n times p times one minus p. 125 00:06:34,260 --> 00:06:36,990 If we plug in the values from our stock market example 126 00:06:36,990 --> 00:06:38,250 that gives us a variance 127 00:06:38,250 --> 00:06:42,303 of five times 0.6 times 0.4, or 1.2. 128 00:06:43,590 --> 00:06:45,330 This would give us a standard deviation 129 00:06:45,330 --> 00:06:48,240 of approximately 1.1. 130 00:06:48,240 --> 00:06:49,620 Knowing the expected value 131 00:06:49,620 --> 00:06:52,470 and the standard deviation allows us to make more 132 00:06:52,470 --> 00:06:54,303 accurate future forecasts. 133 00:06:56,460 --> 00:06:58,650 Fantastic. In the next video, 134 00:06:58,650 --> 00:07:01,920 we are going to discuss Poisson distributions. 135 00:07:01,920 --> 00:07:02,973 Thanks for watching. 10807

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.