Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated:
1
00:00:03,180 --> 00:00:04,680
Instructor: Welcome back.
2
00:00:04,680 --> 00:00:08,340
In the last video we mentioned binomial distributions.
3
00:00:08,340 --> 00:00:10,680
In essence, binomial events are a sequence
4
00:00:10,680 --> 00:00:12,573
of identical Bernoulli events.
5
00:00:13,560 --> 00:00:14,940
Before we get into the differences
6
00:00:14,940 --> 00:00:17,700
and similarities between these two distributions
7
00:00:17,700 --> 00:00:19,980
let us examine the proper notation
8
00:00:19,980 --> 00:00:21,603
for a binomial distribution.
9
00:00:22,920 --> 00:00:27,000
We use the letter B to express a binomial distribution
10
00:00:27,000 --> 00:00:28,680
followed by the number of trials
11
00:00:28,680 --> 00:00:31,113
and the probability of success in each one.
12
00:00:32,100 --> 00:00:34,800
Therefore, we read the following statement as
13
00:00:34,800 --> 00:00:39,600
variable X follows a binomial distribution with 10 trials
14
00:00:39,600 --> 00:00:44,193
and a likelihood of 0.6 to succeed on each individual trial.
15
00:00:45,870 --> 00:00:49,200
Additionally, we can express a Bernoulli distribution
16
00:00:49,200 --> 00:00:51,783
as a binomial distribution with a single trial.
17
00:00:53,970 --> 00:00:56,520
All right, to better understand the differences
18
00:00:56,520 --> 00:00:58,380
between the two types of events
19
00:00:58,380 --> 00:01:00,033
suppose the following scenario.
20
00:01:01,140 --> 00:01:02,760
You go to class and your professor
21
00:01:02,760 --> 00:01:05,190
gives the class a surprise pop quiz
22
00:01:05,190 --> 00:01:07,590
which you have not prepared for.
23
00:01:07,590 --> 00:01:09,270
Luckily for you, the quiz consists
24
00:01:09,270 --> 00:01:12,240
of 10 true or false questions.
25
00:01:12,240 --> 00:01:14,460
In this case, guessing a single true
26
00:01:14,460 --> 00:01:17,670
or false question is a Bernoulli event,
27
00:01:17,670 --> 00:01:21,183
but guessing the entire quiz is a binomial event.
28
00:01:23,220 --> 00:01:24,210
All right, let's go back
29
00:01:24,210 --> 00:01:26,110
to the quiz example we just mentioned.
30
00:01:27,000 --> 00:01:30,390
In it, the expected value of the Bernoulli distribution
31
00:01:30,390 --> 00:01:33,483
suggests which outcome we expect for a single trial.
32
00:01:34,470 --> 00:01:35,820
Now, the expected value
33
00:01:35,820 --> 00:01:38,730
of the binomial distribution would suggest the number
34
00:01:38,730 --> 00:01:42,093
of times we expect to get a specific outcome.
35
00:01:43,800 --> 00:01:45,900
Great. Now, the graph
36
00:01:45,900 --> 00:01:48,870
of the binomial distribution represents the likelihood
37
00:01:48,870 --> 00:01:52,113
of attaining our desired outcome a specific number of times.
38
00:01:52,980 --> 00:01:55,620
If we run n trials, our graph
39
00:01:55,620 --> 00:01:59,220
would consist n plus one many bars
40
00:01:59,220 --> 00:02:02,910
one for each unique value from zero to n.
41
00:02:02,910 --> 00:02:05,790
For instance, we could be flipping the same unfair coin
42
00:02:05,790 --> 00:02:07,830
we had from the last lecture.
43
00:02:07,830 --> 00:02:11,190
If we toss it twice, we need bars for the three
44
00:02:11,190 --> 00:02:15,573
different outcomes; zero, one, or two tails.
45
00:02:17,040 --> 00:02:20,760
Fantastic. If we wish to find the associated likelihood
46
00:02:20,760 --> 00:02:23,580
of getting a given outcome a precise number of times
47
00:02:23,580 --> 00:02:26,910
over the course of n trials, we need to introduce
48
00:02:26,910 --> 00:02:29,943
the probability function of the binomial distribution.
49
00:02:30,796 --> 00:02:35,610
For starters, each individual trial is a Bernoulli trial
50
00:02:35,610 --> 00:02:38,730
so we express the probability of getting our desired outcome
51
00:02:38,730 --> 00:02:43,730
as p and the likelihood of the other one as one minus p
52
00:02:45,240 --> 00:02:49,080
in order to get our favorite outcome exactly y many times
53
00:02:49,080 --> 00:02:52,470
over the n trials, we also need to get the alternative
54
00:02:52,470 --> 00:02:55,923
outcome n minus y many times.
55
00:02:56,970 --> 00:02:58,710
If we don't account for this, we would
56
00:02:58,710 --> 00:03:01,380
be estimating the likelihood of getting our desired outcome
57
00:03:01,380 --> 00:03:03,513
at least y many times.
58
00:03:05,070 --> 00:03:06,990
Additionally, more than one way to
59
00:03:06,990 --> 00:03:09,990
reach our desired outcome could exist.
60
00:03:09,990 --> 00:03:12,180
To account for this, we need to find the number
61
00:03:12,180 --> 00:03:15,090
of scenarios in which y out of the n
62
00:03:15,090 --> 00:03:17,223
many outcomes would be favorable.
63
00:03:18,510 --> 00:03:21,870
But these are actually the combinations we already know.
64
00:03:21,870 --> 00:03:23,700
For instance, if we wish to find
65
00:03:23,700 --> 00:03:25,800
out the number of ways in which four
66
00:03:25,800 --> 00:03:28,290
out of the six trials could be successful,
67
00:03:28,290 --> 00:03:30,330
it is the same as picking four elements
68
00:03:30,330 --> 00:03:32,640
out of a sample space of six.
69
00:03:32,640 --> 00:03:35,760
Now you see why combinatorics are a fundamental part
70
00:03:35,760 --> 00:03:38,700
of probability; thus, we need
71
00:03:38,700 --> 00:03:40,740
to find the number of combinations
72
00:03:40,740 --> 00:03:44,673
in which y out of the n outcomes would be favorable.
73
00:03:45,570 --> 00:03:47,280
For instance, there are three different ways to
74
00:03:47,280 --> 00:03:50,793
get tails exactly twice in three coin flips.
75
00:03:51,630 --> 00:03:53,790
Therefore, the probability function
76
00:03:53,790 --> 00:03:56,670
for a binomial distribution is the product
77
00:03:56,670 --> 00:03:58,680
of the number of combinations of picking
78
00:03:58,680 --> 00:04:01,170
y many elements out of n times
79
00:04:01,170 --> 00:04:02,970
p to the power of y
80
00:04:02,970 --> 00:04:05,250
times one minus p
81
00:04:05,250 --> 00:04:08,313
to the power of n minus y.
82
00:04:10,740 --> 00:04:15,003
Great. To see this in action, let us look at an example.
83
00:04:15,960 --> 00:04:19,350
Imagine you bought a single stock of General Motors.
84
00:04:19,350 --> 00:04:22,200
Historically, you know there is a 60% chance the price
85
00:04:22,200 --> 00:04:25,260
of your stock will go up on any given day
86
00:04:25,260 --> 00:04:27,960
and a 40% chance it will drop.
87
00:04:27,960 --> 00:04:29,850
By the price going up, we mean that
88
00:04:29,850 --> 00:04:32,400
the closing price is higher than the opening price.
89
00:04:33,360 --> 00:04:35,520
With the probability distribution function
90
00:04:35,520 --> 00:04:37,980
you can calculate the likelihood of the stock price
91
00:04:37,980 --> 00:04:41,463
increasing three times during the five workday week.
92
00:04:42,540 --> 00:04:45,240
If we wish to use the probability distribution formula
93
00:04:45,240 --> 00:04:50,240
we need to plug in three for y, five for n and 0.6 for p.
94
00:04:53,550 --> 00:04:57,360
After plugging in, we get number of different possible
95
00:04:57,360 --> 00:05:01,410
combinations of picking three elements out of five times 0.6
96
00:05:01,410 --> 00:05:05,283
to the power of three times 0.4 to the power of two.
97
00:05:06,300 --> 00:05:11,000
This is equivalent to 10 times 0.216 times 0.16
98
00:05:13,185 --> 00:05:14,890
or 0.3456
99
00:05:16,050 --> 00:05:20,790
Thus, we have a 34.56% chance of getting exactly
100
00:05:20,790 --> 00:05:23,643
three increases over the course of a work week.
101
00:05:24,810 --> 00:05:27,540
The big advantage of recognizing the distribution is
102
00:05:27,540 --> 00:05:30,030
that you can simply use these formulas and plug
103
00:05:30,030 --> 00:05:31,863
in the information you already have.
104
00:05:34,620 --> 00:05:37,830
All right, now that we know the probability function
105
00:05:37,830 --> 00:05:40,173
we can move on to the expected value.
106
00:05:41,246 --> 00:05:44,310
By definition, the expected value equals the sum
107
00:05:44,310 --> 00:05:47,310
of all values in the sample space multiplied
108
00:05:47,310 --> 00:05:49,083
by their respective probabilities.
109
00:05:50,280 --> 00:05:53,040
The expected value formula for a binomial event
110
00:05:53,040 --> 00:05:54,180
equals the probability
111
00:05:54,180 --> 00:05:56,280
of success for a given value
112
00:05:56,280 --> 00:05:59,310
multiplied by the number of trials we carry out.
113
00:05:59,310 --> 00:06:00,510
This seems familiar
114
00:06:00,510 --> 00:06:02,910
because this is the exact formula we used when
115
00:06:02,910 --> 00:06:04,500
computing the expected values
116
00:06:04,500 --> 00:06:07,593
for categorical variables in the beginning of the course.
117
00:06:09,120 --> 00:06:10,920
After computing the expected value
118
00:06:10,920 --> 00:06:13,560
we can finally calculate the variance.
119
00:06:13,560 --> 00:06:15,870
We do so by applying the short formula
120
00:06:15,870 --> 00:06:17,550
we learned earlier;
121
00:06:17,550 --> 00:06:20,370
Variance of y equals the expected value
122
00:06:20,370 --> 00:06:25,370
of y squared minus the expected value of y squared.
123
00:06:27,450 --> 00:06:30,090
After some simplifications, this results in
124
00:06:30,090 --> 00:06:33,273
n times p times one minus p.
125
00:06:34,260 --> 00:06:36,990
If we plug in the values from our stock market example
126
00:06:36,990 --> 00:06:38,250
that gives us a variance
127
00:06:38,250 --> 00:06:42,303
of five times 0.6 times 0.4, or 1.2.
128
00:06:43,590 --> 00:06:45,330
This would give us a standard deviation
129
00:06:45,330 --> 00:06:48,240
of approximately 1.1.
130
00:06:48,240 --> 00:06:49,620
Knowing the expected value
131
00:06:49,620 --> 00:06:52,470
and the standard deviation allows us to make more
132
00:06:52,470 --> 00:06:54,303
accurate future forecasts.
133
00:06:56,460 --> 00:06:58,650
Fantastic. In the next video,
134
00:06:58,650 --> 00:07:01,920
we are going to discuss Poisson distributions.
135
00:07:01,920 --> 00:07:02,973
Thanks for watching.
10807
Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.