Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated:
1
00:00:03,060 --> 00:00:04,320
Instructor: Hello again.
2
00:00:04,320 --> 00:00:06,240
In this lecture, we are going to introduce
3
00:00:06,240 --> 00:00:07,950
one of the most important formulas
4
00:00:07,950 --> 00:00:09,930
in the world of probability.
5
00:00:09,930 --> 00:00:12,390
The Bayes' rule.
6
00:00:12,390 --> 00:00:16,830
People also refer to it as Bayes' Theorem, or Bayes' Law.
7
00:00:16,830 --> 00:00:19,233
So we will use all three interchangeably.
8
00:00:20,100 --> 00:00:23,463
For starters, take two events, A and B.
9
00:00:24,330 --> 00:00:26,880
According to the conditional probability formula,
10
00:00:26,880 --> 00:00:28,020
the conditional probability
11
00:00:28,020 --> 00:00:31,200
of A given B equals the probability
12
00:00:31,200 --> 00:00:35,043
of their intersection over the probability of event B.
13
00:00:36,660 --> 00:00:38,400
Using the multiplication rule,
14
00:00:38,400 --> 00:00:39,960
we can transform the numerator
15
00:00:39,960 --> 00:00:43,980
of this fraction to get the probability of the intersection
16
00:00:43,980 --> 00:00:47,610
of A and B equals the conditional probability
17
00:00:47,610 --> 00:00:51,873
of getting B given A times the probability of getting A.
18
00:00:53,100 --> 00:00:55,020
Therefore, the conditional probability
19
00:00:55,020 --> 00:00:59,403
of getting A given B is equal to the following fraction.
20
00:01:00,420 --> 00:01:01,680
The conditional probability
21
00:01:01,680 --> 00:01:06,300
of getting B given A times the probability of A
22
00:01:06,300 --> 00:01:08,553
divided by the probability of B.
23
00:01:10,110 --> 00:01:12,603
This equation is known as Bayes' Theorem.
24
00:01:13,560 --> 00:01:16,590
It is crucial because it allows us to find a relationship
25
00:01:16,590 --> 00:01:18,660
between the different conditional probabilities
26
00:01:18,660 --> 00:01:19,683
of two events.
27
00:01:21,210 --> 00:01:24,300
One of the most prominent examples of using Bayes' Rule
28
00:01:24,300 --> 00:01:27,060
is in medical research when trying to find
29
00:01:27,060 --> 00:01:29,403
a causal relationship between symptoms.
30
00:01:30,780 --> 00:01:32,550
Knowing both conditional probabilities
31
00:01:32,550 --> 00:01:34,140
between the two helps us
32
00:01:34,140 --> 00:01:35,610
make more reasonable arguments
33
00:01:35,610 --> 00:01:37,623
about which one causes the other.
34
00:01:38,940 --> 00:01:42,240
For instance, there is certain correlation between patients
35
00:01:42,240 --> 00:01:45,003
with back problems and patients wearing glasses.
36
00:01:45,930 --> 00:01:49,830
More specifically, 67% of people with spinal problems
37
00:01:49,830 --> 00:01:53,880
wear glasses while only 41% of patients with eyesight
38
00:01:53,880 --> 00:01:55,653
issues have back pains.
39
00:01:56,970 --> 00:02:00,420
These conditional probabilities suggest that it is much more
40
00:02:00,420 --> 00:02:03,300
likely for someone with back problems to wear glasses
41
00:02:03,300 --> 00:02:04,600
than the other way around.
42
00:02:06,420 --> 00:02:09,090
Even though we cannot find a direct causal link
43
00:02:09,090 --> 00:02:12,000
between the two, there exists some arguments
44
00:02:12,000 --> 00:02:13,353
to support such claims.
45
00:02:14,250 --> 00:02:16,980
For instance, most patients with back pain
46
00:02:16,980 --> 00:02:19,800
are either elderly or work a desk job
47
00:02:19,800 --> 00:02:22,860
where they remain stationary for long periods.
48
00:02:22,860 --> 00:02:25,800
Old age and a lot of time in front of the desktop computer
49
00:02:25,800 --> 00:02:28,800
can have a deteriorating effect on an individual's eyesight.
50
00:02:29,670 --> 00:02:32,760
However, many healthy and young individuals wear glasses
51
00:02:32,760 --> 00:02:34,200
from a young age.
52
00:02:34,200 --> 00:02:37,020
In those cases, there is no other underlying factor
53
00:02:37,020 --> 00:02:39,183
that would suggest incoming back pains.
54
00:02:40,830 --> 00:02:42,420
All right, similarly,
55
00:02:42,420 --> 00:02:45,423
we can also apply Bayes' Theorem in business.
56
00:02:46,260 --> 00:02:48,870
Let's explore this fictional scenario.
57
00:02:48,870 --> 00:02:51,270
Your boss wants you to do research about what companies
58
00:02:51,270 --> 00:02:54,000
are looking for in recent college graduates.
59
00:02:54,000 --> 00:02:56,793
Good academic performance or working experience.
60
00:02:58,170 --> 00:03:00,570
You go over the resumes of the last 200 people
61
00:03:00,570 --> 00:03:01,800
who matched the requirements
62
00:03:01,800 --> 00:03:03,483
and got the job they applied for.
63
00:03:04,410 --> 00:03:08,550
Out of those candidates, 45% had the relevant experience.
64
00:03:08,550 --> 00:03:11,820
In addition, 60% had good grades.
65
00:03:11,820 --> 00:03:14,670
Furthermore, we know that out of those 45%
66
00:03:14,670 --> 00:03:16,320
who had relevant experience,
67
00:03:16,320 --> 00:03:19,113
50% also performed well academically.
68
00:03:20,220 --> 00:03:21,900
To avoid any harsh decisions,
69
00:03:21,900 --> 00:03:23,970
we need to compute the conditional probability
70
00:03:23,970 --> 00:03:26,430
of the candidate to have relevant experience
71
00:03:26,430 --> 00:03:28,293
provided they had a high GPA.
72
00:03:29,610 --> 00:03:31,170
If we used Bayes' Theorem,
73
00:03:31,170 --> 00:03:33,150
we get that the conditional probability
74
00:03:33,150 --> 00:03:35,370
of a candidate performing well academically
75
00:03:35,370 --> 00:03:40,320
to have relevant experience is 0.5 times 0.45
76
00:03:40,320 --> 00:03:44,733
over 0.6 or approximately 0.375.
77
00:03:46,050 --> 00:03:50,370
Since 0.5 is greater than 0.375, then it is more likely
78
00:03:50,370 --> 00:03:53,310
for an experienced candidate to excel academically
79
00:03:53,310 --> 00:03:55,080
than for a student with high grades
80
00:03:55,080 --> 00:03:57,630
to have the required working pedigree.
81
00:03:57,630 --> 00:04:00,540
Thus, candidates who had internships are more likely
82
00:04:00,540 --> 00:04:02,253
to also have a high GPA.
83
00:04:03,510 --> 00:04:05,850
Therefore, firms are much more likely
84
00:04:05,850 --> 00:04:08,790
to get their ideal candidate if they go for somebody
85
00:04:08,790 --> 00:04:11,190
who has experience rather than somebody
86
00:04:11,190 --> 00:04:13,230
who thrived academically.
87
00:04:13,230 --> 00:04:15,450
Well, that explains a lot.
88
00:04:15,450 --> 00:04:16,860
Good thing, online courses
89
00:04:16,860 --> 00:04:18,660
are a completely different category.
90
00:04:20,519 --> 00:04:23,760
To wrap this section up, let's briefly go over the odd case
91
00:04:23,760 --> 00:04:26,733
of applied Bayes' Theorem to independent events.
92
00:04:28,440 --> 00:04:30,030
Take these two events,
93
00:04:30,030 --> 00:04:32,853
the weather being sunny and your code not working.
94
00:04:34,170 --> 00:04:37,020
You can always try and blame it on the sun being too bright
95
00:04:37,020 --> 00:04:39,450
but deep down, you know the weather has nothing to do
96
00:04:39,450 --> 00:04:41,550
with your code not compiling.
97
00:04:41,550 --> 00:04:44,820
Similarly, rain, wind, and snow are seldom affected
98
00:04:44,820 --> 00:04:46,743
by how well your algorithm performs.
99
00:04:47,760 --> 00:04:49,740
Let's get more specific.
100
00:04:49,740 --> 00:04:51,810
Imagine, you know the following,
101
00:04:51,810 --> 00:04:53,960
the probability of your code working is 0.3
102
00:04:55,045 --> 00:04:57,903
and the likelihood of it being sunny tomorrow is 0.4.
103
00:04:58,890 --> 00:05:00,780
Since the two events are independent,
104
00:05:00,780 --> 00:05:02,190
the probability of sunshine
105
00:05:02,190 --> 00:05:04,893
provided your code works is also 0.4.
106
00:05:06,090 --> 00:05:08,880
If we plug these values into Bayes' Theorem,
107
00:05:08,880 --> 00:05:10,980
then we get that the likelihood of your code working
108
00:05:10,980 --> 00:05:14,910
in good weather equals 0.4 times 0.3
109
00:05:14,910 --> 00:05:18,243
divided by 0.4 or simply 0.3.
110
00:05:19,140 --> 00:05:21,240
This result aligns with our suspicion
111
00:05:21,240 --> 00:05:24,180
that the chances of your algorithm performing as intended
112
00:05:24,180 --> 00:05:27,213
neither increase, nor decrease based on the weather.
113
00:05:29,190 --> 00:05:30,930
Good job, everybody.
114
00:05:30,930 --> 00:05:32,250
In the next section of the course,
115
00:05:32,250 --> 00:05:35,220
we are going to focus on probability distributions.
116
00:05:35,220 --> 00:05:37,530
We will introduce the most important ones.
117
00:05:37,530 --> 00:05:39,180
Talk about their expected values
118
00:05:39,180 --> 00:05:42,150
and how wide their prediction intervals should be.
119
00:05:42,150 --> 00:05:43,233
Thanks for watching.
9432
Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.