All language subtitles for 15. Regulating encryption, mandating insecurity & legalizing spying

af Afrikaans
ak Akan
sq Albanian
am Amharic
ar Arabic Download
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
el Greek
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranî)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian
pl Polish
pt-BR Portuguese (Brazil)
pt Portuguese (Portugal)
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese
sk Slovak
sl Slovenian
so Somali
es Spanish
es-419 Spanish (Latin American)
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:00,480 --> 00:00:07,280 One of the biggest threats to your security privacy and anonymity online is potentially Unfortunately 2 00:00:07,310 --> 00:00:07,970 your own. 3 00:00:07,970 --> 00:00:12,190 Go home and this threat is coming from a number of fronts. 4 00:00:12,200 --> 00:00:21,000 The two main fronts one is the push to weaken and regulate encryption and the other is to legalize it 5 00:00:21,000 --> 00:00:26,590 I guess legitimize mass surveillance and spying on its citizens. 6 00:00:26,640 --> 00:00:36,270 Many countries are talking about implementing policies that limits encryption and policies to legitimize 7 00:00:36,270 --> 00:00:37,950 and legalize spying. 8 00:00:37,980 --> 00:00:41,420 That has been going on anyway illegally for years. 9 00:00:41,550 --> 00:00:48,750 And in fact by the time you listen to this because it's moving so fast in your particular country things 10 00:00:48,750 --> 00:00:57,210 may have changed and it may now be more legal to spy on you and encryption may be regulated to a further 11 00:00:57,210 --> 00:01:07,140 extent this regulation and mandating of insecurity and legalised spying is going on in many places United 12 00:01:07,140 --> 00:01:13,450 States the U.K. China Russia Brazil India etc.. 13 00:01:13,620 --> 00:01:20,540 The UK has got the data communications bill which includes recording 12 months of Internet history and 14 00:01:20,610 --> 00:01:22,820 the Orwellian measures. 15 00:01:23,030 --> 00:01:27,600 Other examples that show the tide of change in other countries. 16 00:01:27,630 --> 00:01:33,980 You have whatsapp being banned for 48 hours in Brazil because of the encryption by the government. 17 00:01:33,990 --> 00:01:42,240 India India has some very strong ideas on limiting encryption Kazakstan legally requiring back doors 18 00:01:42,720 --> 00:01:45,420 encryption is fundamentally mathematics. 19 00:01:45,440 --> 00:01:47,550 It cannot be banned. 20 00:01:47,610 --> 00:01:51,470 The horse has already left the stable it's been created. 21 00:01:51,520 --> 00:02:00,770 It cannot be weakened just for a terrorist or a criminal or for someone who you want to have weak encryption. 22 00:02:00,930 --> 00:02:07,560 Those people will just use the strong encryption that's already out there and everyone else will be 23 00:02:07,560 --> 00:02:13,800 stuck with weakened security and weakened encryption because they're forced to use the weakened grips. 24 00:02:13,980 --> 00:02:21,840 If it is weakened or back doored is weakened for everyone including the hackers trying to compromise 25 00:02:21,840 --> 00:02:30,210 our systems something like this was actually tried already as the crypto wars in the 1990s something 26 00:02:30,210 --> 00:02:36,600 called the Clipper Chip was proposed by the U.S. government and the floor was found in it. 27 00:02:36,690 --> 00:02:42,180 And luckily there was no widespread adoption because this chip was going to be put into you know every 28 00:02:42,900 --> 00:02:45,660 electronic device that was going to do encryption. 29 00:02:45,660 --> 00:02:51,370 So the government could bypass encryption and look at what it is you were doing. 30 00:02:51,450 --> 00:02:56,700 So if that had happened that would have been a complete disaster because of the vulnerability that was 31 00:02:56,700 --> 00:02:57,230 found. 32 00:02:57,270 --> 00:02:58,320 This is the problem. 33 00:02:58,470 --> 00:03:02,440 If you weaken encryption you can weaken it for everybody. 34 00:03:02,460 --> 00:03:08,910 Terrorists and criminals will continue to use strong encryption even if normal citizens are banned. 35 00:03:08,910 --> 00:03:13,570 There is no evidence that weakening encryption will help at all. 36 00:03:13,650 --> 00:03:21,670 And to add to all that we have no feasible technical way of achieving this. 37 00:03:21,810 --> 00:03:28,670 Unfortunately all of this stuff is perhaps too complicated for people to really grasp. 38 00:03:28,680 --> 00:03:34,440 Those people that make decisions on these things or maybe they do understand it but because of political 39 00:03:34,480 --> 00:03:36,890 agendas they're still pushing forward. 40 00:03:37,020 --> 00:03:44,930 This is Matt Bley speaking to a U.S. congressional committee on the feasibility of these plans. 41 00:03:45,240 --> 00:03:48,510 And it's definitely worth a watch. 42 00:03:48,540 --> 00:03:50,590 So I'm going to play it now. 43 00:03:50,880 --> 00:03:52,860 It's only five minutes long. 44 00:03:52,950 --> 00:03:53,600 Dr. Blix. 45 00:03:53,620 --> 00:03:54,980 Five minutes. 46 00:03:55,500 --> 00:03:57,130 Thank you Mr. Chairman. 47 00:03:57,870 --> 00:04:06,180 As a technologist I'm finding myself in the very curious and participating in a debate over the desirability 48 00:04:06,180 --> 00:04:13,920 of something that sounds wonderful which is security systems that can be bypassed by the good guys but 49 00:04:13,920 --> 00:04:20,540 that also reliably the bad guys out and we could certainly discuss that. 50 00:04:20,540 --> 00:04:30,600 But as a technologist I can't ignore this stark reality which is simply that it can't be done safely. 51 00:04:30,810 --> 00:04:38,850 And if we make wishful policies that assume and pretend that we can there will be terrible consequences 52 00:04:38,850 --> 00:04:42,070 for our economy and for our national security. 53 00:04:43,490 --> 00:04:49,970 So it would be difficult to overstate today the importance of robust reliable computing and communications 54 00:04:49,970 --> 00:04:53,990 to our personal commercial and national security. 55 00:04:54,020 --> 00:05:00,980 Modern computing and network technologies are obviously yielding great benefits to our society and we 56 00:05:01,040 --> 00:05:07,340 are depending on them to be reliable and trustworthy in the same way that we depend on power and water 57 00:05:07,340 --> 00:05:10,170 and the rest of our critical infrastructure today. 58 00:05:10,610 --> 00:05:17,330 But unfortunately software based system is the foundation on which all of this modern communications 59 00:05:17,400 --> 00:05:26,120 technologies is based are also notoriously vulnerable to attack by criminals and by hostile nation states 60 00:05:26,780 --> 00:05:35,000 large scale data breaches of course are are literally a daily occurrence and this problem is getting 61 00:05:35,000 --> 00:05:41,750 worse rather than better as we build larger and more complex systems and it's really not an exaggeration 62 00:05:41,750 --> 00:05:49,560 to characterize the state of software security as an emerging national crisis. 63 00:05:49,910 --> 00:05:58,610 And the sad truth behind this is that computer science my field simply does not know how to build complex 64 00:05:58,610 --> 00:06:03,220 large scale software that has reliably correct behavior. 65 00:06:03,470 --> 00:06:08,930 And this is not a new problem it has nothing to do with encryption or modern technology. 66 00:06:08,930 --> 00:06:15,080 It's been the central focus of computing research since the dawn of the programmable computer. 67 00:06:15,320 --> 00:06:21,050 And as new technology allows us to build larger and more complex systems the problem of ensuring their 68 00:06:21,050 --> 00:06:26,660 reliability becomes actually exponentially harder with more and more components interacting with each 69 00:06:26,660 --> 00:06:27,470 other. 70 00:06:27,470 --> 00:06:34,370 So as we integrate in secure vulnerable systems into the fabric of our economy the consequences of those 71 00:06:34,370 --> 00:06:40,780 systems failing become both more likely and increasingly serious. 72 00:06:42,210 --> 00:06:47,910 Unfortunately there is no magic bullet for securing software based system. 73 00:06:47,910 --> 00:06:55,930 Large systems are fundamentally risky and this is something that we can at best manage rather than than 74 00:06:56,040 --> 00:06:58,170 fix outright. 75 00:06:58,350 --> 00:07:06,090 There are really only two known ways to manage the risk of unreliable and in secure software. 76 00:07:06,090 --> 00:07:15,300 One is the use of encryption which allows us to process sensitive data over in secure media and in secure 77 00:07:15,810 --> 00:07:18,740 software systems to the extent that we can. 78 00:07:18,750 --> 00:07:26,040 And the other is to design our software systems to be as small and as simple as we possibly can to minimize 79 00:07:26,040 --> 00:07:32,080 the number of features that a malicious attacker might be able to find flaws to exploit. 80 00:07:32,830 --> 00:07:39,700 And this is why proposals for law enforcement access features frighten me so much cryptographic systems 81 00:07:39,700 --> 00:07:44,930 are among the most fragile and subtle elements of modern software. 82 00:07:44,940 --> 00:07:51,270 We often discover devastating weaknesses in even very simple cryptographic systems years after they're 83 00:07:51,280 --> 00:07:57,910 designed and fielded with third party access requirements to do is take even very simple problems that 84 00:07:57,940 --> 00:08:03,940 we don't really know how to solve and turn them into far more complex problems that we really have no 85 00:08:03,940 --> 00:08:06,640 chance of reliably solving. 86 00:08:06,700 --> 00:08:14,530 So backdoor cryptography of the kind advocated by by the FBI you know might solve some problems if we 87 00:08:14,530 --> 00:08:15,010 could do it. 88 00:08:15,010 --> 00:08:21,820 But it's a notoriously and well-known difficult problem we've found subtle flaws even in systems designed 89 00:08:21,820 --> 00:08:25,320 by the National Security Agency such as the clipper chip. 90 00:08:25,420 --> 00:08:31,210 Two decades ago and even if we could get the cryptography right we'd be left with the problem of integrating 91 00:08:31,210 --> 00:08:37,840 access features into the software design and requiring designers to design around third party access 92 00:08:37,840 --> 00:08:44,740 requirements will basically undermine our already tenuous ability to defend against attack. 93 00:08:44,760 --> 00:08:48,670 It's tempting to frame this debate as being between personal privacy and law enforcement. 94 00:08:48,670 --> 00:08:51,180 But in fact the stakes are higher than that. 95 00:08:51,400 --> 00:08:58,370 We just can't do what the FBI is asking without seriously weakening our infrastructure. 96 00:08:58,540 --> 00:09:03,510 The ultimate beneficiaries will be criminals and rival nation states. 97 00:09:03,520 --> 00:09:12,010 Congress faces a crucial choice here to effectively legislate mandatory insecurity in our critical infrastructure 98 00:09:12,490 --> 00:09:19,720 or recognize the critical importance of robust security in preventing crime in our increasingly connected 99 00:09:19,720 --> 00:09:20,200 world. 100 00:09:20,350 --> 00:09:21,210 Thank you very much. 101 00:09:22,500 --> 00:09:24,650 This here is a good paper to read. 102 00:09:24,930 --> 00:09:30,900 And it's from a number of top crypto experts including some of the people that actually design the crypt 103 00:09:31,260 --> 00:09:37,500 that we're going to talk about on the course on why mandating insecurity is a bad idea. 104 00:09:37,590 --> 00:09:39,300 So if you want to dig deeper into that. 105 00:09:39,480 --> 00:09:45,030 Give the readers your homework and other quick Darkman that you can read is the case against regulating 106 00:09:45,120 --> 00:09:46,390 encryption technology. 107 00:09:46,410 --> 00:09:47,580 Only a couple of pages. 108 00:09:47,730 --> 00:09:49,640 And to give you more of a background. 109 00:09:49,870 --> 00:09:55,030 And this is a good read this is the nine epic failures of regulating encryption. 110 00:09:55,030 --> 00:10:01,480 It was a great report to give you an idea of the number of crypto products out there by Bruce Schneier. 111 00:10:01,620 --> 00:10:09,310 This is world wide survey of encryption products give the google the PTF version which is here. 112 00:10:09,370 --> 00:10:16,080 There's also an Excel version that excels pretty cool because you can sort by the type so you can look 113 00:10:16,080 --> 00:10:21,060 all the different of crypto products and maybe where we get to those sections and those products you 114 00:10:21,060 --> 00:10:29,430 can see all the ones that are out there it finds 865 hardware and software products and co-option encryption 115 00:10:29,760 --> 00:10:36,390 from 55 different countries so obviously if you have a law in one country something affects all the 116 00:10:36,390 --> 00:10:42,540 other countries and you know people who do use the crypto from whichever country they choose to use 117 00:10:42,540 --> 00:10:47,270 it for them let's move on to the legalization of spying and mass surveillance. 118 00:10:47,270 --> 00:10:55,010 Now I think we can start with some quotes from Edward Snowden So we have here arguing that you don't 119 00:10:55,010 --> 00:10:58,730 care about the right to privacy because you have nothing to hide. 120 00:10:58,940 --> 00:11:04,290 It's no different than saying you don't care about free speech because you have nothing to say. 121 00:11:04,340 --> 00:11:08,700 He continues people who use the I have nothing to hide line. 122 00:11:08,750 --> 00:11:12,050 Don't understand the basic foundation of human rights. 123 00:11:12,050 --> 00:11:15,990 Nobody needs to justify why they need a right. 124 00:11:16,010 --> 00:11:21,890 The burden of justification falls on the ones seeking to infringe upon the right. 125 00:11:21,920 --> 00:11:27,980 If one person chooses to disregard his right to privacy that doesn't automatically mean everyone should 126 00:11:27,980 --> 00:11:28,810 follow suit. 127 00:11:28,850 --> 00:11:34,440 Either you can't give away the rights of others because they're not useful to you. 128 00:11:34,460 --> 00:11:41,450 More simply the majority cannot vote away the natural rights of the minority. 129 00:11:41,450 --> 00:11:42,320 My view is this. 130 00:11:42,320 --> 00:11:46,720 When people know they will be watch they are being spied on. 131 00:11:46,760 --> 00:11:48,740 They alter what they do. 132 00:11:48,740 --> 00:11:50,870 They are no longer free. 133 00:11:50,930 --> 00:11:58,120 Terrorists want us to lose our freedom by creating mass surveillance to prevent terrorism by creating 134 00:11:58,210 --> 00:11:59,990 mass surveillance infrastructure. 135 00:11:59,990 --> 00:12:03,830 We lose the very freedom we are trying to protect. 136 00:12:03,980 --> 00:12:10,100 But the counter-argument to this is that we will be more secure from mass surveillance will be more 137 00:12:10,100 --> 00:12:12,330 safe from mass surveillance. 138 00:12:12,390 --> 00:12:15,310 But the evidence is thin to support that. 139 00:12:15,500 --> 00:12:22,360 The former head of the NSA global intelligence gathering operations Garko old Bill Binney. 140 00:12:22,400 --> 00:12:28,430 He says that mass surveillance interferes with the government's ability to catch bad guys and that the 141 00:12:28,430 --> 00:12:31,400 government's failure in terms of 9/11. 142 00:12:31,400 --> 00:12:38,540 The Boston bombing the Texas shooting and other terrorist attacks because it was overwhelmed with data 143 00:12:38,630 --> 00:12:40,440 from mass surveillance. 144 00:12:40,460 --> 00:12:47,150 For me the issue of mass surveillance is about giving away too much power to a government. 145 00:12:47,150 --> 00:12:55,910 Key questions to consider and ask can you trust all the people government offices agencies companies 146 00:12:55,970 --> 00:13:02,800 and contractors with your personal and private data gathered through this mass surveillance. 147 00:13:02,810 --> 00:13:09,230 Can you trust that they will always have your best interests at heart and that they will act justly 148 00:13:09,320 --> 00:13:10,100 with this power. 149 00:13:10,100 --> 00:13:18,230 This new power that they will have and not just now but in the future and with your children because 150 00:13:18,230 --> 00:13:21,650 your children will inherit a watched world. 151 00:13:21,650 --> 00:13:29,570 This data will be kept and any slight deviation from what is considered acceptable could be used against 152 00:13:29,570 --> 00:13:33,380 you if you oppose those in power. 153 00:13:33,470 --> 00:13:35,700 Consider the civil rights movement. 154 00:13:35,810 --> 00:13:42,470 If mass surveillance is going on during that political movement how would it affect a political change 155 00:13:42,470 --> 00:13:45,560 for black people in the United States. 156 00:13:45,560 --> 00:13:51,530 Would civil rights have happened much slower more violently because of the mass surveillance. 157 00:13:51,530 --> 00:13:58,490 Or would it have been crushed completely so they wouldn't even exist now if we had mass surveillance. 158 00:13:58,760 --> 00:13:59,820 Things to consider. 159 00:14:01,100 --> 00:14:06,210 Also consider donating to some of these privacy causes. 160 00:14:06,230 --> 00:14:12,770 If privacy is something that you are particularly interested in and passionate about regulating encryption 161 00:14:13,190 --> 00:14:22,010 mandating insecurity and legalizing spying is unfortunately an active threat that's potentially on your 162 00:14:22,010 --> 00:14:24,590 threat landscape that you need to be aware of. 163 00:14:24,620 --> 00:14:33,620 If your ability to use encryption is reduced then your security will be reduced as well and you will 164 00:14:33,620 --> 00:14:36,500 need all the mitigating controls. 18341

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.