All language subtitles for 007 Ingesting a CSV dataset into BigQuery and creating a table.en

af Afrikaans
ak Akan
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
el Greek
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranî)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian
pl Polish
pt-BR Portuguese (Brazil)
pt Portuguese (Portugal) Download
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese
sk Slovak
sl Slovenian
so Somali
es Spanish
es-419 Spanish (Latin American)
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:00,110 --> 00:00:05,700 In this section, we're going to start learning the most fundamental sequel commands and clauses that 2 00:00:05,700 --> 00:00:07,770 we can write in the bakery interface. 3 00:00:08,310 --> 00:00:13,620 And we're also going to upload our first dataset and create a table out of it. 4 00:00:14,500 --> 00:00:21,180 Now, if you recall, in Section one, we were exploring the Chicago taxi file and table. 5 00:00:21,600 --> 00:00:26,780 And what I've done is because there were a lot of records, I've created a sample data set and downloaded 6 00:00:26,790 --> 00:00:28,580 that in a CSB format. 7 00:00:28,860 --> 00:00:32,340 And so we're gonna take that CSB format and basically upload it here. 8 00:00:33,630 --> 00:00:36,480 Now, the first thing that you can see here is the left hand side. 9 00:00:36,480 --> 00:00:40,800 We have our project, big course demo and nothing within it. 10 00:00:41,460 --> 00:00:47,700 So I step one, we have to create a dataset so you can simply click create data set here. 11 00:00:48,700 --> 00:00:50,320 And it's fairly simple to do. 12 00:00:50,560 --> 00:00:55,510 We just have to put a data set I.D. that would be taxi. 13 00:00:57,890 --> 00:00:59,210 Data set. 14 00:01:00,080 --> 00:01:01,460 And you have to use underscores. 15 00:01:01,490 --> 00:01:04,240 You cannot have spaces when you're creating those ideas. 16 00:01:04,460 --> 00:01:06,530 That's just good, good practice. 17 00:01:08,030 --> 00:01:12,650 The default table expiration because we're using the sandbox account, it defaults to 60 days. 18 00:01:13,010 --> 00:01:16,610 So we're going to keep that there and create a data set. 19 00:01:16,910 --> 00:01:21,530 And almost instantly, you'll see that the data set has been created by expand here. 20 00:01:21,890 --> 00:01:28,760 We have our data set and we can click through in the data set and now we have the option to create a 21 00:01:28,760 --> 00:01:32,210 table table is what hosts the data. 22 00:01:32,810 --> 00:01:37,250 So we're going to create a table here and we will have a couple of options. 23 00:01:37,460 --> 00:01:39,940 So this menu pops up a create table. 24 00:01:39,950 --> 00:01:44,240 If I click create table from here, there are a few options. 25 00:01:45,200 --> 00:01:47,210 There is the drive. 26 00:01:47,240 --> 00:01:52,610 So you could upload a Google Drive, Google Sheets, a file. 27 00:01:52,670 --> 00:01:59,480 You could use Google cloud storage or in our case, you can also upload a C as V, so click, upload 28 00:02:00,050 --> 00:02:02,780 and I'll click select file here. 29 00:02:03,050 --> 00:02:08,720 Now I've you would have this file available in your resource section so you can download it on your 30 00:02:08,720 --> 00:02:13,400 local machine and you should be able to follow through this section. 31 00:02:14,180 --> 00:02:22,160 So you will simply click, browse and wherever you store that file, you will find it and click on Chicago 32 00:02:22,160 --> 00:02:22,470 trips. 33 00:02:22,610 --> 00:02:24,160 Ten thousand records sample. 34 00:02:25,130 --> 00:02:31,410 And what that does is one site click it immediately to file format changes to see ASV. 35 00:02:31,700 --> 00:02:34,190 Google actually recognized that that's a CSB file. 36 00:02:35,630 --> 00:02:39,800 And the destination here project, we've already have that data set. 37 00:02:39,830 --> 00:02:46,270 We simply created that the table here would be taxi trips. 38 00:02:47,450 --> 00:02:53,450 And I will put sample because all I've done is from the million and million records of that public data 39 00:02:53,450 --> 00:02:53,780 set. 40 00:02:54,650 --> 00:02:59,810 I've taken that and I've just picked ten thousand records that are now, you know, randomized. 41 00:03:01,720 --> 00:03:05,200 Once we have the table ready, then we go to the schema. 42 00:03:05,320 --> 00:03:07,240 So we we talked a lot about the schema. 43 00:03:07,870 --> 00:03:10,270 And there's a couple of options here when you're uploading. 44 00:03:10,690 --> 00:03:14,200 Now, the easy one that we're going to use here is auto detect. 45 00:03:14,440 --> 00:03:20,270 So relieving it up to big query to detect each each value that is within that data set. 46 00:03:20,680 --> 00:03:23,020 What type of what is the format? 47 00:03:24,340 --> 00:03:28,650 And it's using, you know, what are the input parameters. 48 00:03:29,140 --> 00:03:30,860 And that's all happening automatic. 49 00:03:31,690 --> 00:03:35,290 Now, you could do that manually either by editing this text. 50 00:03:35,350 --> 00:03:42,670 So if I click edit this text, I could do this whole process manually, or I can add the fields here 51 00:03:43,060 --> 00:03:45,730 simply by typing the name of the column. 52 00:03:46,450 --> 00:03:52,120 Choosing the type and you have string, integer, float and etc. and in choosing the mode. 53 00:03:52,750 --> 00:03:57,010 But we're not gonna do this in this in this case because we can auto detect. 54 00:03:57,820 --> 00:03:59,000 So we'll click on Detect. 55 00:03:59,800 --> 00:04:01,840 And we're going to move on to partitioning. 56 00:04:02,020 --> 00:04:03,040 So partitioning. 57 00:04:03,910 --> 00:04:06,500 We're gonna explore that in our first project. 58 00:04:06,610 --> 00:04:10,870 But it has to do with how much data do you want to load? 59 00:04:11,260 --> 00:04:16,810 Because when you create tables, you don't really want to query the whole table where there's millions 60 00:04:16,810 --> 00:04:17,680 and millions of records. 61 00:04:17,710 --> 00:04:22,150 If you just want to look at a subset, you can what is called a partition, that subset of data. 62 00:04:22,510 --> 00:04:25,210 But we're gonna see this in action in our project. 63 00:04:26,410 --> 00:04:28,900 Now, last but not least, advanced options here. 64 00:04:29,320 --> 00:04:33,070 When you click, then you'll expanded writing preference. 65 00:04:33,520 --> 00:04:38,020 Now, there's three options here because we're creating a brand new table. 66 00:04:39,070 --> 00:04:39,400 Right. 67 00:04:39,430 --> 00:04:40,120 If empty. 68 00:04:40,330 --> 00:04:43,750 So all the data that you're uploading will be written into the table. 69 00:04:43,780 --> 00:04:47,050 And there's going to be no problem because that table is anyways empty. 70 00:04:47,950 --> 00:04:49,990 Now, you can also choose a pen to table. 71 00:04:50,830 --> 00:04:57,460 And this could be used in a scenario where you have all the January data file and you want to upload 72 00:04:57,460 --> 00:04:58,270 the February data. 73 00:04:58,750 --> 00:05:03,220 So if you have a separate file, if you upload, then file in your keeping the same schema. 74 00:05:03,280 --> 00:05:04,330 And that's very important. 75 00:05:04,360 --> 00:05:09,010 If you're keeping the same schema, you can click append to table and all that will be appended as opposed 76 00:05:09,010 --> 00:05:12,010 to overwritten in some way. 77 00:05:12,790 --> 00:05:14,980 And that obviously brings us to option three. 78 00:05:15,100 --> 00:05:20,390 If we do want to override completely and change the entire table with new data. 79 00:05:20,620 --> 00:05:24,100 We can simply override would our new file. 80 00:05:24,340 --> 00:05:31,180 So if you've actually consolidated your, let's say, data for January and February already in a S.A.C. 81 00:05:31,180 --> 00:05:36,460 file and you just want to override the main table, you could do that by picking the overwrite table 82 00:05:36,460 --> 00:05:36,790 option. 83 00:05:37,570 --> 00:05:40,690 But in our case, right of empty because it's a brand new table. 84 00:05:41,820 --> 00:05:43,130 So we're all set here. 85 00:05:43,590 --> 00:05:45,150 We're going to click create table. 86 00:05:47,180 --> 00:05:53,230 It's going to take a little bit of time to compute the table, and that would show you can see Loja 87 00:05:53,360 --> 00:05:59,330 job created and immediately says taxi trips, sample, create it and you can see it reflected here. 88 00:05:59,920 --> 00:06:00,770 So taxi trips. 89 00:06:00,770 --> 00:06:09,740 And if I click on this, what we're going to see is our familiar type mode and the entire data schema. 90 00:06:09,890 --> 00:06:13,410 So it becomes very similar to what you saw in the public dataset. 91 00:06:14,450 --> 00:06:19,190 Now, in the next part, we're actually going to dive a little bit deeper into this and write our very 92 00:06:19,190 --> 00:06:20,030 first query. 9330

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.