Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated:
1
00:00:00,520 --> 00:00:03,140
Let's first start with state estimation.
2
00:00:03,140 --> 00:00:07,510
Again, the goal here is to be able to
obtain reliable estimates of the position
3
00:00:07,510 --> 00:00:11,450
and the velocity as the vehicle moves
through a three dimensional environment.
4
00:00:13,570 --> 00:00:18,250
So in the lab,
we have motion capture cameras
5
00:00:18,250 --> 00:00:23,064
that allow the robot to
measure its position.
6
00:00:23,064 --> 00:00:27,927
What essentially happens is through
the reflective markers that are mounted on
7
00:00:27,927 --> 00:00:32,792
the robot, the cameras can estimate
the position of each reflective marker and
8
00:00:32,792 --> 00:00:37,585
these cameras can do it at split second
timing exceeding speeds of 100 times
9
00:00:37,585 --> 00:00:39,980
a second to 200 times a second.
10
00:00:39,980 --> 00:00:42,190
Also, these measurements
are incredibly precise.
11
00:00:42,190 --> 00:00:44,499
The accuracies can be well
below one millimeter.
12
00:00:45,580 --> 00:00:48,670
Of course,
this only happens in a laboratory setting.
13
00:00:48,670 --> 00:00:50,400
What happens when you go outdoors?
14
00:00:51,720 --> 00:00:56,930
Well, in larger vehicles, like this X-47B
made by Northrop Grumman which is capable
15
00:00:56,930 --> 00:01:02,480
of landing on a ship deck autonomously,
the vehicle uses GPS and
16
00:01:02,480 --> 00:01:06,870
other kinds of communication
to determine where it is,
17
00:01:06,870 --> 00:01:09,800
relevant to the ship deck, and
is able to perform autonomously.
18
00:01:11,750 --> 00:01:16,240
If you go outdoors,
in general you might not have GPS, or
19
00:01:16,240 --> 00:01:20,620
your estimates of position from
GPS can be very inaccurate.
20
00:01:20,620 --> 00:01:22,950
This is especially true next
in extra tall buildings.
21
00:01:24,200 --> 00:01:28,950
Certainly when you go indoors,
it’s hard to get GPS.
22
00:01:28,950 --> 00:01:33,090
And we’d like to be able to
operate both indoors and outdoors.
23
00:01:33,090 --> 00:01:36,630
There’s no GPS and
because these vehicles are small and
24
00:01:36,630 --> 00:01:40,850
maneuverable, they can find
themselves in settings
25
00:01:40,850 --> 00:01:45,380
where it’s very difficult to
communicate directly with the vehicle.
26
00:01:45,380 --> 00:01:47,980
We also wanted these vehicles
to move very quickly and
27
00:01:47,980 --> 00:01:49,750
manuever through these
complex environments.
28
00:01:51,020 --> 00:01:52,940
How do we navigate without GPS,
29
00:01:52,940 --> 00:01:57,580
without external motion capture cameras or
any other kinds of external sensors?
30
00:01:58,740 --> 00:02:05,120
Well, imagine that the vehicle is
equipped with sensors such as cameras or
31
00:02:05,120 --> 00:02:09,430
color plus depth cameras as
you see on the bottom, or
32
00:02:09,430 --> 00:02:11,670
laser range finders as
you see on the top right.
33
00:02:13,180 --> 00:02:18,090
These sensors allow the vehicles
to infer information
34
00:02:18,090 --> 00:02:23,020
about the environment, and from this
information allow it to localize itself.
35
00:02:24,360 --> 00:02:26,090
How does that work?
36
00:02:26,090 --> 00:02:27,340
Well, let's look at this cartoon.
37
00:02:28,880 --> 00:02:33,890
Imagine you have a robot and imagine there
are three pillars in the environment.
38
00:02:34,890 --> 00:02:40,610
And let's imagine it has rain sensors that
allow it to detect these obstacles or
39
00:02:40,610 --> 00:02:41,920
the pillars.
40
00:02:41,920 --> 00:02:46,580
And imagine these rain sensors give you
estimates of where these pillars are, d1,
41
00:02:46,580 --> 00:02:47,260
d2 and d3.
42
00:02:49,460 --> 00:02:53,480
Now let's assume that the robot has
something like an inertial measurement
43
00:02:53,480 --> 00:02:57,810
unit that allows it to estimate its
movement as it goes from one position
44
00:02:57,810 --> 00:02:58,740
to another position.
45
00:03:00,740 --> 00:03:02,890
So you have some estimate of delta x, and
46
00:03:02,890 --> 00:03:06,750
when it gets to this new
position this range finder
47
00:03:06,750 --> 00:03:11,400
estimates the positions of the pillars
that it had measured previously.
48
00:03:11,400 --> 00:03:15,890
Except now these range estimates,
d1 prime, d2 prime, and d3 prime,
49
00:03:15,890 --> 00:03:20,170
are different from the original depth
estimates which are d1, d2, and d3.
50
00:03:20,170 --> 00:03:25,860
So, the question we wanna ask ourselves,
is it possible for the robot
51
00:03:25,860 --> 00:03:31,410
to concurrently estimate the locations of
the pillars and the displacement delta x.
52
00:03:32,710 --> 00:03:36,369
So if you think about it, you're trying
to estimate these eight variables.
53
00:03:38,140 --> 00:03:44,750
Three pairs of x y coordinates for
the pillars, and delta x.
54
00:03:44,750 --> 00:03:49,620
This problem is referred to as
Simultaneous Localization And Mapping, or
55
00:03:49,620 --> 00:03:50,460
simply SLAM.
56
00:03:51,940 --> 00:03:55,770
And the idea here is that you're
trying to localize yourself,
57
00:03:55,770 --> 00:04:00,830
in other words you're trying to estimate
delta x, while mapping the pillars,
58
00:04:00,830 --> 00:04:03,260
x1, y1, x2, y2, and x3, y3.
59
00:04:04,990 --> 00:04:05,670
In this video,
60
00:04:05,670 --> 00:04:09,730
you will see a video of a robot entering
a building that it hasn't seen before.
61
00:04:10,790 --> 00:04:15,900
It uses the SLAM methodology to map
the three dimensional building,
62
00:04:15,900 --> 00:04:19,651
while estimating its location relative
to the features in the building.
63
00:04:21,410 --> 00:04:25,770
The map that's building,
you can see in the central screen.
64
00:04:25,770 --> 00:04:29,970
The blue colors are the ground floor and
the red colors are the top floor.
65
00:04:31,770 --> 00:04:34,560
You will see there are intermediate
points where the vehicle
66
00:04:34,560 --> 00:04:36,280
actually plans its trajectory.
67
00:04:36,280 --> 00:04:42,230
If you look at the red snaking curve that
emanates from the vehicle to gold coins.
68
00:04:42,230 --> 00:04:47,010
These gold coins have been designated by
an operator that's viewing this map and
69
00:04:47,010 --> 00:04:48,120
tasking the vehicle.
70
00:04:49,350 --> 00:04:51,710
And the vehicle does everything else.
71
00:04:51,710 --> 00:04:53,270
In other words, you can click and
72
00:04:53,270 --> 00:04:57,090
point your way through this building
without entering the building,
73
00:04:57,090 --> 00:04:59,870
while getting information about
what's inside the building.
74
00:05:01,200 --> 00:05:05,400
In our lab we've built many
different types of vehicles.
75
00:05:05,400 --> 00:05:06,870
Here are four examples.
76
00:05:06,870 --> 00:05:12,780
On the top left you see a vehicle that's
powered by lasers, a set of cameras.
77
00:05:12,780 --> 00:05:16,950
It has a GPS unit as well as
an Inertial Measurement Unit.
78
00:05:16,950 --> 00:05:21,380
On the bottom left you see another vehicle
that's only powered by two cameras and
79
00:05:21,380 --> 00:05:22,410
an Inertial Measurement Unit.
80
00:05:23,570 --> 00:05:27,230
On the top right,
a smartphone drives the robot.
81
00:05:27,230 --> 00:05:35,260
On the bottom right, the vehicle is
actually instrumented with a RGBD camera.
82
00:05:35,260 --> 00:05:37,700
Red, green, blue, and depth camera,
83
00:05:37,700 --> 00:05:42,170
which you can now get as part of
an Xbox video entertainment system.
84
00:05:43,260 --> 00:05:45,150
It also has on it a laser scanner.
85
00:05:46,290 --> 00:05:50,240
You can see that each vehicle has
a different mass and a different size.
86
00:05:51,820 --> 00:05:54,400
And the reason for that is very simple.
87
00:05:54,400 --> 00:05:57,570
As you put more hardware on
the vehicle in terms of sensors and
88
00:05:57,570 --> 00:06:01,420
processors, the vehicle
has to become bigger.
89
00:06:01,420 --> 00:06:05,820
And to support this weight you have to
have bigger motors, and to support those
90
00:06:05,820 --> 00:06:10,530
motor you have to have bigger props,
which in turn requires bigger batteries.
91
00:06:12,050 --> 00:06:16,200
Here's another vehicle we use for
instruction in our classroom, and
92
00:06:16,200 --> 00:06:18,990
this vehicle just has a single camera and
an IMU.
93
00:06:21,700 --> 00:06:27,200
Because it only has a single camera, we
actually instrument the room with beacons.
94
00:06:27,200 --> 00:06:28,890
These are AprilTags.
95
00:06:28,890 --> 00:06:31,130
And as you'll see in this video,
96
00:06:31,130 --> 00:06:37,160
the vehicle is able to localize itself
with respect to these beacons, and hover.
97
00:06:37,160 --> 00:06:39,110
So it's estimating its position and
98
00:06:39,110 --> 00:06:43,650
velocity relative to the beacon and
hovering autonomously.
99
00:06:45,280 --> 00:06:48,710
You can switch between small markers and
100
00:06:48,710 --> 00:06:52,790
large markers which then allow
the vehicle to control its height.
101
00:06:53,970 --> 00:06:56,780
And, of course,
if you have lots of these markers,
102
00:06:56,780 --> 00:07:00,000
then you can actually navigate
over larger distances.
103
00:07:01,910 --> 00:07:05,780
So in our laboratory this is
an inexpensive replacement
104
00:07:05,780 --> 00:07:07,440
to the motion camera system.
105
00:07:08,600 --> 00:07:12,650
Here it's a single camera
which is off the shelf,
106
00:07:12,650 --> 00:07:17,910
it's inexpensive, but we do instrument
the environment with beacons.
107
00:07:17,910 --> 00:07:23,010
And because these beacons, which
are the AprilTags you see on the carpet,
108
00:07:23,010 --> 00:07:28,100
are known to the robot, It's able
to recognize them and estimate its
109
00:07:28,100 --> 00:07:33,240
position and orientation relative to
the tags, and therefore the environment.9948
Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.