All language subtitles for 04_state-estimation.en

af Afrikaans
sq Albanian
am Amharic
ar Arabic Download
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bn Bengali
bs Bosnian
bg Bulgarian
ca Catalan
ceb Cebuano
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
tl Filipino
fi Finnish
fr French
fy Frisian
gl Galician
ka Georgian
de German
el Greek
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
km Khmer
ko Korean
ku Kurdish (Kurmanji)
ky Kyrgyz
lo Lao
la Latin
lv Latvian
lt Lithuanian
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mn Mongolian
my Myanmar (Burmese)
ne Nepali
no Norwegian
ps Pashto
fa Persian
pl Polish
pt Portuguese
pa Punjabi
ro Romanian
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
st Sesotho
sn Shona
sd Sindhi
si Sinhala
sk Slovak
sl Slovenian
so Somali
es Spanish
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
te Telugu
th Thai
tr Turkish
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
or Odia (Oriya)
rw Kinyarwanda
tk Turkmen
tt Tatar
ug Uyghur
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:00,520 --> 00:00:03,140 Let's first start with state estimation. 2 00:00:03,140 --> 00:00:07,510 Again, the goal here is to be able to obtain reliable estimates of the position 3 00:00:07,510 --> 00:00:11,450 and the velocity as the vehicle moves through a three dimensional environment. 4 00:00:13,570 --> 00:00:18,250 So in the lab, we have motion capture cameras 5 00:00:18,250 --> 00:00:23,064 that allow the robot to measure its position. 6 00:00:23,064 --> 00:00:27,927 What essentially happens is through the reflective markers that are mounted on 7 00:00:27,927 --> 00:00:32,792 the robot, the cameras can estimate the position of each reflective marker and 8 00:00:32,792 --> 00:00:37,585 these cameras can do it at split second timing exceeding speeds of 100 times 9 00:00:37,585 --> 00:00:39,980 a second to 200 times a second. 10 00:00:39,980 --> 00:00:42,190 Also, these measurements are incredibly precise. 11 00:00:42,190 --> 00:00:44,499 The accuracies can be well below one millimeter. 12 00:00:45,580 --> 00:00:48,670 Of course, this only happens in a laboratory setting. 13 00:00:48,670 --> 00:00:50,400 What happens when you go outdoors? 14 00:00:51,720 --> 00:00:56,930 Well, in larger vehicles, like this X-47B made by Northrop Grumman which is capable 15 00:00:56,930 --> 00:01:02,480 of landing on a ship deck autonomously, the vehicle uses GPS and 16 00:01:02,480 --> 00:01:06,870 other kinds of communication to determine where it is, 17 00:01:06,870 --> 00:01:09,800 relevant to the ship deck, and is able to perform autonomously. 18 00:01:11,750 --> 00:01:16,240 If you go outdoors, in general you might not have GPS, or 19 00:01:16,240 --> 00:01:20,620 your estimates of position from GPS can be very inaccurate. 20 00:01:20,620 --> 00:01:22,950 This is especially true next in extra tall buildings. 21 00:01:24,200 --> 00:01:28,950 Certainly when you go indoors, it’s hard to get GPS. 22 00:01:28,950 --> 00:01:33,090 And we’d like to be able to operate both indoors and outdoors. 23 00:01:33,090 --> 00:01:36,630 There’s no GPS and because these vehicles are small and 24 00:01:36,630 --> 00:01:40,850 maneuverable, they can find themselves in settings 25 00:01:40,850 --> 00:01:45,380 where it’s very difficult to communicate directly with the vehicle. 26 00:01:45,380 --> 00:01:47,980 We also wanted these vehicles to move very quickly and 27 00:01:47,980 --> 00:01:49,750 manuever through these complex environments. 28 00:01:51,020 --> 00:01:52,940 How do we navigate without GPS, 29 00:01:52,940 --> 00:01:57,580 without external motion capture cameras or any other kinds of external sensors? 30 00:01:58,740 --> 00:02:05,120 Well, imagine that the vehicle is equipped with sensors such as cameras or 31 00:02:05,120 --> 00:02:09,430 color plus depth cameras as you see on the bottom, or 32 00:02:09,430 --> 00:02:11,670 laser range finders as you see on the top right. 33 00:02:13,180 --> 00:02:18,090 These sensors allow the vehicles to infer information 34 00:02:18,090 --> 00:02:23,020 about the environment, and from this information allow it to localize itself. 35 00:02:24,360 --> 00:02:26,090 How does that work? 36 00:02:26,090 --> 00:02:27,340 Well, let's look at this cartoon. 37 00:02:28,880 --> 00:02:33,890 Imagine you have a robot and imagine there are three pillars in the environment. 38 00:02:34,890 --> 00:02:40,610 And let's imagine it has rain sensors that allow it to detect these obstacles or 39 00:02:40,610 --> 00:02:41,920 the pillars. 40 00:02:41,920 --> 00:02:46,580 And imagine these rain sensors give you estimates of where these pillars are, d1, 41 00:02:46,580 --> 00:02:47,260 d2 and d3. 42 00:02:49,460 --> 00:02:53,480 Now let's assume that the robot has something like an inertial measurement 43 00:02:53,480 --> 00:02:57,810 unit that allows it to estimate its movement as it goes from one position 44 00:02:57,810 --> 00:02:58,740 to another position. 45 00:03:00,740 --> 00:03:02,890 So you have some estimate of delta x, and 46 00:03:02,890 --> 00:03:06,750 when it gets to this new position this range finder 47 00:03:06,750 --> 00:03:11,400 estimates the positions of the pillars that it had measured previously. 48 00:03:11,400 --> 00:03:15,890 Except now these range estimates, d1 prime, d2 prime, and d3 prime, 49 00:03:15,890 --> 00:03:20,170 are different from the original depth estimates which are d1, d2, and d3. 50 00:03:20,170 --> 00:03:25,860 So, the question we wanna ask ourselves, is it possible for the robot 51 00:03:25,860 --> 00:03:31,410 to concurrently estimate the locations of the pillars and the displacement delta x. 52 00:03:32,710 --> 00:03:36,369 So if you think about it, you're trying to estimate these eight variables. 53 00:03:38,140 --> 00:03:44,750 Three pairs of x y coordinates for the pillars, and delta x. 54 00:03:44,750 --> 00:03:49,620 This problem is referred to as Simultaneous Localization And Mapping, or 55 00:03:49,620 --> 00:03:50,460 simply SLAM. 56 00:03:51,940 --> 00:03:55,770 And the idea here is that you're trying to localize yourself, 57 00:03:55,770 --> 00:04:00,830 in other words you're trying to estimate delta x, while mapping the pillars, 58 00:04:00,830 --> 00:04:03,260 x1, y1, x2, y2, and x3, y3. 59 00:04:04,990 --> 00:04:05,670 In this video, 60 00:04:05,670 --> 00:04:09,730 you will see a video of a robot entering a building that it hasn't seen before. 61 00:04:10,790 --> 00:04:15,900 It uses the SLAM methodology to map the three dimensional building, 62 00:04:15,900 --> 00:04:19,651 while estimating its location relative to the features in the building. 63 00:04:21,410 --> 00:04:25,770 The map that's building, you can see in the central screen. 64 00:04:25,770 --> 00:04:29,970 The blue colors are the ground floor and the red colors are the top floor. 65 00:04:31,770 --> 00:04:34,560 You will see there are intermediate points where the vehicle 66 00:04:34,560 --> 00:04:36,280 actually plans its trajectory. 67 00:04:36,280 --> 00:04:42,230 If you look at the red snaking curve that emanates from the vehicle to gold coins. 68 00:04:42,230 --> 00:04:47,010 These gold coins have been designated by an operator that's viewing this map and 69 00:04:47,010 --> 00:04:48,120 tasking the vehicle. 70 00:04:49,350 --> 00:04:51,710 And the vehicle does everything else. 71 00:04:51,710 --> 00:04:53,270 In other words, you can click and 72 00:04:53,270 --> 00:04:57,090 point your way through this building without entering the building, 73 00:04:57,090 --> 00:04:59,870 while getting information about what's inside the building. 74 00:05:01,200 --> 00:05:05,400 In our lab we've built many different types of vehicles. 75 00:05:05,400 --> 00:05:06,870 Here are four examples. 76 00:05:06,870 --> 00:05:12,780 On the top left you see a vehicle that's powered by lasers, a set of cameras. 77 00:05:12,780 --> 00:05:16,950 It has a GPS unit as well as an Inertial Measurement Unit. 78 00:05:16,950 --> 00:05:21,380 On the bottom left you see another vehicle that's only powered by two cameras and 79 00:05:21,380 --> 00:05:22,410 an Inertial Measurement Unit. 80 00:05:23,570 --> 00:05:27,230 On the top right, a smartphone drives the robot. 81 00:05:27,230 --> 00:05:35,260 On the bottom right, the vehicle is actually instrumented with a RGBD camera. 82 00:05:35,260 --> 00:05:37,700 Red, green, blue, and depth camera, 83 00:05:37,700 --> 00:05:42,170 which you can now get as part of an Xbox video entertainment system. 84 00:05:43,260 --> 00:05:45,150 It also has on it a laser scanner. 85 00:05:46,290 --> 00:05:50,240 You can see that each vehicle has a different mass and a different size. 86 00:05:51,820 --> 00:05:54,400 And the reason for that is very simple. 87 00:05:54,400 --> 00:05:57,570 As you put more hardware on the vehicle in terms of sensors and 88 00:05:57,570 --> 00:06:01,420 processors, the vehicle has to become bigger. 89 00:06:01,420 --> 00:06:05,820 And to support this weight you have to have bigger motors, and to support those 90 00:06:05,820 --> 00:06:10,530 motor you have to have bigger props, which in turn requires bigger batteries. 91 00:06:12,050 --> 00:06:16,200 Here's another vehicle we use for instruction in our classroom, and 92 00:06:16,200 --> 00:06:18,990 this vehicle just has a single camera and an IMU. 93 00:06:21,700 --> 00:06:27,200 Because it only has a single camera, we actually instrument the room with beacons. 94 00:06:27,200 --> 00:06:28,890 These are AprilTags. 95 00:06:28,890 --> 00:06:31,130 And as you'll see in this video, 96 00:06:31,130 --> 00:06:37,160 the vehicle is able to localize itself with respect to these beacons, and hover. 97 00:06:37,160 --> 00:06:39,110 So it's estimating its position and 98 00:06:39,110 --> 00:06:43,650 velocity relative to the beacon and hovering autonomously. 99 00:06:45,280 --> 00:06:48,710 You can switch between small markers and 100 00:06:48,710 --> 00:06:52,790 large markers which then allow the vehicle to control its height. 101 00:06:53,970 --> 00:06:56,780 And, of course, if you have lots of these markers, 102 00:06:56,780 --> 00:07:00,000 then you can actually navigate over larger distances. 103 00:07:01,910 --> 00:07:05,780 So in our laboratory this is an inexpensive replacement 104 00:07:05,780 --> 00:07:07,440 to the motion camera system. 105 00:07:08,600 --> 00:07:12,650 Here it's a single camera which is off the shelf, 106 00:07:12,650 --> 00:07:17,910 it's inexpensive, but we do instrument the environment with beacons. 107 00:07:17,910 --> 00:07:23,010 And because these beacons, which are the AprilTags you see on the carpet, 108 00:07:23,010 --> 00:07:28,100 are known to the robot, It's able to recognize them and estimate its 109 00:07:28,100 --> 00:07:33,240 position and orientation relative to the tags, and therefore the environment.9948

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.