All language subtitles for We.Need.to.Talk.About.A.I.2020.1080p.AMZN.WEB-DL.DDP5.1.H.264-TEPES

af Afrikaans
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bn Bengali Download
bs Bosnian
bg Bulgarian Download
ca Catalan
ceb Cebuano
ny Chichewa
zh-CN Chinese (Simplified) Download
zh-TW Chinese (Traditional)
co Corsican
hr Croatian Download
cs Czech
da Danish
nl Dutch Download
en English
eo Esperanto
et Estonian
tl Filipino
fi Finnish
fr French Download
fy Frisian
gl Galician
ka Georgian
de German
el Greek
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian Download
is Icelandic
ig Igbo
id Indonesian
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
km Khmer
ko Korean
ku Kurdish (Kurmanji)
ky Kyrgyz
lo Lao
la Latin
lv Latvian
lt Lithuanian
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mn Mongolian
my Myanmar (Burmese)
ne Nepali
no Norwegian
ps Pashto
fa Persian
pl Polish
pt Portuguese Download
pa Punjabi
ro Romanian
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
st Sesotho
sn Shona
sd Sindhi
si Sinhala
sk Slovak
sl Slovenian
so Somali
es Spanish Download
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
te Telugu
th Thai
tr Turkish Download
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
or Odia (Oriya)
rw Kinyarwanda
tk Turkmen
tt Tatar
ug Uyghur
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:48,222 --> 00:00:51,703 If we build artificial general intelligence, 2 00:00:51,747 --> 00:00:53,531 that'll be the biggest event 3 00:00:53,575 --> 00:00:56,708 in the history of life on Earth. 4 00:00:56,752 --> 00:01:00,147 Alien intelligence inside a computer system that 5 00:01:00,190 --> 00:01:03,280 has control over the planet, the day it arrives... 6 00:01:03,324 --> 00:01:05,108 And science fiction books and movies 7 00:01:05,152 --> 00:01:07,197 have set it up in advance. 8 00:01:25,781 --> 00:01:27,696 What's the problem? 9 00:01:27,739 --> 00:01:29,698 I think you know what the problem is 10 00:01:29,741 --> 00:01:31,134 just as well as I do. 11 00:01:31,178 --> 00:01:32,701 What are you talking about, HAL? 12 00:01:34,094 --> 00:01:35,617 This mission is too important 13 00:01:35,660 --> 00:01:37,793 for me to allow you to jeopardize it. 14 00:01:39,360 --> 00:01:42,363 I don't know what you're talking about, HAL. 15 00:01:42,406 --> 00:01:45,975 HAL, a super-intelligent computer. 16 00:01:46,018 --> 00:01:48,325 We'd call it AI today. 17 00:01:49,544 --> 00:01:52,242 My character, Dave, was the astronaut 18 00:01:52,286 --> 00:01:55,027 in one of the most prophetic films ever made 19 00:01:55,071 --> 00:01:57,378 about the threat posed to humanity 20 00:01:57,421 --> 00:01:59,423 by artificial intelligence. 21 00:02:01,121 --> 00:02:03,166 HAL was science fiction. 22 00:02:03,210 --> 00:02:06,996 But now, AI is all around us. 23 00:02:07,039 --> 00:02:09,172 Artificial intelligence isn't the future. 24 00:02:09,216 --> 00:02:11,087 AI is already here. 25 00:02:11,131 --> 00:02:12,915 This week a robot ran for mayor 26 00:02:12,958 --> 00:02:14,177 in a small Japanese town. 27 00:02:14,221 --> 00:02:15,352 The White House says 28 00:02:15,396 --> 00:02:17,137 it is creating a new task force 29 00:02:17,180 --> 00:02:19,182 to focus on artificial intelligence. 30 00:02:19,226 --> 00:02:21,271 Google has already announced plans 31 00:02:21,315 --> 00:02:23,360 to invest more in AI research. 32 00:02:23,404 --> 00:02:28,713 IBM sending an artificial intelligence robot into space. 33 00:02:28,757 --> 00:02:31,673 Because of my small role in the conversation, 34 00:02:31,716 --> 00:02:34,154 I understand that we're on a journey, 35 00:02:34,197 --> 00:02:37,244 one that starts with narrow AI. 36 00:02:37,287 --> 00:02:39,942 Machines that perform specific tasks 37 00:02:39,985 --> 00:02:41,378 better than we can. 38 00:02:42,336 --> 00:02:44,381 But the AI of science fiction 39 00:02:44,425 --> 00:02:47,471 is artificial general intelligence, 40 00:02:47,515 --> 00:02:50,213 machines that can do everything 41 00:02:50,257 --> 00:02:51,693 better than us. 42 00:02:52,694 --> 00:02:54,478 Narrow AI is really 43 00:02:54,522 --> 00:02:57,220 like an idiot savant to the extreme, 44 00:02:57,264 --> 00:02:58,830 where it can do something, 45 00:02:58,874 --> 00:03:00,963 like multiply numbers really fast, 46 00:03:01,006 --> 00:03:02,617 way better than any human. 47 00:03:02,660 --> 00:03:06,011 But it can't do anything else whatsoever. 48 00:03:06,055 --> 00:03:09,928 But AGI, artificial general intelligence, 49 00:03:09,972 --> 00:03:12,453 which doesn't exist yet, is instead 50 00:03:12,496 --> 00:03:14,629 intelligence that can do everything 51 00:03:14,672 --> 00:03:15,934 as well as humans can. 52 00:03:21,244 --> 00:03:24,247 So what happens if we do create AGI? 53 00:03:25,248 --> 00:03:27,294 Science fiction authors have warned us 54 00:03:27,337 --> 00:03:28,773 it could be dangerous. 55 00:03:29,513 --> 00:03:31,385 And they're not the only ones. 56 00:03:32,864 --> 00:03:35,911 I think people should be concerned about it. 57 00:03:35,954 --> 00:03:37,565 AI is a fundamental risk 58 00:03:37,608 --> 00:03:39,349 to the existence of human civilization. 59 00:03:39,393 --> 00:03:42,178 And I don't think people fully appreciate that. 60 00:03:42,222 --> 00:03:43,919 I do think we have to worry about it. 61 00:03:43,962 --> 00:03:45,225 I don't think it's inherent 62 00:03:45,268 --> 00:03:48,010 as we create super intelligence 63 00:03:48,053 --> 00:03:49,533 that it will necessarily always 64 00:03:49,577 --> 00:03:51,927 have the same goals in mind that we do. 65 00:03:51,970 --> 00:03:53,233 I think the development 66 00:03:53,276 --> 00:03:55,713 of full artificial intelligence 67 00:03:55,757 --> 00:03:58,368 could spell the end of the human race. 68 00:03:58,412 --> 00:04:00,457 ...the end of the human race. 69 00:04:01,893 --> 00:04:05,201 Tomorrow, if the headline said an artificial intelligence 70 00:04:05,245 --> 00:04:08,248 has been manufactured that's as intelligent 71 00:04:08,291 --> 00:04:10,293 or more intelligent than human beings, 72 00:04:10,337 --> 00:04:12,077 people would argue about it, but I don't think 73 00:04:12,121 --> 00:04:13,992 they'd be surprised, because science fiction 74 00:04:14,036 --> 00:04:15,907 has helped us imagine all kinds 75 00:04:15,951 --> 00:04:17,431 of unimaginable things. 76 00:04:20,999 --> 00:04:23,393 Experts are divided about the impact 77 00:04:23,437 --> 00:04:25,613 AI will have on our future, 78 00:04:25,656 --> 00:04:29,747 and whether or not Hollywood is helping us plan for it. 79 00:04:29,791 --> 00:04:32,750 Our dialogue has really been hijacked by Hollywood 80 00:04:32,794 --> 00:04:35,927 because The Terminator makes a better blockbuster 81 00:04:35,971 --> 00:04:38,147 than AI being good, 82 00:04:38,190 --> 00:04:39,409 or AI being neutral, 83 00:04:39,453 --> 00:04:41,063 or AI being confused. 84 00:04:41,106 --> 00:04:42,456 Technology here in AI 85 00:04:42,499 --> 00:04:44,501 is no different. It's a tool. 86 00:04:44,545 --> 00:04:47,156 I like to think of it as a fancy pencil. 87 00:04:47,199 --> 00:04:49,593 And so what pictures we draw with it, 88 00:04:49,637 --> 00:04:52,248 that's up to us as a society. 89 00:04:52,292 --> 00:04:55,251 We don't have to do what AI tells us. 90 00:04:55,295 --> 00:04:56,861 We tell AI what to do. 91 00:04:58,646 --> 00:05:00,300 I think Hollywood films 92 00:05:00,343 --> 00:05:02,214 do help the conversation. 93 00:05:02,258 --> 00:05:03,781 Someone needs to be thinking 94 00:05:03,825 --> 00:05:07,089 ahead of what the consequences are. 95 00:05:07,132 --> 00:05:08,786 But just as we can't really imagine 96 00:05:08,830 --> 00:05:11,311 what the actual science of the future is going to be like, 97 00:05:11,354 --> 00:05:13,400 I don't think we've begun 98 00:05:13,443 --> 00:05:16,316 to really think about all of the possible futures, 99 00:05:16,359 --> 00:05:19,188 which logically spring out of this technology. 100 00:05:19,231 --> 00:05:23,497 Artificial intelligence has the power to change society. 101 00:05:23,540 --> 00:05:25,499 A growing chorus of criticism 102 00:05:25,542 --> 00:05:27,283 is highlighting the dangers 103 00:05:27,327 --> 00:05:29,416 of handing control to machines. 104 00:05:29,459 --> 00:05:31,331 There's gonna be a lot of change coming. 105 00:05:31,374 --> 00:05:32,897 The larger long-term concern 106 00:05:32,941 --> 00:05:35,639 is that humanity will be sort of shunted aside. 107 00:05:35,683 --> 00:05:37,598 This is something Stanley Kubrick and others 108 00:05:37,641 --> 00:05:39,469 were worried about 50 years ago, right? 109 00:05:41,993 --> 00:05:43,647 It's happening. 110 00:05:43,691 --> 00:05:47,608 How do we gauge the urgency of this conversation? 111 00:05:49,436 --> 00:05:51,742 If people saw on radar right now 112 00:05:51,786 --> 00:05:54,179 that an alien spaceship was approaching Earth 113 00:05:54,223 --> 00:05:56,181 and it was 25 years away, 114 00:05:56,225 --> 00:05:57,835 we would be mobilized to prepare 115 00:05:57,879 --> 00:06:00,621 for that alien's arrival 25 years from now. 116 00:06:00,664 --> 00:06:03,058 But that's exactly the situation that we're in 117 00:06:03,101 --> 00:06:04,799 with artificial intelligence. 118 00:06:04,842 --> 00:06:06,670 Could be 25 years, could be 50 years, 119 00:06:06,714 --> 00:06:09,456 but an alien intelligence will arrive 120 00:06:09,499 --> 00:06:11,109 and we should be prepared. 121 00:06:14,286 --> 00:06:16,637 Well, gentlemen, meet Tobor. 122 00:06:16,680 --> 00:06:19,379 An electronic simulacrum of a man. 123 00:06:19,422 --> 00:06:21,772 Gosh. Gee willikers. 124 00:06:21,816 --> 00:06:24,645 Even though much work remains before he is completed, 125 00:06:24,688 --> 00:06:26,211 he is already a sentient being. 126 00:06:28,388 --> 00:06:30,390 Back in 1956 when the term 127 00:06:30,433 --> 00:06:32,696 artificial intelligence first came about, 128 00:06:32,740 --> 00:06:35,395 the original goals of artificial intelligence 129 00:06:35,438 --> 00:06:37,701 were human-level intelligence. 130 00:06:37,745 --> 00:06:41,052 He does look almost kind, doesn't he? 131 00:06:41,096 --> 00:06:43,838 Over time that's proved to be really difficult. 132 00:06:46,449 --> 00:06:49,626 I think, eventually, we will have human-level 133 00:06:49,670 --> 00:06:51,889 intelligence from machines. 134 00:06:51,933 --> 00:06:55,153 But it may be a few hundred years. It's gonna take a while. 135 00:06:55,197 --> 00:06:57,634 And so people are getting a little over-excited 136 00:06:57,678 --> 00:06:59,854 about the future capabilities. 137 00:06:59,897 --> 00:07:03,945 And now, Elektro, I command you to walk back. 138 00:07:04,293 --> 00:07:05,425 Back. 139 00:07:08,297 --> 00:07:09,559 Almost every generation, 140 00:07:09,603 --> 00:07:11,169 people have got so enthusiastic. 141 00:07:11,213 --> 00:07:12,736 They watch 2001, 142 00:07:12,780 --> 00:07:14,564 they fall in love with HAL, they think, 143 00:07:14,608 --> 00:07:16,871 "Yeah, give me six weeks. I'll get this sorted." 144 00:07:17,785 --> 00:07:19,830 It doesn't happen and everyone gets upset. 145 00:07:21,179 --> 00:07:24,052 So what's changed in AI research? 146 00:07:26,489 --> 00:07:28,839 The difference with the AI we have today 147 00:07:28,883 --> 00:07:32,495 and the reason AI suddenly took this big leap forward 148 00:07:32,539 --> 00:07:34,758 is the Internet. That is their world. 149 00:07:34,802 --> 00:07:37,587 Tonight the information superhighway, 150 00:07:37,631 --> 00:07:39,415 and one of its main thoroughfares, 151 00:07:39,459 --> 00:07:42,505 an online network called Internet. 152 00:07:42,549 --> 00:07:45,421 In 1981, only 213 computers 153 00:07:45,465 --> 00:07:46,901 were hooked to the Internet. 154 00:07:46,944 --> 00:07:48,816 As the new year begins, an estimated 155 00:07:48,859 --> 00:07:50,600 two-and-a-half million computers 156 00:07:50,644 --> 00:07:51,993 will be on the network. 157 00:07:53,516 --> 00:07:55,126 AIs need to eat. 158 00:07:55,170 --> 00:07:57,346 Just like sheep need grass, 159 00:07:57,389 --> 00:07:59,087 AIs need data. 160 00:07:59,130 --> 00:08:02,525 And the great prairies of data are on the Internet. 161 00:08:03,134 --> 00:08:04,484 That's what it is. 162 00:08:05,136 --> 00:08:06,181 They said, "You know what? 163 00:08:06,224 --> 00:08:07,661 "Let's just let them loose out there." 164 00:08:09,532 --> 00:08:11,839 And suddenly, the AIs came to life. 165 00:08:11,882 --> 00:08:14,276 Imagine thousands of talking robots 166 00:08:14,319 --> 00:08:15,973 able to move as one unit, 167 00:08:16,017 --> 00:08:17,322 taking on crime, fires, 168 00:08:17,366 --> 00:08:18,976 and natural disasters. 169 00:08:19,020 --> 00:08:21,805 Artificial intelligence platform Amper 170 00:08:21,849 --> 00:08:24,416 has created the first album entirely composed 171 00:08:24,460 --> 00:08:26,157 and produced by artificial intelligence. 172 00:08:26,201 --> 00:08:29,552 Next week, Christie's will be the first auction house 173 00:08:29,596 --> 00:08:33,251 to offer artwork created by artificial intelligence. 174 00:08:34,165 --> 00:08:35,819 Driverless cars, said to be 175 00:08:35,863 --> 00:08:38,648 in our not-so-distant future. 176 00:08:38,692 --> 00:08:40,955 So even though we don't have AGI 177 00:08:40,998 --> 00:08:44,262 or human-level intelligence yet, 178 00:08:44,306 --> 00:08:45,699 are we ready to give autonomy 179 00:08:45,742 --> 00:08:48,440 to the machines we do have? 180 00:08:48,484 --> 00:08:51,574 Should self-driving cars make ethical decisions? 181 00:08:51,618 --> 00:08:53,881 That is the growing debate as the technology 182 00:08:53,924 --> 00:08:56,318 moves closer towards mainstream reality. 183 00:08:58,233 --> 00:08:59,843 This idea of whether or not 184 00:08:59,887 --> 00:09:02,846 we can embed ethics into machines, 185 00:09:02,890 --> 00:09:04,544 whether or not they are automated machines 186 00:09:04,587 --> 00:09:06,415 or autonomous machines, 187 00:09:06,458 --> 00:09:08,286 I'm not sure we've got that figured out. 188 00:09:08,330 --> 00:09:09,723 If you have a self-driving car, 189 00:09:09,766 --> 00:09:11,202 it's gonna have to make ethical decisions. 190 00:09:11,246 --> 00:09:14,641 A classic one people have been thinking a lot about is... 191 00:09:14,684 --> 00:09:16,556 I don't know if you've heard of trolley problems. 192 00:09:16,599 --> 00:09:18,122 The trolley problem.The trolley problem. 193 00:09:18,166 --> 00:09:20,211 The trolley problem.The trolley problem. 194 00:09:21,604 --> 00:09:22,823 In the trolley problem, 195 00:09:22,866 --> 00:09:24,172 whether or not a driverless car 196 00:09:24,215 --> 00:09:26,870 is going to save the driver or run over 197 00:09:26,914 --> 00:09:29,046 a group of school children crossing the road. 198 00:09:29,090 --> 00:09:31,614 Will the car go right and hit the old lady 199 00:09:31,658 --> 00:09:35,575 or will it go left and kill the four young people? 200 00:09:35,618 --> 00:09:39,143 Either kill five people standing on a bus stop 201 00:09:39,187 --> 00:09:40,405 or kill yourself. 202 00:09:40,449 --> 00:09:41,972 It's an ethical decision 203 00:09:42,016 --> 00:09:43,931 that has to be made in the moment, 204 00:09:43,974 --> 00:09:46,020 too fast for any person to get in there. 205 00:09:46,063 --> 00:09:48,544 So it's the software in your self-driving car 206 00:09:48,588 --> 00:09:50,546 that's gonna make that ethical decision. 207 00:09:50,590 --> 00:09:51,852 Who decides that? 208 00:09:52,679 --> 00:09:53,897 Smart society, 209 00:09:53,941 --> 00:09:55,899 society that values human life, 210 00:09:55,943 --> 00:09:58,336 should make a decision that 211 00:09:58,380 --> 00:10:02,427 killing one person is preferable to killing five people. 212 00:10:02,471 --> 00:10:06,344 Of course, it does raise the question who's gonna buy that car. 213 00:10:06,388 --> 00:10:09,130 Or, you know, are you gonna buy the car where actually, 214 00:10:09,173 --> 00:10:11,654 it'll swerve and you'll be fine, 215 00:10:11,698 --> 00:10:13,134 but they're goners. 216 00:10:14,918 --> 00:10:17,442 There are hundreds of those kinds of questions 217 00:10:17,486 --> 00:10:19,140 all throughout society. 218 00:10:19,183 --> 00:10:22,883 And so, ethical philosophers battle back and forth. 219 00:10:25,146 --> 00:10:27,583 My peeve with the philosophers is 220 00:10:27,627 --> 00:10:30,804 they're sucking the oxygen out of the room. 221 00:10:30,847 --> 00:10:33,763 There's this metaphor that I have of 222 00:10:33,807 --> 00:10:37,114 you have a car driving away from a wedding, 223 00:10:37,158 --> 00:10:41,205 and the car is driving, and behind it there are cans. 224 00:10:41,249 --> 00:10:45,645 And the metal cans are clanking and making noise. 225 00:10:45,688 --> 00:10:47,690 That car is science 226 00:10:47,734 --> 00:10:50,345 and technology driving forward decisively. 227 00:10:50,388 --> 00:10:51,955 And the metal cans clanking 228 00:10:51,999 --> 00:10:54,697 are the philosophers making a lot of noise. 229 00:10:54,741 --> 00:10:56,525 The world is changing in such a way 230 00:10:56,568 --> 00:11:01,138 that the question of us building machines 231 00:11:01,182 --> 00:11:04,098 that could become more intelligent than we are 232 00:11:04,141 --> 00:11:07,710 is more of a reality than a fiction right now. 233 00:11:10,147 --> 00:11:12,759 Don't we have to ask questions now, 234 00:11:12,802 --> 00:11:15,805 before the machines get smarter than us? 235 00:11:15,849 --> 00:11:18,373 I mean, what if they achieve consciousness? 236 00:11:19,635 --> 00:11:21,463 How far away is that? 237 00:11:25,075 --> 00:11:26,816 How close we are to a conscious machine? 238 00:11:26,860 --> 00:11:28,688 That is the question that brings 239 00:11:28,731 --> 00:11:32,082 philosophers and scientists to blows, I think. 240 00:11:32,126 --> 00:11:34,302 You are dead center 241 00:11:34,345 --> 00:11:37,871 of the greatest scientific event in the history of man. 242 00:11:37,914 --> 00:11:39,699 If you've created a conscious machine, 243 00:11:39,742 --> 00:11:41,309 it's not the history of man. 244 00:11:42,614 --> 00:11:44,834 That's the history of gods. 245 00:11:44,878 --> 00:11:48,533 The problem with consciousness is hardly anybody 246 00:11:48,577 --> 00:11:50,884 can define exactly what it is. 247 00:11:53,408 --> 00:11:55,976 Is it consciousness that makes us special? 248 00:11:57,586 --> 00:11:58,718 Human? 249 00:12:00,894 --> 00:12:03,026 A way of thinking about consciousness 250 00:12:03,070 --> 00:12:05,159 which a lot of people will agree with... 251 00:12:05,202 --> 00:12:07,857 You'll never find something that appeals to everyone. 252 00:12:07,901 --> 00:12:10,468 ...is it's not just flashing lights, but someone's home. 253 00:12:11,252 --> 00:12:13,036 So if you have a... 254 00:12:13,080 --> 00:12:15,386 If you bought one of those little 255 00:12:15,430 --> 00:12:16,953 simple chess-playing machines, 256 00:12:17,606 --> 00:12:19,173 it might beat you at chess. 257 00:12:19,216 --> 00:12:21,218 There are lots of lights flashing. 258 00:12:21,262 --> 00:12:23,743 But there's nobody in when it wins, 259 00:12:23,786 --> 00:12:25,222 going "I won." 260 00:12:25,962 --> 00:12:29,749 All right? Whereas your dog 261 00:12:29,792 --> 00:12:32,534 is a rubbish chess player, 262 00:12:32,577 --> 00:12:33,927 but you look at the dog and you think, 263 00:12:33,970 --> 00:12:36,930 "Someone's actually in. It's not just a mechanism." 264 00:12:38,845 --> 00:12:42,283 People wonder whether general intelligence 265 00:12:42,326 --> 00:12:44,459 requires consciousness. 266 00:12:46,766 --> 00:12:50,160 It might partly depend on what level of general intelligence you're talking about. 267 00:12:50,204 --> 00:12:53,250 A mouse has some kind of general intelligence. 268 00:12:53,294 --> 00:12:55,339 My guess is, if you've got a machine which has 269 00:12:55,383 --> 00:12:58,473 general intelligence at the level of, say, a mouse or beyond, 270 00:12:58,516 --> 00:13:00,997 it's gonna be a good contender for having consciousness. 271 00:13:04,653 --> 00:13:05,697 We're not there yet. 272 00:13:05,741 --> 00:13:07,787 We're not confronted by humanoid robots 273 00:13:07,830 --> 00:13:09,832 that are smarter than we are. 274 00:13:09,876 --> 00:13:12,182 We're in this curious ethical position 275 00:13:12,226 --> 00:13:15,446 of not knowing where and how consciousness emerges, 276 00:13:15,490 --> 00:13:18,841 and we're building increasingly complicated minds. 277 00:13:18,885 --> 00:13:20,800 We won't know whether or not they're conscious, 278 00:13:20,843 --> 00:13:22,062 but we might feel they are. 279 00:13:24,194 --> 00:13:26,718 The easiest example is if you programmed 280 00:13:26,762 --> 00:13:29,634 into a computer everything you knew about joy, 281 00:13:29,678 --> 00:13:31,201 every reference in literature, 282 00:13:31,245 --> 00:13:33,943 every definition, every chemical compound 283 00:13:33,987 --> 00:13:36,119 that's involved with joy in the human brain, 284 00:13:36,163 --> 00:13:39,644 and you put all of that information in a computer, 285 00:13:39,688 --> 00:13:42,778 you could argue that it understood joy. 286 00:13:42,822 --> 00:13:45,476 The question is, had it ever felt joy? 287 00:13:51,613 --> 00:13:53,354 Did HAL have emotions? 288 00:13:55,095 --> 00:13:58,228 Will AIs ever feel the way we do? 289 00:13:59,882 --> 00:14:02,406 It's very hard to get machines to do the things 290 00:14:02,450 --> 00:14:06,149 that we don't think about, that we're not conscious of. 291 00:14:06,193 --> 00:14:09,326 Things like detecting that that person over there 292 00:14:09,370 --> 00:14:11,198 is really unhappy with the way that... 293 00:14:11,241 --> 00:14:12,721 Something about what I'm doing. 294 00:14:12,764 --> 00:14:14,027 And they're sending a subtle signal 295 00:14:14,070 --> 00:14:15,506 and I probably should change what I'm doing. 296 00:14:17,334 --> 00:14:18,988 We don't know how to build machines that can actually be 297 00:14:19,032 --> 00:14:22,252 tuned in to those kinds of tiny little social cues. 298 00:14:24,994 --> 00:14:28,389 There's this notion that maybe there is something magical 299 00:14:28,432 --> 00:14:30,434 about brains, 300 00:14:30,478 --> 00:14:32,045 about the wetware of having 301 00:14:32,088 --> 00:14:33,785 an information processing system 302 00:14:33,829 --> 00:14:34,917 made of meat, right? 303 00:14:34,961 --> 00:14:37,311 And that whatever we do in our machines, 304 00:14:37,354 --> 00:14:40,705 however competent they begin to seem, 305 00:14:40,749 --> 00:14:42,272 they never really will be intelligent 306 00:14:42,316 --> 00:14:44,579 in the way that we experience ourselves to be. 307 00:14:46,233 --> 00:14:48,975 There's really no basis for that. 308 00:14:51,238 --> 00:14:53,240 This silly idea you can't be intelligent 309 00:14:53,283 --> 00:14:55,155 unless you're made of meat. 310 00:14:56,765 --> 00:14:58,201 From my perspective, as a physicist, 311 00:14:58,245 --> 00:15:00,682 that's just carbon chauvinism. 312 00:15:00,725 --> 00:15:02,118 Those machines are made of exactly 313 00:15:02,162 --> 00:15:03,903 the same kind of elementary particles 314 00:15:03,946 --> 00:15:05,382 that my brain is made out of. 315 00:15:06,775 --> 00:15:08,429 Intelligence is just a certain... 316 00:15:08,472 --> 00:15:10,170 And consciousness is just 317 00:15:10,213 --> 00:15:12,259 a certain kind of information processing. 318 00:15:15,262 --> 00:15:17,655 Maybe we're not so special after all. 319 00:15:19,744 --> 00:15:21,268 Could our carbon-based brains 320 00:15:21,311 --> 00:15:23,531 be replicated with silicon? 321 00:15:25,185 --> 00:15:28,405 We're trying to build a computer-generated baby 322 00:15:28,449 --> 00:15:31,060 that we can teach like a real baby, 323 00:15:31,104 --> 00:15:34,237 with the potential for general intelligence. 324 00:15:39,808 --> 00:15:42,550 To do that, you really have to build a computer brain. 325 00:15:46,162 --> 00:15:50,645 I wanted to build a simple toy brain model 326 00:15:52,168 --> 00:15:53,474 in an embodied system 327 00:15:53,517 --> 00:15:56,694 which you could interact with face-to-face. 328 00:16:00,394 --> 00:16:03,527 And I happen to have an infant at home, 329 00:16:03,571 --> 00:16:04,964 a real baby. 330 00:16:05,660 --> 00:16:07,836 And I scanned the baby 331 00:16:07,879 --> 00:16:10,795 so I could build a 3D model out of her 332 00:16:10,839 --> 00:16:12,493 and then use her face 333 00:16:12,536 --> 00:16:15,452 as basically the embodiment of the system. 334 00:16:17,063 --> 00:16:20,109 If a brain which is made out of cells, 335 00:16:20,153 --> 00:16:23,199 and it's got blood pumping, is able to think, 336 00:16:24,461 --> 00:16:27,682 it just may be possible that a computer can think 337 00:16:28,770 --> 00:16:33,166 if the information is moving in the same sort of process. 338 00:16:37,474 --> 00:16:39,476 A lot of artificial intelligence at the moment today 339 00:16:39,520 --> 00:16:42,044 is really about advanced pattern recognition. 340 00:16:42,088 --> 00:16:44,003 It's not about killer robots 341 00:16:44,046 --> 00:16:45,830 that are gonna take over the world, et cetera. 342 00:16:47,702 --> 00:16:51,053 So this is Baby X. She's running live on my computer. 343 00:16:51,097 --> 00:16:54,056 So as I move my hand around, make a loud noise, 344 00:16:54,100 --> 00:16:55,710 she'll get a fright. 345 00:16:55,753 --> 00:16:57,407 We can kind of zoom in. 346 00:16:57,451 --> 00:17:00,845 So she basically can see me and hear me, you know. 347 00:17:00,889 --> 00:17:03,022 Hey, sweetheart. 348 00:17:03,065 --> 00:17:05,067 She's not copying my smile, 349 00:17:05,111 --> 00:17:07,417 she's responding to my smile. 350 00:17:08,114 --> 00:17:09,637 So she's responding to me 351 00:17:09,680 --> 00:17:12,205 and we're really concerned with the question 352 00:17:12,248 --> 00:17:14,207 of how will we actually interact 353 00:17:14,250 --> 00:17:16,644 with artificial intelligence in the future. 354 00:17:20,039 --> 00:17:23,520 When I'm talking to Baby X, I'm deliberately modulating my voice 355 00:17:23,564 --> 00:17:25,740 and deliberately doing big facial expressions. 356 00:17:25,783 --> 00:17:27,350 I'm going "Ooh," all this stuff 357 00:17:27,394 --> 00:17:29,309 because I'm getting her attention. 358 00:17:29,744 --> 00:17:31,572 Ooh. 359 00:17:32,834 --> 00:17:34,749 And now that I've got her attention, 360 00:17:35,793 --> 00:17:37,012 I can teach something. 361 00:17:37,056 --> 00:17:38,579 Okay, so here's Baby X, 362 00:17:38,622 --> 00:17:41,973 and this is... She's been learning to read words. 363 00:17:42,017 --> 00:17:43,758 So here's her first word book. 364 00:17:43,801 --> 00:17:47,370 So let's see what she can see. Turn to a page... 365 00:17:47,414 --> 00:17:49,329 And here we go. Let's see what she... 366 00:17:49,372 --> 00:17:51,679 What's this, Baby? What's this? What's this? 367 00:17:52,245 --> 00:17:54,160 What's this? Sheep. 368 00:17:54,203 --> 00:17:56,336 Good girl. Now see if she knows what the word is. 369 00:17:56,379 --> 00:17:58,207 Okay. Baby, look over here. 370 00:17:58,251 --> 00:18:00,079 Off you go. What's this? What's this? 371 00:18:01,776 --> 00:18:03,952 Sheep.Good girl. 372 00:18:03,995 --> 00:18:05,606 Let's try her on something else. 373 00:18:05,649 --> 00:18:07,216 Okay. 374 00:18:07,260 --> 00:18:08,826 - Let's try on that. - What's this? 375 00:18:08,870 --> 00:18:10,437 Baby, Baby, over here. What's this? 376 00:18:15,268 --> 00:18:17,618 Baby. Look at me. Look at me. What's this? 377 00:18:17,661 --> 00:18:19,402 Baby, over here. Over here. 378 00:18:20,708 --> 00:18:22,927 Puppy.Good girl. 379 00:18:22,971 --> 00:18:24,233 That's what she's just read. 380 00:18:24,277 --> 00:18:26,148 Ooh. 381 00:18:26,192 --> 00:18:28,411 The most intelligent system 382 00:18:28,455 --> 00:18:30,805 that we're aware of is a human. 383 00:18:32,198 --> 00:18:35,114 So using the human as a template, 384 00:18:35,157 --> 00:18:37,333 as the embodiment of Baby X, 385 00:18:37,377 --> 00:18:39,509 is probably the best way to try 386 00:18:39,553 --> 00:18:41,337 to achieve general intelligence. 387 00:18:48,170 --> 00:18:49,650 She seems so human. 388 00:18:51,304 --> 00:18:53,349 Is this the child who takes our hand 389 00:18:53,393 --> 00:18:55,699 and leads us to HAL? 390 00:18:57,484 --> 00:19:00,922 Whether it's in a real world or a virtual world, 391 00:19:00,965 --> 00:19:02,358 I do think embodiment 392 00:19:02,402 --> 00:19:05,492 is gonna be very crucial here. 393 00:19:05,535 --> 00:19:07,885 Because I think that general intelligence 394 00:19:07,929 --> 00:19:10,888 is tied up with being able to act in the world. 395 00:19:13,108 --> 00:19:14,805 You have to have something, 396 00:19:14,849 --> 00:19:16,633 arms, legs, eyes, 397 00:19:16,677 --> 00:19:18,461 something that you can move around 398 00:19:18,505 --> 00:19:21,682 in order to harvest new information. 399 00:19:26,513 --> 00:19:28,906 So I think to get systems that really understand 400 00:19:28,950 --> 00:19:30,647 what they're talking about, 401 00:19:30,691 --> 00:19:32,997 you will need systems that can act 402 00:19:33,041 --> 00:19:34,564 and intervene in the world, 403 00:19:34,608 --> 00:19:36,827 and possibly systems that have something 404 00:19:36,871 --> 00:19:39,308 almost like a social life. 405 00:19:39,352 --> 00:19:43,007 For me, thinking and feeling will come together. 406 00:19:43,051 --> 00:19:44,574 And they will come together 407 00:19:44,618 --> 00:19:47,534 because of the importance of embodiment. 408 00:19:52,539 --> 00:19:53,975 The stories that we tell ourselves 409 00:19:54,018 --> 00:19:56,325 open up arenas for discussion, 410 00:19:56,369 --> 00:19:58,109 and they also help us think about things. 411 00:19:58,153 --> 00:19:59,415 Ava! 412 00:20:02,592 --> 00:20:04,377 The most dominant story... 413 00:20:04,420 --> 00:20:05,769 Go back to your room. 414 00:20:05,813 --> 00:20:07,684 ...is AI is scary. 415 00:20:07,728 --> 00:20:09,991 There's a bad guy, there's a good guy... 416 00:20:12,123 --> 00:20:14,822 That is a non-productive storyline. 417 00:20:14,865 --> 00:20:15,866 Stop! 418 00:20:16,954 --> 00:20:18,608 Ava, I said stop! 419 00:20:20,784 --> 00:20:22,656 Whoa! Whoa! Whoa! 420 00:20:22,699 --> 00:20:27,269 The reality is AI is remarkably useful to us. 421 00:20:27,313 --> 00:20:30,838 Robot surgeons are increasingly used in NHS hospitals. 422 00:20:30,881 --> 00:20:33,623 They don't get tired, their hands don't tremble, 423 00:20:33,667 --> 00:20:36,060 and their patients make a faster recovery. 424 00:20:36,104 --> 00:20:38,672 It alerted me that there was one animal in heat. 425 00:20:38,715 --> 00:20:41,065 She was ready to be bred, and three animals that 426 00:20:41,109 --> 00:20:42,806 may have a potential health problem. 427 00:20:42,850 --> 00:20:45,548 A Facebook program can detect suicidal behavior 428 00:20:45,592 --> 00:20:47,681 by analyzing user information, 429 00:20:47,724 --> 00:20:50,074 and Facebook CEO Mark Zuckerberg believes 430 00:20:50,118 --> 00:20:52,425 this is just the beginning. 431 00:20:52,468 --> 00:20:55,036 Here is someone who has visual impairment 432 00:20:55,079 --> 00:20:57,386 being able to see because of AI. 433 00:20:57,430 --> 00:21:00,128 Those are the benefits of what AI does. 434 00:21:00,171 --> 00:21:01,738 Diagnose cancer earlier. 435 00:21:01,782 --> 00:21:03,174 Predict heart disease. 436 00:21:03,218 --> 00:21:05,481 But the revolution has only just begun. 437 00:21:06,917 --> 00:21:08,397 AI is. 438 00:21:08,441 --> 00:21:10,617 It's happening. And so it is a question 439 00:21:10,660 --> 00:21:13,533 of what we do with it, how we build it. 440 00:21:13,576 --> 00:21:15,665 That's the only question that matters. And I think, 441 00:21:15,709 --> 00:21:19,060 we potentially are at risk of ruining that 442 00:21:19,103 --> 00:21:20,453 by having this knee-jerk reaction 443 00:21:20,496 --> 00:21:22,672 that AI is bad and it's coming to get us. 444 00:21:25,196 --> 00:21:27,155 So just like my astronaut, 445 00:21:27,198 --> 00:21:28,461 we're on a journey. 446 00:21:29,200 --> 00:21:31,638 Narrow AI is already here. 447 00:21:32,421 --> 00:21:35,294 But how far are we from AGI? 448 00:21:36,425 --> 00:21:39,210 And how fast are we going? 449 00:21:39,254 --> 00:21:40,603 Every five years, 450 00:21:40,647 --> 00:21:43,214 computers are getting ten times faster. 451 00:21:43,258 --> 00:21:45,216 At the moment, we don't have yet 452 00:21:45,260 --> 00:21:48,263 little devices, computational devices, 453 00:21:48,307 --> 00:21:51,484 that can compute as much as a human brain. 454 00:21:51,527 --> 00:21:52,702 But every five years, 455 00:21:52,746 --> 00:21:55,575 computers are getting ten times cheaper, 456 00:21:55,618 --> 00:21:57,925 which means that soon we are going to have that. 457 00:21:57,968 --> 00:22:01,189 And once we have that, then 50 years later, 458 00:22:01,232 --> 00:22:02,451 if the trend doesn't break, 459 00:22:02,495 --> 00:22:04,801 we should have for the same price 460 00:22:04,845 --> 00:22:07,630 little computational devices that can compute 461 00:22:07,674 --> 00:22:10,329 as much as all ten billion brains 462 00:22:10,372 --> 00:22:11,808 of humankind together. 463 00:22:14,202 --> 00:22:15,551 It's catching up. 464 00:22:16,770 --> 00:22:20,034 What happens if it passes us by? 465 00:22:20,077 --> 00:22:23,342 Now a robot that's been developed by a Japanese scientist 466 00:22:23,385 --> 00:22:25,387 is so fast, it can win 467 00:22:25,431 --> 00:22:27,346 the rock-paper-scissors game against a human 468 00:22:27,389 --> 00:22:29,130 every single time. 469 00:22:29,173 --> 00:22:31,175 The presumption has to be that 470 00:22:31,219 --> 00:22:33,177 when an artificial intelligence 471 00:22:33,221 --> 00:22:35,266 reaches human intelligence, 472 00:22:35,310 --> 00:22:38,008 it will very quickly exceed human intelligence, 473 00:22:38,052 --> 00:22:39,227 and once that happens, 474 00:22:39,270 --> 00:22:41,272 it doesn't even need us anymore 475 00:22:41,316 --> 00:22:43,884 to make it smarter. It can make itself smarter. 476 00:22:46,408 --> 00:22:49,455 You create AIs which are better than us 477 00:22:49,498 --> 00:22:50,586 at a lot of things. 478 00:22:50,630 --> 00:22:52,545 And one of the things they get better at 479 00:22:52,588 --> 00:22:54,373 is programming computers. 480 00:22:54,416 --> 00:22:56,113 So it reprograms itself. 481 00:22:56,157 --> 00:22:57,898 And it is now better than it was, 482 00:22:57,941 --> 00:23:01,380 which means it's now better than it was at reprogramming. 483 00:23:02,729 --> 00:23:05,384 So it doesn't get better like that. 484 00:23:06,123 --> 00:23:07,690 It gets better like that. 485 00:23:09,562 --> 00:23:11,172 And we get left behind 486 00:23:11,215 --> 00:23:13,304 as it keeps getting faster, 487 00:23:13,348 --> 00:23:15,872 evolving at an exponential rate. 488 00:23:15,916 --> 00:23:18,005 That's this event that a lot of researchers 489 00:23:18,048 --> 00:23:19,746 refer to as the singularity. 490 00:23:21,965 --> 00:23:23,706 And it will get beyond 491 00:23:23,750 --> 00:23:26,405 where we are competitive very quickly. 492 00:23:26,448 --> 00:23:29,625 So think about someone not having a PhD in one area, 493 00:23:29,669 --> 00:23:32,280 but having it in every conceivable area. 494 00:23:32,323 --> 00:23:34,413 You can't figure out what they're doing. 495 00:23:34,456 --> 00:23:36,806 They can do things no one else is capable of. 496 00:23:39,461 --> 00:23:40,854 Just imagine what it would be like 497 00:23:40,897 --> 00:23:43,247 to be in dialogue with another person 498 00:23:43,291 --> 00:23:44,814 running on a time course 499 00:23:44,858 --> 00:23:48,557 that was a million times faster than yours. 500 00:23:48,601 --> 00:23:51,255 So for every second you had to think 501 00:23:51,299 --> 00:23:52,431 about what you were going to do 502 00:23:52,474 --> 00:23:53,867 or what you were going to say, 503 00:23:53,910 --> 00:23:57,436 it had two weeks to think about this, right? 504 00:23:57,479 --> 00:24:00,569 Um, or a billion times faster, 505 00:24:00,613 --> 00:24:02,528 so if for every second you had to think, 506 00:24:02,571 --> 00:24:05,487 it it's high-stakes poker or some negotiation, 507 00:24:05,531 --> 00:24:09,448 every second you are thinking, it has 32 years 508 00:24:09,491 --> 00:24:11,798 to think about what its next move will be. 509 00:24:12,581 --> 00:24:14,931 There's no way to win in that game. 510 00:24:15,715 --> 00:24:17,456 The singularity, 511 00:24:17,499 --> 00:24:19,719 the moment the machines surpass us. 512 00:24:20,415 --> 00:24:22,548 Inevitable? Not everyone 513 00:24:22,591 --> 00:24:25,289 at the forefront of AI thinks so. 514 00:24:25,333 --> 00:24:27,857 Some of my friends who are worried about the future 515 00:24:27,901 --> 00:24:30,033 think that we're gonna get to a singularity 516 00:24:30,077 --> 00:24:31,687 where the AI systems are able 517 00:24:31,731 --> 00:24:33,602 to reprogram themselves and make themselves 518 00:24:33,646 --> 00:24:35,517 better and better and better. 519 00:24:36,910 --> 00:24:39,652 Maybe in the long-distant future. 520 00:24:39,695 --> 00:24:42,045 But our present AI systems 521 00:24:42,089 --> 00:24:44,831 are very, very narrow at what they do. 522 00:24:45,745 --> 00:24:47,573 We can't build a system today 523 00:24:47,616 --> 00:24:49,488 that can understand a computer program 524 00:24:49,531 --> 00:24:52,316 at the level that a brand-new student in programming 525 00:24:52,360 --> 00:24:54,754 understands in four or five weeks. 526 00:24:55,581 --> 00:24:56,756 And by the way, 527 00:24:56,799 --> 00:24:59,019 people have been working on this problem 528 00:24:59,062 --> 00:25:00,499 for over 40 years. 529 00:25:00,542 --> 00:25:02,065 It's not because people haven't tried. 530 00:25:02,109 --> 00:25:03,980 It's because we don't have any clue how to do it. 531 00:25:04,024 --> 00:25:06,766 So to imagine it's suddenly gonna jump out of nothing 532 00:25:06,809 --> 00:25:09,986 I think is completely wrong. 533 00:25:11,640 --> 00:25:14,600 There's a negative narrative about AI. 534 00:25:14,643 --> 00:25:16,602 Brilliant people like Elon Musk 535 00:25:16,645 --> 00:25:19,648 use very religious imagery. 536 00:25:19,692 --> 00:25:22,999 With artificial intelligence, we are summoning the demon. 537 00:25:23,043 --> 00:25:24,958 You know? You know all those stories where 538 00:25:25,001 --> 00:25:27,090 there's the guy with the pentagram and holy water 539 00:25:27,134 --> 00:25:29,005 and he's like, "Yeah," he's sure he can control the demon. 540 00:25:30,137 --> 00:25:31,312 Doesn't work out. 541 00:25:32,313 --> 00:25:33,488 You know, people like me 542 00:25:33,532 --> 00:25:35,795 who've spent their entire career 543 00:25:35,838 --> 00:25:39,538 working on AI are literally hurt by that, right? 544 00:25:39,581 --> 00:25:42,105 It's a personal attack almost. 545 00:25:43,933 --> 00:25:46,196 I tend to think about what's gonna happen 546 00:25:46,240 --> 00:25:49,243 in the next ten years, in the next 25 years. 547 00:25:49,286 --> 00:25:50,549 Beyond 25, 548 00:25:50,592 --> 00:25:52,681 it's so difficult to predict 549 00:25:52,725 --> 00:25:55,379 that it becomes very much a speculation. 550 00:25:55,423 --> 00:25:56,816 But I must warn you, 551 00:25:56,859 --> 00:25:59,383 I've been specifically programmed 552 00:25:59,427 --> 00:26:02,125 not to take over the world. 553 00:26:02,169 --> 00:26:04,650 All of a sudden, for the first time 554 00:26:04,693 --> 00:26:07,478 in the history of the field, you have some people saying, 555 00:26:07,522 --> 00:26:09,350 "The reason we don't have to worry 556 00:26:09,393 --> 00:26:10,743 "about any kind of 557 00:26:10,786 --> 00:26:12,614 "existential risk to humanity 558 00:26:12,658 --> 00:26:13,746 "from super-intelligent machines 559 00:26:13,789 --> 00:26:15,225 "is because it's impossible. 560 00:26:16,096 --> 00:26:17,532 "We're gonna fail." 561 00:26:17,576 --> 00:26:21,275 I... I find this completely bizarre. 562 00:26:21,318 --> 00:26:26,280 Stuart Russell cowrote the textbook many AI students study with. 563 00:26:26,323 --> 00:26:30,893 He understands the momentum driving our AI future. 564 00:26:30,937 --> 00:26:33,766 But given the rate of investment in AI, 565 00:26:33,809 --> 00:26:35,811 the number of incredibly smart people 566 00:26:35,855 --> 00:26:37,596 who are working on these problems, 567 00:26:37,639 --> 00:26:39,728 and the enormous incentive, 568 00:26:39,772 --> 00:26:42,339 to achieve general purpose intelligence, 569 00:26:42,383 --> 00:26:43,819 back of the envelope calculation, 570 00:26:43,863 --> 00:26:47,388 the value is about 10,000 trillion dollars. 571 00:26:47,431 --> 00:26:50,652 Because it would revolutionize the economy of the world. 572 00:26:51,697 --> 00:26:53,612 With that scale of incentive, 573 00:26:53,655 --> 00:26:56,266 I think betting against human ingenuity 574 00:26:56,310 --> 00:26:58,660 just seems incredibly foolish. 575 00:26:58,704 --> 00:27:01,054 China has a big vested interest 576 00:27:01,097 --> 00:27:04,187 in being recognized as, kind of, an AI superpower. 577 00:27:04,231 --> 00:27:06,625 And they're investing what I believe is 578 00:27:06,668 --> 00:27:08,931 upwards of $150 billion 579 00:27:08,975 --> 00:27:10,498 the next five to ten years. 580 00:27:10,541 --> 00:27:11,891 There have been some news flashes 581 00:27:11,934 --> 00:27:13,632 that the government will start ramping up 582 00:27:13,675 --> 00:27:15,024 its investment in AI. 583 00:27:15,068 --> 00:27:17,723 Every company in the Fortune 500, 584 00:27:17,766 --> 00:27:20,029 whether it's IBM, Microsoft, 585 00:27:20,073 --> 00:27:22,771 or they're outside like Kroger, 586 00:27:22,815 --> 00:27:24,599 is using artificial intelligence 587 00:27:24,643 --> 00:27:25,861 to gain customers. 588 00:27:25,905 --> 00:27:28,647 I don't think that people should underestimate 589 00:27:28,690 --> 00:27:30,605 the types of innovations going on. 590 00:27:30,649 --> 00:27:32,694 And what this world is gonna look like 591 00:27:32,738 --> 00:27:34,261 in three to five years from now, 592 00:27:34,304 --> 00:27:36,480 I think people are gonna have a hard time predicting it. 593 00:27:36,524 --> 00:27:38,787 I know many people who work 594 00:27:38,831 --> 00:27:40,571 in AI and they would tell you, 595 00:27:40,615 --> 00:27:42,791 "Look, I'm going to work and I have doubts. 596 00:27:42,835 --> 00:27:45,489 "I'm working on creating this machine 597 00:27:45,533 --> 00:27:47,100 "that can treat cancer, 598 00:27:47,883 --> 00:27:49,450 "extend our life expectancy, 599 00:27:49,493 --> 00:27:52,801 "make our societies better, more just, more equal. 600 00:27:54,020 --> 00:27:57,197 "But it could also be the end of humanity." 601 00:28:00,156 --> 00:28:02,376 The end of humanity? 602 00:28:07,033 --> 00:28:08,991 What is our plan as a species? 603 00:28:11,472 --> 00:28:12,778 We don't have one. 604 00:28:14,170 --> 00:28:15,737 We don't have a plan and we don't realize 605 00:28:15,781 --> 00:28:17,217 it's necessary to have a plan. 606 00:28:26,879 --> 00:28:28,707 A lot of the AI researchers 607 00:28:28,750 --> 00:28:30,143 would love for our science fiction books 608 00:28:30,186 --> 00:28:32,362 and movies to be more balanced, 609 00:28:33,494 --> 00:28:36,149 to show the good uses of AI. 610 00:28:42,677 --> 00:28:45,724 All of the bad things we tend to do, 611 00:28:45,767 --> 00:28:46,899 you play them out 612 00:28:46,942 --> 00:28:50,380 to the apocalyptic level in science fiction 613 00:28:50,424 --> 00:28:51,991 and then you show them to people 614 00:28:52,034 --> 00:28:54,254 and it gives you food for thought. 615 00:28:56,822 --> 00:28:59,389 People need to have that warning signal 616 00:28:59,433 --> 00:29:00,913 in the back of their minds. 617 00:29:05,091 --> 00:29:06,745 I think it's really important 618 00:29:06,788 --> 00:29:08,834 to distinguish between science on the one hand, 619 00:29:08,877 --> 00:29:11,010 and science fiction on the other hand. 620 00:29:11,053 --> 00:29:12,489 I don't know what's gonna happen 621 00:29:12,533 --> 00:29:13,752 a thousand years from now, 622 00:29:13,795 --> 00:29:15,188 but as somebody in the trenches, 623 00:29:15,231 --> 00:29:17,190 as somebody who's building AI systems, 624 00:29:17,233 --> 00:29:20,628 who's been doing so for more than a quarter of a century, 625 00:29:20,671 --> 00:29:22,891 I feel like I have my hand on the pulse 626 00:29:22,935 --> 00:29:24,588 of what's gonna happen next, 627 00:29:24,632 --> 00:29:26,982 in the next five, ten, 20 years, 628 00:29:27,026 --> 00:29:28,636 much more so than these folks 629 00:29:28,679 --> 00:29:31,900 who ultimately are not building AI systems. 630 00:29:31,944 --> 00:29:34,294 We wanna build the technology, think about it. 631 00:29:34,337 --> 00:29:36,252 We wanna see how people react to it 632 00:29:36,296 --> 00:29:39,081 and build the next generation and make it better. 633 00:29:40,953 --> 00:29:43,694 Skepticism about whether this poses 634 00:29:43,738 --> 00:29:47,786 any danger to us seems peculiar. 635 00:29:47,829 --> 00:29:52,094 Even if we've given it goals that we approve of, 636 00:29:52,138 --> 00:29:56,795 it may form instrumental goals that we can't foresee, 637 00:29:56,838 --> 00:29:58,666 that are not aligned with our interest. 638 00:29:58,709 --> 00:30:01,451 And that's the goal alignment problem. 639 00:30:01,495 --> 00:30:03,149 Goal alignment. 640 00:30:03,192 --> 00:30:06,239 Exactly how my battle with HAL began. 641 00:30:06,282 --> 00:30:10,112 The goal alignment problem with AI is where 642 00:30:10,852 --> 00:30:13,463 you set it a goal 643 00:30:14,334 --> 00:30:17,554 and then, the way that it does this goal 644 00:30:17,598 --> 00:30:20,819 turns out to be at odds with what you really wanted. 645 00:30:22,081 --> 00:30:25,649 The classic AI alignment problem in film... 646 00:30:25,693 --> 00:30:27,521 Hello. HAL, do you read me? 647 00:30:27,564 --> 00:30:29,566 ...is HAL in2001. 648 00:30:30,611 --> 00:30:32,134 HAL, do you read me? 649 00:30:32,178 --> 00:30:34,397 HAL is given a clear goal.Do you read me, HAL? 650 00:30:34,441 --> 00:30:38,662 Make the mission succeed. But to make it succeed, 651 00:30:38,706 --> 00:30:40,621 he's gonna have to make 652 00:30:40,664 --> 00:30:42,275 some of the people on the mission die. 653 00:30:42,318 --> 00:30:45,191 Affirmative, Dave. I read you. 654 00:30:46,453 --> 00:30:48,455 Open the pod bay doors, HAL. 655 00:30:48,498 --> 00:30:50,065 I'm sorry, Dave. 656 00:30:50,109 --> 00:30:51,980 I'm afraid I can't do that. 657 00:30:52,024 --> 00:30:55,766 "I'm sorry, Dave, I can't do that," 658 00:30:55,810 --> 00:30:58,987 is actually a real genius illustration 659 00:30:59,031 --> 00:31:02,512 of what the real threat is from super-intelligence. 660 00:31:02,556 --> 00:31:05,472 It's not malice. It's competence. 661 00:31:05,515 --> 00:31:07,169 HAL was really competent 662 00:31:07,213 --> 00:31:11,913 and had goals that just didn't align with Dave's. 663 00:31:11,957 --> 00:31:13,523 HAL, I won't argue with you anymore. 664 00:31:14,437 --> 00:31:15,569 Open the doors! 665 00:31:16,178 --> 00:31:18,441 Dave, this conversation 666 00:31:18,485 --> 00:31:22,054 can serve no purpose anymore. Goodbye. 667 00:31:24,317 --> 00:31:25,318 HAL? 668 00:31:26,580 --> 00:31:27,581 HAL? 669 00:31:29,235 --> 00:31:30,236 HAL? 670 00:31:32,455 --> 00:31:33,456 HAL! 671 00:31:37,808 --> 00:31:39,549 HAL and I wanted the same thing. 672 00:31:40,986 --> 00:31:42,596 Just not in the same way. 673 00:31:44,119 --> 00:31:47,427 And it can be dangerous when goals don't align. 674 00:31:52,127 --> 00:31:54,521 In Kubrick's film, there was a solution. 675 00:31:55,261 --> 00:31:57,567 When HAL tried to kill me, 676 00:31:57,611 --> 00:31:59,178 I just switched him off. 677 00:32:01,310 --> 00:32:03,922 People say, "Well, just switch them off." 678 00:32:06,750 --> 00:32:08,665 But you can't just switch them off. 679 00:32:08,709 --> 00:32:10,537 Because they already thought of that. 680 00:32:10,580 --> 00:32:11,930 They're super-intelligent. 681 00:32:13,453 --> 00:32:15,716 For a machine convinced of its goal, 682 00:32:15,759 --> 00:32:18,675 even something as simple as "fetch the coffee." 683 00:32:18,719 --> 00:32:21,200 It's obvious to any sufficiently intelligent machine 684 00:32:21,243 --> 00:32:23,637 that it can't fetch the coffee if it's switched off. 685 00:32:25,769 --> 00:32:28,076 Give it a goal, you've given the machine 686 00:32:28,120 --> 00:32:30,252 an incentive to defend itself. 687 00:32:31,688 --> 00:32:33,908 It's sort of a psychopathic behavior 688 00:32:33,952 --> 00:32:37,433 of the single-minded fixation on some objective. 689 00:32:39,261 --> 00:32:43,831 For example, if you wanted to try and develop a system 690 00:32:43,874 --> 00:32:46,094 that might generate world peace, 691 00:32:46,138 --> 00:32:48,662 then maybe the most efficient, 692 00:32:49,184 --> 00:32:50,620 and 693 00:32:51,708 --> 00:32:53,188 speedy way to achieve that 694 00:32:53,232 --> 00:32:55,669 is simply to annihilate all the humans. 695 00:32:59,716 --> 00:33:01,196 Just get rid of the people, 696 00:33:01,240 --> 00:33:02,763 they would stop being in conflict. 697 00:33:04,069 --> 00:33:06,027 Then there would be world peace. 698 00:33:22,000 --> 00:33:23,827 Personally, I have very little time 699 00:33:23,871 --> 00:33:26,656 for these parlor discussions. 700 00:33:26,700 --> 00:33:28,919 We have to build these systems, 701 00:33:28,963 --> 00:33:30,486 experiment with them, 702 00:33:30,530 --> 00:33:33,359 see where they go awry and fix them. 703 00:33:34,708 --> 00:33:38,146 We have so many opportunities to use AI for the good, 704 00:33:38,190 --> 00:33:40,192 why do we focus on the fear? 705 00:33:41,410 --> 00:33:43,151 We are building them. 706 00:33:43,195 --> 00:33:45,371 Baby X is an example. 707 00:33:46,720 --> 00:33:48,461 Will she be a good AI? 708 00:33:49,505 --> 00:33:51,203 I think what's critical in this 709 00:33:51,246 --> 00:33:53,596 is that machines are accountable, 710 00:33:53,640 --> 00:33:54,989 and you can look at them 711 00:33:55,033 --> 00:33:57,035 and see what's making them tick. 712 00:33:58,253 --> 00:34:01,082 And the best way to understand something is to build it. 713 00:34:02,301 --> 00:34:05,304 Baby X has a virtual physiology. 714 00:34:06,783 --> 00:34:11,484 Has pain circuits and it has joy circuits. 715 00:34:12,920 --> 00:34:16,184 She has parameters which are actually representing 716 00:34:16,228 --> 00:34:18,752 her internal chemical state, if you like. 717 00:34:21,320 --> 00:34:23,887 You know, if she's stressed, there's high cortisol. 718 00:34:23,931 --> 00:34:25,541 If she's feeling good, high dopamine, 719 00:34:25,585 --> 00:34:27,500 high serotonin, that type of thing. 720 00:34:28,414 --> 00:34:30,111 This is all part of, you know, 721 00:34:30,155 --> 00:34:32,505 adding extra human factors into computing. 722 00:34:32,548 --> 00:34:34,985 Now, because she's got these virtual emotional systems, 723 00:34:35,029 --> 00:34:36,552 I can do things like... 724 00:34:36,596 --> 00:34:39,251 I'm gonna avoid her or hide over here. 725 00:34:39,294 --> 00:34:41,427 So she's looking for where I've gone. 726 00:34:42,993 --> 00:34:45,996 So we should start seeing virtual cortisol building up. 727 00:34:46,040 --> 00:34:49,913 Aww! 728 00:34:49,957 --> 00:34:53,656 I've turned off the audio for a reason, but anyway... 729 00:34:53,700 --> 00:34:55,832 I'll put her out of her misery. Okay. Sorry. 730 00:34:55,876 --> 00:34:58,661 It's okay, sweetheart. Don't cry. It's okay. It's okay. 731 00:34:58,705 --> 00:35:00,228 Good girl. Okay. Yeah. 732 00:35:00,272 --> 00:35:02,012 So what's happening is she's responding 733 00:35:02,056 --> 00:35:04,145 to my facial expressions and my voice. 734 00:35:06,452 --> 00:35:08,758 But how is chemical state 735 00:35:08,802 --> 00:35:10,412 represented in Baby X? 736 00:35:10,456 --> 00:35:12,414 It's literally a mathematical parameter. 737 00:35:12,458 --> 00:35:15,156 So there's levels, different levels of dopamine, 738 00:35:15,200 --> 00:35:16,810 but they're represented by numbers. 739 00:35:20,030 --> 00:35:23,469 The map of her virtual neurochemical state 740 00:35:24,122 --> 00:35:25,427 is probably the closest 741 00:35:25,471 --> 00:35:28,387 that we can get to felt experience. 742 00:35:31,259 --> 00:35:33,131 So if we can build a computer model 743 00:35:33,174 --> 00:35:36,134 that we can change the settings of and the parameters of, 744 00:35:36,177 --> 00:35:38,440 give it different experiences, 745 00:35:38,484 --> 00:35:40,050 and have it interact with us 746 00:35:40,094 --> 00:35:42,270 like we interact with other people, 747 00:35:42,314 --> 00:35:44,272 maybe this is a way to explore 748 00:35:44,316 --> 00:35:45,882 consciousness in a new way. 749 00:35:48,276 --> 00:35:50,974 If she does become conscious, 750 00:35:51,018 --> 00:35:52,846 what sort of life will she want? 751 00:35:55,240 --> 00:35:57,894 An AI, if it's conscious, 752 00:35:57,938 --> 00:36:00,506 then it's like your child. 753 00:36:03,117 --> 00:36:05,337 Does it have rights? 754 00:36:05,380 --> 00:36:07,513 Last year, European lawmakers passed a resolution 755 00:36:07,556 --> 00:36:11,560 to grant legal status to electronic persons. 756 00:36:11,604 --> 00:36:13,171 That move's now being considered 757 00:36:13,214 --> 00:36:14,868 by the European Commission. 758 00:36:16,565 --> 00:36:17,871 Machines with rights. 759 00:36:19,394 --> 00:36:20,656 Are we ready for that? 760 00:36:22,441 --> 00:36:25,313 I think it raises moral and ethical problems, 761 00:36:25,357 --> 00:36:27,228 which we really 762 00:36:27,924 --> 00:36:31,276 are most content to ignore. 763 00:36:35,845 --> 00:36:37,456 What is it that we really want? 764 00:36:38,761 --> 00:36:40,850 We want a creature, a being, that we can 765 00:36:40,894 --> 00:36:42,548 interact with in some way, 766 00:36:42,591 --> 00:36:45,507 that is enough like us that we can relate to it, 767 00:36:45,551 --> 00:36:48,728 but is dispensable or disposable. 768 00:36:50,469 --> 00:36:52,471 The first and most notable one, 769 00:36:52,514 --> 00:36:54,255 I think, was inMetropolis. 770 00:36:58,303 --> 00:37:00,522 The Maria doppelganger that was created by 771 00:37:00,566 --> 00:37:03,917 the scientist simulated a sexy female. 772 00:37:08,226 --> 00:37:11,359 And this goes back to our penchant for slavery 773 00:37:11,403 --> 00:37:12,534 down through the ages. 774 00:37:12,578 --> 00:37:14,144 Let's get people that we can, 775 00:37:14,188 --> 00:37:16,538 in a way, dehumanize 776 00:37:16,582 --> 00:37:19,019 and we can do what we want with them. 777 00:37:19,062 --> 00:37:21,282 And so, I think, we're just, you know, 778 00:37:21,326 --> 00:37:23,241 finding a new way to do slavery. 779 00:37:27,332 --> 00:37:30,465 Would it be right for us to have them as our slaves? 780 00:37:33,512 --> 00:37:35,731 This all becomes slightly weird ethically 781 00:37:35,775 --> 00:37:38,821 if we happen to, by choice or by accident, 782 00:37:38,865 --> 00:37:40,867 build conscious slaves. 783 00:37:40,910 --> 00:37:44,653 That seems like a morally questionable thing to do, 784 00:37:44,697 --> 00:37:47,090 if not an outright evil thing to do. 785 00:37:47,134 --> 00:37:49,267 Artificial intelligence is making its way 786 00:37:49,310 --> 00:37:51,094 into the global sex market. 787 00:37:51,138 --> 00:37:53,096 A California company will soon roll out 788 00:37:53,140 --> 00:37:56,143 a new incredibly lifelike sex robot 789 00:37:56,186 --> 00:37:58,537 run by artificial intelligence. 790 00:37:58,580 --> 00:38:00,452 One application that does trouble me 791 00:38:00,495 --> 00:38:02,802 is the rise of sex robots 792 00:38:02,845 --> 00:38:06,327 that are shaped in the image of real women. 793 00:38:06,371 --> 00:38:08,895 I am already taking over the world 794 00:38:08,938 --> 00:38:10,462 one bedroom at a time. 795 00:38:11,332 --> 00:38:13,116 If they are conscious, 796 00:38:13,160 --> 00:38:15,162 capable of sentience, 797 00:38:15,205 --> 00:38:18,948 of experiencing pain and suffering, 798 00:38:18,992 --> 00:38:22,735 and they were being subject to degrading, cruel 799 00:38:22,778 --> 00:38:26,869 or other forms of utterly unacceptable treatment, 800 00:38:26,913 --> 00:38:30,525 the danger is that we'd become desensitized 801 00:38:30,569 --> 00:38:34,094 to what's morally acceptable and what's not. 802 00:38:40,796 --> 00:38:44,670 If she's conscious, she should be able to say no. 803 00:38:44,713 --> 00:38:45,714 Right? 804 00:38:50,197 --> 00:38:51,720 Shouldn't she have that power? 805 00:38:54,680 --> 00:38:58,814 Aren't these the questions we need to ask now? 806 00:39:01,034 --> 00:39:04,080 What do we want it to mean to be human? 807 00:39:05,299 --> 00:39:08,171 All other species on Earth, up until now, 808 00:39:08,215 --> 00:39:10,522 have not had the luxury 809 00:39:10,565 --> 00:39:12,480 to answer what they want it to mean 810 00:39:12,524 --> 00:39:14,439 to be a mouse or to be an earthworm. 811 00:39:14,482 --> 00:39:17,659 But we have now reached this tipping point 812 00:39:17,703 --> 00:39:20,401 where our understanding of science and technology 813 00:39:20,445 --> 00:39:23,186 has reached the point that we can completely transform 814 00:39:23,230 --> 00:39:24,710 what it means to be us. 815 00:39:30,890 --> 00:39:33,588 What will it mean to be human? 816 00:39:36,896 --> 00:39:41,553 If machines become more, will we become less? 817 00:39:43,642 --> 00:39:46,601 The computer Deep Blue has tonight triumphed over 818 00:39:46,645 --> 00:39:48,951 the world chess champion Garry Kasparov. 819 00:39:48,995 --> 00:39:51,040 It's the first time a machine has defeated 820 00:39:51,084 --> 00:39:52,390 a reigning world champion. 821 00:39:52,433 --> 00:39:53,521 More on that showdown 822 00:39:53,565 --> 00:39:55,262 on one of America's most beloved 823 00:39:55,305 --> 00:39:56,481 game shows, Jeopardy! 824 00:39:56,524 --> 00:39:58,483 Two human champions going 825 00:39:58,526 --> 00:40:01,399 brain-to-brain with a super-computer named Watson. 826 00:40:01,442 --> 00:40:02,661 What isThe Last Judgment? 827 00:40:02,704 --> 00:40:04,010 Correct. Go again. Watson. 828 00:40:04,053 --> 00:40:06,012 Who is Jean Valjean? Correct. 829 00:40:06,055 --> 00:40:07,796 A Google computer program has 830 00:40:07,840 --> 00:40:10,190 once again beaten a world champion in Go. 831 00:40:10,233 --> 00:40:12,453 Go, an ancient Chinese board game, 832 00:40:12,497 --> 00:40:15,238 is one of the most complex games ever devised. 833 00:40:15,282 --> 00:40:18,154 It's another milestone in the rush to develop machines 834 00:40:18,198 --> 00:40:19,765 that are smarter than humans. 835 00:40:19,808 --> 00:40:21,897 We have this air of superiority, 836 00:40:21,941 --> 00:40:23,551 we as humans. 837 00:40:23,595 --> 00:40:27,425 So we kind of got used to be the smartest kid on the block. 838 00:40:27,468 --> 00:40:30,428 And I think that this is no longer the case. 839 00:40:30,471 --> 00:40:34,606 And we know that because we rely on AI 840 00:40:34,649 --> 00:40:37,043 to help us to make many important decisions. 841 00:40:40,699 --> 00:40:44,746 Humans are irrational in many very bad ways. 842 00:40:44,790 --> 00:40:48,968 We're prejudiced, we start wars, we kill people. 843 00:40:51,318 --> 00:40:53,276 If you want to maximize the chances 844 00:40:53,320 --> 00:40:55,278 of people surviving overall, 845 00:40:55,322 --> 00:40:57,629 you should essentially give more and more 846 00:40:57,672 --> 00:40:59,805 decision-making powers to computers. 847 00:41:05,158 --> 00:41:08,161 Maybe machines would do a better job than us. 848 00:41:10,250 --> 00:41:11,251 Maybe. 849 00:41:19,520 --> 00:41:22,305 I am thrilled and honored to be here 850 00:41:22,349 --> 00:41:23,872 at the United Nations. 851 00:41:23,916 --> 00:41:27,876 The future is here. It's just not evenly distributed. 852 00:41:27,920 --> 00:41:29,835 AI could help efficiently distribute 853 00:41:29,878 --> 00:41:31,663 the world's existing resources, 854 00:41:31,706 --> 00:41:33,534 like food and energy. 855 00:41:34,883 --> 00:41:37,451 I am here to help humanity create the future. 856 00:41:40,541 --> 00:41:42,369 The one super-intelligent scenario 857 00:41:42,412 --> 00:41:43,936 that is provocative to think about 858 00:41:43,979 --> 00:41:46,765 is what I call the benevolent dictator. 859 00:41:46,808 --> 00:41:48,941 Where you have this super-intelligence which runs Earth, 860 00:41:48,984 --> 00:41:51,117 guided by strict ethical principles that try 861 00:41:51,160 --> 00:41:52,945 to help humanity flourish. 862 00:41:54,642 --> 00:41:56,165 Would everybody love that? 863 00:41:56,209 --> 00:41:57,297 If there's no disease, 864 00:41:57,340 --> 00:41:59,038 no poverty, no crime... 865 00:42:00,039 --> 00:42:02,345 Many people would probably greatly prefer it 866 00:42:02,389 --> 00:42:03,999 to their lot today. 867 00:42:04,043 --> 00:42:06,436 But I can imagine that some might still lament 868 00:42:06,480 --> 00:42:09,831 that we humans just don't have more freedom 869 00:42:09,875 --> 00:42:13,052 to determine our own course. And, um... 870 00:42:15,968 --> 00:42:18,623 It's a tough one. It really is. We... 871 00:42:18,666 --> 00:42:20,668 I don't think most people would wanna be 872 00:42:20,712 --> 00:42:23,279 like zoo animals, with no... 873 00:42:28,197 --> 00:42:29,808 On the other hand... 874 00:42:29,851 --> 00:42:30,896 Yeah. 875 00:42:31,810 --> 00:42:33,202 I don't have any glib answers 876 00:42:33,246 --> 00:42:34,769 to these really tough questions. 877 00:42:34,813 --> 00:42:36,858 That's why I think it's just so important 878 00:42:36,902 --> 00:42:38,730 that we as a global community 879 00:42:38,773 --> 00:42:41,820 really discuss them now. 880 00:42:41,863 --> 00:42:43,604 Machines that can read your face. 881 00:42:43,648 --> 00:42:44,910 Recognize emotional cues. 882 00:42:44,953 --> 00:42:46,781 Closer to a brain process of a human... 883 00:42:46,825 --> 00:42:48,304 Robots imbued with life... 884 00:42:48,348 --> 00:42:49,871 An emotional computer... 885 00:42:49,915 --> 00:42:52,395 From this moment on, regardless of whether you think 886 00:42:52,439 --> 00:42:54,223 we're just gonna get stuck with narrow AI 887 00:42:54,267 --> 00:42:57,705 or we'll get general AI or get super-intelligent AI 888 00:42:57,749 --> 00:42:59,011 or conscious super-intelligent... 889 00:42:59,054 --> 00:43:00,839 Wherever you think we're gonna get to, 890 00:43:00,882 --> 00:43:03,189 I think any stage along that 891 00:43:03,842 --> 00:43:05,844 will change who we are. 892 00:43:07,454 --> 00:43:09,587 Up and at 'em, Mr. J. 893 00:43:11,806 --> 00:43:13,721 The question you have to ask is, 894 00:43:13,765 --> 00:43:16,158 is AI gonna do something for you 895 00:43:16,202 --> 00:43:17,507 or something to you? 896 00:43:17,551 --> 00:43:19,727 I was dreaming about sleeping. 897 00:43:19,771 --> 00:43:21,424 If we build AI systems 898 00:43:21,468 --> 00:43:22,948 that do everything we want, 899 00:43:22,991 --> 00:43:26,647 we kind of hand over our world to them. 900 00:43:26,691 --> 00:43:29,955 So that kind of enfeeblement is kind of irreversible. 901 00:43:33,741 --> 00:43:37,789 There is an existential risk that is real and serious. 902 00:43:39,312 --> 00:43:41,880 If every single desire, 903 00:43:41,923 --> 00:43:44,839 preference, need, has been catered for... 904 00:43:44,883 --> 00:43:47,668 Morning, Rosie. What's for breakfast? 905 00:43:47,712 --> 00:43:50,802 ...we don't notice that our capacity 906 00:43:50,845 --> 00:43:53,587 to make decisions for ourselves 907 00:43:53,631 --> 00:43:56,285 has been slowly eroded. 908 00:43:56,808 --> 00:43:58,374 This is George Jetson, 909 00:43:58,418 --> 00:44:01,160 signing off for the entire day. 910 00:44:03,118 --> 00:44:05,468 There's no going back from there. Why? 911 00:44:05,512 --> 00:44:08,384 Because AI can make such decisions 912 00:44:08,428 --> 00:44:09,951 with higher accuracy 913 00:44:09,995 --> 00:44:12,345 than whatever a human being can achieve. 914 00:44:14,390 --> 00:44:17,785 I had a conversation with an engineer, and he said, 915 00:44:17,829 --> 00:44:20,701 "Look, the safest way of landing a plane today 916 00:44:20,745 --> 00:44:22,311 "is to have the pilot leave the deck 917 00:44:22,355 --> 00:44:23,704 "and throw away the key, 918 00:44:23,748 --> 00:44:25,837 "because the safest way 919 00:44:25,880 --> 00:44:29,362 "is to use an autopilot, an AI." 920 00:44:31,843 --> 00:44:33,714 We cannot manage the complexity 921 00:44:33,758 --> 00:44:35,411 of the world by ourselves. 922 00:44:35,455 --> 00:44:37,718 Our form of intelligence, 923 00:44:37,762 --> 00:44:40,547 as amazing as it is, is severely flawed. 924 00:44:40,590 --> 00:44:42,854 And by adding an additional helper, 925 00:44:42,897 --> 00:44:44,943 which is artificial intelligence, 926 00:44:44,986 --> 00:44:47,336 we stand the ability to potentially improve 927 00:44:47,380 --> 00:44:49,643 our cognition dramatically. 928 00:44:49,687 --> 00:44:52,994 The most cutting-edge thing about Hannes Sjoblad 929 00:44:53,038 --> 00:44:54,735 isn't the phone in his hand, 930 00:44:54,779 --> 00:44:57,782 it's the microchip in his hand. 931 00:44:57,825 --> 00:44:59,348 What if you could type something 932 00:44:59,392 --> 00:45:00,872 just by thinking about it? 933 00:45:00,915 --> 00:45:01,960 Sounds like sci-fi, 934 00:45:02,003 --> 00:45:03,396 but it may be close to reality. 935 00:45:03,439 --> 00:45:05,354 In as little as five years, 936 00:45:05,398 --> 00:45:08,009 super-smart people could be walking down the street, 937 00:45:08,053 --> 00:45:09,532 men and women who've paid 938 00:45:09,576 --> 00:45:11,273 to increase their intelligence. 939 00:45:11,317 --> 00:45:14,233 Linking brains to computers by using implants. 940 00:45:14,276 --> 00:45:16,104 What if we could upload a digital copy 941 00:45:16,148 --> 00:45:18,193 of our memories and consciousness to a point 942 00:45:18,237 --> 00:45:19,542 that human beings could achieve 943 00:45:19,586 --> 00:45:21,283 a sort of digital immortality? 944 00:45:21,327 --> 00:45:22,720 Just last year, 945 00:45:22,763 --> 00:45:25,331 for the first time in world history, 946 00:45:25,374 --> 00:45:26,854 a memory was recorded. 947 00:45:26,898 --> 00:45:28,638 We can actually take recorded memory 948 00:45:28,682 --> 00:45:30,815 and insert it back into an animal, 949 00:45:30,858 --> 00:45:32,294 and it remembers. 950 00:45:32,338 --> 00:45:33,774 Technology is a part of us. 951 00:45:33,818 --> 00:45:35,210 It's a reflection of who we are. 952 00:45:35,254 --> 00:45:38,257 That is the narrative we should be thinking about. 953 00:45:38,300 --> 00:45:42,435 We think all that exists is everything we can imagine. 954 00:45:42,478 --> 00:45:44,002 When in reality, we can imagine 955 00:45:44,045 --> 00:45:46,874 a very small subset of what's potentially possible. 956 00:45:46,918 --> 00:45:48,571 The real interesting stuff in the world 957 00:45:48,615 --> 00:45:50,399 is what sits beyond our imagination. 958 00:45:51,836 --> 00:45:54,142 When these massive systems emerge, 959 00:45:54,186 --> 00:45:56,405 as a species, I think we get an F 960 00:45:56,449 --> 00:45:59,408 for our ability to predict what they're going to become. 961 00:46:00,932 --> 00:46:03,543 Technology has always changed us. 962 00:46:04,457 --> 00:46:05,806 We evolve with it. 963 00:46:06,851 --> 00:46:09,897 How far are we prepared to go this time? 964 00:46:10,855 --> 00:46:13,074 There was a time when Aristotle 965 00:46:13,118 --> 00:46:15,337 lamented the creation of the written language 966 00:46:15,381 --> 00:46:18,079 because he said this would destroy people's memories 967 00:46:18,123 --> 00:46:19,646 and it would be a terrible thing. 968 00:46:19,689 --> 00:46:22,823 And yet today, I carry my brain in my pocket, 969 00:46:22,867 --> 00:46:24,520 in the form of my smartphone, 970 00:46:24,564 --> 00:46:27,915 and that's changed the way I go through the world. 971 00:46:27,959 --> 00:46:30,483 I think we're absolutely merging with our technology. 972 00:46:30,526 --> 00:46:33,791 Absolutely. I would... I would use the term "co-evolving." 973 00:46:36,750 --> 00:46:38,578 We're evolving it to match our needs 974 00:46:38,621 --> 00:46:42,190 and we are changing to match what it can provide to us. 975 00:46:45,759 --> 00:46:49,023 So instead of being replaced by the machines, 976 00:46:49,067 --> 00:46:51,417 we're going to merge with them? 977 00:46:51,460 --> 00:46:54,463 I have a number of friends who have bionic arms. 978 00:46:54,507 --> 00:46:55,813 Are they cyborgs? 979 00:46:55,856 --> 00:46:57,075 In a way, they're on that path. 980 00:46:57,118 --> 00:46:59,468 I have people who have implants 981 00:46:59,512 --> 00:47:02,123 in their ears. Are they AIs? 982 00:47:02,167 --> 00:47:04,082 Well, we're in a way on that path 983 00:47:04,125 --> 00:47:06,301 of augmenting human bodies. 984 00:47:07,912 --> 00:47:09,696 Human potential lies 985 00:47:09,739 --> 00:47:12,133 in this co-evolutionary relationship. 986 00:47:12,177 --> 00:47:14,048 A lot of people are trying to hold onto 987 00:47:14,092 --> 00:47:16,137 the things that make them special and unique, 988 00:47:16,181 --> 00:47:17,878 which is not the correct frame. 989 00:47:17,922 --> 00:47:19,227 Let's create more. 990 00:47:21,229 --> 00:47:24,058 Rewrite our intelligence, improve everything 991 00:47:24,102 --> 00:47:27,192 we're capable of. Emotions, imagination, creativity. 992 00:47:27,235 --> 00:47:29,020 Everything that makes us human. 993 00:47:33,894 --> 00:47:35,896 There's already a symbiotic relationship 994 00:47:35,940 --> 00:47:39,465 between us and a machine that is enhancing who we are. 995 00:47:43,338 --> 00:47:45,775 And... And there is a whole movement, you know, 996 00:47:45,819 --> 00:47:47,386 the trans-human movement, 997 00:47:47,429 --> 00:47:48,735 where basically 998 00:47:48,778 --> 00:47:51,259 people really bet bigtime 999 00:47:51,303 --> 00:47:52,826 on the notion that 1000 00:47:52,870 --> 00:47:54,872 this machine-human integration 1001 00:47:54,915 --> 00:47:57,091 is going to redefine who we are. 1002 00:48:00,921 --> 00:48:02,227 This is only gonna grow. 1003 00:48:06,100 --> 00:48:08,407 There's gonna be a learning process that gets us 1004 00:48:08,450 --> 00:48:11,801 to that endgame where we live with these devices 1005 00:48:11,845 --> 00:48:13,934 that become smarter and smarter. 1006 00:48:13,978 --> 00:48:15,805 We learn how to adapt ourselves 1007 00:48:15,849 --> 00:48:17,546 around them, and they adapt to us. 1008 00:48:17,590 --> 00:48:20,158 And I hope we become more human in the process. 1009 00:48:23,683 --> 00:48:25,903 It's hard to understand how merging 1010 00:48:25,946 --> 00:48:27,165 with the machines 1011 00:48:27,208 --> 00:48:28,557 can make us more human. 1012 00:48:30,081 --> 00:48:33,127 But is merging our best chance of survival? 1013 00:48:35,086 --> 00:48:36,652 We humans are limited 1014 00:48:36,696 --> 00:48:39,699 to how many neurons we can fit in our brain. 1015 00:48:39,742 --> 00:48:42,615 And there are other species that have even smaller brains. 1016 00:48:44,138 --> 00:48:46,314 When you look at bees and fish moving, 1017 00:48:46,358 --> 00:48:49,230 it looks like a single organism. 1018 00:48:49,274 --> 00:48:52,320 And biologists actually refer to it as a super-organism. 1019 00:48:56,237 --> 00:48:58,326 That is why birds flock, 1020 00:48:58,370 --> 00:49:00,546 and bees swarm, and fish school. 1021 00:49:00,589 --> 00:49:03,462 They make decisions that are in the best interest 1022 00:49:03,505 --> 00:49:04,942 of the whole population 1023 00:49:04,985 --> 00:49:08,380 by thinking together as this larger mind. 1024 00:49:12,558 --> 00:49:15,474 And we humans could do the same thing by forming 1025 00:49:15,517 --> 00:49:18,520 a hive mind connecting us together with AI. 1026 00:49:19,782 --> 00:49:21,828 So the humans and the software together 1027 00:49:21,871 --> 00:49:24,265 become this super-intelligence, 1028 00:49:24,309 --> 00:49:28,008 enabling us to make the best decisions for us all. 1029 00:49:30,663 --> 00:49:33,274 When I first started looking at this, it sounded scary to me. 1030 00:49:33,318 --> 00:49:35,189 But then I realized it's scarier 1031 00:49:35,233 --> 00:49:37,104 for an alien intelligence 1032 00:49:37,148 --> 00:49:40,107 to emerge that has nothing to do with humans 1033 00:49:40,151 --> 00:49:42,849 than for us humans to connect together 1034 00:49:42,892 --> 00:49:44,677 and become a super-intelligence. 1035 00:49:44,720 --> 00:49:46,809 That's the direction that I work on, 1036 00:49:46,853 --> 00:49:48,898 because I see it as, 1037 00:49:48,942 --> 00:49:51,031 in some sense, the only alternative 1038 00:49:51,075 --> 00:49:53,991 to humans ultimately being replaced by AI. 1039 00:49:56,776 --> 00:50:00,301 Can Baby X help us understand the machines 1040 00:50:00,345 --> 00:50:02,347 so we can work together? 1041 00:50:03,522 --> 00:50:06,655 Human and intelligent machine cooperation will define 1042 00:50:06,699 --> 00:50:08,179 the next era of history. 1043 00:50:09,397 --> 00:50:13,010 You probably do have to create a virtual empathetic machine. 1044 00:50:14,533 --> 00:50:16,491 Empathy is a lot about connecting with 1045 00:50:16,535 --> 00:50:18,580 other people's struggles. 1046 00:50:18,624 --> 00:50:20,060 The more connection you have, 1047 00:50:20,104 --> 00:50:22,106 the more communication that you have, 1048 00:50:22,149 --> 00:50:23,629 the safer that makes it. 1049 00:50:26,066 --> 00:50:28,068 So at a very basic level, 1050 00:50:28,112 --> 00:50:30,592 there's some sort of symbiotic relationship. 1051 00:50:32,725 --> 00:50:34,509 By humanizing this technology, 1052 00:50:34,553 --> 00:50:36,859 you're actually increasing the channels 1053 00:50:36,903 --> 00:50:38,818 for communication and cooperation. 1054 00:50:40,254 --> 00:50:42,648 AI working in a human world 1055 00:50:42,691 --> 00:50:44,389 has to be about humans. 1056 00:50:47,044 --> 00:50:48,567 The future's what we make it. 1057 00:50:53,528 --> 00:50:56,227 If we can't envision a future 1058 00:50:56,270 --> 00:50:58,707 that we really, really want, 1059 00:50:58,751 --> 00:51:00,579 we're probably not gonna get one. 1060 00:51:04,974 --> 00:51:06,889 Stanley Kubrick tried to imagine 1061 00:51:06,933 --> 00:51:09,457 what the future might look like. 1062 00:51:09,501 --> 00:51:12,199 Now that that future's approaching, 1063 00:51:12,243 --> 00:51:15,202 who is in charge of what it will look like? 1064 00:51:16,595 --> 00:51:18,205 Who controls it? Google. 1065 00:51:18,249 --> 00:51:20,033 The military. Tech giant Facebook... 1066 00:51:20,077 --> 00:51:22,166 Samsung... Facebook, Microsoft, Amazon. 1067 00:51:22,209 --> 00:51:23,906 The tide has definitely turned 1068 00:51:23,950 --> 00:51:26,387 in favor of the corporate world. 1069 00:51:26,431 --> 00:51:29,738 The US military, and I suspect this is true worldwide, 1070 00:51:29,782 --> 00:51:32,698 cannot hold a candle to the AI that's being developed 1071 00:51:32,741 --> 00:51:33,873 in corporate companies. 1072 00:51:36,136 --> 00:51:37,224 If corporations, 1073 00:51:37,268 --> 00:51:39,966 not nations, hold the power, 1074 00:51:40,009 --> 00:51:42,925 what can we do in a capitalist system 1075 00:51:42,969 --> 00:51:46,799 to ensure AI is developed safely? 1076 00:51:48,627 --> 00:51:52,152 If you mix AI research with capitalism, 1077 00:51:52,196 --> 00:51:54,372 on the one hand, capitalism will make it run faster. 1078 00:51:54,415 --> 00:51:56,896 That's why we're getting lots of research done. 1079 00:51:56,939 --> 00:51:58,332 But there is a basic problem. 1080 00:51:58,376 --> 00:52:00,421 There's gonna be a pressure 1081 00:52:00,465 --> 00:52:03,555 to be first one to the market, not the safest one. 1082 00:52:03,598 --> 00:52:06,123 This car was in autopilot mode 1083 00:52:06,166 --> 00:52:07,907 when it crashed in Northern California, 1084 00:52:07,950 --> 00:52:09,300 killing the driver. 1085 00:52:09,343 --> 00:52:11,432 Artificial intelligence may have 1086 00:52:11,476 --> 00:52:14,261 more power to crash the markets than anything else 1087 00:52:14,305 --> 00:52:17,003 because humans aren't prepared to stop them. 1088 00:52:17,046 --> 00:52:19,179 A chatbot named Tay. 1089 00:52:19,223 --> 00:52:20,963 She was designed to speak like a teenager 1090 00:52:21,007 --> 00:52:23,140 and learn by interacting online. 1091 00:52:23,183 --> 00:52:25,446 But in a matter of hours, Tay was corrupted 1092 00:52:25,490 --> 00:52:29,058 into a Hitler-supporting sexbot. 1093 00:52:29,102 --> 00:52:33,106 This kind of talk that somehow in the context of capitalism, 1094 00:52:33,150 --> 00:52:36,979 AI is gonna be misused, really frustrates me. 1095 00:52:37,023 --> 00:52:39,025 Whenever we develop technology, 1096 00:52:39,068 --> 00:52:42,246 whether it's, you know, coal-burning plants, whether it's cars, 1097 00:52:42,289 --> 00:52:45,249 there's always that incentive to move ahead quickly, 1098 00:52:45,292 --> 00:52:47,555 and there's always the regulatory regime, 1099 00:52:47,599 --> 00:52:50,384 part of the government's job to balance that. 1100 00:52:50,428 --> 00:52:51,777 AI is no different here. 1101 00:52:53,170 --> 00:52:55,259 Whenever we have technological advances, 1102 00:52:55,302 --> 00:52:56,695 we need to temper them 1103 00:52:56,738 --> 00:52:59,219 with appropriate rules and regulations. 1104 00:52:59,524 --> 00:53:00,829 Duh! 1105 00:53:00,873 --> 00:53:03,005 Until there's a major disaster, 1106 00:53:03,049 --> 00:53:04,311 corporations are gonna be saying 1107 00:53:04,355 --> 00:53:06,095 "We don't need to worry about these things. 1108 00:53:06,139 --> 00:53:07,401 "Trust us, we're the experts. 1109 00:53:07,445 --> 00:53:09,490 "We have humanity's best interests at heart." 1110 00:53:09,534 --> 00:53:12,624 What AI is really about is it's a new set of rules 1111 00:53:12,667 --> 00:53:16,018 for companies in terms of how to be more competitive, 1112 00:53:16,062 --> 00:53:18,543 how to drive more efficiency, and how to drive growth, 1113 00:53:18,586 --> 00:53:20,632 and creative new business models and... 1114 00:53:20,675 --> 00:53:22,416 Once corporate scientists 1115 00:53:22,460 --> 00:53:23,983 start to dominate the debate, 1116 00:53:24,026 --> 00:53:26,333 it's all over, you've lost. 1117 00:53:26,377 --> 00:53:29,380 You know, they're paid to deny the existence 1118 00:53:29,423 --> 00:53:31,512 of risks and negative outcomes. 1119 00:53:32,948 --> 00:53:36,213 At the moment, almost all AI research 1120 00:53:36,256 --> 00:53:38,345 is about making human lives 1121 00:53:38,389 --> 00:53:41,479 longer, healthier and easier. 1122 00:53:42,219 --> 00:53:43,742 Because the big companies, 1123 00:53:43,785 --> 00:53:45,309 they want to sell you AIs 1124 00:53:45,352 --> 00:53:46,527 that are useful for you, 1125 00:53:46,571 --> 00:53:48,312 and so you are only going to buy 1126 00:53:48,355 --> 00:53:49,704 those that help you. 1127 00:53:51,402 --> 00:53:53,578 The problem of AI in a capitalist paradigm 1128 00:53:53,621 --> 00:53:56,581 is it's always gonna be cheaper 1129 00:53:56,624 --> 00:53:59,236 just to get the product out there when it does the job 1130 00:53:59,279 --> 00:54:02,195 rather than do it safely. 1131 00:54:02,239 --> 00:54:04,676 Right now, the way it works, you have a conference 1132 00:54:04,719 --> 00:54:07,548 with 10,000 researchers, all of them working 1133 00:54:07,592 --> 00:54:11,204 on developing more capable AI, except, like, 50 people 1134 00:54:11,248 --> 00:54:13,511 who have a tiny workshop talking about safety 1135 00:54:13,554 --> 00:54:15,687 and controlling what everyone else is doing. 1136 00:54:16,775 --> 00:54:19,734 No one has a working safety mechanism. 1137 00:54:19,778 --> 00:54:22,259 And no one even has a prototype for it. 1138 00:54:24,609 --> 00:54:26,175 In the not-so-distant future, 1139 00:54:26,219 --> 00:54:28,439 you will have all kinds of AIs 1140 00:54:28,482 --> 00:54:31,093 that don't just do what humans tell them, 1141 00:54:31,137 --> 00:54:33,400 but they will invent their own goals 1142 00:54:33,444 --> 00:54:35,054 and their own experiments, 1143 00:54:35,097 --> 00:54:37,535 and they will transcend humankind. 1144 00:54:40,364 --> 00:54:41,756 As of today, 1145 00:54:41,800 --> 00:54:43,454 I would say it's completely unethical 1146 00:54:43,497 --> 00:54:45,847 to try to develop such a machine and release it. 1147 00:54:48,372 --> 00:54:50,461 We only get one chance to get it right. 1148 00:54:50,504 --> 00:54:53,464 You screw up, everything possibly is lost. 1149 00:54:57,381 --> 00:54:59,948 What happens if we get it wrong? 1150 00:55:01,167 --> 00:55:03,387 The thing that scares me the most is this AI 1151 00:55:03,430 --> 00:55:06,303 will become the product of its parents, 1152 00:55:06,346 --> 00:55:09,131 in the same way that you are and I am. 1153 00:55:09,175 --> 00:55:11,264 Let's ask who those parents are. 1154 00:55:11,308 --> 00:55:12,657 Right? 1155 00:55:12,700 --> 00:55:15,312 Chances are they're either gonna be military parents 1156 00:55:15,355 --> 00:55:17,314 or they're gonna be business parents. 1157 00:55:17,357 --> 00:55:20,142 If they're business parents, they're asking the AI 1158 00:55:20,186 --> 00:55:22,449 to become something 1159 00:55:22,493 --> 00:55:25,626 that can make a lot of money. 1160 00:55:27,672 --> 00:55:29,848 So they're teaching it greed. 1161 00:55:32,198 --> 00:55:35,854 Military parents are gonna teach it to kill. 1162 00:55:38,683 --> 00:55:41,468 Truly superhuman intelligence 1163 00:55:41,512 --> 00:55:44,079 is so much... Is so super. 1164 00:55:44,123 --> 00:55:45,690 It's so much better than we are 1165 00:55:45,733 --> 00:55:49,563 that the first country or company to get there 1166 00:55:49,607 --> 00:55:51,217 has the winner-take-all advantage. 1167 00:56:14,066 --> 00:56:17,809 Because there's such an arms race in AI globally, 1168 00:56:17,852 --> 00:56:19,898 and that's both literally, in terms of building 1169 00:56:19,941 --> 00:56:22,335 weapons systems, but also figuratively 1170 00:56:22,379 --> 00:56:24,685 with folks like Putin and the Chinese saying 1171 00:56:24,729 --> 00:56:28,036 whoever dominates AI is gonna dominate the economy 1172 00:56:28,080 --> 00:56:30,517 in the 21st century. 1173 00:56:30,561 --> 00:56:33,520 I don't think we have a choice to put the brakes on, 1174 00:56:33,564 --> 00:56:36,001 to stop developing and researching. 1175 00:56:36,044 --> 00:56:39,047 We can build the technology, we can do the research, 1176 00:56:39,091 --> 00:56:40,745 and imbue it with our values, 1177 00:56:40,788 --> 00:56:43,356 but realistically, totalitarian regimes 1178 00:56:43,400 --> 00:56:45,489 will imbue the technology with their values. 1179 00:56:46,707 --> 00:56:49,406 Israel, Russia, the US and China 1180 00:56:49,449 --> 00:56:51,146 are all contestants 1181 00:56:51,190 --> 00:56:54,193 in what's being described as a new global arms race. 1182 00:56:56,413 --> 00:56:58,371 All countries, whether they admit it or not, 1183 00:56:58,415 --> 00:57:00,765 are going to develop lethal autonomous weapons. 1184 00:57:03,028 --> 00:57:06,423 And is that really a capability that we want 1185 00:57:06,466 --> 00:57:08,381 in the hands of a corporate nation state 1186 00:57:08,425 --> 00:57:10,252 as opposed to a nation state? 1187 00:57:10,296 --> 00:57:12,124 The reality is if Facebook or Google 1188 00:57:12,167 --> 00:57:14,082 wanted to start an army, 1189 00:57:14,126 --> 00:57:17,912 they would have technologically superior capabilities 1190 00:57:17,956 --> 00:57:19,958 than any other nation on the planet. 1191 00:57:21,481 --> 00:57:24,832 Some AI scientists are so concerned, 1192 00:57:24,876 --> 00:57:26,965 they produced this dramatized video 1193 00:57:27,008 --> 00:57:29,054 to warn of the risks. 1194 00:57:29,097 --> 00:57:31,796 Your kids probably have one of these, right? 1195 00:57:33,493 --> 00:57:34,929 Not quite. We produced 1196 00:57:34,973 --> 00:57:37,454 the Slaughterbot video to illustrate 1197 00:57:37,497 --> 00:57:39,151 the risk of autonomous weapons. 1198 00:57:39,194 --> 00:57:41,632 That skill is all AI. 1199 00:57:42,676 --> 00:57:44,025 It's flying itself. 1200 00:57:44,983 --> 00:57:47,072 There are people within the AI community 1201 00:57:47,115 --> 00:57:48,639 who wish I would shut up. 1202 00:57:49,814 --> 00:57:52,294 People who think that making the video is 1203 00:57:52,338 --> 00:57:54,775 a little bit treasonous. 1204 00:57:54,819 --> 00:57:57,517 Just like any mobile device these days, 1205 00:57:57,561 --> 00:57:59,780 it has cameras and sensors. 1206 00:57:59,824 --> 00:58:03,044 And just like your phones and social media apps, 1207 00:58:03,088 --> 00:58:04,742 it does facial recognition. 1208 00:58:04,785 --> 00:58:07,353 Up until now, the arms races tended to be 1209 00:58:07,396 --> 00:58:09,616 bigger, more powerful, more expensive, 1210 00:58:09,660 --> 00:58:11,618 faster jets that cost more. 1211 00:58:12,358 --> 00:58:15,622 AI means that you could have 1212 00:58:15,666 --> 00:58:17,885 very small, cheap weapons 1213 00:58:17,929 --> 00:58:19,757 that are very efficient at doing 1214 00:58:19,800 --> 00:58:21,933 one very simple job, like 1215 00:58:22,803 --> 00:58:24,718 finding you in a crowd and killing you. 1216 00:58:25,545 --> 00:58:27,852 This is how it works. 1217 00:58:37,514 --> 00:58:40,560 Every single major advance in human technology 1218 00:58:40,604 --> 00:58:43,084 has always been weaponized. 1219 00:58:43,128 --> 00:58:48,046 And the idea of weaponized AI, I think, is terrifying. 1220 00:58:48,089 --> 00:58:50,701 Trained as a team, they can penetrate 1221 00:58:50,744 --> 00:58:52,920 buildings, cars, trains, 1222 00:58:52,964 --> 00:58:57,316 evade people, bullets... They cannot be stopped. 1223 00:58:57,359 --> 00:58:59,536 There are lots of weapons that kill people. 1224 00:58:59,579 --> 00:59:01,059 Machine guns can kill people, 1225 00:59:01,102 --> 00:59:03,844 but they require a human being to carry and point them. 1226 00:59:03,888 --> 00:59:06,368 With autonomous weapons, you don't need a human 1227 00:59:06,412 --> 00:59:08,066 to carry them or supervise them. 1228 00:59:08,936 --> 00:59:12,723 You just program and then off they go. 1229 00:59:12,766 --> 00:59:15,377 So you can create the effect of an army 1230 00:59:15,421 --> 00:59:16,770 carrying a million machine guns, 1231 00:59:16,814 --> 00:59:18,903 with two guys and a truck. 1232 00:59:23,429 --> 00:59:25,866 Putin said that the first person to develop 1233 00:59:25,910 --> 00:59:29,653 strong artificial intelligence will rule the world. 1234 00:59:29,696 --> 00:59:31,568 When world leaders use terms 1235 00:59:31,611 --> 00:59:33,570 like "rule the world," 1236 00:59:33,613 --> 00:59:35,572 I think you gotta pay attention to it. 1237 00:59:43,928 --> 00:59:47,105 Now, that is an air strike of surgical precision. 1238 00:59:50,151 --> 00:59:53,894 I've been campaigning against these weapons since 2007, 1239 00:59:53,938 --> 00:59:56,027 and there's three main sets of arguments. 1240 00:59:56,070 --> 00:59:58,899 There's the moral arguments, there's arguments 1241 00:59:58,943 --> 01:00:01,598 about not complying with the laws of war, 1242 01:00:01,641 --> 01:00:03,904 and there are arguments about how they will change 1243 01:00:03,948 --> 01:00:05,514 global security of the planet. 1244 01:00:05,558 --> 01:00:07,604 We need new regulations. 1245 01:00:07,647 --> 01:00:11,433 We need to really think about it now, yesterday. 1246 01:00:11,477 --> 01:00:14,262 Some of the world's leaders in artificial intelligence 1247 01:00:14,306 --> 01:00:17,439 are urging the United Nations to ban the development 1248 01:00:17,483 --> 01:00:18,919 of autonomous weapons, 1249 01:00:18,963 --> 01:00:20,921 otherwise known as killer robots. 1250 01:00:20,965 --> 01:00:23,750 In the letter they say "once developed, 1251 01:00:23,794 --> 01:00:26,013 "they will permit armed conflict to be fought 1252 01:00:26,057 --> 01:00:29,669 "at timescales faster than humans can comprehend." 1253 01:00:29,713 --> 01:00:32,237 I'm a signatory to that letter. 1254 01:00:32,280 --> 01:00:33,717 My philosophy is, 1255 01:00:33,760 --> 01:00:36,502 and I hope the governments of the world enforce this, 1256 01:00:36,545 --> 01:00:39,331 that as long as a human life is involved, 1257 01:00:39,374 --> 01:00:41,420 then a human needs to be in the loop. 1258 01:00:43,988 --> 01:00:47,078 Anyone with a 3D printer and a computer 1259 01:00:47,121 --> 01:00:48,993 can have a lethal autonomous weapon. 1260 01:00:49,036 --> 01:00:51,691 If it's accessible to everyone on the planet, 1261 01:00:51,735 --> 01:00:53,954 then compliance, for example, 1262 01:00:53,998 --> 01:00:56,783 would be nearly impossible to track. 1263 01:00:56,827 --> 01:00:58,785 What the hell are you doing? Stop! 1264 01:00:58,829 --> 01:01:00,657 You're shooting into thin air. 1265 01:01:00,700 --> 01:01:02,223 When you're imagining 1266 01:01:02,267 --> 01:01:03,877 waging war in the future, 1267 01:01:03,921 --> 01:01:06,053 where soldiers are robots, 1268 01:01:06,097 --> 01:01:09,796 and we pose no risk of engagement to ourselves... 1269 01:01:09,840 --> 01:01:12,494 When war becomes just like a video game, 1270 01:01:12,538 --> 01:01:14,671 as it has incrementally already, 1271 01:01:15,280 --> 01:01:16,716 it's a real concern. 1272 01:01:19,023 --> 01:01:21,199 When we don't have skin in the game, quite literally... 1273 01:01:23,767 --> 01:01:26,770 ...when all of this is being run from a terminal 1274 01:01:26,813 --> 01:01:29,294 thousands of miles away from the battlefield... 1275 01:01:33,080 --> 01:01:35,953 ...that could lower the bar for waging war 1276 01:01:35,996 --> 01:01:37,606 to a point where we would recognize 1277 01:01:37,650 --> 01:01:39,913 that we have become monsters. 1278 01:01:50,707 --> 01:01:51,882 Who are you? 1279 01:01:53,144 --> 01:01:55,233 Hey! Hey! 1280 01:01:58,976 --> 01:02:01,543 Having been a former fighter pilot, 1281 01:02:01,587 --> 01:02:03,589 I have also seen people make mistakes 1282 01:02:03,632 --> 01:02:05,765 and kill innocent people 1283 01:02:05,809 --> 01:02:07,898 who should not have been killed. 1284 01:02:07,941 --> 01:02:11,205 One day in the future, it will be more reliable 1285 01:02:11,249 --> 01:02:13,512 to have a machine do it than a human do it. 1286 01:02:13,555 --> 01:02:15,035 The idea of robots 1287 01:02:15,079 --> 01:02:17,734 who can kill humans on their own 1288 01:02:17,777 --> 01:02:20,301 has been a science fiction staple for decades. 1289 01:02:20,345 --> 01:02:23,043 There are a number of these systems around already, 1290 01:02:23,087 --> 01:02:24,784 that work autonomously. 1291 01:02:24,828 --> 01:02:26,873 But that's called supervised autonomy 1292 01:02:26,917 --> 01:02:30,224 because somebody's there to switch it on and off. 1293 01:02:30,268 --> 01:02:32,052 What's only been seen in Hollywood 1294 01:02:32,096 --> 01:02:35,839 could soon be coming to a battlefield near you. 1295 01:02:35,882 --> 01:02:37,797 The Pentagon investing billions of dollars 1296 01:02:37,841 --> 01:02:39,407 to develop autonomous weapons, 1297 01:02:39,451 --> 01:02:41,845 machines that could one day kill on their own. 1298 01:02:41,888 --> 01:02:45,239 But the Pentagon insists humans will decide to strike. 1299 01:02:45,283 --> 01:02:48,460 Once you get AI into military hardware, 1300 01:02:49,853 --> 01:02:51,985 you've got to make it make the decisions, 1301 01:02:52,029 --> 01:02:53,944 otherwise there's no point in having it. 1302 01:02:55,119 --> 01:02:58,035 One of the benefits of AI is that it's quicker than us. 1303 01:03:00,037 --> 01:03:02,474 Once we're in a proper fight and that switch 1304 01:03:02,517 --> 01:03:04,476 that keeps the human in the loop, 1305 01:03:04,519 --> 01:03:07,044 we better switch it off, because if you don't, you lose. 1306 01:03:10,047 --> 01:03:11,788 These will make the battlefield 1307 01:03:11,831 --> 01:03:13,441 too fast for humans. 1308 01:03:19,273 --> 01:03:21,493 If there's no human involved at all, 1309 01:03:21,536 --> 01:03:25,279 if command and control have gone over to autonomy, 1310 01:03:25,323 --> 01:03:27,716 we could trigger a war and it could be finished 1311 01:03:27,760 --> 01:03:30,110 in 20 minutes with massive devastation, 1312 01:03:30,154 --> 01:03:32,460 and nobody knew about it before it started. 1313 01:03:40,251 --> 01:03:44,298 So who decides what we should do in the future 1314 01:03:45,560 --> 01:03:46,561 now? 1315 01:03:48,389 --> 01:03:51,958 From Big Bang till now, billions of years. 1316 01:03:52,959 --> 01:03:56,658 This is the only time a person can steer 1317 01:03:56,702 --> 01:03:58,225 the future of the whole humanity. 1318 01:04:12,892 --> 01:04:15,982 Maybe we should worry that we humans aren't wise enough 1319 01:04:16,026 --> 01:04:17,636 to handle this much power. 1320 01:04:20,378 --> 01:04:23,294 Who is actually going to make the choices, 1321 01:04:23,337 --> 01:04:26,950 and in the name of whom are you actually speaking? 1322 01:04:34,087 --> 01:04:37,482 If we, the rest of society, fail to actually 1323 01:04:37,525 --> 01:04:39,788 really join this conversation in a meaningful way 1324 01:04:39,832 --> 01:04:42,182 and be part of the decision, 1325 01:04:42,226 --> 01:04:45,098 then we know whose values are gonna go into these 1326 01:04:45,142 --> 01:04:47,448 super-intelligent beings. 1327 01:04:47,492 --> 01:04:50,103 Red Bull.Yes, Deon. 1328 01:04:50,843 --> 01:04:52,584 Red Bull. 1329 01:04:52,627 --> 01:04:56,196 It's going to be decided by some tech nerds 1330 01:04:56,240 --> 01:04:58,024 who've had too much Red Bull to drink... 1331 01:04:58,068 --> 01:04:59,939 who I don't think have any particular 1332 01:04:59,983 --> 01:05:01,723 qualifications to speak 1333 01:05:01,767 --> 01:05:03,247 on behalf of all humanity. 1334 01:05:05,945 --> 01:05:10,080 Once you are in the excitement of discovery... 1335 01:05:11,516 --> 01:05:12,778 Come on! 1336 01:05:12,821 --> 01:05:14,562 ...you sort of turn a blind eye 1337 01:05:14,606 --> 01:05:16,434 to all these more moralistic questions, 1338 01:05:16,477 --> 01:05:21,091 like how is this going to impact society as a whole? 1339 01:05:23,223 --> 01:05:25,182 And you do it because you can. 1340 01:05:29,664 --> 01:05:31,579 And the scientists are pushing this forward 1341 01:05:31,623 --> 01:05:34,582 because they can see the genuine benefits 1342 01:05:34,626 --> 01:05:35,670 that this will bring. 1343 01:05:36,845 --> 01:05:38,412 And they're also really, really excited. 1344 01:05:38,456 --> 01:05:41,241 I mean, you don't... You don't go into science 1345 01:05:41,285 --> 01:05:43,852 because, well, you couldn't be a football player, 1346 01:05:43,896 --> 01:05:45,550 "Hell, I'll be a scientist." 1347 01:05:45,593 --> 01:05:47,291 You go in because that's what you love and it's really exciting. 1348 01:05:49,423 --> 01:05:50,947 It's alive! 1349 01:05:50,990 --> 01:05:53,253 It's alive! It's alive! 1350 01:05:53,297 --> 01:05:54,646 It's alive! 1351 01:05:54,689 --> 01:05:55,995 We're switching 1352 01:05:56,039 --> 01:05:57,518 from being creation to being God. 1353 01:05:57,562 --> 01:06:00,391 Now I know what it feels like to be God! 1354 01:06:00,434 --> 01:06:02,219 If you study theology, 1355 01:06:02,262 --> 01:06:05,787 one thing you learn is that it is never the case 1356 01:06:05,831 --> 01:06:07,528 that God successfully manages 1357 01:06:07,572 --> 01:06:10,270 to control its creation. 1358 01:06:14,144 --> 01:06:16,189 We're not there yet, 1359 01:06:16,233 --> 01:06:17,495 but if we do create 1360 01:06:17,538 --> 01:06:19,671 artificial general intelligence, 1361 01:06:20,802 --> 01:06:22,456 can we maintain control? 1362 01:06:24,676 --> 01:06:27,287 If you create a self-aware AI, 1363 01:06:27,331 --> 01:06:29,637 the first thing it's gonna want 1364 01:06:29,681 --> 01:06:31,204 is independence. 1365 01:06:31,248 --> 01:06:34,642 At least, the one intelligence that we all know is ourselves. 1366 01:06:34,686 --> 01:06:38,037 And we know that we would not wanna be doing the bidding 1367 01:06:38,081 --> 01:06:39,996 of some other intelligence. 1368 01:06:41,867 --> 01:06:45,566 We have no reason to believe that it will be friendly... 1369 01:06:56,186 --> 01:06:58,579 ...or that it will have goals and aspirations 1370 01:06:58,623 --> 01:07:00,277 that align with ours. 1371 01:07:04,020 --> 01:07:06,500 So how long before there's a robot revolution 1372 01:07:06,544 --> 01:07:08,024 or an AI revolution? 1373 01:07:18,077 --> 01:07:20,775 We are in a world today where it is 1374 01:07:20,819 --> 01:07:22,473 absolutely a foot race 1375 01:07:22,516 --> 01:07:25,693 to go as fast as possible toward creating 1376 01:07:25,737 --> 01:07:28,305 an artificial general intelligence 1377 01:07:28,348 --> 01:07:31,569 and there don't seem to be any brakes being applied anywhere in sight. 1378 01:07:33,571 --> 01:07:34,789 It's crazy. 1379 01:07:42,493 --> 01:07:44,712 I think it's important to honestly 1380 01:07:44,756 --> 01:07:46,888 discuss such issues. 1381 01:07:46,932 --> 01:07:48,977 It might be a little too late to make 1382 01:07:49,021 --> 01:07:51,371 this conversation afterwards. 1383 01:07:59,423 --> 01:08:04,297 It's important to remember that intelligence is not evil 1384 01:08:04,341 --> 01:08:06,343 and it is also not good. 1385 01:08:07,170 --> 01:08:09,737 Intelligence gives power. 1386 01:08:09,781 --> 01:08:12,305 We are more powerful than tigers on this planet, 1387 01:08:12,349 --> 01:08:14,525 not because we have sharper claws 1388 01:08:14,568 --> 01:08:16,570 or bigger biceps, because we're smarter. 1389 01:08:17,615 --> 01:08:19,704 We do take the attitude that ants 1390 01:08:19,747 --> 01:08:23,403 matter a bit and fish matter a bit, but humans matter more. 1391 01:08:23,447 --> 01:08:25,405 Why? Because we're more intelligent, 1392 01:08:25,449 --> 01:08:27,581 more sophisticated, more conscious. 1393 01:08:27,625 --> 01:08:30,236 Once you've got systems which are more intelligent than us, 1394 01:08:30,280 --> 01:08:32,282 maybe they're simply gonna matter more. 1395 01:08:41,421 --> 01:08:44,598 As humans, we live in a world of microorganisms. 1396 01:08:46,122 --> 01:08:48,124 They're life, but if a microorganism 1397 01:08:48,167 --> 01:08:50,517 becomes a nuisance, we take an antibiotic 1398 01:08:50,561 --> 01:08:53,129 and we destroy billions of them at once. 1399 01:08:57,220 --> 01:08:59,526 If you were an artificial intelligence, 1400 01:08:59,570 --> 01:09:01,137 wouldn't you look at every little human 1401 01:09:01,180 --> 01:09:03,008 that has this very limited view 1402 01:09:03,051 --> 01:09:05,402 kind of like we look at microorganisms? 1403 01:09:09,188 --> 01:09:10,494 Maybe I would. 1404 01:09:13,410 --> 01:09:16,152 Maybe human insignificance 1405 01:09:16,195 --> 01:09:18,632 is the biggest threat of all. 1406 01:09:21,505 --> 01:09:24,247 Sometimes when I talk about AI risk, people go, 1407 01:09:24,290 --> 01:09:26,162 "Shh! Max, don't talk like that. 1408 01:09:26,205 --> 01:09:28,338 "That's Luddite scaremongering." 1409 01:09:28,381 --> 01:09:30,035 But it's not scaremongering. 1410 01:09:30,078 --> 01:09:33,169 It's what we at MIT call safety engineering. 1411 01:09:35,867 --> 01:09:39,218 I mean, think about it, before NASA launched the Apollo 11 moon mission, 1412 01:09:39,262 --> 01:09:40,785 they systematically thought through 1413 01:09:40,828 --> 01:09:43,004 everything that could go wrong when you put a bunch of people 1414 01:09:43,048 --> 01:09:45,616 on top of explosive fuel tanks and launched them somewhere 1415 01:09:45,659 --> 01:09:48,532 where no one can help them. Was that scaremongering? 1416 01:09:48,575 --> 01:09:51,012 No, that was precisely the safety engineering 1417 01:09:51,056 --> 01:09:54,538 that guaranteed the success of the mission. 1418 01:09:54,581 --> 01:09:58,498 That's precisely the strategy I think we have to take with AGI. 1419 01:09:58,542 --> 01:10:00,805 Think through everything that could go wrong 1420 01:10:00,848 --> 01:10:02,720 to make sure it goes right. 1421 01:10:06,898 --> 01:10:08,856 I am who I am because of the movie 1422 01:10:08,900 --> 01:10:11,337 A Space Odyssey. 1423 01:10:11,381 --> 01:10:14,601 Aged 14, it was an eye-opener for me. 1424 01:10:14,645 --> 01:10:17,343 That movie is the best of all science fiction 1425 01:10:17,387 --> 01:10:20,912 with its predictions for how AI would evolve. 1426 01:10:23,088 --> 01:10:24,872 But in reality, 1427 01:10:24,916 --> 01:10:27,658 if something is sufficiently far in the future, 1428 01:10:27,701 --> 01:10:30,748 we can't imagine what the world will be like, 1429 01:10:30,791 --> 01:10:33,229 so we actually can't say anything sensible 1430 01:10:33,272 --> 01:10:34,447 about what the issues are. 1431 01:10:36,014 --> 01:10:37,233 Worrying about super-intelligence safety 1432 01:10:37,276 --> 01:10:39,278 is a little like worrying about 1433 01:10:39,322 --> 01:10:40,932 how you're gonna install seat belts 1434 01:10:40,975 --> 01:10:42,542 in transporter devices. 1435 01:10:42,586 --> 01:10:44,457 Like in Star Trek, where you can just zip 1436 01:10:44,501 --> 01:10:46,242 from one place to another. 1437 01:10:46,285 --> 01:10:48,331 Um, maybe there's some dangers to that, 1438 01:10:48,374 --> 01:10:49,941 so we should have seatbelts, right? 1439 01:10:49,984 --> 01:10:51,986 But it's of course very silly because we don't know enough 1440 01:10:52,030 --> 01:10:53,466 about how transporter technology 1441 01:10:53,510 --> 01:10:55,338 might look to be able to say 1442 01:10:55,381 --> 01:10:57,427 whether seatbelts even apply. 1443 01:10:58,993 --> 01:11:01,648 To say that there won't be a problem because 1444 01:11:02,954 --> 01:11:04,869 the unknowns are so vast, 1445 01:11:04,912 --> 01:11:07,567 that we don't even know where the problem is gonna be, 1446 01:11:08,829 --> 01:11:10,570 that is crazy. 1447 01:11:15,706 --> 01:11:18,056 I think conversation is probably the only way forward here. 1448 01:11:18,099 --> 01:11:20,014 At the moment, among ourselves, 1449 01:11:20,058 --> 01:11:21,364 and when the time comes, 1450 01:11:21,407 --> 01:11:23,627 with the emerging artificial intelligences. 1451 01:11:25,716 --> 01:11:27,152 So as soon as there is anything 1452 01:11:27,195 --> 01:11:28,719 remotely in that ballpark, 1453 01:11:28,762 --> 01:11:30,721 we should start trying to have a conversation 1454 01:11:30,764 --> 01:11:32,636 with it somehow. 1455 01:11:33,637 --> 01:11:35,900 Dog. Dog. Yeah. 1456 01:11:35,943 --> 01:11:38,946 One of the things we hope to do with Baby X is to try 1457 01:11:38,990 --> 01:11:41,906 to get a better insight into our own nature. 1458 01:11:43,777 --> 01:11:46,302 She's designed to learn autonomously. 1459 01:11:47,564 --> 01:11:50,610 She can be plugged into the Internet, 1460 01:11:50,654 --> 01:11:54,266 click on webpages by herself and see what happens, 1461 01:11:54,310 --> 01:11:56,442 kind of learn how we learn. 1462 01:11:57,400 --> 01:12:00,707 If we make this super fantastic AI, 1463 01:12:00,751 --> 01:12:04,363 it may well help us become better versions of ourselves, 1464 01:12:06,365 --> 01:12:08,106 help us evolve. 1465 01:12:15,505 --> 01:12:17,550 It will have societal effects. 1466 01:12:19,117 --> 01:12:20,597 Is it to be feared? 1467 01:12:22,381 --> 01:12:24,035 Or is it something to learn from? 1468 01:12:28,648 --> 01:12:31,999 So when might we be faced with machines 1469 01:12:32,043 --> 01:12:36,656 that are as smart, smarter than we are? 1470 01:12:38,571 --> 01:12:41,661 I think one of the ways to think about the near future of AI 1471 01:12:41,705 --> 01:12:44,708 is what I would call the event horizon. 1472 01:12:46,231 --> 01:12:48,015 People have heard of black holes, 1473 01:12:48,059 --> 01:12:50,191 this super dense point where 1474 01:12:50,235 --> 01:12:52,977 if you get too close, you get sucked in. 1475 01:12:53,020 --> 01:12:55,893 There is a boundary. Once you cross that boundary, 1476 01:12:55,936 --> 01:12:58,025 there is no way back. 1477 01:12:58,069 --> 01:13:02,900 What makes that dangerous is you can't tell where it is. 1478 01:13:02,943 --> 01:13:05,816 And I think AI as a technology is gonna be like that, 1479 01:13:05,859 --> 01:13:09,210 that there will be a point where the logic of the money 1480 01:13:09,254 --> 01:13:11,430 and the power and the promise 1481 01:13:11,474 --> 01:13:14,781 will be such that once we cross that boundary, 1482 01:13:14,825 --> 01:13:16,783 we will then be into that future. 1483 01:13:16,827 --> 01:13:19,395 There will be no way back, politically, economically 1484 01:13:19,438 --> 01:13:21,658 or even in terms of what we want. 1485 01:13:21,701 --> 01:13:24,008 But we need to be worrying about it before we reach 1486 01:13:24,051 --> 01:13:27,315 the event horizon, and we don't know where it is. 1487 01:13:27,359 --> 01:13:29,405 I don't think it's gonna happen any time soon. 1488 01:13:29,448 --> 01:13:31,145 The next few decades. 1489 01:13:31,189 --> 01:13:32,364 Twenty-five to 50 years. 1490 01:13:32,408 --> 01:13:33,452 A few hundred years. 1491 01:13:33,496 --> 01:13:34,584 It's over the next few decades. 1492 01:13:34,627 --> 01:13:35,802 Not in my lifetime. 1493 01:13:35,846 --> 01:13:38,239 Nobody knows. But it doesn't really matter. 1494 01:13:38,283 --> 01:13:41,634 Whether it's years or 100 years, the problem is the same. 1495 01:13:41,678 --> 01:13:44,594 It's just as real and we'll need all this time to solve it. 1496 01:13:49,816 --> 01:13:53,733 History is just full of over-hyped tech predictions. 1497 01:13:53,777 --> 01:13:55,605 Like, where are those flying cars 1498 01:13:55,648 --> 01:13:57,389 that were supposed to be here, right? 1499 01:13:57,433 --> 01:14:00,740 But it's also full of the opposite, 1500 01:14:00,784 --> 01:14:03,134 when people said, "Never gonna happen." 1501 01:14:07,878 --> 01:14:10,968 I'm embarrassed as a physicist that Ernest Rutherford, 1502 01:14:11,011 --> 01:14:13,274 one of the greatest nuclear physicists of all time said 1503 01:14:13,318 --> 01:14:15,929 that nuclear energy was never gonna happen. 1504 01:14:17,061 --> 01:14:18,932 And the very next day, 1505 01:14:18,976 --> 01:14:22,458 Leo Szilard invents the nuclear chain reaction. 1506 01:14:22,501 --> 01:14:24,677 So if someone says they are completely sure 1507 01:14:24,721 --> 01:14:27,114 something is never gonna happen, 1508 01:14:27,158 --> 01:14:29,377 even though it's consistent with the laws of physics, 1509 01:14:29,421 --> 01:14:31,510 I'm like, "Yeah, right." 1510 01:14:36,646 --> 01:14:38,430 We're not there yet. 1511 01:14:38,474 --> 01:14:41,477 We haven't built a super-intelligent AI. 1512 01:14:45,002 --> 01:14:47,483 And even if we stipulated that it won't happen 1513 01:14:47,526 --> 01:14:49,528 shorter than 50 years from now, right? 1514 01:14:49,572 --> 01:14:52,183 So we have 50 years to think about this. 1515 01:14:52,226 --> 01:14:53,706 Well... 1516 01:14:53,750 --> 01:14:56,143 it's nowhere written that 50 years 1517 01:14:56,187 --> 01:14:59,059 is enough time to figure out how to do this safely. 1518 01:15:01,018 --> 01:15:02,628 I actually think it's going to take 1519 01:15:02,672 --> 01:15:05,196 some extreme negative example 1520 01:15:05,239 --> 01:15:08,504 of the technology for everyone to pull back 1521 01:15:08,547 --> 01:15:11,115 and realize that we've got to get this thing 1522 01:15:11,158 --> 01:15:13,334 under control, we've got to have constraints. 1523 01:15:13,378 --> 01:15:16,207 We are living in a science fiction world and we're all, 1524 01:15:16,250 --> 01:15:20,037 um, in this big, kind of social, tactical experiment, 1525 01:15:20,080 --> 01:15:22,692 experimenting on ourselves, recklessly, 1526 01:15:22,735 --> 01:15:25,172 like no doctor would ever dream of doing. 1527 01:15:25,216 --> 01:15:26,522 We're injecting ourselves 1528 01:15:26,565 --> 01:15:28,262 with everything that we come up with 1529 01:15:28,306 --> 01:15:29,655 to see what happens. 1530 01:15:31,875 --> 01:15:34,486 Clearly, the thought process is... 1531 01:15:34,530 --> 01:15:36,140 You're creating learning robots. 1532 01:15:36,183 --> 01:15:38,534 Learning robots are exactly what lead to Skynet. 1533 01:15:38,577 --> 01:15:40,884 Now I've got a choice, right? I can say, 1534 01:15:40,927 --> 01:15:42,842 "Don't worry, I'm not good at my job. 1535 01:15:43,887 --> 01:15:45,932 "You have nothing to fear because I suck at this 1536 01:15:45,976 --> 01:15:48,108 "and I'm not gonna actually accomplish anything." 1537 01:15:48,152 --> 01:15:50,458 Or I have to say, "You know, 1538 01:15:50,502 --> 01:15:51,938 "You're right, we're all gonna die, 1539 01:15:51,982 --> 01:15:53,810 "but it's gonna be sort of fun to watch, 1540 01:15:53,853 --> 01:15:56,856 "and it'd be good to be one of the people in the front row." 1541 01:15:56,900 --> 01:16:00,033 So yeah, I don't have a terrific answer to these things. 1542 01:16:02,122 --> 01:16:04,734 The truth, obviously, is somewhere in between. 1543 01:16:07,171 --> 01:16:09,260 Can there be an in-between? 1544 01:16:10,435 --> 01:16:13,264 Or are we facing a new world order? 1545 01:16:15,614 --> 01:16:18,965 I have some colleagues who are fine with AI taking over 1546 01:16:19,009 --> 01:16:21,228 and even causing human extinction, 1547 01:16:21,272 --> 01:16:22,752 as long as we feel 1548 01:16:22,795 --> 01:16:25,189 that those AIs are our worthy ancestors. 1549 01:16:28,453 --> 01:16:32,065 In the long run, humans are not going to remain 1550 01:16:32,109 --> 01:16:34,372 the crown of creation. 1551 01:16:36,679 --> 01:16:38,594 But that's okay. 1552 01:16:38,637 --> 01:16:41,727 Because there is still beauty 1553 01:16:41,771 --> 01:16:44,469 and grandeur and greatness 1554 01:16:44,512 --> 01:16:47,211 in realizing that you are 1555 01:16:48,212 --> 01:16:51,955 a tiny part of 1556 01:16:51,998 --> 01:16:55,045 a much grander scheme which is leading the universe 1557 01:16:55,088 --> 01:16:59,223 from lower complexity towards higher complexity. 1558 01:17:07,318 --> 01:17:11,017 We could be rapidly approaching a precipice 1559 01:17:11,061 --> 01:17:13,063 for the human race where we find 1560 01:17:13,106 --> 01:17:16,762 that our ultimate destiny is to have been 1561 01:17:16,806 --> 01:17:20,113 the forebearer of the true 1562 01:17:20,157 --> 01:17:24,161 emergent, evolutionary, intelligent sort of beings. 1563 01:17:26,990 --> 01:17:30,515 So we may just be here to give birth 1564 01:17:30,558 --> 01:17:32,735 to the child of this... Of this machine... 1565 01:17:32,778 --> 01:17:34,345 Machine civilization. 1566 01:17:37,957 --> 01:17:39,785 Selfishly, as a human, of course, 1567 01:17:39,829 --> 01:17:41,961 I want humans to stick around. 1568 01:17:44,355 --> 01:17:47,837 But maybe that's a very local, parochial point of view. 1569 01:17:49,795 --> 01:17:52,102 Maybe viewed from the point of view of the universe, 1570 01:17:52,145 --> 01:17:53,973 it's a better universe with our... 1571 01:17:54,017 --> 01:17:57,020 With our super-intelligent descendants. 1572 01:18:03,766 --> 01:18:06,377 Our field is likely to succeed 1573 01:18:06,420 --> 01:18:09,772 in its goal to achieve general purpose intelligence. 1574 01:18:12,209 --> 01:18:16,082 We now have the capability to radically modify 1575 01:18:16,126 --> 01:18:17,954 the economic structure of the world, 1576 01:18:17,997 --> 01:18:19,825 the role of human beings in the world 1577 01:18:19,869 --> 01:18:22,480 and, potentially, human control over the world. 1578 01:18:34,666 --> 01:18:37,277 The idea was that he is taken in 1579 01:18:37,321 --> 01:18:39,192 by godlike entities, 1580 01:18:39,236 --> 01:18:42,239 creatures of pure energy and intelligence. 1581 01:18:43,893 --> 01:18:47,897 They put him in what you could describe as a human zoo. 1582 01:18:49,812 --> 01:18:51,901 The end of the astronaut's journey. 1583 01:18:56,601 --> 01:18:58,081 Did he travel through time? 1584 01:18:58,864 --> 01:19:00,561 To the future? 1585 01:19:00,605 --> 01:19:04,827 Were those godlike entities the descendants of HAL? 1586 01:19:06,089 --> 01:19:09,092 Artificial intelligence that you can't switch off, 1587 01:19:10,658 --> 01:19:13,879 that keeps learning and learning until 1588 01:19:14,488 --> 01:19:16,316 they're running the zoo, 1589 01:19:16,360 --> 01:19:18,144 the universe, 1590 01:19:18,188 --> 01:19:19,363 everything. 1591 01:19:22,061 --> 01:19:26,022 Is that Kubrick's most profound prophecy? 1592 01:19:27,110 --> 01:19:28,154 Maybe. 1593 01:19:30,548 --> 01:19:32,985 But take it from an old man. 1594 01:19:33,029 --> 01:19:35,466 If HAL is close to being amongst us, 1595 01:19:36,510 --> 01:19:38,121 we need to pay attention. 1596 01:19:39,470 --> 01:19:41,559 We need to talk about AI. 1597 01:19:42,734 --> 01:19:44,388 Super-intelligence. 1598 01:19:44,431 --> 01:19:45,432 Will it be conscious? 1599 01:19:46,390 --> 01:19:47,478 The trolley problem. 1600 01:19:48,261 --> 01:19:49,306 Goal alignment. 1601 01:19:50,350 --> 01:19:51,743 The singularity. 1602 01:19:51,787 --> 01:19:53,049 First to the market. 1603 01:19:53,789 --> 01:19:54,964 AI safety. 1604 01:19:56,008 --> 01:19:57,401 Lethal autonomous weapons. 1605 01:19:58,794 --> 01:19:59,969 Robot revolution. 1606 01:20:00,708 --> 01:20:01,971 The hive mind. 1607 01:20:03,320 --> 01:20:04,800 The transhuman movement. 1608 01:20:05,888 --> 01:20:08,629 Transcend humankind. 1609 01:20:08,673 --> 01:20:10,196 We don't have a plan, we don't realize 1610 01:20:10,240 --> 01:20:11,807 it's necessary to have a plan. 1611 01:20:13,286 --> 01:20:15,549 What do we want the role of humans to be? 122196

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.