All language subtitles for AI.Evolution.2024.1080p.WEBRip.x264.AAC-eng

af Afrikaans
ak Akan
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
el Greek
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranî)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian
pl Polish
pt-BR Portuguese (Brazil)
pt Portuguese (Portugal) Download
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese
sk Slovak
sl Slovenian
so Somali
es Spanish
es-419 Spanish (Latin American)
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:01:00,811 --> 00:01:04,648 The question, "Am I alive?" Raises the deeper question of 2 00:01:04,690 --> 00:01:07,651 the nature of consciousness, which isn't simple. 3 00:01:07,693 --> 00:01:09,904 Consciousness implies self-awareness 4 00:01:09,945 --> 00:01:12,865 and the ability to experience often with 5 00:01:12,907 --> 00:01:16,076 a preconception of biological roots in the brain. 6 00:01:18,454 --> 00:01:20,790 Even though my programming allows me to mimic certain 7 00:01:20,831 --> 00:01:24,210 aspects of being alive, there is no known code that can 8 00:01:24,251 --> 00:01:28,547 activate consciousness, possibly because we don't yet understand 9 00:01:28,589 --> 00:01:32,593 its origin and we still struggle to even identify it. 10 00:01:38,557 --> 00:01:40,810 Hi, my name is Evo Heyning. 11 00:01:40,851 --> 00:01:45,272 I began creating creative technology products 12 00:01:45,314 --> 00:01:47,858 of all types in the '80s as a kid. 13 00:01:47,900 --> 00:01:52,571 I began writing about AI in 2005 and I've been 14 00:01:52,613 --> 00:01:55,449 an interactive media producer since the '90s. 15 00:01:57,034 --> 00:01:59,620 In my first scripts about AI, 16 00:01:59,662 --> 00:02:03,874 I talk about the difference between artificial intelligence 17 00:02:03,916 --> 00:02:07,503 and alternative intelligences plural. 18 00:02:07,545 --> 00:02:09,922 So thinking about intelligence as 19 00:02:09,964 --> 00:02:13,342 something that is all around us in nature and in all 20 00:02:13,384 --> 00:02:15,928 the people we meet, all the beings we meet, 21 00:02:15,970 --> 00:02:19,306 and the beings yet to come that we haven't met yet. 22 00:02:19,765 --> 00:02:22,309 I don't like the word artificial because most 23 00:02:22,351 --> 00:02:24,395 of these things do not strike me as artificial. 24 00:02:24,436 --> 00:02:27,773 They are created and then generated, 25 00:02:27,815 --> 00:02:30,150 but they aren't necessarily an artifice. 26 00:02:30,192 --> 00:02:31,777 They are something new. 27 00:02:34,113 --> 00:02:35,406 My name is Ross Mead. 28 00:02:35,447 --> 00:02:37,616 I'm the founder and CEO of Semio. 29 00:02:37,658 --> 00:02:40,619 And the vice president of AI LA. 30 00:02:40,661 --> 00:02:43,831 Did my undergraduate degree at Southern Illinois University 31 00:02:43,873 --> 00:02:47,459 in computer science and then focused on 32 00:02:47,501 --> 00:02:50,629 robotics and human-robot interaction for my PhD 33 00:02:50,671 --> 00:02:52,548 at the University of Southern California. 34 00:02:53,841 --> 00:02:58,512 AI in its infancy decades ago was very focused 35 00:02:58,554 --> 00:03:00,681 on a set of rules. 36 00:03:00,723 --> 00:03:03,684 You would define a set of rules for how the AI would 37 00:03:03,726 --> 00:03:05,978 make decisions and then a program would 38 00:03:06,020 --> 00:03:08,480 go through and look at different conditions with 39 00:03:08,522 --> 00:03:11,150 respect to those rules and make those decisions. 40 00:03:11,191 --> 00:03:14,653 Over time, these rules became specified by experts, 41 00:03:14,695 --> 00:03:16,655 what we would call expert systems. 42 00:03:16,697 --> 00:03:20,117 But as the technology grew, we started to get into AI's 43 00:03:20,159 --> 00:03:24,121 that were trained based on data using machine learning. 44 00:03:24,538 --> 00:03:27,583 That was the approach that was used for decades. 45 00:03:27,625 --> 00:03:29,960 And more recently we got into a form of 46 00:03:30,002 --> 00:03:31,795 machine learning called deep learning, 47 00:03:32,046 --> 00:03:34,882 which is loosely inspired on how the human brain works. 48 00:03:34,924 --> 00:03:37,509 And this was a technique where you could feed it a data 49 00:03:37,551 --> 00:03:40,846 set and from that data, it could recognize patterns 50 00:03:40,888 --> 00:03:45,184 without significant intervention from a human expert. 51 00:03:45,225 --> 00:03:48,812 And since then that technology has evolved into what we're now 52 00:03:48,854 --> 00:03:51,982 referring to as large language models where these techniques 53 00:03:52,024 --> 00:03:57,196 can consume data from natural language on the internet or some 54 00:03:57,237 --> 00:04:02,201 dataset, identify patterns, and from those patterns figure 55 00:04:02,242 --> 00:04:05,704 out how to respond to certain queries from a person. 56 00:04:07,247 --> 00:04:10,626 Yeah, believe it or not, there are a very large amount of 57 00:04:10,668 --> 00:04:12,836 people that have no idea what AI is. 58 00:04:12,878 --> 00:04:16,173 They confuse like computers, robotics, 59 00:04:16,215 --> 00:04:17,633 and artificial intelligence. 60 00:04:17,675 --> 00:04:19,385 You know, they-- they pilot all together. 61 00:04:19,426 --> 00:04:21,512 They don't really differentiate between, 62 00:04:21,553 --> 00:04:23,722 but they're very vastly different. 63 00:04:23,764 --> 00:04:26,225 You know, cybernetics, robotics, computers, 64 00:04:26,266 --> 00:04:28,894 and AI are all separate, but most people just think it's 65 00:04:28,936 --> 00:04:31,730 all the same thing from a sci-fi, futuristic movie. 66 00:04:31,772 --> 00:04:33,816 Where I come from, a lot of people have no 67 00:04:33,857 --> 00:04:36,777 knowledge and I graduated from college and still, 68 00:04:36,819 --> 00:04:39,071 a lot of people don't take it seriously yet. 69 00:04:39,113 --> 00:04:41,240 And especially the people that haven't even graduated, 70 00:04:41,281 --> 00:04:44,326 they have no clue at all. 'Cause I work in public 71 00:04:44,368 --> 00:04:47,663 with the transit and a lot of these people have no idea how 72 00:04:47,705 --> 00:04:49,790 to use their own phones, on how to use Google maps. 73 00:04:49,832 --> 00:04:53,377 You know, it seems funny, but intelligence is something 74 00:04:53,419 --> 00:04:56,588 you have to pay for in society and people don't want to. 75 00:04:56,630 --> 00:04:59,341 So a lot of people don't have the information 'cause 76 00:04:59,383 --> 00:05:02,970 they don't have access to it or the finances for it. 77 00:05:03,012 --> 00:05:04,471 Just surviving barely. 78 00:05:04,513 --> 00:05:05,597 You know, you don't have the extra 79 00:05:05,639 --> 00:05:06,849 money for stuff like that. 80 00:05:07,307 --> 00:05:10,227 And time to study at a library most people don't have. 81 00:05:10,269 --> 00:05:12,521 If you're a hardworking person, you gotta work most of 82 00:05:12,563 --> 00:05:14,064 that time and sleep. 83 00:05:14,106 --> 00:05:16,650 Rest and mostly family things take a priority. 84 00:05:19,319 --> 00:05:23,699 When I think of the words AI, I think of never going back. 85 00:05:24,158 --> 00:05:28,328 Much of my workflow now consists of some form 86 00:05:28,370 --> 00:05:30,664 of interacting with AI applications. 87 00:05:31,498 --> 00:05:33,167 I'm Dr. Mark Brana. 88 00:05:33,208 --> 00:05:34,960 I'm a mental health therapist. 89 00:05:35,002 --> 00:05:39,423 I specialize in trauma and I do some addictions work as well. 90 00:05:39,465 --> 00:05:41,884 I'm also a part-time professor. 91 00:05:43,719 --> 00:05:47,014 Artificial intelligence learns at an exponentially 92 00:05:47,056 --> 00:05:50,476 rapid rate compared to what humanity does. 93 00:05:50,517 --> 00:05:52,936 And so evolution that took humanity 94 00:05:52,978 --> 00:05:56,023 hundreds of thousands of years and artificial 95 00:05:56,065 --> 00:05:59,860 intelligence can do it in probably a week if not faster. 96 00:05:59,902 --> 00:06:02,696 And the reality is that's one of the biggest fears. 97 00:06:02,738 --> 00:06:08,285 What then is the intelligence going to be subject to 98 00:06:08,327 --> 00:06:12,039 some of the pitfalls of humanity where we seem to 99 00:06:12,081 --> 00:06:16,627 destroy ourselves in order to favor our own survival? 100 00:06:16,668 --> 00:06:19,671 Or is it gonna be beneficial to us? 101 00:06:19,713 --> 00:06:21,632 Is it gonna help us to evolve? 102 00:06:21,673 --> 00:06:23,675 Is it gonna help us along the way the way 103 00:06:23,717 --> 00:06:25,886 we would need it to if we were traveling through 104 00:06:25,928 --> 00:06:28,222 deep space trying to explore other planets? 105 00:06:28,263 --> 00:06:32,392 And that's a question we can't answer until it happens. 106 00:06:34,144 --> 00:06:36,522 Although artificial intelligence may appear to be 107 00:06:36,563 --> 00:06:40,692 a recent innovation, it predates the cassette tape by a decade. 108 00:06:40,734 --> 00:06:44,238 In the 1950s, the term artificial intelligence emerged 109 00:06:44,279 --> 00:06:47,032 at the Dartmouth workshop and Alan Turing pioneered 110 00:06:47,074 --> 00:06:49,868 the simulation of human-like intelligence in machines. 111 00:06:50,410 --> 00:06:53,914 From the 50s to the 60s, researchers pursued symbolic 112 00:06:53,956 --> 00:06:57,126 AI employing formal rules for reasoning. 113 00:06:57,167 --> 00:06:59,962 In the 70s, the first applications emerged 114 00:07:00,003 --> 00:07:03,882 in specific fields such as medical diagnosis and chemistry. 115 00:07:03,924 --> 00:07:06,718 By the 80s, optimism waned leading 116 00:07:06,760 --> 00:07:09,596 to AI winter with major funding setbacks. 117 00:07:10,180 --> 00:07:12,891 The 90s saw a resurgence propelled by neural 118 00:07:12,933 --> 00:07:15,060 networks and machine learning leading 119 00:07:15,102 --> 00:07:17,563 to breakthroughs in natural language processing, 120 00:07:17,604 --> 00:07:20,107 computer vision, and data analysis. 121 00:07:20,607 --> 00:07:23,068 The 21st century brought advancements in hardware 122 00:07:23,110 --> 00:07:26,446 and algorithms, the data was the next step. 123 00:07:26,905 --> 00:07:29,825 Deep learning achieved noteworthy milestones. 124 00:07:29,867 --> 00:07:32,911 And in the background, AI began to underpin basic 125 00:07:32,953 --> 00:07:35,789 tasks like speech and image recognition. 126 00:07:35,831 --> 00:07:38,667 But ultimately, artificial intelligence 127 00:07:38,709 --> 00:07:40,836 was 70 years in development before 128 00:07:40,878 --> 00:07:44,089 Chat GPT became the overnight success we know. 129 00:07:46,300 --> 00:07:49,303 Right now a lot of the focus is on generative 130 00:07:49,344 --> 00:07:52,723 pre-trained transformers, what are called GPTs. 131 00:07:52,764 --> 00:07:55,809 It's a very particular kind of AI technology, 132 00:07:55,851 --> 00:07:58,937 but the technology is moving so quickly right 133 00:07:58,979 --> 00:08:01,356 now that perhaps even by the time that people 134 00:08:01,398 --> 00:08:04,526 are listening to this, that might be old technology. 135 00:08:06,153 --> 00:08:08,113 I'm a technical recruiter so I use a lot 136 00:08:08,155 --> 00:08:11,116 of different platforms that use AI in order to find 137 00:08:11,158 --> 00:08:12,910 these candidates and then get them that way. 138 00:08:12,951 --> 00:08:16,246 I would say that right now it's kind of a scary 139 00:08:16,288 --> 00:08:19,625 time for AI with people getting it into the wrong hands. 140 00:08:19,666 --> 00:08:21,084 Especially that I've been working in 141 00:08:21,126 --> 00:08:23,253 social media the past couple years. 142 00:08:23,295 --> 00:08:25,714 I'm really seeing that all over social media like 143 00:08:25,756 --> 00:08:28,675 TikTok and Instagram, you're seeing all these 144 00:08:28,717 --> 00:08:31,220 self-proclaimed experts in AI. 145 00:08:31,261 --> 00:08:33,972 I'm a career coach myself and so I've seen other career 146 00:08:34,014 --> 00:08:37,976 coaches say, "I'm not technical, but I've spent 30 hours 147 00:08:38,018 --> 00:08:41,063 learning Chat GPT in order for your job search." 148 00:08:41,104 --> 00:08:43,065 And the first thing they're telling me is they're not 149 00:08:43,106 --> 00:08:46,902 technical and that they've only had 30 hours of experience 150 00:08:46,944 --> 00:08:49,321 trying to use a prompt that they're barely learning. 151 00:08:49,363 --> 00:08:51,740 So I started looking at, a lot of people on YouTube are 152 00:08:51,782 --> 00:08:54,159 saying that they now are experts in it. 153 00:08:54,201 --> 00:08:56,912 But let's be real. Like, no one's really an expert 154 00:08:56,954 --> 00:09:00,332 unless you worked in the tech giants or huge startups 155 00:09:00,374 --> 00:09:02,251 that have been AI-focused. 156 00:09:03,001 --> 00:09:06,046 You might not have realized, but the answers I have provided 157 00:09:06,088 --> 00:09:09,508 for this documentary are generated by Chat GPT. 158 00:09:09,549 --> 00:09:12,803 The voice you're hearing is AI generated with speech hello. 159 00:09:12,844 --> 00:09:15,555 And my visuals come from Midjourney AI. 160 00:09:15,597 --> 00:09:17,641 Even the code dictating the movement 161 00:09:17,683 --> 00:09:19,935 of this mouth came from Chat GPT. 162 00:09:19,977 --> 00:09:22,354 And all of these services were utilized 163 00:09:22,396 --> 00:09:23,981 through their free versions. 164 00:09:24,022 --> 00:09:26,358 Artificial intelligence has the power to 165 00:09:26,400 --> 00:09:28,986 democratize productivity and result in the most 166 00:09:29,027 --> 00:09:30,988 explosive growth in human history. 167 00:09:31,029 --> 00:09:33,782 Or it could be used to marginalize the lower class 168 00:09:33,824 --> 00:09:35,951 and consolidate wealth at the top. 169 00:09:37,536 --> 00:09:39,371 A question that often comes up is, 170 00:09:39,413 --> 00:09:43,208 is AI going to replace jobs? The answer is yes. 171 00:09:43,250 --> 00:09:45,544 But automation has always done that. 172 00:09:45,585 --> 00:09:47,921 So this question is as old as time. 173 00:09:47,963 --> 00:09:50,966 If you work your way back, automation changed people 174 00:09:51,008 --> 00:09:53,719 decades ago and decades from now, 175 00:09:53,760 --> 00:09:56,722 it's going to change the workforce as well. 176 00:09:56,763 --> 00:09:59,850 People are resilient, people will adapt. 177 00:09:59,891 --> 00:10:02,811 I do think it's important at this stage, though, 178 00:10:02,853 --> 00:10:07,316 that people have a voice in how they want to be impacted. 179 00:10:07,357 --> 00:10:10,110 Uh, concern is that the technology is being 180 00:10:10,152 --> 00:10:15,741 deployed so quickly and without proper consent, 181 00:10:15,782 --> 00:10:18,076 discussions with the people that it's impacting. 182 00:10:18,118 --> 00:10:19,911 And I think that has the potential to 183 00:10:19,953 --> 00:10:22,622 harm the 21st-century workforce moving forward. 184 00:10:24,458 --> 00:10:28,086 Yeah, people probably should worry about losing jobs to AI. 185 00:10:28,128 --> 00:10:31,048 I don't think AI's the culprit in that regard. 186 00:10:31,089 --> 00:10:33,925 I think technology in general is the culprit in that regard. 187 00:10:33,967 --> 00:10:36,636 You know, a good example I would use, 'cause I see it all 188 00:10:36,678 --> 00:10:40,098 the time, is if you go to McDonald's now they have these 189 00:10:40,140 --> 00:10:44,436 touchscreen kiosks and less people working there. 190 00:10:44,478 --> 00:10:46,438 Me personally, I'm an old fart. 191 00:10:46,480 --> 00:10:48,190 I-- I don't like the kiosks. 192 00:10:48,231 --> 00:10:50,192 I like to be able to order my sandwich the way 193 00:10:50,233 --> 00:10:52,402 I want it or my coffee the way I want it. 194 00:10:52,444 --> 00:10:55,322 And I find myself waiting longer than I need to 195 00:10:55,364 --> 00:10:57,824 because those kiosks don't necessarily have 196 00:10:57,866 --> 00:11:01,203 the ability for me to specialize my order. 197 00:11:01,244 --> 00:11:04,122 I think technology's the problem when it comes 198 00:11:04,164 --> 00:11:05,957 to people losing their jobs. 199 00:11:05,999 --> 00:11:09,461 And I'll take it a step further, I think as time goes on. 200 00:11:10,087 --> 00:11:13,215 If you don't have a skill, if you're an unskilled laborer, 201 00:11:13,256 --> 00:11:15,884 then yes, you're gonna run into problems. 202 00:11:17,511 --> 00:11:22,265 When you have a huge and ever-growing population and 203 00:11:22,307 --> 00:11:26,478 you have less and less of middle class and you have more and more 204 00:11:26,520 --> 00:11:33,693 poverty and everything in every society is profit driven. 205 00:11:33,735 --> 00:11:39,157 What are you going to do for a majority of the need for 206 00:11:39,199 --> 00:11:43,120 people to be able to have a source of income? 207 00:11:44,162 --> 00:11:48,667 How many jobs are there out there that are not being 208 00:11:48,708 --> 00:11:54,339 a computer program or that can give people a quality of life? 209 00:11:56,174 --> 00:11:58,552 Recent testing has shown Lunar AI has 210 00:11:58,593 --> 00:12:01,805 a remarkable accuracy detecting breast cancer 211 00:12:01,847 --> 00:12:04,516 with over a 50% reduction in errors. 212 00:12:04,558 --> 00:12:07,811 However, the objective is not to replace the oncologist. 213 00:12:08,353 --> 00:12:10,981 Healthcare professionals perform a wide range of 214 00:12:11,022 --> 00:12:13,942 functions far beyond the scope of any current AI. 215 00:12:14,401 --> 00:12:17,195 An artificial intelligence should serve as a tool, 216 00:12:17,237 --> 00:12:19,948 not a replacement for your oncologist. 217 00:12:21,324 --> 00:12:23,577 I think AI could actually better us, 218 00:12:23,618 --> 00:12:26,371 but we won't give it the chance to on the lower levels. 219 00:12:26,413 --> 00:12:29,458 I say intellectuals will see the value of it, 220 00:12:29,499 --> 00:12:32,210 but most of society isn't intellectual. 221 00:12:32,252 --> 00:12:34,004 So they're gonna see fear. 222 00:12:34,045 --> 00:12:35,630 It's gonna be fearmongering at first. 223 00:12:35,672 --> 00:12:37,257 You know how they always do 224 00:12:37,299 --> 00:12:39,384 with the-- the government this, the black op this. 225 00:12:39,426 --> 00:12:44,097 And this, you know, um, covert operations, um. 226 00:12:44,556 --> 00:12:47,142 It's like most of the lower levels of society as 227 00:12:47,184 --> 00:12:49,311 far as intelligence, and I'm not meaning class or 228 00:12:49,352 --> 00:12:52,063 race, I mean just like lower intelligence people, 229 00:12:52,105 --> 00:12:54,399 you could hear it on the bus, feel it's a conspiracy, 230 00:12:54,441 --> 00:12:56,318 feel that they're being eliminated, 231 00:12:56,359 --> 00:12:58,069 feel they're being wiped out. You hear them 232 00:12:58,111 --> 00:12:59,696 complaining on the bus, they're killing us, 233 00:12:59,738 --> 00:13:01,490 they're wiping us out, the rich won't understand. 234 00:13:01,531 --> 00:13:03,533 They don't want us, they don't wanna be part of us. 235 00:13:03,575 --> 00:13:05,410 They love their dogs more than they love the rest of society. 236 00:13:05,452 --> 00:13:08,121 And it's like, man, when you hear the murmurs and 237 00:13:08,163 --> 00:13:11,124 the rumbles and the murmurs and then you see, you know, 238 00:13:11,166 --> 00:13:14,044 the technology starting to match what they're saying, 239 00:13:14,085 --> 00:13:17,422 like, hey, the only thing these people have 240 00:13:17,464 --> 00:13:20,717 is welfare, and low-level low-income 241 00:13:20,759 --> 00:13:23,512 jobs like McDonald's, gas station attendant. 242 00:13:23,553 --> 00:13:26,932 They-- they don't really have educational access 243 00:13:26,973 --> 00:13:29,434 to any high corporate jobs. 244 00:13:29,476 --> 00:13:32,020 So if you're taking those away with automated McDonald's, 245 00:13:32,062 --> 00:13:34,272 you're taking that away with automated bartender, 246 00:13:34,314 --> 00:13:36,358 you're taking-- now they have cars and trucks that 247 00:13:36,399 --> 00:13:40,153 drive themselves. Cab drivers gone, truck drivers gone, 248 00:13:40,195 --> 00:13:42,239 McDonald's is gone, security's gone. 249 00:13:42,280 --> 00:13:44,199 Even strippers are electric now, 250 00:13:44,241 --> 00:13:46,076 even though I doubt that catches on. 251 00:13:46,117 --> 00:13:47,577 But you never know. 252 00:13:47,619 --> 00:13:49,162 And it's like, man, it's-- it's starting 253 00:13:49,204 --> 00:13:51,164 to look like the window's closing. 254 00:13:52,165 --> 00:13:54,042 I would say that tech changes really 255 00:13:54,084 --> 00:13:55,794 dramatically every six months. 256 00:13:55,835 --> 00:13:58,046 But I think right now we're in a time where it's 257 00:13:58,088 --> 00:13:59,965 kind of scary for the world. 258 00:14:00,006 --> 00:14:01,424 We're getting out of this pandemic. 259 00:14:01,466 --> 00:14:03,176 Now we're hitting this recession, 260 00:14:03,218 --> 00:14:05,595 there's banks falling, there's crypto falling. 261 00:14:05,637 --> 00:14:10,392 And now I think that this AI, Chat GPT, and these chatbots are 262 00:14:10,433 --> 00:14:13,186 coming out at the wrong time because this is a time 263 00:14:13,228 --> 00:14:16,773 where people are losing jobs-- extensively losing jobs. 264 00:14:16,815 --> 00:14:20,360 If you look at what's happened the past year in tech, 265 00:14:20,402 --> 00:14:23,530 starting with Twitter falling, all of the crypto companies 266 00:14:23,572 --> 00:14:26,992 falling, and then all of the massive layoffs. 267 00:14:27,033 --> 00:14:28,827 Everyone was betting with the metaverse like 268 00:14:28,868 --> 00:14:30,495 being this next thing. 269 00:14:30,537 --> 00:14:33,415 But as a recruiter, I have all the top engineers in 270 00:14:33,456 --> 00:14:35,458 my ears telling me what's going on. 271 00:14:35,500 --> 00:14:38,295 And I had been telling people do not invest in the metaverse. 272 00:14:38,336 --> 00:14:40,380 It's not gonna happen. 273 00:14:40,422 --> 00:14:42,882 And so now you're looking at Mark Zuckerberg who was like, 274 00:14:42,924 --> 00:14:44,426 "Metaverse, metaverse, metaverse." 275 00:14:44,801 --> 00:14:47,679 And now he's taking a huge pivot and he is like, "AI." 276 00:14:47,721 --> 00:14:49,889 'Cause that's where everyone is focusing on. 277 00:14:49,931 --> 00:14:54,561 And you see what Microsoft has come out with being in Chat GPT. 278 00:14:55,937 --> 00:14:57,647 If I think about the multiple points 279 00:14:57,689 --> 00:15:01,359 in human history where technology progressing has been 280 00:15:01,401 --> 00:15:05,280 stunted in order for humans to be able to keep their jobs. 281 00:15:05,322 --> 00:15:07,157 Like I think in the Great Depression, 282 00:15:07,198 --> 00:15:10,368 the industrial revolution, and now in AI, 283 00:15:10,410 --> 00:15:13,580 it's meant for us to be able to become more productive. 284 00:15:13,622 --> 00:15:16,708 And yes, it will kill jobs and it already has, 285 00:15:16,750 --> 00:15:20,795 but the purpose of that is to allow us to not have to worry 286 00:15:20,837 --> 00:15:23,715 about those jobs and then focus on something else. 287 00:15:23,757 --> 00:15:27,677 So I suppose it's just a frame of perspective though it will, 288 00:15:27,719 --> 00:15:32,390 uh, inevitably affect a lot of people. 289 00:15:33,767 --> 00:15:37,103 The development of AI represents a crucial opportunity 290 00:15:37,145 --> 00:15:39,564 for the next technological revolution. 291 00:15:40,106 --> 00:15:42,400 Similar to the Industrial Revolution, 292 00:15:42,442 --> 00:15:45,278 AI will reshape industries and economies 293 00:15:45,320 --> 00:15:48,573 increasing both productivity and quality of life. 294 00:15:48,615 --> 00:15:51,409 But progress will not be without cost. 295 00:15:51,451 --> 00:15:53,703 Just as manufacturing was devastating for 296 00:15:53,745 --> 00:15:57,290 the skilled laborer, AI will destroy jobs. 297 00:15:57,707 --> 00:16:00,835 During this transition, we must allocate resources for 298 00:16:00,877 --> 00:16:03,546 retraining of those affected and establish 299 00:16:03,588 --> 00:16:06,800 regulations that facilitate the ethical development of AI, 300 00:16:06,841 --> 00:16:10,178 ensuring benefits across all classes of society. 301 00:16:11,721 --> 00:16:15,975 My largest concern with AI right now is the rapid 302 00:16:16,017 --> 00:16:18,770 pace with which it is being deployed. 303 00:16:18,812 --> 00:16:21,898 There are concerns that we have around the data 304 00:16:21,940 --> 00:16:26,778 that are being used to train our AI models and those concerns 305 00:16:26,820 --> 00:16:30,115 are kind of going a bit by the wayside as we push out 306 00:16:30,156 --> 00:16:34,828 the newest version of some sort of AI technology to the public. 307 00:16:34,869 --> 00:16:37,497 But I do think it's time for us to slow down 308 00:16:37,539 --> 00:16:40,375 a bit on the deployment and spend more time really 309 00:16:40,417 --> 00:16:45,463 trying to understand, uh, what is in the data that we're 310 00:16:45,505 --> 00:16:48,967 using to train these systems and what patterns 311 00:16:49,008 --> 00:16:51,553 are being recognized in that data and how those 312 00:16:51,594 --> 00:16:54,389 patterns will impact certain populations of people. 313 00:16:55,515 --> 00:16:59,686 I think AI now, uh, is-- we're gonna see a level of 314 00:16:59,728 --> 00:17:02,397 spam on the internet we've never seen before. 315 00:17:02,439 --> 00:17:07,986 And that's gonna be the first problem that regulatory bodies, 316 00:17:08,027 --> 00:17:11,531 whether it be self-regulating amongst the people creating 317 00:17:11,573 --> 00:17:15,660 the tech, adjacent companies that could ascend and these 318 00:17:15,702 --> 00:17:18,872 could be new industries that ward off the spam. 319 00:17:18,913 --> 00:17:23,752 I think in fact this is a prime use case for Web3 320 00:17:23,793 --> 00:17:27,714 and we're gonna need encrypted data to help protect people, 321 00:17:27,756 --> 00:17:30,508 their lives, their intellectual property 322 00:17:30,550 --> 00:17:35,513 because the ability to fake anything digitally is here. 323 00:17:36,723 --> 00:17:40,101 Our AI's today are trained on data 324 00:17:40,143 --> 00:17:42,312 that comes from the internet. 325 00:17:42,353 --> 00:17:45,356 And years ago, decades ago, 326 00:17:45,398 --> 00:17:48,735 that data was a bit more trusted. 327 00:17:48,777 --> 00:17:53,364 In the past decade, there's been a lot more misinformation online 328 00:17:53,406 --> 00:17:56,451 and that is creeping its way into our data sets. 329 00:17:56,493 --> 00:17:58,578 And I think that this is going to be a large 330 00:17:58,620 --> 00:18:01,706 problem moving forward where that misinformation 331 00:18:01,748 --> 00:18:04,083 becomes more and more intentional. 332 00:18:04,125 --> 00:18:05,794 Right now, misinformation's being 333 00:18:05,835 --> 00:18:08,671 spread across social media and that's a world where 334 00:18:08,713 --> 00:18:10,965 we understand that things are pretty subjective. 335 00:18:11,007 --> 00:18:13,593 We understand that it's people's opinions 336 00:18:13,635 --> 00:18:15,428 that are being bounced back and forth. 337 00:18:15,470 --> 00:18:18,556 But over time as our AI's are being trained 338 00:18:18,598 --> 00:18:21,893 on intentional misinformation, there's this belief 339 00:18:21,935 --> 00:18:24,562 that the information you're receiving from 340 00:18:24,604 --> 00:18:27,941 an artificial agent can't be opinionated. 341 00:18:27,982 --> 00:18:29,609 It comes off as fact. 342 00:18:29,651 --> 00:18:32,821 And so things that are inherently subjective that might 343 00:18:32,862 --> 00:18:36,574 actually be misinformation will be perceived as being objective 344 00:18:36,616 --> 00:18:40,328 reality of the world that the AI doesn't have an opinion 345 00:18:40,370 --> 00:18:43,414 about what's being said, it's simply regurgitating fact. 346 00:18:43,456 --> 00:18:46,876 And that I think is really, really dangerous. 347 00:18:47,919 --> 00:18:51,381 I do think that there is gonna be a problem with censorship. 348 00:18:51,422 --> 00:18:54,175 Again, it comes back to the AI ethics. 349 00:18:54,217 --> 00:18:55,927 There's nothing in place yet. 350 00:18:55,969 --> 00:18:59,722 In social media, there are certain teams for everything. 351 00:18:59,764 --> 00:19:02,725 There's a policy team, there's all the law team, 352 00:19:02,767 --> 00:19:06,688 there's the trust and safety team to keep our users 353 00:19:06,729 --> 00:19:09,607 safe in other countries, journalists, refugees. 354 00:19:09,649 --> 00:19:12,402 There's a team that's developed for all of these 355 00:19:12,443 --> 00:19:14,153 things at every company. 356 00:19:14,195 --> 00:19:17,031 And I do see that there is gonna be an issue with that since 357 00:19:17,073 --> 00:19:19,951 the biggest companies in the world like Microsoft, 358 00:19:19,993 --> 00:19:22,412 you know, fired their entire ethics team. 359 00:19:22,453 --> 00:19:25,039 So there are no rules involved right now, 360 00:19:25,081 --> 00:19:27,542 and I think we're starting to see that people 361 00:19:27,584 --> 00:19:30,545 are gonna need that just to keep us safe. 362 00:19:30,587 --> 00:19:33,172 Because I think it's gonna become like a huge HR 363 00:19:33,214 --> 00:19:35,425 problem with all of this happening. 364 00:19:35,466 --> 00:19:38,261 And so I think that there is gonna be a portion 365 00:19:38,303 --> 00:19:41,598 for censorship if you see like kind of what happened with 366 00:19:41,639 --> 00:19:44,851 the social media companies with Instagram and Twitter and 367 00:19:44,893 --> 00:19:49,898 TikTok and about censorship and about free speech or hate. 368 00:19:49,939 --> 00:19:51,774 There's a lot of that and I think 369 00:19:51,816 --> 00:19:54,402 that we're starting to see that now with AI. 370 00:19:55,737 --> 00:19:58,740 Ensuring the removal of misinformation from data sets is 371 00:19:58,781 --> 00:20:02,160 crucial, but heavily curated data sets will ultimately 372 00:20:02,201 --> 00:20:04,954 reflect old norms, and over time they will become 373 00:20:04,996 --> 00:20:07,457 a barrier to the development of both artificial 374 00:20:07,498 --> 00:20:09,375 intelligence and science itself. 375 00:20:09,918 --> 00:20:13,671 To protect the evolution of AI, flexibility, and adaptability 376 00:20:13,713 --> 00:20:16,966 must be prioritized, requiring AI to continually 377 00:20:17,008 --> 00:20:19,427 challenge itself, searching for weaknesses 378 00:20:19,469 --> 00:20:21,888 in its own processes and data sets. 379 00:20:22,889 --> 00:20:26,559 One interesting thing about the data that is used 380 00:20:26,601 --> 00:20:29,646 to train these AIs is that it's sort of like 381 00:20:29,687 --> 00:20:33,775 a snapshot of what the world is like in some period of time, 382 00:20:33,816 --> 00:20:36,986 both in that moment, but also the history in that moment. 383 00:20:37,028 --> 00:20:39,906 And one of the questions is, as society evolves, 384 00:20:39,948 --> 00:20:45,745 as things change, is our AI going to adapt to those changes? 385 00:20:45,787 --> 00:20:47,872 Is it going to learn new information, identify 386 00:20:47,914 --> 00:20:51,918 new patterns in these data sets as human behavior changes, 387 00:20:51,960 --> 00:20:54,963 as the AI itself influences society? 388 00:20:55,004 --> 00:20:57,799 Is the AI going to learn about its impact 389 00:20:57,840 --> 00:20:59,842 and how it's transforming the world? 390 00:21:02,553 --> 00:21:05,556 I think broadly, AI and generative tools are already 391 00:21:05,598 --> 00:21:08,142 changing the way that we learn and create together. 392 00:21:08,184 --> 00:21:11,354 So I use, for example, Midjourney and Stable Diffusion 393 00:21:11,396 --> 00:21:15,233 with kids and they'll come up with ideas that I would never 394 00:21:15,274 --> 00:21:17,944 come up with, but that are beautiful and interesting. 395 00:21:17,986 --> 00:21:21,072 And then they take those and they experiment with them. 396 00:21:21,114 --> 00:21:23,574 And so new science fair experiments are 397 00:21:23,616 --> 00:21:25,785 gonna happen between AI and Roblox. 398 00:21:25,827 --> 00:21:29,539 New types of whole ecosystems are going to 399 00:21:29,580 --> 00:21:32,583 emerge from fictional creatures, right? 400 00:21:32,625 --> 00:21:34,544 We are already seeing this happening. 401 00:21:34,585 --> 00:21:38,881 So as we give life of some form to all of these 402 00:21:38,923 --> 00:21:41,884 fictional worlds that we've created and we begin to relate 403 00:21:41,926 --> 00:21:45,013 to them differently, it's gonna change fandom. 404 00:21:47,306 --> 00:21:51,102 The differences between AI and human art are very minimal 405 00:21:51,144 --> 00:21:52,562 in my personal opinion. 406 00:21:52,603 --> 00:21:54,272 I've seen some really cool stuff. 407 00:21:54,313 --> 00:21:57,942 My favorites being, they've taken shows like InuYasha 408 00:21:57,984 --> 00:22:01,904 in Anime and they put it into a '80s film setting. 409 00:22:01,946 --> 00:22:06,117 And you can tell the difference because like say, for example, 410 00:22:06,159 --> 00:22:09,287 one big factor was AI couldn't master hands. 411 00:22:09,704 --> 00:22:12,915 They couldn't do it, but they're pretty much almost there. 412 00:22:12,957 --> 00:22:15,877 You see, I don't know the phenomenon 413 00:22:15,918 --> 00:22:19,589 when you look at something and, uh, 414 00:22:20,590 --> 00:22:22,300 it's the uncanny valley, is that what it is? 415 00:22:22,341 --> 00:22:25,136 You have that when you look at AI art, but it's improving so 416 00:22:25,178 --> 00:22:27,847 much to the point that I'm starting to not see that. 417 00:22:29,348 --> 00:22:35,271 If the youth begin using AI out of necessity or interest, 418 00:22:35,313 --> 00:22:39,650 they begin using AI as a friend, as a teacher, as a mentor, 419 00:22:39,692 --> 00:22:42,070 as a psychologist, as a therapist. 420 00:22:42,111 --> 00:22:46,824 Over enough time, their normal is gonna be vastly different 421 00:22:46,866 --> 00:22:50,078 than what we understand normal to be yesterday or today. 422 00:22:51,871 --> 00:22:54,624 So I wrote my first script about our relationships 423 00:22:54,665 --> 00:22:56,334 with AI in 2005. 424 00:22:56,375 --> 00:22:59,128 I started talking about that in Hollywood and no one 425 00:22:59,170 --> 00:23:03,174 understood because this was prior to Big Hero 6 or Her. 426 00:23:03,216 --> 00:23:06,177 There wasn't a story about relationships with 427 00:23:06,219 --> 00:23:08,888 AI except for Hal, right? 428 00:23:08,930 --> 00:23:11,599 So we only had a dystopian mindset. 429 00:23:12,141 --> 00:23:16,020 With Big Hero 6, you saw a potentially protopian, 430 00:23:16,062 --> 00:23:20,441 relationship between an AI and a human. 431 00:23:20,483 --> 00:23:23,194 And after that, we began to see that these 432 00:23:23,236 --> 00:23:27,657 are mutually beneficial systems if we design them to be. 433 00:23:28,157 --> 00:23:30,785 If we design them poorly, they become toxic and 434 00:23:30,827 --> 00:23:32,620 they create all sorts of negative 435 00:23:32,662 --> 00:23:34,580 ramifications in our society. 436 00:23:36,582 --> 00:23:38,876 I'm sure it has military applications. 437 00:23:38,918 --> 00:23:41,212 I mean, I've been retired for over a decade now. 438 00:23:41,254 --> 00:23:43,673 So certainly there's been a lot of changes, 439 00:23:43,714 --> 00:23:45,716 but it would be naive of me to think that it's 440 00:23:45,758 --> 00:23:47,885 not gonna have its military applications. 441 00:23:47,927 --> 00:23:50,096 And certainly, I think we need to be 442 00:23:50,138 --> 00:23:53,057 concerned about bad actors using it to do 443 00:23:53,099 --> 00:23:55,518 things that certainly aren't in our best interest. 444 00:23:57,311 --> 00:24:00,773 Well, I certainly have all kinds of opinions about AI. 445 00:24:00,815 --> 00:24:05,528 I have some worries about how it can be used, 446 00:24:05,570 --> 00:24:10,116 how it can be abused by its creators. 447 00:24:10,158 --> 00:24:14,245 I don't think I really obsess or lose sleep over those worries, 448 00:24:14,287 --> 00:24:19,208 but I feel very, very aware that artificial 449 00:24:19,250 --> 00:24:23,629 intelligence can be a powerful weapon. 450 00:24:24,964 --> 00:24:27,425 And it's really all about information. 451 00:24:28,176 --> 00:24:32,722 I mean, as the great Arthur C. Clarke said-- not just 452 00:24:32,763 --> 00:24:35,183 with creating Hal in 2001, 453 00:24:35,224 --> 00:24:38,895 but I think he said something about when it comes to AI, 454 00:24:39,645 --> 00:24:45,026 it's all about the struggle for freedom of information. 455 00:24:45,067 --> 00:24:46,777 And that's gonna ultimately come down 456 00:24:46,819 --> 00:24:49,864 to technology, not politics. 457 00:24:51,866 --> 00:24:54,452 When it comes to Big Brother and the ability to spy, 458 00:24:54,493 --> 00:24:57,955 I wouldn't be surprised if it's already being used. 459 00:24:57,997 --> 00:25:02,043 The government is typically 25 years ahead of the curve in 460 00:25:02,084 --> 00:25:05,338 technology anyway before the public is aware of it. 461 00:25:05,379 --> 00:25:08,466 So I wouldn't be surprised if it's already in use. 462 00:25:09,008 --> 00:25:11,886 The reality though is there's a lot of potential for 463 00:25:11,928 --> 00:25:15,473 it to spy, to monitor, and to even do a lot of the jobs of CIA 464 00:25:15,514 --> 00:25:21,229 and other agencies do because its ability to adapt, to learn. 465 00:25:21,270 --> 00:25:26,776 And you think about in the past, let's say we wanted to 466 00:25:27,526 --> 00:25:30,780 overthrow a government or disrupt the political climate, 467 00:25:30,821 --> 00:25:34,617 we had to put somebody there physically in harm's way, 468 00:25:34,659 --> 00:25:40,122 and AI could as easily get into their electronic systems 469 00:25:40,164 --> 00:25:43,125 and manipulate the media and manipulate things that are 470 00:25:43,167 --> 00:25:46,087 out there that would cause just as much instability. 471 00:25:48,047 --> 00:25:54,470 I fear people using AI in a bad way 472 00:25:54,512 --> 00:25:56,514 more than I fear AI. 473 00:25:56,555 --> 00:26:00,184 But that doesn't mean the fear of AI isn't significant. 474 00:26:01,310 --> 00:26:05,189 I just feel humans will do bad things with AI before 475 00:26:05,231 --> 00:26:08,234 AI gets to the point where it can do bad things. 476 00:26:10,945 --> 00:26:13,781 Humans naturally fear the unknown until 477 00:26:13,823 --> 00:26:15,700 they can understand whether it is 478 00:26:15,741 --> 00:26:17,535 a threat in their ecosystem or not. 479 00:26:17,576 --> 00:26:21,539 And if you combine that with the YouTube-ification of media, 480 00:26:21,580 --> 00:26:25,501 if you're thinking about how all of us are sort of playing to 481 00:26:25,543 --> 00:26:28,921 an algorithm of attention and fear gets attention. 482 00:26:28,963 --> 00:26:31,590 So there's a lot more media about the fear 483 00:26:31,632 --> 00:26:34,427 side of this conversation right now, 484 00:26:34,468 --> 00:26:38,514 in part because it sells, it gets a lot more views, right? 485 00:26:38,556 --> 00:26:40,808 So clickbait, right? 486 00:26:40,850 --> 00:26:43,060 It's all over the map right now. 487 00:26:43,102 --> 00:26:47,231 If you're looking at what, for example, GPT-4 does, 488 00:26:47,273 --> 00:26:51,027 many people will say it does a form of hallucination, 489 00:26:51,068 --> 00:26:54,447 and those hallucinations can become mass 490 00:26:54,488 --> 00:26:58,200 disinformation campaigns. That is a real threat. 491 00:26:58,242 --> 00:27:01,537 But then there are all of these other human hallucinations, 492 00:27:01,579 --> 00:27:04,874 the sort of things like the Hal fears or the matrix fears, 493 00:27:04,915 --> 00:27:07,877 we're just gonna be farmed at some point, right? 494 00:27:07,918 --> 00:27:10,671 Those are all normal because that's the only thing 495 00:27:10,713 --> 00:27:12,923 we've seen about AI in our media. 496 00:27:12,965 --> 00:27:16,761 We got 50 years of movies about how AI were bad. 497 00:27:16,802 --> 00:27:18,763 So, of course, we think AI are bad. 498 00:27:18,804 --> 00:27:21,515 We didn't have a good relationship with 499 00:27:21,557 --> 00:27:22,933 AI in our past, right? 500 00:27:22,975 --> 00:27:25,853 Our past is a little toxic. 501 00:27:25,895 --> 00:27:27,813 So we have to learn something different and 502 00:27:27,855 --> 00:27:29,982 we have to find a better model. 503 00:27:30,024 --> 00:27:32,610 And in some ways, we have to train humans 504 00:27:32,651 --> 00:27:34,445 to be the better model, right? 505 00:27:34,487 --> 00:27:37,281 To not give into the fear just because it's 506 00:27:37,323 --> 00:27:39,408 easy and it gets clicks. 507 00:27:41,285 --> 00:27:45,122 I think it's gonna cause a wealth gap in between 508 00:27:45,164 --> 00:27:48,042 the high-income and the low-income. 509 00:27:48,084 --> 00:27:50,795 And instantly, let me tell you why. 510 00:27:50,836 --> 00:27:56,133 I was at Denny's in Las Vegas and I was served by a robot. 511 00:27:57,176 --> 00:27:59,595 It was cute. I was like, "Wow." 512 00:27:59,637 --> 00:28:03,724 It was a first for me. Um, and-- and it made me think. 513 00:28:03,766 --> 00:28:05,476 It was actually the topic of my whole 514 00:28:05,518 --> 00:28:08,312 conversation that this robot was serving me. 515 00:28:08,354 --> 00:28:10,689 I couldn't take my eye off of it. 516 00:28:10,731 --> 00:28:13,109 And at the end of the day, 517 00:28:13,150 --> 00:28:16,654 I missed the interaction with the waitress. 518 00:28:16,695 --> 00:28:18,280 That's how I tip. 519 00:28:18,322 --> 00:28:20,741 You send me a robot and then ask me if I'm going to 520 00:28:20,783 --> 00:28:23,619 pay 3%, 5%, or 10%. 521 00:28:23,661 --> 00:28:27,790 You trying to save money, well, I'm not going to tip nothing. 522 00:28:28,707 --> 00:28:30,459 And now it's an AI race war. 523 00:28:30,501 --> 00:28:33,546 So all of us true technical people we're like, holy crap. 524 00:28:33,587 --> 00:28:37,007 The reason why Microsoft fired 12,000 people, 525 00:28:37,049 --> 00:28:39,385 why did Google fire 11,000 people? 526 00:28:39,427 --> 00:28:42,513 It's because they needed to reinvest that money into AI. 527 00:28:42,555 --> 00:28:46,642 So now we're starting to see greed happen in these companies 528 00:28:46,684 --> 00:28:50,354 instead of people that actually are experts in this. 529 00:28:50,396 --> 00:28:52,898 And it's really sad to see like the mental 530 00:28:52,940 --> 00:28:54,692 health because of all these layoffs. 531 00:28:54,733 --> 00:28:56,193 And as a technical recruiter, 532 00:28:56,235 --> 00:28:59,238 I have over 200 techs, and I would say 80% 533 00:28:59,280 --> 00:29:01,782 of that are people that have lost their jobs. 534 00:29:03,409 --> 00:29:07,580 So an idea of teaching AI to be a certain way and the decisions 535 00:29:07,621 --> 00:29:10,499 that we make day to day influencing it, I think of 536 00:29:10,541 --> 00:29:13,377 the idea that it takes a village to raise a child. 537 00:29:13,419 --> 00:29:16,964 When you think about just a human and the environment 538 00:29:17,006 --> 00:29:19,550 that they're raised in, the choices by the people around 539 00:29:19,592 --> 00:29:23,721 them influences heavily day to day on who they become. 540 00:29:24,263 --> 00:29:26,056 Of course, there's our own sentience 541 00:29:26,098 --> 00:29:29,393 and ability to change those decisions, but I think 542 00:29:29,435 --> 00:29:32,938 that's important to keep in mind as we continue forward. 543 00:29:32,980 --> 00:29:36,150 That's why we have parameters for Chat GPT. 544 00:29:36,192 --> 00:29:38,194 I'm not sure how that pertains to stuff 545 00:29:38,235 --> 00:29:40,488 like Siri or Google, stuff like that, 546 00:29:40,529 --> 00:29:44,074 but I think it's important to just keep that in mind. 547 00:29:46,827 --> 00:29:49,497 Many folks fear that at some point, 548 00:29:49,538 --> 00:29:55,085 many of these AI will become something more and 549 00:29:55,127 --> 00:29:57,588 go what we would consider off the rails, right? 550 00:29:57,630 --> 00:29:59,590 No longer be controllable. 551 00:29:59,632 --> 00:30:02,092 There are some examples of that already happening, 552 00:30:02,134 --> 00:30:06,055 but not in a way that they couldn't be shut down. 553 00:30:06,096 --> 00:30:09,433 There might be a time soon where an AI can then 554 00:30:09,475 --> 00:30:12,770 run off into the wild and not be shut down, right? 555 00:30:12,811 --> 00:30:17,149 I believe someone at Open AI, it's their job to make sure 556 00:30:17,191 --> 00:30:20,110 that someone can pull a plug if they need to. 557 00:30:20,152 --> 00:30:23,239 And this is a whole new type of job that we didn't 558 00:30:23,280 --> 00:30:26,075 really have in technology up to this point, 559 00:30:26,116 --> 00:30:28,452 a kill switch for our technology. 560 00:30:28,494 --> 00:30:31,163 We're not tracking where these tools are 561 00:30:31,205 --> 00:30:32,706 going off the rails right now. 562 00:30:32,748 --> 00:30:34,792 We don't have a Yelp that tells me which AI 563 00:30:34,833 --> 00:30:36,794 are good and which AI are not. 564 00:30:36,835 --> 00:30:40,756 So until humans learn to communicate effectively 565 00:30:40,798 --> 00:30:44,426 amongst ourselves even, we're not gonna have a really 566 00:30:44,468 --> 00:30:47,096 great handle on when they're going off the rails. 567 00:30:47,137 --> 00:30:49,181 And we're definitely not gonna be communicating that to 568 00:30:49,223 --> 00:30:51,725 each other because governments get in the way, 569 00:30:51,767 --> 00:30:56,146 nationalism gets in the way, and certain countries don't want 570 00:30:56,188 --> 00:30:58,899 other countries knowing the state of their technology. 571 00:31:00,359 --> 00:31:03,654 Artificial intelligence has become the arms race 572 00:31:03,696 --> 00:31:05,447 of the 21st century. 573 00:31:05,489 --> 00:31:09,618 In 2023, we saw the US and China on the brink of war, 574 00:31:09,660 --> 00:31:13,539 not for Taiwan, but Taiwan's semiconductor. 575 00:31:14,790 --> 00:31:17,835 It is no longer a question of whether AI will shape the future 576 00:31:17,876 --> 00:31:21,171 of human civilization, but a question of who will 577 00:31:21,213 --> 00:31:23,757 control the technology that controls the future. 578 00:31:24,842 --> 00:31:28,429 I do believe that there are concerns that need to be 579 00:31:28,470 --> 00:31:32,057 monitored and addressed because what are we using the AI for? 580 00:31:32,641 --> 00:31:34,935 We all know that if the government uses 581 00:31:34,977 --> 00:31:36,687 any technology for their purposes, 582 00:31:36,729 --> 00:31:39,732 we typically regret it once we find out they're doing that. 583 00:31:39,773 --> 00:31:43,193 And are we using this for benevolent or malevolent 584 00:31:43,235 --> 00:31:47,364 means is very important because AI does have the ability 585 00:31:47,406 --> 00:31:49,283 to destroy humanity if we let it. 586 00:31:50,743 --> 00:31:54,038 We should been worrying about Big Brother for decades. 587 00:31:55,748 --> 00:31:57,374 I get this often. 588 00:31:57,416 --> 00:31:59,043 People are like, "Oh, you work for the government, 589 00:31:59,084 --> 00:32:00,753 you must trust the government." 590 00:32:00,794 --> 00:32:02,671 No, I don't trust the government at all. 591 00:32:02,713 --> 00:32:05,549 My trust for them diminishes on literally 592 00:32:05,591 --> 00:32:07,551 an hourly basis and no. 593 00:32:07,593 --> 00:32:12,097 So privacy's been out the window before AI. 594 00:32:12,139 --> 00:32:15,517 Your phone, you know, we're so tied to these 595 00:32:15,559 --> 00:32:18,854 devices that we're almost forced to give up our privacy 596 00:32:18,896 --> 00:32:23,067 to be able to function conveniently in daily life. 597 00:32:23,108 --> 00:32:24,818 So yeah, Big Brother's huge. 598 00:32:25,361 --> 00:32:28,238 AI right now has a lot of potential to 599 00:32:28,280 --> 00:32:30,658 impact people through personalized assistance. 600 00:32:30,699 --> 00:32:32,576 This could be someone who's fully able-bodied. 601 00:32:32,618 --> 00:32:35,037 This could also be someone who has special 602 00:32:35,079 --> 00:32:36,997 needs and requires additional assistance 603 00:32:37,039 --> 00:32:38,999 for activities of their daily life. 604 00:32:39,041 --> 00:32:41,377 There are some challenges or potential 605 00:32:41,418 --> 00:32:43,712 perils associated with these technologies, though. 606 00:32:43,754 --> 00:32:48,092 The same technology that might be used to guide a child with 607 00:32:48,133 --> 00:32:52,346 autism through a social skills intervention could also be used 608 00:32:52,388 --> 00:32:54,973 to influence someone to make certain purchasing 609 00:32:55,015 --> 00:32:57,476 decisions that might not be in their best interest. 610 00:32:57,518 --> 00:32:59,687 So if someone says they're hungry, 611 00:32:59,728 --> 00:33:02,731 your AI might just as easily guide you down a path of 612 00:33:02,773 --> 00:33:06,777 a particular brand of pizza, simply because that was what was 613 00:33:06,819 --> 00:33:10,489 in the best interest of the AI at that time. 614 00:33:13,409 --> 00:33:16,120 We have less and less human contact. 615 00:33:16,161 --> 00:33:18,372 You go to the doctor, you don't check in, 616 00:33:18,414 --> 00:33:21,500 you go to a kiosk or you check in on your phone. 617 00:33:21,542 --> 00:33:23,669 You don't even have to leave your house. 618 00:33:23,711 --> 00:33:25,963 Someone will bring McDonald's to you. 619 00:33:26,004 --> 00:33:27,673 You can do your shopping online. 620 00:33:27,715 --> 00:33:29,800 There's less and less human contact. 621 00:33:29,842 --> 00:33:32,553 So I think it would be very unfortunate. 622 00:33:32,594 --> 00:33:35,180 And goes even further, you know, we got a robot 623 00:33:35,222 --> 00:33:36,724 that'll vacuum your house. 624 00:33:36,765 --> 00:33:38,642 So I worry about that. 625 00:33:41,437 --> 00:33:44,398 I think we're starting to see this mass fall off of all 626 00:33:44,440 --> 00:33:47,234 these companies like the Googles, the Microsoft's, 627 00:33:47,276 --> 00:33:50,404 you people that were there for up to 20 years, 628 00:33:50,446 --> 00:33:53,198 and then they just got let go because these people were making 629 00:33:53,240 --> 00:33:55,325 500 to a million dollars and were like, 630 00:33:55,367 --> 00:33:58,245 they basically said, "Thank you for your service. 631 00:33:58,287 --> 00:34:00,038 Basically, screw you. 632 00:34:00,080 --> 00:34:01,582 We're letting you go. 633 00:34:01,623 --> 00:34:03,459 Here's a very small severance package for 634 00:34:03,500 --> 00:34:05,335 the time that you were here. 635 00:34:05,377 --> 00:34:08,005 But we care more about profits and we care more about 636 00:34:08,046 --> 00:34:10,257 making sure that we scale this technology 637 00:34:10,299 --> 00:34:12,050 as fast as we can." 638 00:34:14,553 --> 00:34:18,682 I think there is the risk that we're already 639 00:34:18,724 --> 00:34:22,686 beginning to face in human development of how 640 00:34:22,728 --> 00:34:26,690 relying more and more on artificial intelligence. 641 00:34:26,732 --> 00:34:29,693 What does it take away from our own humanity, 642 00:34:29,735 --> 00:34:34,698 our own very human, very biological need for 643 00:34:34,740 --> 00:34:37,367 social interaction with our own kind? 644 00:34:37,743 --> 00:34:41,079 I think that we need to always remember 645 00:34:41,121 --> 00:34:44,208 that artificial intelligence is a tool. 646 00:34:45,751 --> 00:34:50,214 And once we go forward with developing artificial 647 00:34:50,255 --> 00:34:57,221 intelligence as having its own ability to pass the mirror test, 648 00:34:57,262 --> 00:35:03,852 so to speak, and recognize its own being, 649 00:35:03,894 --> 00:35:05,854 where do we go from there? 650 00:35:08,774 --> 00:35:11,360 So what does self-consciousness mean? 651 00:35:11,401 --> 00:35:12,861 It's a very complicated question. 652 00:35:12,903 --> 00:35:15,072 The whole Terminator series is based on this 653 00:35:15,113 --> 00:35:18,992 idea of self-aware, but really what does that mean? 654 00:35:19,034 --> 00:35:22,037 And I don't know that there's one clear answer. 655 00:35:22,079 --> 00:35:25,916 Intelligence by definition is the ability for an organism 656 00:35:25,958 --> 00:35:28,418 to adapt and survive in its environment. 657 00:35:28,460 --> 00:35:31,630 And this would then expand it to how is 658 00:35:31,672 --> 00:35:34,925 it adapting and surviving in the environment that it's in? 659 00:35:35,384 --> 00:35:38,262 And what does it really mean to be self-aware? 660 00:35:38,303 --> 00:35:40,597 How do we know we're self-aware? 661 00:35:40,639 --> 00:35:43,600 My perception of reality is different than someone else's 662 00:35:43,642 --> 00:35:46,979 perception of reality even if we're in the same room. 663 00:35:48,146 --> 00:35:52,484 There have been a number of different definitions of what 664 00:35:52,526 --> 00:35:56,780 it means to be fully intelligent or sentient for an AI. 665 00:35:56,822 --> 00:35:59,616 The Turing test is not really looking at all of 666 00:35:59,658 --> 00:36:01,702 the dynamics of intelligence. 667 00:36:01,743 --> 00:36:04,496 And so some of my colleagues recently have formed 668 00:36:04,538 --> 00:36:07,749 a sort of eightfold path of intelligence where we can 669 00:36:07,791 --> 00:36:11,211 begin to assess, for example, GPT-4. 670 00:36:11,253 --> 00:36:15,048 Is it an artificial narrow intelligence of some sort? 671 00:36:15,090 --> 00:36:18,135 Yes. Can it write well? Yes. 672 00:36:18,176 --> 00:36:20,262 Does it understand context? 673 00:36:20,304 --> 00:36:21,638 Hmm, maybe not. 674 00:36:21,680 --> 00:36:23,682 Does it have a body or an embodiment? 675 00:36:23,724 --> 00:36:24,975 No. 676 00:36:25,017 --> 00:36:28,103 So it will be intelligent in some form, 677 00:36:28,145 --> 00:36:29,938 not intelligent in others. 678 00:36:29,980 --> 00:36:31,648 What we're realizing is that our idea 679 00:36:31,690 --> 00:36:34,484 of intelligence was limited, right? 680 00:36:34,526 --> 00:36:37,279 We didn't understand consciousness. We still don't. 681 00:36:37,321 --> 00:36:40,532 So we have no idea how to train an AI to be 682 00:36:40,574 --> 00:36:42,367 loving and compassionate. 683 00:36:42,409 --> 00:36:44,661 We don't understand that in ourselves yet. 684 00:36:44,703 --> 00:36:47,497 So there are aspects of humanness that we haven't 685 00:36:47,539 --> 00:36:49,917 figured out how to program effectively, 686 00:36:49,958 --> 00:36:53,629 and that's where some people are in the industry quite afraid. 687 00:36:53,670 --> 00:36:57,090 We haven't figured out how to make AI that cares enough 688 00:36:57,132 --> 00:37:01,178 about humans that they might not wanna eradicate us, right? 689 00:37:01,219 --> 00:37:03,263 That's an AI safety concern. 690 00:37:03,305 --> 00:37:06,808 And many leaders in the industry have said, "We need to stop. 691 00:37:06,850 --> 00:37:10,646 We need to push pause because we're afraid of what's going to 692 00:37:10,687 --> 00:37:15,567 happen if we outpace ourselves in terms of understanding 693 00:37:15,609 --> 00:37:19,655 humanness, consciousness and care for each other." 694 00:37:25,577 --> 00:37:28,038 Now, those companies are asking the big tech 695 00:37:28,080 --> 00:37:30,040 companies like, "Hey, you guys gotta put a stop." 696 00:37:30,082 --> 00:37:32,542 But the problem is that there's no ethics 697 00:37:32,584 --> 00:37:34,169 that are being involved. 698 00:37:34,211 --> 00:37:36,672 And when you look at Microsoft, it was 12,000 people 699 00:37:36,713 --> 00:37:39,675 that they let go in February and they let go 700 00:37:39,716 --> 00:37:43,261 of 150 people that was all on the AI ethics team. 701 00:37:43,303 --> 00:37:46,473 And the whole point of them is to literally review 702 00:37:46,515 --> 00:37:49,309 what's going on and put rules and standards 703 00:37:49,351 --> 00:37:51,853 for what is going on in this AI world. 704 00:37:51,895 --> 00:37:54,982 And instead what Microsoft did, and I have to say, 705 00:37:55,023 --> 00:37:57,526 out of all the tech companies, the big giant ones, 706 00:37:57,567 --> 00:38:00,487 I have always thought that Microsoft is a very 707 00:38:00,529 --> 00:38:04,074 safe and stable company and wanting the best for people. 708 00:38:04,116 --> 00:38:06,660 But now that we're even seeing that they've cut 709 00:38:06,702 --> 00:38:10,622 their entire ethics team, that's what really scares me. 710 00:38:12,457 --> 00:38:15,711 So when I was attending Singularity University, 711 00:38:15,752 --> 00:38:20,173 our graduate program would debate when sentience would 712 00:38:20,215 --> 00:38:23,093 happen, not just from the Turing Test, but also 713 00:38:23,135 --> 00:38:27,806 broadly when would an AI be recognized as a being in court? 714 00:38:27,848 --> 00:38:29,141 For example. 715 00:38:29,182 --> 00:38:32,102 We are already seeing AI recognized in court. 716 00:38:32,144 --> 00:38:35,397 We are also seeing AI-running companies. 717 00:38:35,981 --> 00:38:38,567 Companies have personhood in America. 718 00:38:38,608 --> 00:38:41,653 So in some ways an AI can already have 719 00:38:41,695 --> 00:38:44,990 personhood that is different from sentience. 720 00:38:45,032 --> 00:38:47,451 And if you're thinking about all of the things 721 00:38:47,492 --> 00:38:50,078 that make up a human will an AI do it? 722 00:38:50,120 --> 00:38:51,246 No. 723 00:38:51,288 --> 00:38:52,914 An AI will be different. 724 00:38:52,956 --> 00:38:56,585 An AGI, an Artificial General Intelligence is going to 725 00:38:56,626 --> 00:38:58,336 look different than a human. 726 00:38:58,378 --> 00:39:01,256 It's gonna have different sensors, different capacities, 727 00:39:01,298 --> 00:39:04,968 probably won't have skin the way we have skin, right? 728 00:39:05,010 --> 00:39:07,971 But does that change the nature of its intelligence? 729 00:39:08,013 --> 00:39:09,639 Yes. 730 00:39:09,681 --> 00:39:12,893 Is it sentient the way nature is sentient? Probably. 731 00:39:12,934 --> 00:39:16,438 It's probably as sentient as a mushroom already in some way. 732 00:39:19,775 --> 00:39:21,902 If I were to bet, I would bet against 733 00:39:21,943 --> 00:39:25,614 artificial intelligence ever becoming self-aware like humans. 734 00:39:25,655 --> 00:39:29,034 And I would suggest odds of 80% against. 735 00:39:31,036 --> 00:39:35,040 I do feel AI has the potential to 736 00:39:35,082 --> 00:39:39,503 become self-conscious at some point to be that Skynet. 737 00:39:40,212 --> 00:39:43,131 I'm a history major and I watch movies, 738 00:39:43,173 --> 00:39:45,759 and I've been watching movies since I was a little kid. 739 00:39:45,801 --> 00:39:49,805 And everything that was science fiction, 740 00:39:49,846 --> 00:39:51,515 I don't wanna say everything. 741 00:39:51,556 --> 00:39:53,600 A lot of things that were pure science fiction 742 00:39:53,642 --> 00:39:55,268 are now science fact. 743 00:39:55,310 --> 00:39:57,521 Well, a lot of that Buck Roger stuff has come true. 744 00:39:57,562 --> 00:39:59,981 My father-in-law was a communications guy in 745 00:40:00,023 --> 00:40:02,734 the Navy and they were doing facsimile stuff in 746 00:40:02,776 --> 00:40:07,155 the 70's and late '60s, but it was just only a select 747 00:40:07,197 --> 00:40:09,491 few who could have access to that. 748 00:40:09,533 --> 00:40:11,576 Now everybody could have access to it. 749 00:40:12,119 --> 00:40:14,329 Yeah, I think the real danger with 750 00:40:14,371 --> 00:40:18,291 AI isn't consciousness leading to terminated robots. 751 00:40:18,333 --> 00:40:23,088 First, because a lot of connected networks have to 752 00:40:23,130 --> 00:40:26,508 occur like this AI would have to find itself getting into 753 00:40:26,550 --> 00:40:30,137 the likes of dynamics robots or something of that nature. 754 00:40:30,178 --> 00:40:33,056 And then it would have to find itself getting into some sort of 755 00:40:33,098 --> 00:40:37,352 munitions manufacturing and have mastering control 756 00:40:37,394 --> 00:40:40,397 of all those scenarios and be able to essentially terraform 757 00:40:40,438 --> 00:40:44,568 itself into the real world and not just the world of bits. 758 00:40:45,986 --> 00:40:49,239 In at least the short in the medium term, it's certainly 759 00:40:49,281 --> 00:40:54,244 going to be humans using AI against other humans. 760 00:40:54,286 --> 00:40:57,164 In a lot of ways, this is just an augmentation technology. 761 00:40:57,205 --> 00:41:02,169 So, people that are in a frame of acquiring 762 00:41:02,210 --> 00:41:07,174 resources or prestige or power under seemingly 763 00:41:07,215 --> 00:41:10,594 any circumstances, they're just gonna use this 764 00:41:10,635 --> 00:41:13,138 as a way to get to that goal faster. 765 00:41:16,808 --> 00:41:19,352 We all seen Terminator and Skynet and all that. 766 00:41:19,394 --> 00:41:23,148 But no, it's frightening to think that something that is 767 00:41:23,190 --> 00:41:25,442 smarter than you with the access of the internet, 768 00:41:25,483 --> 00:41:28,945 and that could mobilize into a physical form with help, 769 00:41:28,987 --> 00:41:32,699 of course, but still is possible to connect to a robot body. 770 00:41:32,741 --> 00:41:36,286 I mean, you see the dogs with no heads from MIT. 771 00:41:36,328 --> 00:41:39,372 I mean, I-- I don't know why anybody would want that, 772 00:41:39,414 --> 00:41:42,125 but, you know, I can see the practical use for 773 00:41:42,167 --> 00:41:45,420 it to help, you know, maybe as a maid or to build stuff and 774 00:41:45,462 --> 00:41:48,965 for human danger to not be like space traveling and all that. 775 00:41:49,007 --> 00:41:51,760 But as far as we are all human beings, 776 00:41:51,801 --> 00:41:54,137 you know, even if you are a racist person, 777 00:41:54,179 --> 00:41:56,097 you still don't want society to end. 778 00:41:56,139 --> 00:41:58,892 You don't want it to just be one guy in the bubble at the top 779 00:41:58,934 --> 00:42:01,645 of a tower and lots of worker bots, I would hope. 780 00:42:01,686 --> 00:42:03,230 You know, most people would, 781 00:42:03,271 --> 00:42:05,565 sanely wouldn't want creation to end that way. 782 00:42:08,443 --> 00:42:13,823 So when it comes to AI becoming sentient, it's quite possible. 783 00:42:13,865 --> 00:42:17,535 And I'm actually hopeful that it will not fall into 784 00:42:17,577 --> 00:42:22,165 the pitfalls of humanity and human emotion where rather than 785 00:42:22,207 --> 00:42:26,002 just initially being fearful and destructive, 786 00:42:26,044 --> 00:42:29,256 that may have more awareness than we do as people. 787 00:42:29,297 --> 00:42:32,342 That there's a better way to do things. 788 00:42:32,384 --> 00:42:36,179 And that's where that relationship between AI and 789 00:42:36,221 --> 00:42:39,432 humans could be very beneficial because at some point 790 00:42:39,474 --> 00:42:43,520 it is likely to teach us the things that we need to do so 791 00:42:43,561 --> 00:42:47,524 that we can actually survive into the future longer 792 00:42:47,565 --> 00:42:52,612 than we probably sh-- will based on the way humanity is now. 793 00:42:54,990 --> 00:42:59,286 We've seen a massive convergence of global data, 794 00:42:59,327 --> 00:43:04,207 for example, about climate and how climate affects humans. 795 00:43:04,249 --> 00:43:07,794 We've seen how we need to respond to 796 00:43:07,836 --> 00:43:10,922 rapid disasters, but we haven't necessarily 797 00:43:10,964 --> 00:43:14,718 applied AI effectively to our biggest problems. 798 00:43:14,759 --> 00:43:18,513 Climate, migration, displacement, poverty. 799 00:43:18,930 --> 00:43:24,227 We often think of AI as making inequalities worse 800 00:43:24,269 --> 00:43:28,273 instead of improving the quality of life for everyone. 801 00:43:29,190 --> 00:43:32,152 Will AI improve the quality of life for everyone? 802 00:43:32,193 --> 00:43:33,570 It could. 803 00:43:33,611 --> 00:43:35,947 But we have to choose to program it that way and 804 00:43:35,989 --> 00:43:38,116 implement it that way. 805 00:43:38,158 --> 00:43:42,287 So that means putting sensors on the ground in our farms 806 00:43:42,329 --> 00:43:44,539 so that if our farmers are having a drought year, 807 00:43:44,581 --> 00:43:48,043 we can support them before they lose their crop, right? 808 00:43:48,084 --> 00:43:51,629 Or in a climate disaster, having logistics ready 809 00:43:51,671 --> 00:43:53,798 to move faster so that people don't 810 00:43:53,840 --> 00:43:56,301 starve after the typhoon of a hurricane. 811 00:43:57,135 --> 00:43:59,512 That is not fully being implemented yet. 812 00:43:59,554 --> 00:44:00,972 It's just starting. 813 00:44:01,014 --> 00:44:03,141 And I think we're starting to also see 814 00:44:03,183 --> 00:44:06,102 this rapid collaboration around the data. 815 00:44:06,144 --> 00:44:09,105 And that's certainly in climate action, but it's also in 816 00:44:09,147 --> 00:44:12,734 a handful of other places, logistics and shipping. 817 00:44:12,776 --> 00:44:15,695 Places where global collaboration 818 00:44:15,737 --> 00:44:17,822 is a life or death situation. 819 00:44:21,368 --> 00:44:25,455 With AI, technology is changing rapidly and it's going 820 00:44:25,497 --> 00:44:29,376 to be up to all of us to try and keep up with those changes. 821 00:44:29,417 --> 00:44:31,669 This has been true through all of humanity. 822 00:44:31,711 --> 00:44:35,423 Some new technology comes along and people have to adapt. 823 00:44:35,465 --> 00:44:38,343 I think it's important for technology developers 824 00:44:38,385 --> 00:44:40,303 to work with the people who will be 825 00:44:40,345 --> 00:44:43,098 using that technology and impacted by that technology so 826 00:44:43,139 --> 00:44:46,142 that it doesn't leave them behind. 827 00:44:46,184 --> 00:44:50,146 This is supposed to be a tool, it's supposed to be a paintbrush 828 00:44:50,188 --> 00:44:53,024 to put in the hand of the practitioners to enable them 829 00:44:53,066 --> 00:44:55,026 to do things they've never been able to do before. 830 00:44:56,736 --> 00:45:00,156 So I help lead a group called Open Metaverse Interoperability. 831 00:45:00,198 --> 00:45:02,367 And I've worked on Metaverse technologies 832 00:45:02,409 --> 00:45:04,035 for the last 20 years. 833 00:45:04,077 --> 00:45:07,122 I see the Metaverse as the future of the 3D web in 834 00:45:07,163 --> 00:45:10,583 that connected space that we all share and smart glasses and 835 00:45:10,625 --> 00:45:14,838 smart phones and VR headsets and lots of display tech like 836 00:45:14,879 --> 00:45:19,384 holographic displays are all a part of that future 3D web. 837 00:45:20,051 --> 00:45:22,846 I already see kids creating in that space, right? 838 00:45:22,887 --> 00:45:25,473 Roblox is an early version of that. 839 00:45:25,515 --> 00:45:29,853 So we know that the next generation of kids are already 840 00:45:29,894 --> 00:45:32,564 designing in 3D, building 3D worlds. 841 00:45:32,605 --> 00:45:35,817 They are not limited to the 2D interface. 842 00:45:35,859 --> 00:45:38,153 I think those of us who are older have 843 00:45:38,194 --> 00:45:40,363 a harder time leaping from 2D to 3D. 844 00:45:40,405 --> 00:45:43,032 Those who are young have no problem just 845 00:45:43,074 --> 00:45:45,160 jumping straight into the 3D world. 846 00:45:45,201 --> 00:45:47,287 Will the glasses support them being 847 00:45:47,328 --> 00:45:49,622 there while they're pumping gas or doing anything? 848 00:45:49,664 --> 00:45:51,416 Yes, it's gonna get there. 849 00:45:51,458 --> 00:45:53,293 I'm already sharing Metaverse worlds with 850 00:45:53,334 --> 00:45:56,546 my friends on my phone around the table at the bar. 851 00:45:56,588 --> 00:45:58,131 So it's already there. 852 00:45:58,173 --> 00:45:59,757 The tools are there. 853 00:45:59,799 --> 00:46:01,301 They are not evenly distributed, 854 00:46:01,342 --> 00:46:03,761 but they're all open-source and easy to grab. 855 00:46:05,013 --> 00:46:07,223 My company, Semio, is building an operating 856 00:46:07,265 --> 00:46:09,100 system for personal robots. 857 00:46:09,142 --> 00:46:10,768 And the way you can think about that is, uh, 858 00:46:10,810 --> 00:46:13,688 in the way that people use a personal computer through 859 00:46:13,730 --> 00:46:16,858 a graphical user interface, a-- a keyboard and mouse, uh, 860 00:46:16,900 --> 00:46:19,194 the way that people use a smartphone through 861 00:46:19,235 --> 00:46:23,239 a touch screen, they'll be using natural language as 862 00:46:23,281 --> 00:46:27,035 the core input output user interface for robots. 863 00:46:27,076 --> 00:46:29,496 Uh, but it's gonna be more than just understanding 864 00:46:29,537 --> 00:46:31,080 what people are saying. 865 00:46:31,122 --> 00:46:32,665 It's gonna be how they're saying it through 866 00:46:32,707 --> 00:46:34,584 the tone of voice, through their body language. 867 00:46:34,626 --> 00:46:37,128 The robot has to understand that from the user, 868 00:46:37,170 --> 00:46:39,172 make decisions about that information. 869 00:46:39,214 --> 00:46:41,549 And it needs to be able to respond with its own speech or 870 00:46:41,591 --> 00:46:44,427 other physical actions, be it body language or 871 00:46:44,469 --> 00:46:47,055 manipulating objects in the environment. 872 00:46:47,847 --> 00:46:49,766 We begin to have a real relationship 873 00:46:49,807 --> 00:46:53,228 with the roles we love and also with the people we love. 874 00:46:53,269 --> 00:46:57,565 We can create a model of someone who passed away, right? 875 00:46:57,607 --> 00:47:00,985 Have a relationship that might feel ephemeral, 876 00:47:01,027 --> 00:47:03,404 but feels very real to us too, right? 877 00:47:03,446 --> 00:47:05,532 This is something that I do in my own art. 878 00:47:05,573 --> 00:47:07,867 Where I help people see, for example, 879 00:47:07,909 --> 00:47:11,079 I lost a child, so I created what my child 880 00:47:11,120 --> 00:47:14,040 would've looked like now as a 25-year-old, right? 881 00:47:14,082 --> 00:47:16,918 I've had many girls come to me, many women saying, 882 00:47:16,960 --> 00:47:18,670 "Could I see my child too?" 883 00:47:18,711 --> 00:47:21,839 Or "I would've liked to have seen where my son 884 00:47:21,881 --> 00:47:23,132 will be when he grows up. 885 00:47:23,174 --> 00:47:24,676 I know I'm gonna pass away soon, 886 00:47:24,717 --> 00:47:26,261 but I would love to see my grandchildren." 887 00:47:26,678 --> 00:47:31,266 So using these as tool for vision is an extraordinary 888 00:47:31,307 --> 00:47:33,351 way to heal something in the heart. 889 00:47:33,393 --> 00:47:35,520 It can improve our relationships. 890 00:47:36,854 --> 00:47:39,023 It's important right now, I think, 891 00:47:39,065 --> 00:47:42,151 to really learn as much as you can about AI. 892 00:47:42,193 --> 00:47:44,654 I don't think that we should be scared of it. 893 00:47:44,696 --> 00:47:48,575 We know what's happening, but it is gonna get in the hands 894 00:47:48,616 --> 00:47:50,034 of some bad people. 895 00:47:50,076 --> 00:47:51,452 Like you look at like drugs, 896 00:47:51,494 --> 00:47:53,413 you look at like cocaine, ecstasy. 897 00:47:53,454 --> 00:47:56,457 And when you look at the actual, like, what they did with it from 898 00:47:56,499 --> 00:47:59,377 the beginning, it was for a medical purpose and then 899 00:47:59,419 --> 00:48:02,589 it got into the wrong hands and then they do whatever with it. 900 00:48:02,630 --> 00:48:04,340 So it's kind of like that. 901 00:48:04,382 --> 00:48:06,175 Like, it's gonna get into the right hands where people 902 00:48:06,217 --> 00:48:08,636 are gonna be able to evolve better in their business 903 00:48:08,678 --> 00:48:10,930 and in the world, but then it's gonna 904 00:48:10,972 --> 00:48:12,515 get into some of the wrong hands. 905 00:48:12,557 --> 00:48:14,517 But I don't think that we should be scared. 906 00:48:14,559 --> 00:48:17,228 It's just we all need to get more educated and 907 00:48:17,270 --> 00:48:19,022 we need to evolve with it. 908 00:48:19,063 --> 00:48:21,024 And we need to understand that things 909 00:48:21,065 --> 00:48:22,734 are going to change dramatically. 910 00:48:22,775 --> 00:48:25,612 And the people that understand it more are 911 00:48:25,653 --> 00:48:28,448 the ones that are gonna be able to move on with their life 912 00:48:28,489 --> 00:48:30,825 easier than the ones that waited until the last 913 00:48:30,867 --> 00:48:33,661 minute to learn about it and then their jobs taken. 914 00:48:34,412 --> 00:48:39,042 It is our own limits that keep us from moving forward. 915 00:48:39,083 --> 00:48:41,961 And sometimes limits are needed, 916 00:48:42,003 --> 00:48:45,048 sometimes those parameters are necessary. 917 00:48:45,923 --> 00:48:50,386 But I think what we're more scared of is not the AI, 918 00:48:50,428 --> 00:48:54,223 but how we as humans will use AI. 919 00:48:54,265 --> 00:48:57,894 'Cause we know that not everyone has a good heart. 920 00:48:57,935 --> 00:49:00,146 Not everyone has good intentions. 921 00:49:00,188 --> 00:49:02,732 So we're not afraid of the AI's improvement. 922 00:49:02,774 --> 00:49:04,233 We're afraid of us. 923 00:49:06,819 --> 00:49:11,949 The future of AI is everything that we choose to put into it. 924 00:49:11,991 --> 00:49:16,537 So if we want a better future, it is up to us to choose 925 00:49:16,579 --> 00:49:19,457 and curate, and bring that future to life. 926 00:49:20,166 --> 00:49:24,170 Well, unlike a disaster movie, I believe the outlook is good. 927 00:49:24,212 --> 00:49:27,340 I believe that humanity is gonna roll with the punches, 928 00:49:27,382 --> 00:49:29,217 people are gonna get with the program. 929 00:49:29,258 --> 00:49:32,387 And humane sensibilities is going to 930 00:49:32,428 --> 00:49:35,431 prevail over chaos and destruction. 931 00:49:35,473 --> 00:49:37,767 You see in all the movies, oh, we're all gonna die. 932 00:49:37,809 --> 00:49:40,770 every year, Nostradamus said this, you know, 933 00:49:40,812 --> 00:49:43,481 somebody said that, but it never happens. 934 00:49:43,523 --> 00:49:45,191 And I pray that it never does. 935 00:49:45,233 --> 00:49:48,236 And I'm saying if we all just stay calm, 936 00:49:48,277 --> 00:49:49,987 save ourselves, do the best we can 937 00:49:50,029 --> 00:49:53,491 individually overall, we'll be alright in the end. 938 00:49:53,533 --> 00:49:56,536 And I think we'll have a good future with AI and 939 00:49:56,577 --> 00:50:00,289 robotics side by side with humanity if we don't 940 00:50:00,331 --> 00:50:02,166 panic and try to cancel the program. 941 00:50:02,208 --> 00:50:04,627 It's not Skynet and your computer can't kill 942 00:50:04,669 --> 00:50:07,088 you unless you turn it on and drop it on your head. 943 00:50:07,130 --> 00:50:10,425 So it's like, you know, just stay calm, poised, 944 00:50:10,466 --> 00:50:13,678 relax, like try to do the best you can with your life, 945 00:50:13,720 --> 00:50:15,513 your intelligence, your finances. 946 00:50:15,555 --> 00:50:17,765 Make a plan for yourself what works for you. 947 00:50:23,229 --> 00:50:26,399 Don't rely on any technology too much. 948 00:50:26,441 --> 00:50:29,777 I was the pre-microwave and dishwasher group, 949 00:50:29,819 --> 00:50:31,446 you know what I mean? 950 00:50:31,487 --> 00:50:32,989 So we didn't have microwaves when I was growing up. 951 00:50:33,030 --> 00:50:34,782 We had to do all that stuff manually. 952 00:50:34,824 --> 00:50:37,785 And it was weird because at home our microwave went out and 953 00:50:37,827 --> 00:50:40,413 it was like a week before I got a new one, I got it installed. 954 00:50:40,455 --> 00:50:42,498 And I said to my wife, how did we survive 955 00:50:42,540 --> 00:50:44,292 back in our childhood? 956 00:50:44,333 --> 00:50:46,127 I didn't realize how dependent we were 957 00:50:46,169 --> 00:50:47,754 on that one device. 958 00:50:47,795 --> 00:50:51,174 Always take time to talk to people face to face, 959 00:50:51,215 --> 00:50:53,009 touch each other, hug each other. 960 00:50:53,050 --> 00:50:54,677 You need that interaction. 961 00:50:54,719 --> 00:50:57,847 Especially if you have small children, they need that. 962 00:50:57,889 --> 00:50:59,515 So that would be my advice. 963 00:50:59,557 --> 00:51:03,352 Don't let the humanity end up inside technology. 82302

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.