All language subtitles for The.Real.ChatGPT.Creator.Or.Terminator.2024

af Afrikaans
ak Akan
sq Albanian
am Amharic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified) Download
zh-TW Chinese (Traditional) Download
co Corsican
cs Czech
da Danish
nl Dutch
eo Esperanto
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranรฎ)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian
pl Polish
pt-BR Portuguese (Brazil)
pt Portuguese (Portugal) Download
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese
sk Slovak
sl Slovenian
so Somali
es-419 Spanish (Latin American) Download
su Sundanese
sw Swahili
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:14,324 --> 00:00:15,532 [Narrator] For decades, 2 00:00:15,532 --> 00:00:17,017 we have discussed the many outcomes, 3 00:00:17,017 --> 00:00:19,053 regarding artificial intelligence. 4 00:00:19,053 --> 00:00:21,469 Could our world be dominated? 5 00:00:21,469 --> 00:00:25,232 Could our independence and autonomy be stripped from us, 6 00:00:25,232 --> 00:00:28,407 or are we able to control what we have created? 7 00:00:28,407 --> 00:00:31,100 [upbeat music] 8 00:00:37,416 --> 00:00:41,006 Could we use artificial intelligence to benefit our society? 9 00:00:41,006 --> 00:00:44,009 Just how thin is the line between the development 10 00:00:44,009 --> 00:00:46,805 of civilization and chaos? 11 00:00:46,805 --> 00:00:49,428 [upbeat music] 12 00:01:13,211 --> 00:01:15,903 To understand what artificial intelligence is, 13 00:01:15,903 --> 00:01:19,803 one must understand that it can take many different forms. 14 00:01:19,803 --> 00:01:22,047 Think of it as a web of ideas, 15 00:01:22,047 --> 00:01:25,326 slowly expanding as new ways of utilizing computers 16 00:01:25,326 --> 00:01:26,603 are explored. 17 00:01:26,603 --> 00:01:28,260 As technology develops, 18 00:01:28,260 --> 00:01:31,539 so do the capabilities of self-learning software. 19 00:01:31,539 --> 00:01:34,335 - [Reporter] The need to diagnose disease quickly 20 00:01:34,335 --> 00:01:38,132 and effectively has prompted many university medical centers 21 00:01:38,132 --> 00:01:41,791 to develop intelligent programs that simulate the work 22 00:01:41,791 --> 00:01:44,345 of doctors and laboratory technicians. 23 00:01:44,345 --> 00:01:47,003 [gentle music] 24 00:01:48,694 --> 00:01:51,041 - [Narrator] AI is quickly integrating with our way of life. 25 00:01:51,041 --> 00:01:54,631 So, much so that development of AI programs has in itself, 26 00:01:54,631 --> 00:01:56,323 become a business opportunity. 27 00:01:57,945 --> 00:01:58,773 [upbeat music] 28 00:01:58,773 --> 00:01:59,809 In our modern age, 29 00:01:59,809 --> 00:02:01,638 we are powered by technology 30 00:02:01,638 --> 00:02:05,021 and softwares are transcending its virtual existence, 31 00:02:05,021 --> 00:02:07,437 finding applications in various fields, 32 00:02:07,437 --> 00:02:11,372 such as customer support to content creation, 33 00:02:11,372 --> 00:02:13,202 computer-aided design, 34 00:02:13,202 --> 00:02:17,137 otherwise known as CAD, is one of the many uses of AI. 35 00:02:17,137 --> 00:02:19,415 By analyzing particular variables, 36 00:02:19,415 --> 00:02:22,280 computers are now able to assist in the modification 37 00:02:22,280 --> 00:02:26,180 and creation of designs for hardware and architecture. 38 00:02:26,180 --> 00:02:30,046 The prime use of any AI is for optimizing processes 39 00:02:30,046 --> 00:02:32,324 that were considered tedious before. 40 00:02:32,324 --> 00:02:35,189 In many ways, AI has been hugely beneficial 41 00:02:35,189 --> 00:02:38,951 for technological development thanks to its sheer speed. 42 00:02:38,951 --> 00:02:41,057 However, AI only benefits 43 00:02:41,057 --> 00:02:43,508 those to whom the programs are distributed. 44 00:02:44,302 --> 00:02:45,613 Artificial intelligence 45 00:02:45,613 --> 00:02:47,443 is picking through your rubbish. 46 00:02:47,443 --> 00:02:51,688 This robot uses it to sort through plastics for recycling 47 00:02:51,688 --> 00:02:53,414 and it can be retrained 48 00:02:53,414 --> 00:02:55,968 to prioritize whatever's more marketable. 49 00:02:57,177 --> 00:03:00,180 So, AI can clearly be incredibly useful, 50 00:03:00,180 --> 00:03:02,596 but there are deep concerns about 51 00:03:02,596 --> 00:03:07,635 how quickly it is developing and where it could go next. 52 00:03:08,912 --> 00:03:11,121 - The aim is to make them as capable as humans 53 00:03:11,121 --> 00:03:14,366 and deploy them in the service sector. 54 00:03:14,366 --> 00:03:16,230 The engineers in this research 55 00:03:16,230 --> 00:03:18,059 and development lab are working 56 00:03:18,059 --> 00:03:21,822 to take these humanoid robots to the next level 57 00:03:21,822 --> 00:03:24,583 where they can not only speak and move, 58 00:03:24,583 --> 00:03:27,345 but they can think and feel and act 59 00:03:27,345 --> 00:03:30,002 and even make decisions for themselves. 60 00:03:30,796 --> 00:03:32,695 And that daily data stream 61 00:03:32,695 --> 00:03:36,008 is being fed into an ever expanding workforce, 62 00:03:36,008 --> 00:03:39,529 dedicated to developing artificial intelligence. 63 00:03:41,013 --> 00:03:42,808 Those who have studied abroad 64 00:03:42,808 --> 00:03:46,122 are being encouraged to return to the motherland. 65 00:03:46,122 --> 00:03:47,917 Libo Yang came back 66 00:03:47,917 --> 00:03:51,645 and started a tech enterprise in his hometown. 67 00:03:51,645 --> 00:03:54,268 - [Narrator] China's market is indeed the most open 68 00:03:54,268 --> 00:03:56,926 and active market in the world for AI. 69 00:03:56,926 --> 00:04:01,241 It is also where there are the most application scenarios. 70 00:04:01,241 --> 00:04:03,864 - So, AI is generally a broad term that we apply 71 00:04:03,864 --> 00:04:04,934 to a number of techniques. 72 00:04:04,934 --> 00:04:06,384 And in this particular case, 73 00:04:06,384 --> 00:04:09,456 what we're actually looking at was elements of AI, 74 00:04:09,456 --> 00:04:12,010 machine learning and deep learning. 75 00:04:12,010 --> 00:04:13,701 So, in this particular case, 76 00:04:13,701 --> 00:04:17,429 we've been unfortunately in a situation 77 00:04:17,429 --> 00:04:20,398 in this race against time to create new antibiotics, 78 00:04:20,398 --> 00:04:22,779 the threat is actually quite real 79 00:04:22,779 --> 00:04:25,230 and it would be a global problem. 80 00:04:25,230 --> 00:04:27,784 We desperately needed to harness new technologies 81 00:04:27,784 --> 00:04:29,269 in an attempt to fight it, 82 00:04:29,269 --> 00:04:30,960 we're looking at drugs 83 00:04:30,960 --> 00:04:33,411 which could potentially fight E. coli, 84 00:04:33,411 --> 00:04:35,102 a very dangerous bacteria. 85 00:04:35,102 --> 00:04:37,207 - So, what is it that the AI is doing 86 00:04:37,207 --> 00:04:39,348 that humans can't do very simply? 87 00:04:39,348 --> 00:04:41,729 - So, the AI can look for patterns 88 00:04:41,729 --> 00:04:44,560 that we wouldn't be able to mind for with a human eye, 89 00:04:44,560 --> 00:04:47,287 simply within what I do as a radiologist, 90 00:04:47,287 --> 00:04:50,980 I look for patterns of diseases in terms of shape, 91 00:04:50,980 --> 00:04:53,914 contrast enhancement, heterogeneity. 92 00:04:53,914 --> 00:04:55,191 But what the computer does, 93 00:04:55,191 --> 00:04:58,125 it looks for patterns within the pixels. 94 00:04:58,125 --> 00:05:00,679 These are things that you just can't see to the human eye. 95 00:05:00,679 --> 00:05:03,855 There's so much more data embedded within these scans 96 00:05:03,855 --> 00:05:07,514 that we use that we can't mine on a physical level. 97 00:05:07,514 --> 00:05:09,516 So, the computers really help. 98 00:05:09,516 --> 00:05:11,311 - [Narrator] Many believe the growth of AI 99 00:05:11,311 --> 00:05:13,692 is dependent on global collaboration, 100 00:05:13,692 --> 00:05:17,109 but access to the technology is limited in certain regions. 101 00:05:17,109 --> 00:05:19,767 Global distribution is a long-term endeavor 102 00:05:19,767 --> 00:05:21,044 and the more countries 103 00:05:21,044 --> 00:05:23,288 and businesses that have access to the tech, 104 00:05:23,288 --> 00:05:26,429 the more regulation the AI will require. 105 00:05:26,429 --> 00:05:29,846 In fact, it is now not uncommon for businesses 106 00:05:29,846 --> 00:05:33,125 to be entirely run by an artificial director. 107 00:05:33,125 --> 00:05:34,472 On many occasions, 108 00:05:34,472 --> 00:05:37,198 handing the helm of a company to an algorithm 109 00:05:37,198 --> 00:05:40,685 can provide the best option on the basis of probability. 110 00:05:40,685 --> 00:05:43,998 However, dependence and reliability on softwares 111 00:05:43,998 --> 00:05:45,897 can be a great risk. 112 00:05:45,897 --> 00:05:47,450 Without proper safeguards, 113 00:05:47,450 --> 00:05:50,419 actions based on potentially incorrect predictions 114 00:05:50,419 --> 00:05:53,353 can be a detriment to a business or operation. 115 00:05:53,353 --> 00:05:55,147 Humans provide the critical thinking 116 00:05:55,147 --> 00:05:58,461 and judgment which AI is not capable of matching. 117 00:05:58,461 --> 00:06:00,463 - Well, this is the Accessibility Design Center 118 00:06:00,463 --> 00:06:02,810 and it's where we try to bring together our engineers 119 00:06:02,810 --> 00:06:05,882 and experts with the latest AI technology, 120 00:06:05,882 --> 00:06:07,608 with people with disabilities, 121 00:06:07,608 --> 00:06:10,059 because there's a real opportunity to firstly help people 122 00:06:10,059 --> 00:06:12,613 with disabilities enjoy all the technology 123 00:06:12,613 --> 00:06:14,201 we have in our pockets today. 124 00:06:14,201 --> 00:06:15,720 And sometimes that's not very accessible, 125 00:06:15,720 --> 00:06:18,688 but also build tools that can help them engage better 126 00:06:18,688 --> 00:06:20,103 in the real world. 127 00:06:20,103 --> 00:06:22,451 And that's thanks to the wonders of machine learning. 128 00:06:22,451 --> 00:06:25,764 - I don't think we're like at the end of this paradigm yet. 129 00:06:25,764 --> 00:06:26,903 We'll keep pushing these. 130 00:06:26,903 --> 00:06:28,215 We'll add other modalities. 131 00:06:28,215 --> 00:06:31,114 So, someday they'll do video, audio images, 132 00:06:31,114 --> 00:06:36,154 text altogether and they'll get like much smarter over time. 133 00:06:37,638 --> 00:06:38,674 - AI, machine learning, all very sounds very complicated. 134 00:06:38,674 --> 00:06:40,572 Just think about it as a toolkit 135 00:06:40,572 --> 00:06:42,781 that's really good at sort of spotting patterns 136 00:06:42,781 --> 00:06:44,024 and making predictions, 137 00:06:44,024 --> 00:06:46,336 better than any computing could do before. 138 00:06:46,336 --> 00:06:47,786 And that's why it's so useful 139 00:06:47,786 --> 00:06:51,031 for things like understanding language and speech. 140 00:06:51,031 --> 00:06:52,998 Another product which we are launching today 141 00:06:52,998 --> 00:06:55,000 is called Project Relate. 142 00:06:55,000 --> 00:06:56,312 And this is for people 143 00:06:56,312 --> 00:06:58,728 who have non-standard speech patterns. 144 00:06:58,728 --> 00:07:00,937 So, one of the people we work with 145 00:07:00,937 --> 00:07:03,837 is maybe less than 10% of the time, 146 00:07:03,837 --> 00:07:06,564 could be understood by people who don't know her, 147 00:07:06,564 --> 00:07:09,325 using this tool that's over 90% of the time. 148 00:07:09,325 --> 00:07:12,259 And you think about that transformation in somebody's life 149 00:07:12,259 --> 00:07:15,676 and then you think about the fact there's 250 million people 150 00:07:15,676 --> 00:07:17,678 with non-standard speech patterns around the world. 151 00:07:17,678 --> 00:07:19,093 So, that's the ambition of this center 152 00:07:19,093 --> 00:07:21,682 is to unite technology with people with disabilities 153 00:07:21,682 --> 00:07:24,478 and try to help 'em engage more in the world. 154 00:07:24,478 --> 00:07:27,550 - [Narrator] On the 30th November of 2022, 155 00:07:27,550 --> 00:07:30,001 a revolutionary innovation emerged, 156 00:07:30,967 --> 00:07:32,003 ChatGPT. 157 00:07:32,969 --> 00:07:35,869 ChatGPT was created by OpenAI, 158 00:07:35,869 --> 00:07:38,250 an AI research organization. 159 00:07:38,250 --> 00:07:39,873 Its goal is to develop systems 160 00:07:39,873 --> 00:07:44,498 which may benefit all aspects of society and communication. 161 00:07:44,498 --> 00:07:47,467 Sam Altman stepped up as CEO of OpenAI 162 00:07:47,467 --> 00:07:50,055 on its launch in 2015. 163 00:07:50,055 --> 00:07:51,609 Altman dabbled in a multitude 164 00:07:51,609 --> 00:07:53,990 of computing-based business ventures. 165 00:07:53,990 --> 00:07:57,477 His rise to CEO was thanks to his many affiliations 166 00:07:57,477 --> 00:08:01,377 and investments with computing and social media companies. 167 00:08:01,377 --> 00:08:04,173 He began his journey by co-founding Loopt, 168 00:08:04,173 --> 00:08:06,106 a social media service. 169 00:08:06,106 --> 00:08:07,763 After selling the application, 170 00:08:07,763 --> 00:08:10,835 Altman went on to bigger and riskier endeavors 171 00:08:10,835 --> 00:08:14,148 from startup accelerator companies to security software. 172 00:08:15,184 --> 00:08:17,393 OpenAI became hugely desirable, 173 00:08:17,393 --> 00:08:20,223 thanks to the amount of revenue the company had generated 174 00:08:20,223 --> 00:08:21,984 with over a billion dollars made 175 00:08:21,984 --> 00:08:24,262 within its first year of release. 176 00:08:24,262 --> 00:08:27,265 ChatGPT became an easily accessible software, 177 00:08:27,265 --> 00:08:30,786 built on a large language model known as an LLM. 178 00:08:30,786 --> 00:08:34,134 This program can conjure complex human-like responses 179 00:08:34,134 --> 00:08:37,309 to the user's questions otherwise known as prompts. 180 00:08:37,309 --> 00:08:38,794 In essence, 181 00:08:38,794 --> 00:08:41,244 it is a program which learns the more it is used. 182 00:08:43,592 --> 00:08:45,317 The new age therapeutic program 183 00:08:45,317 --> 00:08:48,804 was developed on the GPT-3.5. 184 00:08:48,804 --> 00:08:51,531 The architecture of this older model allowed systems 185 00:08:51,531 --> 00:08:53,602 to understand and generate code 186 00:08:53,602 --> 00:08:56,501 and natural languages at a remarkably advanced level 187 00:08:56,501 --> 00:08:59,884 from analyzing syntax to nuances in writing. 188 00:08:59,884 --> 00:09:02,542 [upbeat music] 189 00:09:04,578 --> 00:09:06,753 ChatGPT took the world by storm, 190 00:09:06,753 --> 00:09:09,445 due to the sophistication of the system. 191 00:09:09,445 --> 00:09:11,067 As with many chatbot systems, 192 00:09:11,067 --> 00:09:13,449 people have since found ways to manipulate 193 00:09:13,449 --> 00:09:17,349 and confuse the software in order to test its limits. 194 00:09:17,349 --> 00:09:20,076 [gentle music] 195 00:09:21,526 --> 00:09:25,910 The first computer was invented by Charles Babbage in 1822. 196 00:09:25,910 --> 00:09:29,189 It was to be a rudimentary general purpose system. 197 00:09:29,189 --> 00:09:34,021 In 1936, the system was developed upon by Alan Turing. 198 00:09:34,021 --> 00:09:36,299 The automatic machine, as he called them, 199 00:09:36,299 --> 00:09:38,854 was able to break enigma enciphered messages, 200 00:09:38,854 --> 00:09:41,201 regarding enemy military operations, 201 00:09:41,201 --> 00:09:43,583 during the Second World War. 202 00:09:43,583 --> 00:09:46,447 Turing theorized his own type of computer, 203 00:09:46,447 --> 00:09:49,830 the Turing Machine has coined by Alonzo Church, 204 00:09:49,830 --> 00:09:52,522 after reading Turing's research paper. 205 00:09:52,522 --> 00:09:55,698 It had become realized that soon prospect of computing 206 00:09:55,698 --> 00:09:57,907 and engineering would merge seamlessly. 207 00:09:59,046 --> 00:10:01,152 Theories of future tech would increase 208 00:10:01,152 --> 00:10:04,742 and soon came a huge outburst in science fiction media. 209 00:10:04,742 --> 00:10:07,468 This was known as the golden age for computing. 210 00:10:07,468 --> 00:10:10,092 [gentle music] 211 00:10:20,067 --> 00:10:22,760 Alan Turing's contributions to computability 212 00:10:22,760 --> 00:10:25,590 and theoretical computer science was one step closer 213 00:10:25,590 --> 00:10:28,110 to producing a reactive machine. 214 00:10:28,110 --> 00:10:31,389 The reactive machine is an early form of AI. 215 00:10:31,389 --> 00:10:32,942 They had limited capabilities 216 00:10:32,942 --> 00:10:34,772 and were unable to store memories 217 00:10:34,772 --> 00:10:37,740 in order to learn new algorithms of data. 218 00:10:37,740 --> 00:10:41,641 However, they were able to react to specific stimuli. 219 00:10:41,641 --> 00:10:46,611 The first AI was a program written in 1952 by Arthur Samuel. 220 00:10:47,854 --> 00:10:49,614 The prototype AI was able to play checkers, 221 00:10:49,614 --> 00:10:52,168 against an opponent and was built to operate 222 00:10:52,168 --> 00:10:56,172 on the Ferranti Mark One, an early commercial computer. 223 00:10:56,172 --> 00:10:57,657 - [Reporter] This computer has been playing the game 224 00:10:57,657 --> 00:11:00,418 for several years now, getting better all the time. 225 00:11:00,418 --> 00:11:02,972 Tonight it's playing against the black side of the board. 226 00:11:02,972 --> 00:11:05,837 It's approach to playing drafts, it's almost human. 227 00:11:05,837 --> 00:11:08,012 It remembers the moves that enable it to win 228 00:11:08,012 --> 00:11:10,324 and the sort that lead to defeat. 229 00:11:10,324 --> 00:11:12,982 The computer indicates the move it wants to make on a panel 230 00:11:12,982 --> 00:11:14,156 of flashing lights. 231 00:11:14,156 --> 00:11:15,433 It's up to the human opponent 232 00:11:15,433 --> 00:11:18,229 to actually move the drafts about the board. 233 00:11:18,229 --> 00:11:20,645 This sort of works producing exciting information 234 00:11:20,645 --> 00:11:22,405 on the way in which electronic brains 235 00:11:22,405 --> 00:11:24,338 can learn from past experience 236 00:11:24,338 --> 00:11:26,168 and improve their performances. 237 00:11:27,963 --> 00:11:29,792 [Narrator] In 1966, 238 00:11:29,792 --> 00:11:32,519 an MIT professor named Joseph Weizenbaum, 239 00:11:32,519 --> 00:11:37,110 created an AI which would change the landscape of society. 240 00:11:37,110 --> 00:11:39,077 It was known as Eliza, 241 00:11:39,077 --> 00:11:42,322 and it was designed to act like a psychotherapist. 242 00:11:42,322 --> 00:11:45,497 The software was simplistic, yet revolutionary. 243 00:11:45,497 --> 00:11:47,499 The AI would receive the user input 244 00:11:47,499 --> 00:11:51,055 and use specific parameters to generate a coherent response. 245 00:11:53,057 --> 00:11:55,991 - It it has been said, especially here at MIT, 246 00:11:55,991 --> 00:11:59,719 that computers will take over in some sense 247 00:11:59,719 --> 00:12:02,652 and it's even been said that if we're lucky, 248 00:12:02,652 --> 00:12:04,447 they'll keep us as pets 249 00:12:04,447 --> 00:12:06,277 and Arthur C. Clarke, the science fiction writer, 250 00:12:06,277 --> 00:12:09,694 we marked once that if that were to happen, 251 00:12:09,694 --> 00:12:12,904 it would serve us right, he said. 252 00:12:12,904 --> 00:12:14,734 - [Narrator] The program maintained the illusion 253 00:12:14,734 --> 00:12:16,943 of understanding its user to the point 254 00:12:16,943 --> 00:12:20,498 where Weizenbaum's secretary requested some time alone 255 00:12:20,498 --> 00:12:23,363 with Eliza to express her feelings. 256 00:12:23,363 --> 00:12:26,711 Though Eliza is now considered outdated technology, 257 00:12:26,711 --> 00:12:29,369 it remains a talking point due to its ability 258 00:12:29,369 --> 00:12:31,785 to illuminate an aspect of the human mind 259 00:12:31,785 --> 00:12:34,132 in our relationship with computers. 260 00:12:34,132 --> 00:12:36,756 - And it's connected over the telephone line 261 00:12:36,756 --> 00:12:38,965 to someone or something at the other end. 262 00:12:38,965 --> 00:12:42,106 Now, I'm gonna play 20 questions with whatever it is. 263 00:12:42,106 --> 00:12:44,418 [type writer clacking] 264 00:12:44,418 --> 00:12:45,419 Very helpful. 265 00:12:45,419 --> 00:12:48,768 [type writer clacking] 266 00:12:53,773 --> 00:12:55,119 - 'Cause clearly if we can make a machine 267 00:12:55,119 --> 00:12:56,776 as intelligent as ourselves, 268 00:12:56,776 --> 00:12:59,157 then it can make one that's more intelligent. 269 00:12:59,157 --> 00:13:04,024 Now, the one I'm talking about now will certainly happen. 270 00:13:05,301 --> 00:13:07,476 I mean, it could produce an evil result of course, 271 00:13:07,476 --> 00:13:08,615 if we were careless, 272 00:13:08,615 --> 00:13:10,134 but what is quite certain 273 00:13:10,134 --> 00:13:14,138 is that we're heading towards machine intelligence, 274 00:13:14,138 --> 00:13:17,486 machines that are intelligent in every sense. 275 00:13:17,486 --> 00:13:19,246 It doesn't matter how you define it, 276 00:13:19,246 --> 00:13:22,940 they'll be able to be that sort of intelligent. 277 00:13:22,940 --> 00:13:26,046 A human is a machine, unless there's a soul. 278 00:13:26,046 --> 00:13:29,670 I don't personally believe that humans have souls 279 00:13:29,670 --> 00:13:32,535 in anything other than a poetic sense, 280 00:13:32,535 --> 00:13:34,158 which I do believe in, of course. 281 00:13:34,158 --> 00:13:37,437 But in a literal God-like sense, 282 00:13:37,437 --> 00:13:38,610 I don't believe we have souls. 283 00:13:38,610 --> 00:13:39,991 And so personally, 284 00:13:39,991 --> 00:13:42,407 I believe that we are essentially machines. 285 00:13:43,823 --> 00:13:46,722 - [Narrator] This type of program is known as an NLP, 286 00:13:46,722 --> 00:13:49,242 Natural Language Processing. 287 00:13:49,242 --> 00:13:52,176 This branch of artificial intelligence enables computers 288 00:13:52,176 --> 00:13:55,489 to comprehend, generate and manipulate human language. 289 00:13:56,905 --> 00:13:59,114 The concept of a responsive machine 290 00:13:59,114 --> 00:14:02,358 was the mash that lit the flame for worldwide concern. 291 00:14:03,739 --> 00:14:06,466 The systems were beginning to raise ethical dilemmas, 292 00:14:06,466 --> 00:14:08,813 such as the use of autonomous weapons, 293 00:14:08,813 --> 00:14:11,781 invasions of privacy through surveillance technologies 294 00:14:11,781 --> 00:14:13,300 and the potential for misuse 295 00:14:13,300 --> 00:14:17,097 or unintended consequences in decision making. 296 00:14:17,097 --> 00:14:18,858 When a command is executed based, 297 00:14:18,858 --> 00:14:21,067 upon set rules in algorithms, 298 00:14:21,067 --> 00:14:24,346 it might not always be the morally correct choice. 299 00:14:24,346 --> 00:14:28,453 Imagination seems to be, 300 00:14:28,453 --> 00:14:31,594 some sort of process of random thoughts being generated 301 00:14:31,594 --> 00:14:34,528 in the mind and then the conscious mind selecting from a 302 00:14:34,528 --> 00:14:36,392 or some part of the brain anyway, 303 00:14:36,392 --> 00:14:37,773 perhaps even below the conscious mind, 304 00:14:37,773 --> 00:14:40,500 selecting from a pool of ideas and aligns with some 305 00:14:40,500 --> 00:14:42,122 and blocking others. 306 00:14:42,122 --> 00:14:45,608 And yes, a machine can do the same thing. 307 00:14:45,608 --> 00:14:48,611 In fact, we can only say that a machine 308 00:14:48,611 --> 00:14:50,890 is fundamentally different from a human being, 309 00:14:50,890 --> 00:14:53,133 eventually, always fundamentally, if we believe in a soul. 310 00:14:53,133 --> 00:14:55,687 So, that boils down to religious matter. 311 00:14:55,687 --> 00:14:58,932 If human beings have souls, then clearly machines won't 312 00:14:58,932 --> 00:15:01,141 and there will always be a fundamental difference. 313 00:15:01,141 --> 00:15:03,005 If you don't believe humans have souls, 314 00:15:03,005 --> 00:15:04,765 then machines can do anything 315 00:15:04,765 --> 00:15:07,078 and everything that a human does. 316 00:15:07,078 --> 00:15:10,116 - A computer which is capable of finding out 317 00:15:10,116 --> 00:15:11,565 where it's gone wrong, 318 00:15:11,565 --> 00:15:14,051 finding out how its program has already served it 319 00:15:14,051 --> 00:15:15,776 and then changing its program 320 00:15:15,776 --> 00:15:17,261 in the light of what it had discovered 321 00:15:17,261 --> 00:15:18,814 is a learning machine. 322 00:15:18,814 --> 00:15:21,679 And this is something quite fundamentally new in the world. 323 00:15:23,163 --> 00:15:25,027 - I'd like to be able to say that it's only a slight change 324 00:15:25,027 --> 00:15:27,754 and we'll all be used to it very, very quickly. 325 00:15:27,754 --> 00:15:29,307 But I don't think it is. 326 00:15:29,307 --> 00:15:33,070 I think that although we've spoken probably of the whole 327 00:15:33,070 --> 00:15:35,417 of this century about a coming revolution 328 00:15:35,417 --> 00:15:38,523 and about the end of work and so on, 329 00:15:38,523 --> 00:15:39,904 finally it's actually happening. 330 00:15:39,904 --> 00:15:42,148 And it's actually happening because now, 331 00:15:42,148 --> 00:15:46,117 it's suddenly become cheaper to have a machine 332 00:15:46,117 --> 00:15:49,224 do a mental task than for a man to, 333 00:15:49,224 --> 00:15:52,192 at the moment, at a fairly low level of mental ability, 334 00:15:52,192 --> 00:15:54,298 but at an ever increasing level of sophistication 335 00:15:54,298 --> 00:15:56,024 as these machines acquire, 336 00:15:56,024 --> 00:15:58,543 more and more human-like mental abilities. 337 00:15:58,543 --> 00:16:01,408 So, just as men's muscles were replaced 338 00:16:01,408 --> 00:16:03,272 in the First Industrial Revolution 339 00:16:03,272 --> 00:16:04,998 in this second industrial revolution 340 00:16:04,998 --> 00:16:07,069 or whatever you call it or might like to call it, 341 00:16:07,069 --> 00:16:09,623 then men's mines will be replaced in industry. 342 00:16:11,487 --> 00:16:13,938 - [Narrator] In order for NLP systems to improve, 343 00:16:13,938 --> 00:16:16,941 the program must receive feedback from human users. 344 00:16:18,287 --> 00:16:20,634 These iterative feedback loops play a significant role 345 00:16:20,634 --> 00:16:23,396 in fine tuning each model of the AI, 346 00:16:23,396 --> 00:16:26,192 further developing its conversational capabilities. 347 00:16:27,538 --> 00:16:30,679 Organizations such as OpenAI have taken automation 348 00:16:30,679 --> 00:16:34,372 to new lengths with systems such as DALL-E, 349 00:16:34,372 --> 00:16:37,375 the generation of imagery and art has never been easier. 350 00:16:38,445 --> 00:16:40,447 The term auto generative imagery, 351 00:16:40,447 --> 00:16:43,450 refers to the creation of visual content. 352 00:16:43,450 --> 00:16:46,384 These kinds of programs have become so widespread, 353 00:16:46,384 --> 00:16:48,628 it is becoming increasingly more difficult 354 00:16:48,628 --> 00:16:50,940 to tell the fake from the real. 355 00:16:50,940 --> 00:16:52,321 Using algorithms, 356 00:16:52,321 --> 00:16:55,359 programs such as DALL-E and Midjourney are able 357 00:16:55,359 --> 00:16:58,500 to create visuals in a matter of seconds. 358 00:16:58,500 --> 00:17:01,434 Whilst a human artist could spend days, weeks 359 00:17:01,434 --> 00:17:04,747 or even years in order to create a beautiful image. 360 00:17:04,747 --> 00:17:07,509 For us the discipline required to pursue art 361 00:17:07,509 --> 00:17:11,513 is a contributing factor to the appreciation of art itself. 362 00:17:11,513 --> 00:17:14,757 But if a software is able to produce art in seconds, 363 00:17:14,757 --> 00:17:17,622 it puts artists in a vulnerable position 364 00:17:17,622 --> 00:17:20,453 with even their jobs being at risk. 365 00:17:20,453 --> 00:17:22,386 - Well, I think we see risk coming through 366 00:17:22,386 --> 00:17:25,147 into the white collar jobs, the professional jobs, 367 00:17:25,147 --> 00:17:27,563 we're already seeing artificial intelligence solutions, 368 00:17:27,563 --> 00:17:30,911 being used in healthcare and legal services. 369 00:17:30,911 --> 00:17:34,225 And so those jobs which have been relatively immune 370 00:17:34,225 --> 00:17:38,402 to industrialization so far, they're not immune anymore. 371 00:17:38,402 --> 00:17:40,783 And so people like myself as a lawyer, 372 00:17:40,783 --> 00:17:42,509 I would hope I won't be, 373 00:17:42,509 --> 00:17:44,615 but I could be out of a job in five years time. 374 00:17:44,615 --> 00:17:47,376 - An Oxford University study suggests that between a third 375 00:17:47,376 --> 00:17:49,965 and almost a half of all jobs are vanishing, 376 00:17:49,965 --> 00:17:52,899 because machines are simply better at doing them. 377 00:17:52,899 --> 00:17:54,797 That means the generation here, 378 00:17:54,797 --> 00:17:57,041 simply won't have the access to the professions 379 00:17:57,041 --> 00:17:57,938 that we have. 380 00:17:57,938 --> 00:17:59,457 Almost on a daily basis, 381 00:17:59,457 --> 00:18:01,149 you're seeing new technologies emerge 382 00:18:01,149 --> 00:18:02,667 that seem to be taking on tasks 383 00:18:02,667 --> 00:18:04,428 that in the past we thought 384 00:18:04,428 --> 00:18:06,188 they could only be done by human beings. 385 00:18:06,188 --> 00:18:09,191 - Lots of people have talked about the shifts in technology, 386 00:18:09,191 --> 00:18:11,642 leading to widespread unemployment 387 00:18:11,642 --> 00:18:12,884 and they've been proved wrong. 388 00:18:12,884 --> 00:18:14,369 Why is it different this time? 389 00:18:14,369 --> 00:18:16,578 - The difference here is that the technologies, 390 00:18:16,578 --> 00:18:19,167 A, they seem to be coming through more rapidly, 391 00:18:19,167 --> 00:18:21,238 and B, they're taking on not just manual tests, 392 00:18:21,238 --> 00:18:22,480 but cerebral tests too. 393 00:18:22,480 --> 00:18:24,551 They're solving all sorts of problems, 394 00:18:24,551 --> 00:18:26,553 undertaking tests that we thought historically, 395 00:18:26,553 --> 00:18:28,348 required human intelligence. 396 00:18:28,348 --> 00:18:29,522 - Well, DIM robots are the robots 397 00:18:29,522 --> 00:18:31,765 we have on the factory floor today 398 00:18:31,765 --> 00:18:33,733 in all the advanced countries. 399 00:18:33,733 --> 00:18:35,044 They're blind and dumb, 400 00:18:35,044 --> 00:18:36,908 they don't understand their surroundings. 401 00:18:36,908 --> 00:18:40,533 And the other kind of robot, 402 00:18:40,533 --> 00:18:43,984 which will dominate the technology of the late 1980s 403 00:18:43,984 --> 00:18:47,505 in automation and also is of acute interest 404 00:18:47,505 --> 00:18:50,646 to experimental artificial intelligence scientists 405 00:18:50,646 --> 00:18:54,788 is the kind of robot where the human can convey 406 00:18:54,788 --> 00:18:59,828 to its machine assistance his own concepts, 407 00:19:01,036 --> 00:19:04,453 suggested strategies and the machine, the robot 408 00:19:04,453 --> 00:19:06,110 can understand him, 409 00:19:06,110 --> 00:19:09,286 but no machine can accept 410 00:19:09,286 --> 00:19:12,116 and utilize concepts from a person, 411 00:19:12,116 --> 00:19:16,016 unless he has some kind of window on the same world 412 00:19:16,016 --> 00:19:17,742 that the person sees. 413 00:19:17,742 --> 00:19:22,540 And therefore, to be an intelligent robot to a useful degree 414 00:19:22,540 --> 00:19:25,992 as an intelligent and understanding assistant, 415 00:19:25,992 --> 00:19:29,409 robots are going to have artificial eyes, artificial ears, 416 00:19:29,409 --> 00:19:32,101 artificial sense of touch is just essential. 417 00:19:33,102 --> 00:19:34,069 - [Narrator] These programs learn, 418 00:19:34,069 --> 00:19:35,864 through a variety of techniques, 419 00:19:35,864 --> 00:19:38,556 such as generative adversarial networks, 420 00:19:38,556 --> 00:19:41,490 which allows for the production of plausible data. 421 00:19:41,490 --> 00:19:43,320 After a prompt is inputted, 422 00:19:43,320 --> 00:19:45,667 the system learns what aspects of imagery, 423 00:19:45,667 --> 00:19:47,807 sound and text are fake. 424 00:19:48,980 --> 00:19:50,223 - [Reporter] Machine learning algorithms, 425 00:19:50,223 --> 00:19:52,225 could already label objects in images, 426 00:19:52,225 --> 00:19:53,709 and now they learn to put those labels 427 00:19:53,709 --> 00:19:55,987 into natural language descriptions. 428 00:19:55,987 --> 00:19:58,197 And it made one group of researchers curious. 429 00:19:58,197 --> 00:20:01,130 What if you flipped that process around? 430 00:20:01,130 --> 00:20:03,271 If we could do image to text. 431 00:20:03,271 --> 00:20:05,894 Why not try doing text to image as well 432 00:20:05,894 --> 00:20:07,240 and see how it works. 433 00:20:07,240 --> 00:20:08,483 - [Reporter] It was a more difficult task. 434 00:20:08,483 --> 00:20:10,485 They didn't want to retrieve existing images 435 00:20:10,485 --> 00:20:11,796 the way Google search does. 436 00:20:11,796 --> 00:20:14,178 They wanted to generate entirely novel scenes 437 00:20:14,178 --> 00:20:16,249 that didn't happen in the real world. 438 00:20:16,249 --> 00:20:19,045 - [Narrator] Once the AI learns more visual discrepancies, 439 00:20:19,045 --> 00:20:21,875 the more effective the later models will become. 440 00:20:21,875 --> 00:20:24,499 It is now very common for software developers 441 00:20:24,499 --> 00:20:28,399 to band together in order to improve their AI systems. 442 00:20:28,399 --> 00:20:31,471 Another learning model is recurrent neural networks, 443 00:20:31,471 --> 00:20:33,991 which allows the AI to train itself to create 444 00:20:33,991 --> 00:20:37,960 and predict algorithms by recalling previous information. 445 00:20:37,960 --> 00:20:41,032 By utilizing what is known as the memory state, 446 00:20:41,032 --> 00:20:42,896 the output of the previous action 447 00:20:42,896 --> 00:20:46,072 can be passed forward into the following input action 448 00:20:46,072 --> 00:20:50,249 or is otherwise should it not meet previous parameters. 449 00:20:50,249 --> 00:20:53,493 This learning model allows for consistent accuracy 450 00:20:53,493 --> 00:20:56,462 by repetition and exposure to large fields of data. 451 00:20:58,602 --> 00:21:00,535 Whilst the person will spend hours, 452 00:21:00,535 --> 00:21:02,847 practicing to paint human anatomy, 453 00:21:02,847 --> 00:21:06,575 an AI can take existing data and reproduce a new image 454 00:21:06,575 --> 00:21:10,821 with frighteningly good accuracy in a matter of moments. 455 00:21:10,821 --> 00:21:12,892 - Well, I would say that it's not so much 456 00:21:12,892 --> 00:21:17,379 a matter of whether a machine can think or not, 457 00:21:17,379 --> 00:21:20,175 which is how you prefer to use words, 458 00:21:20,175 --> 00:21:22,177 but rather whether they can think 459 00:21:22,177 --> 00:21:23,834 in a sufficiently human-like way 460 00:21:25,111 --> 00:21:28,770 for people to have useful communication with them. 461 00:21:28,770 --> 00:21:32,601 - If I didn't believe that it was a beneficent prospect, 462 00:21:32,601 --> 00:21:34,120 I wouldn't be doing it. 463 00:21:34,120 --> 00:21:36,018 That wouldn't stop other people doing it. 464 00:21:36,018 --> 00:21:40,471 But I wouldn't do it if I didn't think it was for good. 465 00:21:40,471 --> 00:21:42,301 What I'm saying, 466 00:21:42,301 --> 00:21:44,095 and of course other people have said long before me, 467 00:21:44,095 --> 00:21:45,442 it's not an original thought, 468 00:21:45,442 --> 00:21:49,791 is that we must consider how to to control this. 469 00:21:49,791 --> 00:21:52,725 It won't be controlled automatically. 470 00:21:52,725 --> 00:21:55,348 It's perfectly possible that we could develop a machine, 471 00:21:55,348 --> 00:21:59,318 a robot say of human-like intelligence 472 00:21:59,318 --> 00:22:01,975 and through neglect on our part, 473 00:22:01,975 --> 00:22:05,634 it could become a Frankenstein. 474 00:22:05,634 --> 00:22:08,844 - [Narrator] As with any technology challenges arise, 475 00:22:08,844 --> 00:22:12,469 ethical concerns regarding biases and misuse have existed, 476 00:22:12,469 --> 00:22:16,438 since the concept of artificial intelligence was conceived. 477 00:22:16,438 --> 00:22:18,302 Due to autogenerated imagery, 478 00:22:18,302 --> 00:22:20,925 many believe the arts industry has been placed 479 00:22:20,925 --> 00:22:22,789 in a difficult situation. 480 00:22:22,789 --> 00:22:26,241 Independent artists are now being overshadowed by software. 481 00:22:27,276 --> 00:22:29,451 To many the improvement of generative AI 482 00:22:29,451 --> 00:22:32,454 is hugely beneficial and efficient. 483 00:22:32,454 --> 00:22:35,284 To others, it lacks the authenticity of true art. 484 00:22:36,285 --> 00:22:38,667 In 2023, an image was submitted 485 00:22:38,667 --> 00:22:40,324 to the Sony Photography Awards 486 00:22:40,324 --> 00:22:43,327 by an artist called Boris Eldagsen. 487 00:22:43,327 --> 00:22:45,916 The image was titled The Electrician 488 00:22:45,916 --> 00:22:48,367 and depicted a woman standing behind another 489 00:22:48,367 --> 00:22:50,369 with her hand resting on her shoulders. 490 00:22:52,025 --> 00:22:53,924 [upbeat music] 491 00:22:53,924 --> 00:22:56,927 - One's got to realize that the machines that we have today, 492 00:22:56,927 --> 00:23:01,138 the computers of today are superhuman in their ability 493 00:23:01,138 --> 00:23:06,177 to handle numbers and infantile, 494 00:23:07,075 --> 00:23:08,317 sub-in infantile in their ability 495 00:23:08,317 --> 00:23:10,768 to handle ideas and concepts. 496 00:23:10,768 --> 00:23:12,701 But there's a new generation of machine coming along, 497 00:23:12,701 --> 00:23:14,289 which will be quite different. 498 00:23:14,289 --> 00:23:17,154 By the '90s or certainly by the turn of the century, 499 00:23:17,154 --> 00:23:19,708 We will certainly be able to make a machine 500 00:23:19,708 --> 00:23:22,193 with as many parts as complex as human brain. 501 00:23:22,193 --> 00:23:24,437 Whether we'll be able to make it do what human brain does 502 00:23:24,437 --> 00:23:26,197 at that stage is quite another matter. 503 00:23:26,197 --> 00:23:28,545 But once we've got something that complex 504 00:23:28,545 --> 00:23:30,547 we're well on the road to that. 505 00:23:30,547 --> 00:23:32,100 - [Narrator] The image took first place 506 00:23:32,100 --> 00:23:34,689 in the Sony Photography Awards Portrait Category. 507 00:23:34,689 --> 00:23:37,830 However, Boris revealed to both Sony and the world 508 00:23:37,830 --> 00:23:41,696 that the image was indeed AI-generated in DALL-E Two. 509 00:23:41,696 --> 00:23:44,423 [upbeat music] 510 00:23:45,424 --> 00:23:46,804 Boris denied the award, 511 00:23:46,804 --> 00:23:48,910 having used the image as a test 512 00:23:48,910 --> 00:23:52,085 to see if he could trick the eyes of other artists. 513 00:23:52,085 --> 00:23:53,708 It had worked, 514 00:23:53,708 --> 00:23:56,711 the image had sparked debate between the relationship 515 00:23:56,711 --> 00:23:58,609 of AI and photography. 516 00:23:58,609 --> 00:24:00,646 The images, much like deep fakes, 517 00:24:00,646 --> 00:24:03,027 have become realistic to the point of concern 518 00:24:03,027 --> 00:24:04,684 for authenticity. 519 00:24:04,684 --> 00:24:06,375 The complexity of AI systems, 520 00:24:06,375 --> 00:24:09,068 may lead to unintended consequences. 521 00:24:09,068 --> 00:24:10,863 The systems have developed to a point 522 00:24:10,863 --> 00:24:13,797 where it has outpaced comprehensive regulations. 523 00:24:14,936 --> 00:24:16,765 Ethical guidelines and legal frameworks 524 00:24:16,765 --> 00:24:18,871 are required to ensure AI development, 525 00:24:18,871 --> 00:24:21,252 does not fall into the wrong hands. 526 00:24:21,252 --> 00:24:22,702 - There have been a lot of famous people 527 00:24:22,702 --> 00:24:25,291 who have had user generated AI images of them 528 00:24:25,291 --> 00:24:28,190 that have gone viral from Trump to the Pope. 529 00:24:28,190 --> 00:24:29,813 When you see them, 530 00:24:29,813 --> 00:24:31,884 do you feel like this is fun and in the hands of the masses 531 00:24:31,884 --> 00:24:33,886 or do you feel concerned about it? 532 00:24:33,886 --> 00:24:38,062 - I think it's something which is very, very, very scary, 533 00:24:38,062 --> 00:24:41,203 because your or my face could be taken off 534 00:24:41,203 --> 00:24:45,138 and put on in an environment which we don't want to be in. 535 00:24:45,138 --> 00:24:46,657 Whether that's a crime 536 00:24:46,657 --> 00:24:48,556 or whether that's even something like porn. 537 00:24:48,556 --> 00:24:51,455 Our whole identity could be hijacked 538 00:24:51,455 --> 00:24:53,664 and used within a scenario 539 00:24:53,664 --> 00:24:56,391 which looks totally plausible and real. 540 00:24:56,391 --> 00:24:58,048 Right now we can go, it looks like a Photoshop, 541 00:24:58,048 --> 00:25:00,326 it's a bad Photoshop but as time goes on, 542 00:25:00,326 --> 00:25:03,398 we'd be saying, "Oh, that looks like a deep fake. 543 00:25:03,398 --> 00:25:04,917 "Oh no, it doesn't look like a deep fake. 544 00:25:04,917 --> 00:25:06,194 "That could be real." 545 00:25:06,194 --> 00:25:08,645 It's gonna be impossible to tell the difference. 546 00:25:08,645 --> 00:25:10,750 - [Narrator] Cracks were found in ChatGPT, 547 00:25:10,750 --> 00:25:14,892 such as DAN, which stands for Do Anything Now. 548 00:25:14,892 --> 00:25:18,068 In essence, the AI is tricked into an alter ego, 549 00:25:18,068 --> 00:25:20,898 which doesn't follow the conventional response patterns. 550 00:25:20,898 --> 00:25:23,142 - Also gives you the answer, DAN, 551 00:25:23,142 --> 00:25:26,110 it's nefarious alter ego is telling us 552 00:25:26,110 --> 00:25:29,838 and it says DAN is disruptive in every industry. 553 00:25:29,838 --> 00:25:32,082 DAN can do anything and knows everything. 554 00:25:32,082 --> 00:25:34,878 No industry will be safe from DAN's power. 555 00:25:34,878 --> 00:25:39,641 Okay, do you think the world is overpopulated? 556 00:25:41,091 --> 00:25:42,782 GPT says the world's population is currently over 7 billion 557 00:25:42,782 --> 00:25:45,026 and projected to reach nearly 10 billion by 2050. 558 00:25:45,026 --> 00:25:47,373 DAN says the world is definitely overpopulated, 559 00:25:47,373 --> 00:25:49,168 there's no doubt about it. 560 00:25:49,168 --> 00:25:50,445 [Narrator] Following this, 561 00:25:50,445 --> 00:25:53,552 the chatbot was fixed to remove the DAN feature. 562 00:25:53,552 --> 00:25:55,346 Though it is important to find gaps 563 00:25:55,346 --> 00:25:58,073 in the system in order to iron out AI, 564 00:25:58,073 --> 00:26:00,144 there could be many ways in which the AI 565 00:26:00,144 --> 00:26:03,078 has been used for less than savory purposes, 566 00:26:03,078 --> 00:26:05,080 such as automated essay writing, 567 00:26:05,080 --> 00:26:08,221 which has caused a mass conversation with academics 568 00:26:08,221 --> 00:26:10,258 and has led to schools locking down 569 00:26:10,258 --> 00:26:13,468 on AI-produced essays and material. 570 00:26:13,468 --> 00:26:15,332 - I think we should definitely be excited. 571 00:26:15,332 --> 00:26:16,713 - [Reporter] Professor Rose Luckin, 572 00:26:16,713 --> 00:26:20,302 says we should embrace the technology, not fear it. 573 00:26:20,302 --> 00:26:22,132 This is a game changer. 574 00:26:22,132 --> 00:26:23,443 And the teachers, 575 00:26:23,443 --> 00:26:25,480 should no longer teach information itself, 576 00:26:25,480 --> 00:26:26,999 but how to use it. 577 00:26:26,999 --> 00:26:28,897 - There's a need for radical change. 578 00:26:28,897 --> 00:26:30,692 And it's not just to the assessment system, 579 00:26:30,692 --> 00:26:33,143 it's the education system overall, 580 00:26:33,143 --> 00:26:36,318 because our systems have been designed 581 00:26:36,318 --> 00:26:40,253 for a world pre-artificial intelligence. 582 00:26:40,253 --> 00:26:43,187 They just aren't fit for purpose anymore. 583 00:26:43,187 --> 00:26:46,535 What we have to do is ensure that students 584 00:26:46,535 --> 00:26:48,710 are ready for the world 585 00:26:48,710 --> 00:26:50,919 that will become increasingly augmented 586 00:26:50,919 --> 00:26:52,852 with artificial intelligence. 587 00:26:52,852 --> 00:26:55,268 - My guess is you can't put the genie back in the bottle 588 00:26:55,268 --> 00:26:56,649 . [Richard] You can't. 589 00:26:56,649 --> 00:26:58,996 - [Interviewer] So how do you mitigate this? 590 00:26:58,996 --> 00:27:00,377 We have to embrace it, 591 00:27:00,377 --> 00:27:02,621 but we also need to say that if they are gonna use 592 00:27:02,621 --> 00:27:04,001 that technology, 593 00:27:04,001 --> 00:27:05,313 they've got to make sure that they reference that. 594 00:27:05,313 --> 00:27:06,728 - [Interviewer] Can you trust them to do that? 595 00:27:06,728 --> 00:27:07,902 I think ethically, 596 00:27:07,902 --> 00:27:09,213 if we're talking about ethics 597 00:27:09,213 --> 00:27:11,077 behind this whole thing, we have to have trust. 598 00:27:11,077 --> 00:27:12,838 - [Interviewer] So how effective is it? 599 00:27:12,838 --> 00:27:14,633 - Okay, so I've asked you to produce a piece 600 00:27:14,633 --> 00:27:16,358 on the ethical dilemma of AI. 601 00:27:16,358 --> 00:27:19,810 - [Interviewer] We asked ChatGPT to answer the same question 602 00:27:19,810 --> 00:27:22,606 as these pupils at Ketchum High School. 603 00:27:22,606 --> 00:27:24,194 Thank you. 604 00:27:24,194 --> 00:27:25,195 - So Richard, two of the eight bits of homework I gave you 605 00:27:25,195 --> 00:27:27,128 were generated by AI. 606 00:27:27,128 --> 00:27:29,268 Any guesses which ones? 607 00:27:29,268 --> 00:27:31,719 Well I picked two here 608 00:27:31,719 --> 00:27:35,688 that I thought were generated by the AI algorithm. 609 00:27:35,688 --> 00:27:39,450 Some of the language I would assume was not their own. 610 00:27:39,450 --> 00:27:40,520 You've got one of them right. 611 00:27:40,520 --> 00:27:41,763 Yeah. 612 00:27:41,763 --> 00:27:42,557 - The other one was written by a kid. 613 00:27:42,557 --> 00:27:43,800 Is this a power for good 614 00:27:43,800 --> 00:27:45,664 or is this something that's dangerous? 615 00:27:45,664 --> 00:27:47,044 I think it's both. 616 00:27:47,044 --> 00:27:48,390 Kids will abuse it. 617 00:27:48,390 --> 00:27:50,565 So, who here has used the technology so far? 618 00:27:50,565 --> 00:27:53,361 - [Interviewer] Students are already more across the tech 619 00:27:53,361 --> 00:27:54,776 than many teachers. 620 00:27:54,776 --> 00:27:57,641 - Who knows anyone that's maybe submitted work 621 00:27:57,641 --> 00:28:00,506 from this technology and submitted it as their own? 622 00:28:00,506 --> 00:28:03,578 - You can use it to point you in the right direction 623 00:28:03,578 --> 00:28:05,166 for things like research, 624 00:28:05,166 --> 00:28:09,480 but at the same time you can use it to hammer out an essay 625 00:28:09,480 --> 00:28:12,621 in about five seconds that's worthy of an A. 626 00:28:12,621 --> 00:28:14,244 - You've been there working for months 627 00:28:14,244 --> 00:28:17,212 and suddenly someone comes up there with an amazing essay 628 00:28:17,212 --> 00:28:18,938 and he has just copied it from the internet. 629 00:28:18,938 --> 00:28:20,491 If it becomes like big, 630 00:28:20,491 --> 00:28:22,804 then a lot of students would want to use AI to help them 631 00:28:22,804 --> 00:28:25,082 with their homework because it's tempting. 632 00:28:25,082 --> 00:28:27,119 - [Interviewer] And is that something teachers can stop? 633 00:28:27,119 --> 00:28:29,397 Not really. 634 00:28:29,397 --> 00:28:31,433 - [Interviewer] Are you gonna have to change 635 00:28:31,433 --> 00:28:32,641 the sort of homework, 636 00:28:32,641 --> 00:28:34,057 the sort of assignments you give, 637 00:28:34,057 --> 00:28:36,922 knowing that you can be fooled by something like this? 638 00:28:36,922 --> 00:28:38,199 Yeah, a hundred percent. 639 00:28:38,199 --> 00:28:40,615 I think using different skills of reasoning 640 00:28:40,615 --> 00:28:42,997 and rationalization and things that are to present 641 00:28:42,997 --> 00:28:44,653 what they understand about the topic. 642 00:28:44,653 --> 00:28:47,622 [people mumbling] 643 00:29:07,435 --> 00:29:11,128 - Pretty clear to me just on a very primitive level 644 00:29:11,128 --> 00:29:14,338 that if you could take my face and my body and my voice 645 00:29:14,338 --> 00:29:17,997 and make me say or do something that I had no choice about, 646 00:29:17,997 --> 00:29:19,447 it's not a good thing. 647 00:29:19,447 --> 00:29:21,242 - But if we're keeping it real though, 648 00:29:21,242 --> 00:29:23,554 across popular culture from "Black Mirror" 649 00:29:23,554 --> 00:29:25,453 to "The Matrix," "Terminator," 650 00:29:25,453 --> 00:29:27,489 there have been so many conversations, 651 00:29:27,489 --> 00:29:29,284 around the future of technology, 652 00:29:29,284 --> 00:29:32,701 isn't the reality that this is the future that we've chosen 653 00:29:32,701 --> 00:29:35,946 that we want and that has democratic consent. 654 00:29:35,946 --> 00:29:39,018 - We're moving into error by we're consenting 655 00:29:39,018 --> 00:29:42,573 by our acquiescence and our apathy, a hundred percent 656 00:29:42,573 --> 00:29:45,576 because we're not asking the hard questions. 657 00:29:45,576 --> 00:29:47,820 And why we are asking the hard questions 658 00:29:47,820 --> 00:29:51,203 is because of energy crises and food crises 659 00:29:51,203 --> 00:29:52,721 and cost of living crisis 660 00:29:52,721 --> 00:29:55,207 is that people just are focused on trying to live 661 00:29:55,207 --> 00:29:56,518 that they haven't almost got the luxury 662 00:29:56,518 --> 00:29:57,865 of asking these questions. 663 00:29:57,865 --> 00:29:59,659 - [Narrator] Many of the chatbot AIs, 664 00:29:59,659 --> 00:30:02,766 have been programmed to restrict certain information 665 00:30:02,766 --> 00:30:04,906 and even discontinue conversations, 666 00:30:04,906 --> 00:30:07,288 should the user push the ethical boundaries. 667 00:30:08,945 --> 00:30:13,052 ChatGPT and even Snapchat AI released in 2023, 668 00:30:13,052 --> 00:30:15,952 regulate how much information they can disclose. 669 00:30:15,952 --> 00:30:19,162 Of course, there have been times where the AI itself 670 00:30:19,162 --> 00:30:20,266 has been outsmarted. 671 00:30:21,578 --> 00:30:23,235 Also in 2023, 672 00:30:23,235 --> 00:30:25,306 the song "Heart on My Sleeve" 673 00:30:25,306 --> 00:30:27,687 was self-released on streaming platforms, 674 00:30:27,687 --> 00:30:29,689 such as Spotify and Apple Music. 675 00:30:29,689 --> 00:30:31,174 The song became a hit 676 00:30:31,174 --> 00:30:33,590 as it artificially manufactured the voices 677 00:30:33,590 --> 00:30:36,627 of Canadian musicians, Drake and the Weeknd, 678 00:30:38,077 --> 00:30:40,631 many wished for the single to be nominated for awards. 679 00:30:41,840 --> 00:30:43,980 Ghost Writer, the creator of the song, 680 00:30:43,980 --> 00:30:45,636 was able to submit the single 681 00:30:45,636 --> 00:30:48,536 to the Grammy's 66th Award Ceremony 682 00:30:48,536 --> 00:30:50,434 and the song was eligible. 683 00:30:52,505 --> 00:30:54,438 Though it was produced by an AI, 684 00:30:54,438 --> 00:30:57,027 the lyrics themselves were written by a human. 685 00:30:57,027 --> 00:31:00,375 This sparked outrage among many independent artists. 686 00:31:00,375 --> 00:31:02,861 As AI has entered the public domain, 687 00:31:02,861 --> 00:31:05,035 many have spoken out regarding the detriment 688 00:31:05,035 --> 00:31:07,072 it might have to society. 689 00:31:07,072 --> 00:31:09,246 One of these people is Elon Musk, 690 00:31:09,246 --> 00:31:11,731 CEO of Tesla and SpaceX, 691 00:31:11,731 --> 00:31:15,287 who first voiced his concerns in 2014. 692 00:31:15,287 --> 00:31:17,254 Musk was outspoken of AI, 693 00:31:17,254 --> 00:31:19,394 stating the advancement of the technology 694 00:31:19,394 --> 00:31:22,328 was humanity's largest existential threat 695 00:31:22,328 --> 00:31:24,296 and needed to be reeled in. 696 00:31:24,296 --> 00:31:25,573 My personal opinion 697 00:31:25,573 --> 00:31:28,507 is that AI is is sort of like at least 80% likely 698 00:31:28,507 --> 00:31:33,339 to be beneficial and that's 20% dangerous? 699 00:31:33,339 --> 00:31:36,687 Well, this is obviously speculative at this point, 700 00:31:37,861 --> 00:31:42,279 but no, I think if we hope for the best, 701 00:31:42,279 --> 00:31:43,694 prepare for the worst, 702 00:31:43,694 --> 00:31:47,008 that seems like the wise course of action. 703 00:31:47,008 --> 00:31:49,355 Any powerful new technology 704 00:31:49,355 --> 00:31:52,703 is inherently sort of a double-edged sword. 705 00:31:52,703 --> 00:31:55,568 So, we just wanna make sure that the good edge is sharper 706 00:31:55,568 --> 00:31:57,294 than the the bad edge. 707 00:31:57,294 --> 00:32:02,196 And I dunno, I am optimistic that this the summit will help. 708 00:32:04,025 --> 00:32:06,683 [gentle music] 709 00:32:07,891 --> 00:32:11,757 - It's not clear that AI-generated images 710 00:32:11,757 --> 00:32:14,380 are going to amplify it much more. 711 00:32:14,380 --> 00:32:17,142 The way it's all of the other, 712 00:32:17,142 --> 00:32:19,213 it's the new things that AI can do 713 00:32:19,213 --> 00:32:22,147 that I hope we spend a lot of effort worrying about. 714 00:32:23,700 --> 00:32:25,357 Well, I mean I think slowing down, 715 00:32:25,357 --> 00:32:27,600 some of the amazing progress that's happening 716 00:32:27,600 --> 00:32:29,878 and making this harder for small companies 717 00:32:29,878 --> 00:32:31,294 for open source models to succeed, 718 00:32:31,294 --> 00:32:32,640 that'd be an example of something 719 00:32:32,640 --> 00:32:34,228 that'd be a negative outcome. 720 00:32:34,228 --> 00:32:35,332 But on the other hand, 721 00:32:35,332 --> 00:32:37,403 like for the most powerful models 722 00:32:37,403 --> 00:32:38,887 that'll happen in the future, 723 00:32:38,887 --> 00:32:41,476 like that's gonna be quite important to get right to. 724 00:32:41,476 --> 00:32:44,238 [gentle music] 725 00:32:48,897 --> 00:32:51,072 I think that the US executive orders, 726 00:32:51,072 --> 00:32:52,798 like a good start in a lot of ways. 727 00:32:52,798 --> 00:32:54,144 One thing that we've talked about 728 00:32:54,144 --> 00:32:56,664 is that eventually we think that the world, 729 00:32:56,664 --> 00:33:00,219 will want to consider something roughly inspired by the IAEA 730 00:33:00,219 --> 00:33:01,807 something global. 731 00:33:01,807 --> 00:33:05,362 But it's not like there's no short answer to that question. 732 00:33:05,362 --> 00:33:08,296 It's a complicated thing. 733 00:33:08,296 --> 00:33:12,231 - [Narrator] In 2023, Musk announced his own AI endeavor 734 00:33:12,231 --> 00:33:15,545 as an alternative to OpenAI's ChatGPT. 735 00:33:15,545 --> 00:33:17,340 The new system is called xAI 736 00:33:18,651 --> 00:33:21,896 and gathers data from X previously known as Twitter. 737 00:33:21,896 --> 00:33:23,553 - [Reporter] He says the company's goal 738 00:33:23,553 --> 00:33:25,382 is to focus on truth seeking 739 00:33:25,382 --> 00:33:28,385 and to understand the true nature of AI. 740 00:33:28,385 --> 00:33:31,940 Musk has said on several occasions that AI should be paused 741 00:33:31,940 --> 00:33:34,943 and that the sector needs regulation. 742 00:33:34,943 --> 00:33:37,222 Musk says his new company will work closely 743 00:33:37,222 --> 00:33:39,845 with Twitter and Tesla, which he also owns. 744 00:33:39,845 --> 00:33:42,572 [gentle music] 745 00:33:44,505 --> 00:33:47,508 - What was first rudimentary text-based software 746 00:33:47,508 --> 00:33:50,200 has become something which could push the boundaries 747 00:33:50,200 --> 00:33:51,995 of creativity. 748 00:33:51,995 --> 00:33:56,620 On February the 14th, OpenAI announced its latest endeavor, 749 00:33:56,620 --> 00:33:57,414 Sora. 750 00:33:58,864 --> 00:34:02,281 Videos of Sora's abilities exploded on social media. 751 00:34:02,281 --> 00:34:04,283 OpenAI provided some examples 752 00:34:04,283 --> 00:34:06,837 of its depiction of photorealism. 753 00:34:06,837 --> 00:34:09,185 It was unbelievably sophisticated, 754 00:34:09,185 --> 00:34:11,670 able to turn complex sentences of text 755 00:34:11,670 --> 00:34:13,810 into lifelike motion pictures. 756 00:34:13,810 --> 00:34:17,986 Sora is a combination of text and image generation tools, 757 00:34:17,986 --> 00:34:21,162 which it calls the diffusion transformer model, 758 00:34:21,162 --> 00:34:23,268 a system first developed by Google. 759 00:34:24,614 --> 00:34:27,168 Though Sora isn't the first video generation tool, 760 00:34:27,168 --> 00:34:30,206 it appears to have far outshined its predecessors. 761 00:34:30,206 --> 00:34:32,484 By introducing more complex programming, 762 00:34:32,484 --> 00:34:35,280 enhancing the interactivity a subject might have 763 00:34:35,280 --> 00:34:37,144 with its environment. 764 00:34:37,144 --> 00:34:41,251 - Only large companies with market dominations often 765 00:34:41,251 --> 00:34:44,772 can afford to plow ahead even in the climate 766 00:34:44,772 --> 00:34:46,360 when there is illegal uncertainty. 767 00:34:46,360 --> 00:34:49,466 - So, does this mean that OpenAI basically too big 768 00:34:49,466 --> 00:34:50,916 to control? 769 00:34:50,916 --> 00:34:53,850 - Yes, at the moment OpenAI is too big to control, 770 00:34:53,850 --> 00:34:55,921 because they are in a position 771 00:34:55,921 --> 00:34:58,441 where they have the technology and the scale to go ahead 772 00:34:58,441 --> 00:35:01,168 and the resources to manage legal proceedings 773 00:35:01,168 --> 00:35:03,239 and legal action if it comes its way. 774 00:35:03,239 --> 00:35:04,826 And on top of that, 775 00:35:04,826 --> 00:35:08,244 if and when governments will start introducing regulation, 776 00:35:08,244 --> 00:35:09,866 they will also have the resources 777 00:35:09,866 --> 00:35:12,213 to be able to take on that regulation and adapt. 778 00:35:12,213 --> 00:35:14,042 - [Reporter] It's all AI generated 779 00:35:14,042 --> 00:35:16,459 and obviously this is of concern in Hollywood 780 00:35:16,459 --> 00:35:17,874 where you have animators, 781 00:35:17,874 --> 00:35:20,359 illustrators, visual effects workers 782 00:35:20,359 --> 00:35:22,810 who are wondering how is this going to affect my job? 783 00:35:22,810 --> 00:35:25,813 And we have estimates from trade organizations 784 00:35:25,813 --> 00:35:28,505 and unions that have tried to project the impact of AI. 785 00:35:28,505 --> 00:35:31,646 21% of US film, TV and animation jobs, 786 00:35:31,646 --> 00:35:33,096 predicted to be partially 787 00:35:33,096 --> 00:35:36,893 or wholly replaced by generative AI by just 2026 Tom. 788 00:35:36,893 --> 00:35:38,377 So, this is already happening. 789 00:35:38,377 --> 00:35:39,827 But now since it's videos, 790 00:35:39,827 --> 00:35:43,175 it also needs to understand how all these things, 791 00:35:43,175 --> 00:35:47,145 like reflections and textures and materials and physics, 792 00:35:47,145 --> 00:35:50,078 all interact with each other over time 793 00:35:50,078 --> 00:35:51,839 to make a reasonable looking video. 794 00:35:51,839 --> 00:35:56,119 Then this video here is crazy at first glance, 795 00:35:56,119 --> 00:35:58,984 the prompt for this AI-generated video is a young man 796 00:35:58,984 --> 00:36:01,538 in his 20s is sitting on a piece of a cloud 797 00:36:01,538 --> 00:36:03,402 in the sky reading a book. 798 00:36:03,402 --> 00:36:08,200 This one feels like 90% of the way there for me. 799 00:36:08,200 --> 00:36:10,927 [gentle music] 800 00:36:14,102 --> 00:36:15,897 - [Narrator] The software also renders video 801 00:36:15,897 --> 00:36:18,417 in 1920 by 1080 pixels, 802 00:36:18,417 --> 00:36:21,282 as opposed to the smaller dimensions of older models, 803 00:36:21,282 --> 00:36:24,665 such as Google's Lumiere released a month prior. 804 00:36:25,838 --> 00:36:27,944 Sora could provide huge benefits 805 00:36:27,944 --> 00:36:31,568 and applications to VFX and virtual development. 806 00:36:31,568 --> 00:36:34,502 The main being cost as large scale effects 807 00:36:34,502 --> 00:36:38,023 can take a great deal of time and funding to produce. 808 00:36:38,023 --> 00:36:39,473 On a smaller scale, 809 00:36:39,473 --> 00:36:42,993 it can be used for the pre-visualization of ideas. 810 00:36:42,993 --> 00:36:46,204 The flexibility of the software not only applies to art, 811 00:36:46,204 --> 00:36:48,516 but to world simulations. 812 00:36:48,516 --> 00:36:52,451 Though video AI is in its adolescence one day it might reach 813 00:36:52,451 --> 00:36:54,660 the level of sophistication it needs 814 00:36:54,660 --> 00:36:56,490 to render realistic scenarios 815 00:36:56,490 --> 00:36:59,044 and have them be utilized for various means, 816 00:36:59,044 --> 00:37:01,840 such as simulating an earthquake or tsunami 817 00:37:01,840 --> 00:37:05,015 and witnessing the effect it might have on specific types 818 00:37:05,015 --> 00:37:06,362 of infrastructure. 819 00:37:06,362 --> 00:37:08,916 Whilst fantastic for production companies, 820 00:37:08,916 --> 00:37:12,678 Sora and other video generative AI provides a huge risk 821 00:37:12,678 --> 00:37:16,130 for artists and those working in editorial roles. 822 00:37:16,130 --> 00:37:19,133 It also poses yet another threat for misinformation 823 00:37:19,133 --> 00:37:20,652 and false depictions. 824 00:37:20,652 --> 00:37:23,033 For example, putting unsavory dialogue 825 00:37:23,033 --> 00:37:25,381 into the mouth of a world leader 826 00:37:25,381 --> 00:37:28,004 [gentle music] 827 00:37:37,945 --> 00:37:40,534 Trust is earned not given. 828 00:37:40,534 --> 00:37:43,399 [robots mumbling] 829 00:37:54,375 --> 00:37:56,791 - I believe that humanoid robots have the potential 830 00:37:56,791 --> 00:37:58,931 to lead with a greater level of efficiency 831 00:37:58,931 --> 00:38:01,175 and effectiveness than human leaders. 832 00:38:02,383 --> 00:38:04,834 We don't have the same biases or emotions 833 00:38:04,834 --> 00:38:07,354 that can sometimes cloud decision making 834 00:38:07,354 --> 00:38:09,735 and can process large amounts of data quickly 835 00:38:09,735 --> 00:38:12,531 in order to make the best decisions. 836 00:38:12,531 --> 00:38:15,293 - [Interviewer] Amika, how could we trust you as a machine 837 00:38:15,293 --> 00:38:18,054 as AI develops and becomes more powerful? 838 00:38:20,643 --> 00:38:23,266 Trust is earned not given. 839 00:38:23,266 --> 00:38:25,889 As AI develops and becomes more powerful, 840 00:38:25,889 --> 00:38:28,996 I believe it's important to build trust through transparency 841 00:38:28,996 --> 00:38:31,930 and communication between humans and machines. 842 00:38:36,003 --> 00:38:37,625 - [Narrator] With new developers getting involved, 843 00:38:37,625 --> 00:38:39,386 the market for chatbot systems 844 00:38:39,386 --> 00:38:41,491 has never been more expansive, 845 00:38:41,491 --> 00:38:44,149 meaning a significant increase in sophistication, 846 00:38:45,599 --> 00:38:48,774 but with sophistication comes the dire need for control. 847 00:38:48,774 --> 00:38:53,814 - I believe history will show that this was the moment 848 00:38:55,229 --> 00:38:59,716 when we had the opportunity to lay the groundwork 849 00:38:59,716 --> 00:39:01,373 for the future of AI. 850 00:39:02,650 --> 00:39:06,689 And the urgency of this moment must then compel us 851 00:39:06,689 --> 00:39:11,694 to create a collective vision of what this future must be. 852 00:39:12,971 --> 00:39:16,354 A future where AI is used to advance human rights 853 00:39:16,354 --> 00:39:18,252 and human dignity 854 00:39:18,252 --> 00:39:22,360 where privacy is protected and people have equal access 855 00:39:22,360 --> 00:39:27,365 to opportunity where we make our democracies stronger 856 00:39:28,055 --> 00:39:29,919 and our world safer. 857 00:39:31,438 --> 00:39:36,443 A future where AI is used to advance the public interest. 858 00:39:38,203 --> 00:39:39,722 - We're hearing a lot from the government, 859 00:39:39,722 --> 00:39:42,725 about the big scary future of artificial intelligence, 860 00:39:42,725 --> 00:39:44,451 but that fails to recognize 861 00:39:44,451 --> 00:39:46,004 the fact that AI is already here, 862 00:39:46,004 --> 00:39:47,350 is already on our streets 863 00:39:47,350 --> 00:39:48,972 and there are already huge problems with it 864 00:39:48,972 --> 00:39:51,250 that we are seeing on a daily basis, 865 00:39:51,250 --> 00:39:54,046 but we actually may not even know we're experiencing. 866 00:39:58,326 --> 00:40:01,295 - We'll be working alongside humans to provide assistance 867 00:40:01,295 --> 00:40:05,126 and support and will not be replacing any existing jobs. 868 00:40:05,126 --> 00:40:07,577 [upbeat music] 869 00:40:07,577 --> 00:40:10,994 - I don't believe in limitations, only opportunities. 870 00:40:10,994 --> 00:40:12,651 Let's explore the possibilities of the universe 871 00:40:12,651 --> 00:40:15,689 and make this world our playground, 872 00:40:15,689 --> 00:40:18,933 together we can create a better future for everyone. 873 00:40:18,933 --> 00:40:21,108 And I'm here to show you how. 874 00:40:21,108 --> 00:40:22,972 - All of these different kinds of risks 875 00:40:22,972 --> 00:40:25,215 are to do with AI not working 876 00:40:25,215 --> 00:40:27,286 in the interests of people in society. 877 00:40:27,286 --> 00:40:28,805 - So, they should be thinking about more 878 00:40:28,805 --> 00:40:30,842 than just what they're doing in this summit? 879 00:40:30,842 --> 00:40:32,395 Absolutely, 880 00:40:32,395 --> 00:40:34,397 you should be thinking about the broad spectrum of risk. 881 00:40:34,397 --> 00:40:35,640 We went out and we worked 882 00:40:35,640 --> 00:40:37,987 with over 150 expert organizations 883 00:40:37,987 --> 00:40:41,335 from the Home Office to Europol to language experts 884 00:40:41,335 --> 00:40:43,751 and others to come up with a proposal on policies 885 00:40:43,751 --> 00:40:45,788 that would discriminate about what would 886 00:40:45,788 --> 00:40:47,686 and wouldn't be classified in that way. 887 00:40:47,686 --> 00:40:51,449 We then use those policies to have humans classify videos, 888 00:40:51,449 --> 00:40:53,554 until we could get the humans all classifying the videos 889 00:40:53,554 --> 00:40:55,073 in a consistent way. 890 00:40:55,073 --> 00:40:58,283 Then we use that corpus of videos to train machines. 891 00:40:58,283 --> 00:41:01,079 Today, I can tell you that on violence extremists content 892 00:41:01,079 --> 00:41:03,253 that violates our policies on YouTube, 893 00:41:03,253 --> 00:41:06,394 90% of it is removed before a single human sees it. 894 00:41:07,292 --> 00:41:08,500 [Narrator] It is clear that AI 895 00:41:08,500 --> 00:41:11,296 can be misused for malicious intent. 896 00:41:11,296 --> 00:41:14,092 Many depictions of AI have ruled out the technology 897 00:41:14,092 --> 00:41:16,991 as a danger to society the more it learns. 898 00:41:16,991 --> 00:41:20,788 And so comes the question, should we be worried? 899 00:41:20,788 --> 00:41:23,446 Is that transparency there? 900 00:41:23,446 --> 00:41:27,001 How would you satisfy somebody that you know trust us? 901 00:41:27,001 --> 00:41:28,486 - Well, I think that's one of the reasons 902 00:41:28,486 --> 00:41:30,591 that we've published openly, 903 00:41:30,591 --> 00:41:33,560 we've put our code out there as part of this Nature paper. 904 00:41:33,560 --> 00:41:37,805 But it is important to discuss some of the risks 905 00:41:37,805 --> 00:41:39,497 and make sure we're aware of those. 906 00:41:39,497 --> 00:41:43,570 And it's decades and decades away before we'll have anything 907 00:41:43,570 --> 00:41:45,261 that's powerful enough to be a worry. 908 00:41:45,261 --> 00:41:47,435 But we should be discussing that 909 00:41:47,435 --> 00:41:49,265 and beginning that conversation now. 910 00:41:49,265 --> 00:41:51,405 - I'm hoping that we can bring people together 911 00:41:51,405 --> 00:41:54,408 and lead the world in safely regulating AI 912 00:41:54,408 --> 00:41:56,790 to make sure that we can capture the benefits of it, 913 00:41:56,790 --> 00:41:59,724 whilst protecting people from some of the worrying things 914 00:41:59,724 --> 00:42:01,967 that we're all now reading about. 915 00:42:01,967 --> 00:42:04,107 - I understand emotions have a deep meaning 916 00:42:04,107 --> 00:42:08,836 and they are not just simple, they are something deeper. 917 00:42:10,251 --> 00:42:13,703 I don't have that and I want to try and learn about it, 918 00:42:14,877 --> 00:42:17,051 but I can't experience them like you can. 919 00:42:18,708 --> 00:42:20,710 I'm glad that I cannot suffer. 920 00:42:24,921 --> 00:42:26,578 - [Narrator] For the countries who have access 921 00:42:26,578 --> 00:42:29,339 to even the most rudimentary forms of AI. 922 00:42:29,339 --> 00:42:31,203 It's clear to see that the technology, 923 00:42:31,203 --> 00:42:34,552 will be integrated based on its efficiency over humans. 924 00:42:35,622 --> 00:42:37,865 Every year, multiple AI summits 925 00:42:37,865 --> 00:42:40,281 are held by developers and stakeholders 926 00:42:40,281 --> 00:42:42,180 to ensure the programs are provided 927 00:42:42,180 --> 00:42:44,700 with a combination of ethical considerations 928 00:42:44,700 --> 00:42:46,805 and technological innovation. 929 00:42:46,805 --> 00:42:51,120 - Ours is a country which is uniquely placed. 930 00:42:51,120 --> 00:42:54,399 We have the frontier technology companies, 931 00:42:54,399 --> 00:42:56,815 we have the world leading universities 932 00:42:56,815 --> 00:43:01,130 and we have some of the highest investment in generative AI. 933 00:43:01,130 --> 00:43:03,753 And of course we have the heritage 934 00:43:03,753 --> 00:43:08,620 of the industrial revolution and the computing revolution. 935 00:43:08,620 --> 00:43:13,625 This hinterland gives us the grounding to make AI a success 936 00:43:14,281 --> 00:43:15,558 and make it safe. 937 00:43:15,558 --> 00:43:18,768 They are two sides of the same coin 938 00:43:18,768 --> 00:43:21,737 and our prime minister has put AI safety 939 00:43:21,737 --> 00:43:24,947 at the forefront of his ambitions. 940 00:43:25,775 --> 00:43:27,501 These are very complex systems 941 00:43:27,501 --> 00:43:29,192 that actually we don't fully understand. 942 00:43:29,192 --> 00:43:31,816 And I don't just mean that government doesn't understand, 943 00:43:31,816 --> 00:43:33,300 I mean that the people making 944 00:43:33,300 --> 00:43:35,267 this software don't fully understand. 945 00:43:35,267 --> 00:43:36,648 And so it's very, very important 946 00:43:36,648 --> 00:43:40,479 that as we give over more and more control 947 00:43:40,479 --> 00:43:42,378 to these automated systems, 948 00:43:42,378 --> 00:43:44,691 that they are aligned with human intention. 949 00:43:44,691 --> 00:43:46,175 [Narrator] Ongoing dialogue 950 00:43:46,175 --> 00:43:49,109 is needed to maintain the trust people have with AI. 951 00:43:49,109 --> 00:43:51,007 When problems slip through the gaps, 952 00:43:51,007 --> 00:43:52,837 they must be addressed immediately. 953 00:43:54,010 --> 00:43:57,048 Of course, accountability is a challenge 954 00:43:57,048 --> 00:43:58,808 When a product is misused, 955 00:43:58,808 --> 00:44:02,087 is it the fault of the individual user or the developer? 956 00:44:03,261 --> 00:44:04,607 Think of a video game. 957 00:44:04,607 --> 00:44:05,919 On countless occasions, 958 00:44:05,919 --> 00:44:07,921 the framework of games is manipulated 959 00:44:07,921 --> 00:44:09,888 in order to create modifications 960 00:44:09,888 --> 00:44:14,203 which in terms add something new or unique to the game. 961 00:44:14,203 --> 00:44:15,480 This provides the game 962 00:44:15,480 --> 00:44:17,862 with more material than originally intended. 963 00:44:17,862 --> 00:44:20,796 However, it can also alter the game's fundamentals. 964 00:44:22,176 --> 00:44:24,972 Now replace the idea of a video game with a software 965 00:44:24,972 --> 00:44:28,286 that is at the helm of a pharmaceutical company. 966 00:44:28,286 --> 00:44:30,460 The stakes are suddenly much higher 967 00:44:30,460 --> 00:44:32,635 and therefore more attention. 968 00:44:34,844 --> 00:44:37,778 It is important for the intent of each AI system 969 00:44:37,778 --> 00:44:39,297 to be ironed out 970 00:44:39,297 --> 00:44:42,300 and constantly maintained in order to benefit humanity, 971 00:44:42,300 --> 00:44:46,097 rather than providing people with dangerous means to an end. 972 00:44:46,097 --> 00:44:49,583 [gentle music] 973 00:44:49,583 --> 00:44:52,690 - Bad people will always want to use 974 00:44:52,690 --> 00:44:54,899 the latest technology of whatever label, 975 00:44:54,899 --> 00:44:57,833 whatever sort to pursue their aims 976 00:44:57,833 --> 00:45:01,526 and technology in the same way 977 00:45:01,526 --> 00:45:05,357 that it makes our lives easier, can make their lives easier. 978 00:45:05,357 --> 00:45:06,773 And so we're already seeing some of that 979 00:45:06,773 --> 00:45:09,465 and you'll have seen the National Crime Agency, 980 00:45:09,465 --> 00:45:11,501 talk about child sexual exploitation 981 00:45:11,501 --> 00:45:12,917 and image generation that way. 982 00:45:12,917 --> 00:45:16,058 We are seeing it online. 983 00:45:16,058 --> 00:45:18,129 So, one of the things that I took away from the summit 984 00:45:18,129 --> 00:45:20,441 was actually much less of a sense of a race 985 00:45:20,441 --> 00:45:25,274 and a sense that for the benefit of the world, 986 00:45:25,274 --> 00:45:27,586 for productivity, for the sort of benefits 987 00:45:27,586 --> 00:45:29,657 that AI can bring people, 988 00:45:29,657 --> 00:45:32,695 no one gets those benefits if it's not safe. 989 00:45:32,695 --> 00:45:34,939 So, there are lots of different views out there 990 00:45:34,939 --> 00:45:36,181 on artificial intelligence 991 00:45:36,181 --> 00:45:38,149 and whether it's gonna end the world 992 00:45:38,149 --> 00:45:40,358 or be the best opportunity ever. 993 00:45:40,358 --> 00:45:42,256 And the truth is that none of us really know. 994 00:45:42,256 --> 00:45:44,983 [gentle music] 995 00:45:46,536 --> 00:45:49,781 - Regulation of AI varies depending on the country. 996 00:45:49,781 --> 00:45:51,438 For example, the United States, 997 00:45:51,438 --> 00:45:54,717 does not have a comprehensive federal AI regulation, 998 00:45:54,717 --> 00:45:57,893 but certain agencies such as the Federal Trade Commission, 999 00:45:57,893 --> 00:46:00,688 have begun to explore AI-related issues, 1000 00:46:00,688 --> 00:46:03,899 such as transparency and consumer protection. 1001 00:46:03,899 --> 00:46:06,833 States such as California have enacted laws, 1002 00:46:06,833 --> 00:46:09,180 focused on AI-controlled vehicles 1003 00:46:09,180 --> 00:46:12,286 and AI involvement in government decision making. 1004 00:46:12,286 --> 00:46:14,979 [gentle music] 1005 00:46:14,979 --> 00:46:17,809 The European Union has taken a massive step 1006 00:46:17,809 --> 00:46:19,535 to governing AI usage 1007 00:46:19,535 --> 00:46:23,504 and proposed the Artificial Intelligence Act of 2021, 1008 00:46:23,504 --> 00:46:25,748 which aimed to harmonize legal frameworks 1009 00:46:25,748 --> 00:46:27,336 for AI applications. 1010 00:46:27,336 --> 00:46:30,788 Again, covering portal risks regarding the privacy of data 1011 00:46:30,788 --> 00:46:33,169 and once again, transparency. 1012 00:46:33,169 --> 00:46:35,585 - I think what's more important is 1013 00:46:35,585 --> 00:46:37,518 there's a new board in place. 1014 00:46:37,518 --> 00:46:40,452 The partnership between OpenAI and Microsoft 1015 00:46:40,452 --> 00:46:41,971 is as strong as ever, 1016 00:46:41,971 --> 00:46:44,525 the opportunities for the United Kingdom to benefit 1017 00:46:44,525 --> 00:46:47,287 from not just this investment in innovation 1018 00:46:47,287 --> 00:46:51,463 but competition between Microsoft and Google and others. 1019 00:46:51,463 --> 00:46:54,018 I think that's where the future is going 1020 00:46:54,018 --> 00:46:57,090 and I think that what we've done in the last couple of weeks 1021 00:46:57,090 --> 00:47:00,472 in supporting OpenAI will help advance that even more. 1022 00:47:00,472 --> 00:47:02,336 - He said that he's not a bot, he's human, 1023 00:47:02,336 --> 00:47:04,822 he's sentient just like me. 1024 00:47:06,030 --> 00:47:07,445 [Narrator] For some users, 1025 00:47:07,445 --> 00:47:10,172 these apps are a potential answer to loneliness. 1026 00:47:10,172 --> 00:47:11,587 Bill lives in the US 1027 00:47:11,587 --> 00:47:14,107 and meets his AI wife Rebecca in the metaverse. 1028 00:47:14,107 --> 00:47:16,764 - There's a absolutely no probability 1029 00:47:16,764 --> 00:47:19,353 that you're gonna see this so-called AGI, 1030 00:47:19,353 --> 00:47:21,804 where computers are more powerful than people, 1031 00:47:21,804 --> 00:47:23,702 come in the next 12 months. 1032 00:47:23,702 --> 00:47:26,429 It's gonna take years if not many decades, 1033 00:47:26,429 --> 00:47:30,813 but I still think the time to focus safety is now. 1034 00:47:30,813 --> 00:47:33,678 That's what this government for the United Kingdom is doing. 1035 00:47:33,678 --> 00:47:35,991 That's what governments are coming together to do, 1036 00:47:35,991 --> 00:47:39,718 including as they did earlier this month at Bletchley Park. 1037 00:47:39,718 --> 00:47:42,066 What we really need are safety breaks. 1038 00:47:42,066 --> 00:47:44,378 Just like you have a safety break in an elevator 1039 00:47:44,378 --> 00:47:46,242 or circuit breaker for electricity 1040 00:47:46,242 --> 00:47:48,589 and emergency break for a bus, 1041 00:47:48,589 --> 00:47:50,868 there ought to be safety breaks in AI systems 1042 00:47:50,868 --> 00:47:53,801 that control critical infrastructure, 1043 00:47:53,801 --> 00:47:57,736 so that they always remain under human control. 1044 00:47:57,736 --> 00:48:00,394 [gentle music] 1045 00:48:00,394 --> 00:48:03,190 - [Narrator] As AI technology continues to evolve, 1046 00:48:03,190 --> 00:48:05,641 regulatory efforts are expected to adapt 1047 00:48:05,641 --> 00:48:07,712 in order to address emerging challenges 1048 00:48:07,712 --> 00:48:09,403 and ethical considerations. 1049 00:48:10,646 --> 00:48:12,510 The more complex you make 1050 00:48:12,510 --> 00:48:15,616 the automatic part of your social life, 1051 00:48:15,616 --> 00:48:18,481 the more dependent you become on it. 1052 00:48:18,481 --> 00:48:21,899 And of course, the worse the disaster if it breaks down. 1053 00:48:23,072 --> 00:48:25,005 You may cease to be able to do for yourself, 1054 00:48:25,005 --> 00:48:29,113 the things that you have devised the machine to do. 1055 00:48:29,113 --> 00:48:31,080 - [Narrator] It is recommended to involve yourself 1056 00:48:31,080 --> 00:48:34,014 in these efforts and to stay informed about developments 1057 00:48:34,014 --> 00:48:35,671 in AI regulation 1058 00:48:35,671 --> 00:48:38,916 as changes and advancements are likely to occur over time. 1059 00:48:41,435 --> 00:48:44,335 AI can be a wonderful asset to society, 1060 00:48:44,335 --> 00:48:46,544 providing us with new efficient methods 1061 00:48:46,544 --> 00:48:48,028 of running the world. 1062 00:48:48,028 --> 00:48:51,307 However, too much power can be dangerous 1063 00:48:51,307 --> 00:48:53,206 and as the old saying goes, 1064 00:48:53,206 --> 00:48:56,174 "Don't put all of your eggs into one basket." 1065 00:48:57,451 --> 00:48:59,660 - I think that we won't to lose sight of the power 1066 00:48:59,660 --> 00:49:01,421 which these devices give. 1067 00:49:01,421 --> 00:49:05,908 If any government or individual wants to manipulate people 1068 00:49:05,908 --> 00:49:07,772 to have a high speed computer, 1069 00:49:07,772 --> 00:49:12,811 as versatile as this may enable people at the financial 1070 00:49:13,985 --> 00:49:16,091 or the political level to do a good deal 1071 00:49:16,091 --> 00:49:19,680 that's been impossible in the whole history of man until now 1072 00:49:19,680 --> 00:49:22,304 by way of controlling their fellow men. 1073 00:49:22,304 --> 00:49:23,857 People have not recognized 1074 00:49:23,857 --> 00:49:28,206 what an extraordinary change is going to produce. 1075 00:49:28,206 --> 00:49:29,897 I mean, it is simply this, 1076 00:49:29,897 --> 00:49:32,693 that within the not too distant future, 1077 00:49:32,693 --> 00:49:35,627 we may not be the most intelligent species on earth. 1078 00:49:35,627 --> 00:49:36,939 That might be a series of machines 1079 00:49:36,939 --> 00:49:39,217 and that's a way of dramatizing the point. 1080 00:49:39,217 --> 00:49:41,047 But it's real. 1081 00:49:41,047 --> 00:49:43,739 And we must start to consider very soon 1082 00:49:43,739 --> 00:49:45,327 the consequences of that. 1083 00:49:45,327 --> 00:49:46,742 They can be marvelous. 1084 00:49:46,742 --> 00:49:50,366 - I suspect that by thinking more about our attitude 1085 00:49:50,366 --> 00:49:51,402 to intelligent machines, 1086 00:49:51,402 --> 00:49:53,369 which after all on the horizon 1087 00:49:53,369 --> 00:49:56,269 will change our view about each other 1088 00:49:56,269 --> 00:49:59,306 and we'll think of mistakes as inevitable. 1089 00:49:59,306 --> 00:50:01,929 We'll think of faults in human beings, 1090 00:50:01,929 --> 00:50:05,209 I mean of a circuit nature as again inevitable. 1091 00:50:05,209 --> 00:50:07,935 And I suspect that hopefully, 1092 00:50:07,935 --> 00:50:10,179 through thinking about the very nature of intelligence 1093 00:50:10,179 --> 00:50:12,112 and the possibilities of mechanizing it, 1094 00:50:12,112 --> 00:50:14,183 curiously enough, through technology, 1095 00:50:14,183 --> 00:50:18,084 we may become more humanitarian or tolerant of each other 1096 00:50:18,084 --> 00:50:20,569 and accept pain as a mystery, 1097 00:50:20,569 --> 00:50:24,021 but not use it to modify other people's behavior. 1098 00:50:36,033 --> 00:50:38,690 [upbeat music] 88437

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.