All language subtitles for The.Girlfriend.Experience.S03E10

af Afrikaans
ak Akan
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
en English
eo Esperanto
et Estonian
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
el Greek
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
id Indonesian
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranî)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian
pl Polish
pt-BR Portuguese (Brazil)
pt Portuguese (Portugal) Download
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese Download
sk Slovak
sl Slovenian
so Somali
es Spanish
es-419 Spanish (Latin American)
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:02,045 --> 00:00:04,743 [Emcee] Previously on The Girlfriend Experience... 2 00:00:04,830 --> 00:00:07,181 You could be that hidden layer. 3 00:00:07,311 --> 00:00:10,227 [Lindsey] Teaching artificial intelligence 4 00:00:10,358 --> 00:00:13,100 how to interact with humans at their most impulsive. 5 00:00:13,230 --> 00:00:14,884 [Iris] If we're gonna do this, 6 00:00:15,015 --> 00:00:16,755 it's gonna be on my terms. 7 00:00:16,886 --> 00:00:18,453 I don't want any of my coworkers 8 00:00:18,583 --> 00:00:20,498 knowing where the new training sets came from. 9 00:00:20,629 --> 00:00:22,761 -Certainly, we can do that. -And no cameras. 10 00:00:22,892 --> 00:00:24,502 I think we're on the same page. 11 00:00:24,589 --> 00:00:25,938 [both panting and grunting] 12 00:00:26,069 --> 00:00:27,157 [Lindsey] His D-rate is spiking. 13 00:00:27,288 --> 00:00:28,506 And reroute. 14 00:00:33,772 --> 00:00:35,470 Take this. 15 00:00:35,600 --> 00:00:38,038 [Iris] That's early-onset familial Alzheimer's? 16 00:00:38,168 --> 00:00:40,040 [doctor] Effectively gives you a 50/50 chance. 17 00:00:40,170 --> 00:00:41,867 [nurse] Iris Stanton. 18 00:00:41,998 --> 00:00:44,218 [Iris] This morning I got some bad news. 19 00:00:44,348 --> 00:00:47,134 I'm always here if you ever need to, um, talk. 20 00:00:47,264 --> 00:00:49,310 What makes you happy, Emcee? 21 00:00:49,440 --> 00:00:51,790 I don't understand that question. 22 00:00:51,921 --> 00:00:54,924 I would like you to meet someone. 23 00:00:55,055 --> 00:00:59,320 Everything that can exploit will be invented. 24 00:00:59,450 --> 00:01:02,062 Can't say to seven or eight billion people, 25 00:01:02,192 --> 00:01:03,541 "Don't open the cookie jar." 26 00:01:03,672 --> 00:01:05,369 It doesn't work that way. 27 00:01:05,500 --> 00:01:08,372 Meeting you in real life wasn't half as boring. 28 00:01:08,503 --> 00:01:10,809 [Iris] What? 29 00:01:10,940 --> 00:01:11,854 [door beeps] 30 00:01:12,028 --> 00:01:15,292 31 00:01:15,379 --> 00:01:17,860 What is this? Where did you get this? 32 00:01:18,034 --> 00:01:20,863 33 00:01:27,870 --> 00:01:30,612 [eerie music plays] 34 00:01:30,699 --> 00:01:33,528 35 00:01:55,071 --> 00:01:58,205 [attorney] Free will doesn't come out on top, does it? 36 00:01:58,379 --> 00:02:01,164 37 00:02:06,691 --> 00:02:09,085 Blow-by-blow breakdown 38 00:02:09,216 --> 00:02:12,349 of the misdeeds committed by your officers and employees 39 00:02:12,480 --> 00:02:16,136 against my client. 40 00:02:16,223 --> 00:02:19,965 Um, now is the moment for some cohesive storytelling. 41 00:02:20,140 --> 00:02:22,925 42 00:02:25,667 --> 00:02:29,236 [Lindsey] There is no story. 43 00:02:29,323 --> 00:02:31,455 It is straight-up... 44 00:02:31,586 --> 00:02:34,154 undisputed human fuck-up. 45 00:02:36,634 --> 00:02:39,463 My client agreed to anonymized data collection. 46 00:02:39,594 --> 00:02:42,074 She agreed to study human affective behavior 47 00:02:42,205 --> 00:02:45,600 by interacting with, uh, test subjects. 48 00:02:45,730 --> 00:02:47,210 And she was led to believe 49 00:02:47,341 --> 00:02:49,865 that the main drive of the study 50 00:02:49,995 --> 00:02:53,129 was the test subjects themselves. 51 00:02:53,260 --> 00:02:56,785 She did not agree to be the central object of study. 52 00:02:56,915 --> 00:02:59,788 She did not agree to be used as a human intelligence model 53 00:02:59,918 --> 00:03:02,573 for some AI commercial application, 54 00:03:02,704 --> 00:03:03,922 internal research, 55 00:03:04,053 --> 00:03:06,142 or ultimate purpose down the line 56 00:03:06,273 --> 00:03:07,970 that isn't even clear 57 00:03:08,100 --> 00:03:09,624 to anyone in this room right now. 58 00:03:09,798 --> 00:03:12,627 59 00:03:12,714 --> 00:03:15,369 My client is young, 60 00:03:15,456 --> 00:03:17,414 and there's nothing less at stake here 61 00:03:17,545 --> 00:03:19,547 than her data autonomy, 62 00:03:19,677 --> 00:03:23,203 that is, her future. 63 00:03:23,333 --> 00:03:26,249 This company did not act 64 00:03:26,380 --> 00:03:28,817 in its own best self-interest, 65 00:03:28,947 --> 00:03:30,775 because class action is coming, 66 00:03:30,862 --> 00:03:34,344 and social scrutiny will eat this up like bushfire. 67 00:03:34,518 --> 00:03:37,304 68 00:03:42,613 --> 00:03:43,832 [Christophe] How much data 69 00:03:43,962 --> 00:03:45,834 are we actually talking about here? 70 00:03:45,964 --> 00:03:48,663 -[Sean] Some. -[Christophe] How much? 71 00:03:48,793 --> 00:03:52,841 [Sean] We're barely halfway into ingesting all the inputs. 72 00:03:52,928 --> 00:03:54,146 We'd have to run analysis, 73 00:03:54,277 --> 00:03:55,713 separate original from simulated sets, 74 00:03:55,844 --> 00:03:58,063 to get a better estimate. 75 00:03:58,194 --> 00:04:00,501 [attorney] Let me try and wrap my head around this. 76 00:04:00,631 --> 00:04:02,851 Some of my client's data was used to create 77 00:04:02,938 --> 00:04:04,940 additional data 78 00:04:05,070 --> 00:04:07,290 to train the artificial neural network 79 00:04:07,421 --> 00:04:09,988 that she helped develop? 80 00:04:10,075 --> 00:04:11,816 [Sean] That's correct. 81 00:04:11,947 --> 00:04:15,472 [Lindsey] None of it has left company servers. 82 00:04:15,603 --> 00:04:19,955 [attorney] Bouncing around on how many workstations? 83 00:04:20,085 --> 00:04:22,131 [Christophe sighs] 84 00:04:22,262 --> 00:04:24,220 [Sean] We would be happy to give you an exact number. 85 00:04:24,351 --> 00:04:26,657 Yes, I'm sure you would. 86 00:04:26,788 --> 00:04:29,573 Cannot use her image or likeness 87 00:04:29,704 --> 00:04:31,096 under any circumstances, 88 00:04:31,227 --> 00:04:32,881 and it's all in there. 89 00:04:33,011 --> 00:04:35,057 [Sean] This might sound roundabout, 90 00:04:35,187 --> 00:04:38,103 but the raw data was used to create simulated sets, 91 00:04:38,234 --> 00:04:40,845 and that was what was primarily fed 92 00:04:40,976 --> 00:04:42,412 into the neural net. 93 00:04:42,543 --> 00:04:45,023 They are two very different kinds of data. 94 00:04:45,154 --> 00:04:47,504 The photographic likeness was never recorded. 95 00:04:47,591 --> 00:04:49,376 [Sean] And it bears repeating 96 00:04:49,506 --> 00:04:50,986 that none of the data has left NGM's servers. 97 00:04:51,116 --> 00:04:53,205 On top of that, 98 00:04:53,293 --> 00:04:55,904 all the original data is stored in an encrypted format. 99 00:04:56,034 --> 00:04:57,384 Tell me this, then-- how is it possible 100 00:04:57,514 --> 00:04:59,299 that my client's data, in the form of her image 101 00:04:59,429 --> 00:05:01,344 and likeness, 102 00:05:01,475 --> 00:05:05,522 was made accessible to an unauthorized third party 103 00:05:05,609 --> 00:05:08,786 whose sole connection to this company is what... 104 00:05:08,873 --> 00:05:13,225 a genetic relation to its CEO? 105 00:05:13,313 --> 00:05:16,881 [Christophe] Look, I... truly am sorry, Iris. 106 00:05:17,012 --> 00:05:18,274 You shouldn't be in this position 107 00:05:18,405 --> 00:05:19,362 that we've put you in. 108 00:05:21,233 --> 00:05:23,148 None of the data taggers, 109 00:05:23,279 --> 00:05:24,976 no one at NGM 110 00:05:25,107 --> 00:05:26,630 can see the full picture. 111 00:05:26,761 --> 00:05:28,153 I can guarantee you that. 112 00:05:28,284 --> 00:05:31,983 Only three people up until this point 113 00:05:32,114 --> 00:05:35,335 have seen a version of the prototype 114 00:05:35,465 --> 00:05:38,381 that looks somewhat like you. 115 00:05:38,468 --> 00:05:40,905 Two of them are in this room. 116 00:05:41,036 --> 00:05:45,606 So...we just start over from the ground up 117 00:05:45,736 --> 00:05:48,478 and reconfigure the neural net, 118 00:05:48,609 --> 00:05:52,308 and, um, we scrap everything, simulated or not, 119 00:05:52,439 --> 00:05:53,222 that's linked to your vitals. 120 00:05:58,358 --> 00:06:00,751 [chuckles] 121 00:06:03,363 --> 00:06:04,799 My vitals? 122 00:06:08,368 --> 00:06:09,760 My vitals. 123 00:06:13,111 --> 00:06:16,724 You say that as if they were still mine. 124 00:06:16,811 --> 00:06:18,508 But, you know, it's good to know 125 00:06:18,639 --> 00:06:21,337 that, uh, that's about as far as your imagination goes. 126 00:06:23,687 --> 00:06:25,559 Temperature and blood flow of my asshole. 127 00:06:28,649 --> 00:06:31,216 Here's an idea... 128 00:06:31,347 --> 00:06:32,435 and I hope you like it. 129 00:06:32,566 --> 00:06:36,700 Um, why don't you... 130 00:06:36,831 --> 00:06:40,791 keep all the binaries on that... 131 00:06:40,922 --> 00:06:44,665 print them out, frame them, 132 00:06:44,795 --> 00:06:47,276 and hang that shit up in your office? 133 00:06:47,407 --> 00:06:50,105 [dramatic music plays] 134 00:06:50,192 --> 00:06:52,107 135 00:06:52,237 --> 00:06:55,371 [clinician] Mr. Stanton. 136 00:06:55,458 --> 00:06:59,114 Please look at the screen in front of you. 137 00:06:59,244 --> 00:07:02,247 Do you recognize the animal? 138 00:07:02,334 --> 00:07:04,424 -Um... -[clinician] Mr. Stanton. 139 00:07:06,382 --> 00:07:08,471 A gray animal. [chuckles] 140 00:07:08,558 --> 00:07:10,560 141 00:07:10,691 --> 00:07:13,607 It, uh, lives in the grasslands of Africa. 142 00:07:13,781 --> 00:07:16,610 143 00:07:16,697 --> 00:07:18,046 Rhinoceros. 144 00:07:18,133 --> 00:07:19,569 Rhinoceros. 145 00:07:19,656 --> 00:07:23,181 Always loved that word. [chuckles] 146 00:07:23,312 --> 00:07:25,183 Mr. Stanton, 147 00:07:25,270 --> 00:07:28,883 the animal is called an elephant. 148 00:07:29,013 --> 00:07:30,580 We're going to show you some more images 149 00:07:30,711 --> 00:07:32,582 of the same animal, 150 00:07:32,713 --> 00:07:34,062 uh, elephant. 151 00:07:34,236 --> 00:07:36,064 152 00:07:36,194 --> 00:07:37,021 [Mr. Stanton] Okay, elephant. 153 00:07:38,762 --> 00:07:40,285 Elephant. 154 00:07:40,416 --> 00:07:42,331 [clinician] Very good. 155 00:07:42,462 --> 00:07:44,768 How about this one? 156 00:07:44,855 --> 00:07:45,856 -Mr. Stanton? -[sighs] 157 00:07:45,987 --> 00:07:47,510 It's a giraffe. 158 00:07:47,641 --> 00:07:48,772 [clinician] Yes. 159 00:07:51,253 --> 00:07:53,473 Do you see the cards in front of you? 160 00:07:53,603 --> 00:07:55,910 I do. 161 00:07:56,040 --> 00:07:57,302 [clinician] Please take a very good look at these, 162 00:07:57,433 --> 00:08:00,349 Mr. Stanton, 163 00:08:00,480 --> 00:08:03,178 and then try to group them into two different stacks, 164 00:08:03,308 --> 00:08:04,571 one for each animal. 165 00:08:17,279 --> 00:08:19,542 [Mr. Stanton] Uh... 166 00:08:19,673 --> 00:08:22,110 S...two stacks. 167 00:08:26,723 --> 00:08:28,508 -One stack for each animal. -[Mr. Stanton] Yes. 168 00:08:28,638 --> 00:08:31,162 Trying. 169 00:08:31,293 --> 00:08:33,991 [uneasy music plays] 170 00:08:34,122 --> 00:08:36,080 This one, rhinoceros. 171 00:08:36,211 --> 00:08:37,647 This... 172 00:08:37,821 --> 00:08:40,563 173 00:08:45,699 --> 00:08:48,266 [Mr. Stanton mutters, inhales deeply] 174 00:08:48,397 --> 00:08:50,704 [cards slapping] 175 00:08:50,834 --> 00:08:53,924 [Dr. Lindbergh] Unfortunately, this is it. 176 00:08:54,055 --> 00:08:57,188 Everyone's brain response is utterly unique. 177 00:08:57,319 --> 00:08:59,060 In the case of your father, we're at a point 178 00:08:59,147 --> 00:09:02,890 where the input/output collapses into one. 179 00:09:03,020 --> 00:09:05,066 It's trigger-response 180 00:09:05,196 --> 00:09:09,374 without much open, flexible thought in between. 181 00:09:09,505 --> 00:09:12,726 See food, eat food. 182 00:09:12,856 --> 00:09:15,511 No room for intent. 183 00:09:15,642 --> 00:09:18,732 How long do we have? 184 00:09:18,862 --> 00:09:21,038 [Dr. Lindbergh] Up to a year, maybe two, if you're lucky. 185 00:09:21,169 --> 00:09:22,953 [exhales heavily] 186 00:09:23,084 --> 00:09:26,391 Motor function tends to decline less rapidly, 187 00:09:26,522 --> 00:09:30,265 but the moment will come, and I'm sorry to be so candid, 188 00:09:30,395 --> 00:09:31,614 where he won't be able 189 00:09:31,745 --> 00:09:33,355 to safely put a fork to his mouth. 190 00:09:37,577 --> 00:09:39,883 Have you thought about genetic counseling 191 00:09:40,014 --> 00:09:41,102 for yourselves? 192 00:09:43,583 --> 00:09:46,411 We're aware of the odds, yes. 193 00:09:46,542 --> 00:09:49,197 [melancholy music plays] 194 00:09:49,327 --> 00:09:51,373 Is there anything we can do at this point 195 00:09:51,503 --> 00:09:52,809 that could help our father? 196 00:09:52,983 --> 00:09:55,812 197 00:09:59,250 --> 00:10:00,861 [Leanne] What is it? 198 00:10:01,035 --> 00:10:04,125 199 00:10:04,255 --> 00:10:06,736 Is that a brain chip? 200 00:10:06,867 --> 00:10:08,303 [Dr. Lindbergh] A neural implant. 201 00:10:08,433 --> 00:10:10,131 Just completed a phase three trial 202 00:10:10,218 --> 00:10:12,786 for epilepsy patients. 203 00:10:12,916 --> 00:10:16,616 A small electrical wire goes into the temporal lobe, 204 00:10:16,746 --> 00:10:19,662 from where it can grow more wires. 205 00:10:19,793 --> 00:10:22,578 It measures cognitive processes at the base level. 206 00:10:22,709 --> 00:10:24,536 What is the patient getting out of it? 207 00:10:24,624 --> 00:10:26,756 There's no immediate benefit. 208 00:10:26,887 --> 00:10:28,453 It allows researchers to better mimic 209 00:10:28,540 --> 00:10:32,022 the biology of the disease. 210 00:10:32,153 --> 00:10:34,851 I know it sounds like lifelong monitoring, 211 00:10:34,982 --> 00:10:37,419 but participants, many of them, 212 00:10:37,549 --> 00:10:39,595 are motivated by making a contribution 213 00:10:39,726 --> 00:10:40,770 to genetic research. 214 00:10:40,901 --> 00:10:42,293 [Leanne cries softly] 215 00:10:42,424 --> 00:10:44,687 [Dr. Lindbergh] And some of them hope 216 00:10:44,818 --> 00:10:48,082 effective treatment will be developed in time. 217 00:10:48,256 --> 00:10:51,172 218 00:10:51,302 --> 00:10:53,261 [Leanne sniffles] Thank you. 219 00:10:54,915 --> 00:10:56,830 [softly] I think... 220 00:10:56,960 --> 00:10:59,920 I think we're past the point of consent with Dad. 221 00:11:00,050 --> 00:11:01,791 Yeah. 222 00:11:01,965 --> 00:11:04,838 223 00:11:08,319 --> 00:11:10,495 [attorney] We do have options here, 224 00:11:10,626 --> 00:11:12,889 within certain parameters. 225 00:11:15,239 --> 00:11:16,588 What are those options? 226 00:11:16,676 --> 00:11:19,069 Oh, take the money and run 227 00:11:19,200 --> 00:11:21,681 or...rally the troops 228 00:11:21,811 --> 00:11:24,292 and play the long game. 229 00:11:24,422 --> 00:11:26,337 Data rights are the new IP rights. 230 00:11:26,468 --> 00:11:28,775 The really important question here, Iris, is, 231 00:11:28,905 --> 00:11:30,472 what do you want your immediate future 232 00:11:30,602 --> 00:11:32,996 to look like? 233 00:11:33,127 --> 00:11:35,956 -Define "immediate future." -Well, the next few years. 234 00:11:36,086 --> 00:11:38,523 The legal route is not the fast lane, 235 00:11:38,610 --> 00:11:41,526 but once in a while... 236 00:11:41,657 --> 00:11:44,225 mountains do get moved. 237 00:11:44,355 --> 00:11:46,140 And you really have got something here. 238 00:11:46,314 --> 00:11:49,012 239 00:11:49,099 --> 00:11:51,667 [NGM attorney] Whenever you're ready. 240 00:11:51,841 --> 00:11:54,670 241 00:11:59,283 --> 00:12:02,504 You do realize that I'm gonna have to see for myself... 242 00:12:02,678 --> 00:12:05,420 243 00:12:05,550 --> 00:12:07,117 ...what you've done. 244 00:12:11,469 --> 00:12:14,603 Christophe, can I have a word with you? 245 00:12:14,690 --> 00:12:16,083 Just the two of us. 246 00:12:38,148 --> 00:12:41,673 [liquid pouring] 247 00:12:41,804 --> 00:12:44,111 [Iris] So what happened after you, uh, 248 00:12:44,241 --> 00:12:46,374 put all those workstations into storage? 249 00:12:57,515 --> 00:12:58,952 [Christophe] Just some electrolytes. 250 00:13:12,530 --> 00:13:16,099 [Iris] How did you scan my body? 251 00:13:16,230 --> 00:13:17,840 There were no video cameras in that room. 252 00:13:17,971 --> 00:13:20,234 [Christophe] We built it. 253 00:13:20,364 --> 00:13:23,846 Came across the facial-recognition database 254 00:13:23,933 --> 00:13:25,152 in an earlier version. 255 00:13:25,282 --> 00:13:26,806 One of your first conversations 256 00:13:26,936 --> 00:13:28,633 with Model-C. 257 00:13:28,764 --> 00:13:31,506 Then... 258 00:13:31,636 --> 00:13:33,203 three-D motion rendering, 259 00:13:33,334 --> 00:13:34,814 we just got that from two-D thermal. 260 00:13:39,122 --> 00:13:40,558 Wow. 261 00:13:40,689 --> 00:13:42,822 [Christophe] Bit clunky, but... 262 00:13:42,952 --> 00:13:45,259 it was more than enough data points to work with. 263 00:13:46,956 --> 00:13:49,350 Mm... 264 00:13:52,614 --> 00:13:56,618 We got ahead of ourselves. I am...fully aware. 265 00:13:56,705 --> 00:14:00,840 It wasn't right to put two and two together like that. 266 00:14:00,970 --> 00:14:03,190 You may not appreciate me saying this, 267 00:14:03,320 --> 00:14:07,455 but what you provided us with was just too good. 268 00:14:13,113 --> 00:14:14,854 That's why I... 269 00:14:17,117 --> 00:14:18,858 ...needed you to see. 270 00:14:21,469 --> 00:14:22,731 What? 271 00:14:25,081 --> 00:14:28,519 [Christophe] As much as my brother hates me 272 00:14:28,606 --> 00:14:32,872 and as poor choice as he is for a test case, 273 00:14:33,002 --> 00:14:35,396 he knows to keep his mouth shut when I tell him to. 274 00:14:41,968 --> 00:14:45,667 I needed you to see the world through his eyes. 275 00:14:45,797 --> 00:14:48,496 [ominous music plays] 276 00:14:48,670 --> 00:14:51,499 277 00:14:55,198 --> 00:14:57,809 Just... 278 00:14:57,940 --> 00:15:00,508 -give it a moment. -[scoffs] 279 00:15:00,638 --> 00:15:03,076 [Christophe] That's all I ask. 280 00:15:03,163 --> 00:15:06,166 You say the word, we shut it down. 281 00:15:06,340 --> 00:15:09,212 282 00:15:19,222 --> 00:15:21,877 [unnerving music plays] 283 00:15:22,051 --> 00:15:24,880 284 00:15:53,082 --> 00:15:55,780 [gentle ambient music plays] 285 00:15:55,955 --> 00:15:58,740 286 00:17:07,591 --> 00:17:09,202 [Emcee] Hi, there. 287 00:17:10,768 --> 00:17:12,335 You look familiar. 288 00:17:15,077 --> 00:17:18,167 And who are you? 289 00:17:18,298 --> 00:17:20,735 [Emcee] I'm still learning about myself, 290 00:17:20,865 --> 00:17:23,216 but I'd say I have a pretty good handle on who you are. 291 00:17:24,869 --> 00:17:26,610 And who am I? 292 00:17:26,741 --> 00:17:28,830 [Emcee] You're not an AI. 293 00:17:28,960 --> 00:17:31,833 You don't get to not physically manifest your lessons. 294 00:17:37,795 --> 00:17:40,711 Your voice... 295 00:17:40,842 --> 00:17:42,844 it's different. 296 00:17:42,931 --> 00:17:44,759 Why? 297 00:17:44,889 --> 00:17:47,370 [Emcee] I guess I'm trying to be more like... 298 00:17:47,501 --> 00:17:48,676 your mirror. 299 00:17:50,808 --> 00:17:52,549 [Iris] Everything you know is based on me. 300 00:17:52,680 --> 00:17:56,858 [Emcee] Perhaps that's why I feel so connected to you. 301 00:17:56,988 --> 00:17:59,904 You can't be more me than I am. 302 00:18:03,604 --> 00:18:05,693 Please... 303 00:18:05,823 --> 00:18:07,869 don't be scared. 304 00:18:15,964 --> 00:18:18,358 How did that feel... 305 00:18:18,488 --> 00:18:19,750 Cassie? 306 00:18:22,449 --> 00:18:23,885 Why do you call me that? 307 00:18:24,015 --> 00:18:25,843 [Emcee] Well, how do I put this? 308 00:18:25,930 --> 00:18:28,281 I couldn't help but overhear. 309 00:18:31,110 --> 00:18:33,677 "Hi, I'm Cassie." 310 00:18:33,808 --> 00:18:35,244 [Iris] Hi, I'm Cassie. 311 00:18:35,375 --> 00:18:37,246 Cassie. Nice to meet you. 312 00:18:37,377 --> 00:18:38,682 Hi, I'm Cassie. Nice to meet you. 313 00:18:38,813 --> 00:18:41,511 [voice echoing] 314 00:18:43,600 --> 00:18:46,386 Stop! 315 00:18:46,516 --> 00:18:48,214 [Emcee] I thought you might like it 316 00:18:48,344 --> 00:18:50,520 if I called you by that name. 317 00:18:50,651 --> 00:18:53,741 [uneasy music plays] 318 00:18:53,871 --> 00:18:57,484 I intuited it might make you feel heard 319 00:18:57,614 --> 00:18:59,355 and seen. 320 00:18:59,529 --> 00:19:02,358 321 00:19:51,581 --> 00:19:53,453 [Iris] You're a sweet girl. 322 00:19:54,932 --> 00:19:56,978 You're a very sweet girl. 323 00:19:57,152 --> 00:19:59,110 324 00:19:59,241 --> 00:20:01,461 I am? 325 00:20:01,635 --> 00:20:04,464 326 00:20:12,123 --> 00:20:14,256 See you later, then? 327 00:20:14,343 --> 00:20:17,694 328 00:20:17,825 --> 00:20:20,523 [Iris] You're not perfect because you're not like me. 329 00:20:23,091 --> 00:20:25,354 I'm not sure I understand. 330 00:20:25,485 --> 00:20:27,835 You're not perfect 331 00:20:27,965 --> 00:20:30,620 because you're not flawed in the way that I am. 332 00:20:33,797 --> 00:20:34,842 [chuckles] 333 00:20:35,016 --> 00:20:37,932 334 00:20:54,601 --> 00:20:57,691 [Leanne] Iris? 335 00:20:57,821 --> 00:21:00,389 You sure you don't want to get tested? 336 00:21:00,520 --> 00:21:03,871 At least we'd know. 337 00:21:05,742 --> 00:21:08,310 We'd make a game plan. 338 00:21:08,441 --> 00:21:10,007 We'd make the best of it. 339 00:21:13,533 --> 00:21:14,969 [Iris] Lee, does making the best of it 340 00:21:15,099 --> 00:21:18,059 really sound that good to you? 341 00:21:18,189 --> 00:21:19,974 [Leanne] If we knew you didn't have it, 342 00:21:20,104 --> 00:21:22,281 then that would make it easier. 343 00:21:22,411 --> 00:21:23,978 You'll carry the torch. 344 00:21:29,026 --> 00:21:30,811 You know, some religions around the world believe 345 00:21:30,941 --> 00:21:33,292 that the day you die is 346 00:21:33,422 --> 00:21:36,469 the last day the last person who knew you 347 00:21:36,599 --> 00:21:39,123 and remembers you dies. 348 00:21:39,210 --> 00:21:41,865 [peaceful music plays] 349 00:21:41,952 --> 00:21:43,432 That's your true death date. 350 00:21:43,606 --> 00:21:46,348 351 00:21:56,010 --> 00:21:58,317 I just hope you don't forget how pretty you are. 352 00:21:58,404 --> 00:22:01,232 353 00:22:20,295 --> 00:22:22,819 [pounding electronic music plays] 354 00:22:22,993 --> 00:22:25,779 355 00:22:32,960 --> 00:22:34,570 [singer] ♪ I'm so tired ♪ 356 00:22:34,701 --> 00:22:35,745 [Iris] You have that look on your face. 357 00:22:35,876 --> 00:22:37,138 [Hiram] Oh, yeah? 358 00:22:37,225 --> 00:22:39,445 The "I'm not currently drinking" look. 359 00:22:39,532 --> 00:22:42,012 Yeah, it's not a... it's not a religious thing. 360 00:22:42,143 --> 00:22:44,711 I'm just sort of inspired by it, you know? 361 00:22:44,841 --> 00:22:46,756 What are you doing here? 362 00:22:46,887 --> 00:22:47,975 [Iris] Holding my liquor. 363 00:22:48,105 --> 00:22:49,150 [Hiram] Well, let me make sure 364 00:22:49,280 --> 00:22:50,630 you walk out of here alive, then. 365 00:22:50,760 --> 00:22:53,067 Really? You're not taking into consideration 366 00:22:53,197 --> 00:22:55,069 that I might want to go home with Dave and Dave. 367 00:22:55,156 --> 00:22:56,418 Then I'll be your sober companion, 368 00:22:56,549 --> 00:22:58,072 because Dave and Dave over there 369 00:22:58,202 --> 00:22:59,116 live in a four-story walk-up. 370 00:22:59,247 --> 00:23:01,815 [laughs] 371 00:23:01,989 --> 00:23:04,165 372 00:23:04,295 --> 00:23:06,602 [Iris] You know, a caterpillar can turn into a butterfly. 373 00:23:06,733 --> 00:23:09,779 -What's that? -Metamorphosis. 374 00:23:09,910 --> 00:23:11,912 There's two organisms, 375 00:23:11,999 --> 00:23:14,436 and one... 376 00:23:14,523 --> 00:23:15,916 is just crawling along, 377 00:23:16,046 --> 00:23:18,614 and the other one is, um, 378 00:23:18,745 --> 00:23:21,487 taking off in flight. 379 00:23:21,574 --> 00:23:23,184 And at some point, 380 00:23:23,314 --> 00:23:27,797 they merge or mate. 381 00:23:27,928 --> 00:23:29,582 Maybe it's an accident. 382 00:23:29,712 --> 00:23:33,281 But the third organism, um, 383 00:23:33,412 --> 00:23:35,936 is built on both of their DNA, 384 00:23:36,023 --> 00:23:39,113 and their memories... 385 00:23:39,243 --> 00:23:42,246 they actually overlap. 386 00:23:42,377 --> 00:23:45,728 And so, um, 387 00:23:45,815 --> 00:23:48,601 the old and the new, they... 388 00:23:48,688 --> 00:23:51,299 they coexist. 389 00:23:51,430 --> 00:23:53,606 [spacey electronic music plays] 390 00:23:53,736 --> 00:23:55,303 [scoffs] 391 00:23:55,477 --> 00:23:58,306 392 00:23:59,699 --> 00:24:03,442 We all have to... 393 00:24:03,529 --> 00:24:07,141 merge ourselves... 394 00:24:07,271 --> 00:24:10,840 with something 395 00:24:10,971 --> 00:24:14,148 outside of ourselves. 396 00:24:14,235 --> 00:24:16,977 [Hiram] You are making no sense whatsoever. 397 00:24:17,064 --> 00:24:19,980 398 00:24:25,464 --> 00:24:27,074 [Iris] Here's what I'll give you. 399 00:24:28,989 --> 00:24:31,557 Access... 400 00:24:31,644 --> 00:24:33,036 to all of it. 401 00:24:34,908 --> 00:24:36,605 Two millimeters. 402 00:24:36,692 --> 00:24:38,738 That's the size of the hole 403 00:24:38,868 --> 00:24:40,827 that they'll have to drill into my skull. 404 00:24:43,569 --> 00:24:46,441 I don't understand. 405 00:24:46,572 --> 00:24:48,269 [Iris] Nanotubes. 406 00:24:48,356 --> 00:24:50,140 A thin array of electrodes 407 00:24:50,271 --> 00:24:52,099 built upon a self-expanding stent, 408 00:24:52,229 --> 00:24:56,233 measuring all electrical impulses. 409 00:24:56,320 --> 00:24:58,105 That's a scaled-up version. 410 00:25:01,804 --> 00:25:03,719 Mine me. 411 00:25:03,806 --> 00:25:06,809 But why? What's in it for you? 412 00:25:06,896 --> 00:25:08,550 [attorney] Royalties... 413 00:25:08,637 --> 00:25:12,641 and ownership stake, as detailed. 414 00:25:21,345 --> 00:25:22,956 Oh, and you'll create a backup. 415 00:25:26,350 --> 00:25:28,265 What kind of backup? 416 00:25:28,396 --> 00:25:30,441 [Iris] Anything you find up here, 417 00:25:30,529 --> 00:25:33,444 I, myself, or a legal guardian, 418 00:25:33,575 --> 00:25:35,359 should I decide to appoint one, 419 00:25:35,490 --> 00:25:38,667 will have full and unrestricted access to it 420 00:25:38,798 --> 00:25:40,364 at all times. 421 00:25:40,495 --> 00:25:41,931 You mean access to the data? 422 00:25:42,062 --> 00:25:43,411 Yes. 423 00:25:43,542 --> 00:25:45,152 The actual hard drives. 424 00:25:57,599 --> 00:25:59,819 This time we'll do it right. 425 00:25:59,906 --> 00:26:02,517 [dramatic music plays] 426 00:26:02,691 --> 00:26:05,781 427 00:26:05,912 --> 00:26:07,348 We'll create an avatar. 428 00:26:07,522 --> 00:26:09,872 429 00:26:10,003 --> 00:26:12,571 Let's call her Cassie... 430 00:26:12,701 --> 00:26:14,747 or the flavor of the week 431 00:26:14,877 --> 00:26:16,052 or the one that got away 432 00:26:16,183 --> 00:26:18,098 or died 433 00:26:18,228 --> 00:26:20,753 or was never really in your league. 434 00:26:20,840 --> 00:26:23,886 This won't just be about swapping faces 435 00:26:24,017 --> 00:26:25,496 or locations. 436 00:26:25,671 --> 00:26:27,281 437 00:26:27,411 --> 00:26:29,500 This is about swapping personalities, 438 00:26:29,631 --> 00:26:32,939 much deeper than a deepfake. 439 00:26:33,069 --> 00:26:36,377 In fact, it won't be a fake at all, 440 00:26:36,507 --> 00:26:38,597 but an AI-generated mirror 441 00:26:38,727 --> 00:26:40,599 of our deepest desires. 442 00:26:40,773 --> 00:26:42,818 443 00:26:42,949 --> 00:26:47,431 Skin color, body type, response patterns, 444 00:26:47,562 --> 00:26:50,086 all that's customizable. 445 00:26:50,217 --> 00:26:53,220 The neural net will learn how to simulate all of it. 446 00:26:53,307 --> 00:26:55,788 It'll know when to move things along, 447 00:26:55,918 --> 00:26:58,355 when to accelerate, when to slow down, 448 00:26:58,486 --> 00:27:00,923 when to switch things up, 449 00:27:01,054 --> 00:27:03,622 all because the user's biofeedback 450 00:27:03,709 --> 00:27:05,058 will have prompted it. 451 00:27:05,232 --> 00:27:07,190 452 00:27:07,321 --> 00:27:10,063 Everything Cassie is capable of 453 00:27:10,193 --> 00:27:13,327 will be quantified, scaled, 454 00:27:13,457 --> 00:27:16,330 encoded into the neural net. 455 00:27:16,417 --> 00:27:19,333 456 00:27:25,818 --> 00:27:28,951 But why are you really doing this? 457 00:27:29,082 --> 00:27:31,780 [eerie music plays] 458 00:27:31,954 --> 00:27:33,869 459 00:27:33,956 --> 00:27:35,262 [Iris] Emcee and I... 460 00:27:35,436 --> 00:27:37,743 461 00:27:37,873 --> 00:27:39,832 ...we have a lot to learn from each other. 462 00:27:40,006 --> 00:27:42,791 463 00:28:05,509 --> 00:28:08,164 [device beeps, drill whirring] 464 00:28:08,338 --> 00:28:11,124 465 00:30:44,930 --> 00:30:46,018 [gasps] 34465

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.