All language subtitles for To.Err.Is.Human.A.Patient.Safety.Documentary.2019.1080p.WEBRip.x264.AAC-[YTS.MX]

af Afrikaans
ak Akan
sq Albanian
am Amharic
ar Arabic
hy Armenian
az Azerbaijani
eu Basque
be Belarusian
bem Bemba
bn Bengali
bh Bihari
bs Bosnian
br Breton
bg Bulgarian
km Cambodian
ca Catalan
ceb Cebuano
chr Cherokee
ny Chichewa
zh-CN Chinese (Simplified)
zh-TW Chinese (Traditional)
co Corsican
hr Croatian
cs Czech
da Danish
nl Dutch
eo Esperanto
et Estonian
ee Ewe
fo Faroese
tl Filipino
fi Finnish
fr French
fy Frisian
gaa Ga
gl Galician
ka Georgian
de German
el Greek
gn Guarani
gu Gujarati
ht Haitian Creole
ha Hausa
haw Hawaiian
iw Hebrew
hi Hindi
hmn Hmong
hu Hungarian
is Icelandic
ig Igbo
ia Interlingua
ga Irish
it Italian
ja Japanese
jw Javanese
kn Kannada
kk Kazakh
rw Kinyarwanda
rn Kirundi
kg Kongo
ko Korean
kri Krio (Sierra Leone)
ku Kurdish
ckb Kurdish (Soranรฎ)
ky Kyrgyz
lo Laothian
la Latin
lv Latvian
ln Lingala
lt Lithuanian
loz Lozi
lg Luganda
ach Luo
lb Luxembourgish
mk Macedonian
mg Malagasy
ms Malay
ml Malayalam
mt Maltese
mi Maori
mr Marathi
mfe Mauritian Creole
mo Moldavian
mn Mongolian
my Myanmar (Burmese)
sr-ME Montenegrin
ne Nepali
pcm Nigerian Pidgin
nso Northern Sotho
no Norwegian
nn Norwegian (Nynorsk)
oc Occitan
or Oriya
om Oromo
ps Pashto
fa Persian
pl Polish
pt-BR Portuguese (Brazil) Download
pt Portuguese (Portugal) Download
pa Punjabi
qu Quechua
ro Romanian
rm Romansh
nyn Runyakitara
ru Russian
sm Samoan
gd Scots Gaelic
sr Serbian
sh Serbo-Croatian
st Sesotho
tn Setswana
crs Seychellois Creole
sn Shona
sd Sindhi
si Sinhalese
sk Slovak
sl Slovenian
so Somali
es Spanish
es-419 Spanish (Latin American)
su Sundanese
sw Swahili
sv Swedish
tg Tajik
ta Tamil
tt Tatar
te Telugu
th Thai
ti Tigrinya
to Tonga
lua Tshiluba
tum Tumbuka
tr Turkish
tk Turkmen
tw Twi
ug Uighur
uk Ukrainian
ur Urdu
uz Uzbek
vi Vietnamese
cy Welsh
wo Wolof
xh Xhosa
yi Yiddish
yo Yoruba
zu Zulu
Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated: 1 00:00:02,000 --> 00:00:07,000 Downloaded from YTS.MX 2 00:00:08,000 --> 00:00:13,000 Official YIFY movies site: YTS.MX 3 00:00:17,810 --> 00:00:22,520 It was devastating. It was very hard. 4 00:00:22,520 --> 00:00:25,990 It was hard for us to understand and believe that this could happen 5 00:00:26,280 --> 00:00:31,700 in a developed country like the United States. 6 00:00:33,080 --> 00:00:37,620 Both my husband, Pat, and my son, Cal, experienced 7 00:00:37,620 --> 00:00:41,290 what I would say classic diagnostic errors. 8 00:00:41,290 --> 00:00:44,670 Cal suffered brain damage from his newborn jaundice when it was misdiagnosed 9 00:00:44,670 --> 00:00:49,050 and it was never tested or treated appropriately, 10 00:00:49,050 --> 00:00:51,350 and today he has significant cerebral palsy. 11 00:00:51,350 --> 00:00:54,520 At about 16 hours, the nurse charted that he was yellow, 12 00:00:55,810 --> 00:00:59,360 it was no big deal. We were basically discharged 13 00:00:59,360 --> 00:01:02,980 with a very sick baby, but we were told he was a well baby. 14 00:01:02,980 --> 00:01:06,070 I was familiar with jaundice and it was communicated to us 15 00:01:06,070 --> 00:01:10,200 that it was no big deal and not to worry about it. 16 00:01:10,200 --> 00:01:12,030 And they asked me if I was a first-time mom. 17 00:01:12,040 --> 00:01:14,200 I said I was, and they reminded me that first-time moms are 18 00:01:14,210 --> 00:01:19,080 often over-reactive, and they didn't seem worried at all. 19 00:01:19,090 --> 00:01:23,710 I didn't really know at the time, 20 00:01:23,720 --> 00:01:26,970 but I learned later on that Cal was in the process of dying. 21 00:01:26,970 --> 00:01:32,060 We actually watched our son suffer brain damage in the hospital before our eyes, 22 00:01:33,230 --> 00:01:38,730 and... Quite honestly, that will haunt me forever. 23 00:01:40,070 --> 00:01:44,400 And Pat, my husband, died when he was 45 from cancer, 24 00:01:45,110 --> 00:01:49,780 a cancer that was appropriately diagnosed, 25 00:01:49,790 --> 00:01:52,870 but the pathology failed to get communicated to the doctor or Pat. 26 00:01:52,870 --> 00:01:57,920 They did an MRI and they discovered that there was a mass 27 00:01:57,920 --> 00:02:00,590 in his neck at the base of the skull, and so Pat had surgery. 28 00:02:00,590 --> 00:02:04,220 Six months later, the pain returned in Pat's neck. 29 00:02:04,220 --> 00:02:07,890 A whole series of doctors came through asking Pat 30 00:02:07,890 --> 00:02:10,810 why he never got treatment after his first surgery, 31 00:02:10,810 --> 00:02:14,440 and I had all the documents, and I said, well, because it was benign. 32 00:02:14,440 --> 00:02:17,400 And then by the time the third doctor came through, I said, wait a second, 33 00:02:17,400 --> 00:02:22,030 what was his pathology on the first surgery? 34 00:02:22,030 --> 00:02:25,160 And the final pathology was a high-grade malignant synovial cell sarcoma. 35 00:02:25,160 --> 00:02:27,100 And that document either never arrived or was 36 00:02:27,100 --> 00:02:29,620 placed in his chart without the doctor seeing. 37 00:02:30,830 --> 00:02:36,040 And I remember showing it to Pat, and I remember Pat crying. 38 00:02:36,040 --> 00:02:41,260 You know, to think that another error had taken place, and this time with him, 39 00:02:41,260 --> 00:02:46,090 that was difficult for us to witness in our healthcare system. 40 00:02:46,100 --> 00:02:49,720 [somber music] 41 00:02:49,730 --> 00:02:51,980 Stories like Sue Sheridan and what happened to her, 42 00:02:51,980 --> 00:02:55,900 where a small mistake can really be a life-altering event that remind us the human cost 43 00:02:56,150 --> 00:03:01,490 of what we're talking about. These are not theoretical events. 44 00:03:01,490 --> 00:03:04,280 These are not just things that happen to other people. They happen to us. 45 00:03:04,280 --> 00:03:07,540 They happen to our families. And they are things that we need to work on. 46 00:03:07,540 --> 00:03:11,750 [tense music] 47 00:03:12,210 --> 00:03:17,880 [narrator]In 1999, the first significant report on medical mistakes 48 00:03:17,880 --> 00:03:22,010 was released by the Institute of Medicine. They called it To Err Is Human. 49 00:03:22,010 --> 00:03:24,710 This report claimed that as manyas 98,000 people 50 00:03:24,710 --> 00:03:27,850 die every yearnas a result of medical mistakes. 51 00:03:28,980 --> 00:03:34,610 Over the next 15 years, efforts to better understand this number increased, 52 00:03:34,900 --> 00:03:39,820 but so did the number itself. 53 00:03:41,200 --> 00:03:44,540 Recent studies have raised the projected number of deaths to as high as 440,000. To 54 00:03:44,540 --> 00:03:48,470 put this in perspective, that's more than the number of graves in Arlington Cemetery. 55 00:03:48,750 --> 00:03:53,960 It's the equivalent of 2-3 jumbonjets crashing every single day. 56 00:03:53,960 --> 00:03:58,880 So, where does that rank medical mistakes on the leading causes of death? 57 00:03:58,880 --> 00:04:03,100 Number three. Right behind cancer and heart disease. 58 00:04:04,930 --> 00:04:08,480 Now suddenly, whoa. This isn't just some egg-headed study. 59 00:04:10,100 --> 00:04:14,860 This is a big deal. This could be you, and they're right. 60 00:04:14,860 --> 00:04:17,610 Wait a second, you mean those hospitals, my local hospital was killing people? 61 00:04:17,610 --> 00:04:21,870 Is that what you're really saying? 62 00:04:21,870 --> 00:04:24,160 We could prevent many, many, many of these deaths immediately 63 00:04:24,160 --> 00:04:28,660 if we just put in the effort. 64 00:04:28,670 --> 00:04:30,830 Things are happening. Let's take a look at this. 65 00:04:30,840 --> 00:04:33,630 I just think this is like a massive epidemic that we have underestimated, 66 00:04:33,630 --> 00:04:37,840 and the reason is because it's happening to people who are already sick. 67 00:04:37,840 --> 00:04:43,260 But, like, they were sick, that doesn't mean they were going to die. 68 00:04:43,260 --> 00:04:45,850 And their death is no less of a tragedy because they already had a medical problem. 69 00:04:45,850 --> 00:04:50,860 Every time you get on a plane, you don't expect that plane to crash. 70 00:04:50,860 --> 00:04:53,980 And everybody who dies in a plane crash, you say, "Well, those people were healthy. 71 00:04:53,990 --> 00:04:57,150 They were going to do fine otherwise." I think the problem with patient safety 72 00:04:57,160 --> 00:05:00,620 is you say, "Oh, well, these people were sick anyway." 73 00:05:00,620 --> 00:05:02,700 And I think it's a very problematic way to look at the world. 74 00:05:02,700 --> 00:05:04,110 Maybe they didn't die, but they spend the rest 75 00:05:04,110 --> 00:05:05,580 of their life in a wheelchair or a nursing home 76 00:05:05,580 --> 00:05:09,460 and that accelerates their death and obviously harms their quality of life. 77 00:05:09,460 --> 00:05:12,710 So, the numbers about deaths are a big deal, 78 00:05:12,710 --> 00:05:16,010 but in some ways they underestimate the overall toll of preventable harm. 79 00:05:16,010 --> 00:05:20,140 We don't have a stable, agreed way to measure safety or injuries. 80 00:05:20,390 --> 00:05:25,390 Actually, the number you get depends on how you look. 81 00:05:25,390 --> 00:05:28,520 One rule is the harder you look, the more you find. 82 00:05:28,520 --> 00:05:31,110 So, when you really throw the book at it and you do everything 83 00:05:31,110 --> 00:05:35,070 you can to look for injuries, you're going to find a ton of them. 84 00:05:35,070 --> 00:05:38,120 When people start debating, you know, is it 40,000 or 90,000 or 100,000? 85 00:05:38,120 --> 00:05:43,830 Uh, it's a lot. It's a ton. And our job is to make it zero. 86 00:05:43,910 --> 00:05:48,880 This is urgent. It's a public health emergency. 87 00:05:48,880 --> 00:05:53,220 [narrator] While the number ofdeaths related to medical error is staggering, 88 00:05:54,510 --> 00:05:58,220 the number of patients who experience non-fatal errors is even bigger. 89 00:05:58,220 --> 00:06:00,400 Recent studies suggest one-third of all hospital 90 00:06:00,400 --> 00:06:02,770 admissions experience a medical mistake, 91 00:06:03,230 --> 00:06:08,690 and 1.7 million hospital-acquired infections occur every year. 92 00:06:08,690 --> 00:06:13,450 69% of those infections could have been prevented 93 00:06:13,950 --> 00:06:17,280 through methods that already exist, like hand washing. 94 00:06:17,280 --> 00:06:20,910 But healthcare workers wash their hands less than 50% of the time, 95 00:06:20,910 --> 00:06:25,210 with some research suggesting it's as low as 30%. 96 00:06:25,210 --> 00:06:28,880 There are even more dramatic examples. In a five-year span, 97 00:06:29,510 --> 00:06:33,970 surgeons operated on the wrong body part over 2,000 times, 98 00:06:33,970 --> 00:06:38,640 left nearly 5,000 tools inside patients, 99 00:06:38,930 --> 00:06:42,310 and in 27 cases operated on the wrong patient entirely. 100 00:06:42,690 --> 00:06:47,360 But diagnostic errors, like the ones that left Cal Sheridan with cerebral palsy 101 00:06:48,360 --> 00:06:50,640 and delayed the detection ofPat Sheridan's 102 00:06:50,640 --> 00:06:53,030 cancer contribute to 1 in 10 patient deaths. 103 00:06:53,030 --> 00:06:55,950 But whether it's a diagnostic error or any 104 00:06:55,950 --> 00:06:59,200 other preventable harm, the only way to fix it 105 00:06:59,200 --> 00:07:04,210 is to first understand what causes it. 106 00:07:04,210 --> 00:07:07,300 [Boaz Keysar] When we study communication in my lab, we look at how people communicate 107 00:07:07,300 --> 00:07:12,180 and what are the reasons for miscommunication. 108 00:07:12,180 --> 00:07:13,980 In very simple experiments, when we ask people 109 00:07:13,980 --> 00:07:16,160 to communicate something to somebody else, 110 00:07:17,100 --> 00:07:22,020 about 50% of the time when they thought the other person understood them, 111 00:07:22,600 --> 00:07:26,730 they were wrong. Now, I don't know the extent of miscommunication in medicine, 112 00:07:26,730 --> 00:07:32,110 but I am sure it is more than, uh, physicians think. 113 00:07:32,280 --> 00:07:38,370 Part of the problem is that when you, when, when, say a doctor miscommunicates, 114 00:07:38,370 --> 00:07:43,000 he or she might not know. That's the core of the problem, right? 115 00:07:43,710 --> 00:07:47,800 They might not get immediate feedback that they miscommunicated. 116 00:07:47,800 --> 00:07:51,130 And if that happens, then that error 117 00:07:51,140 --> 00:07:56,430 could amplify without anybody realizing that the source was 118 00:07:56,560 --> 00:08:01,060 just a minor miscommunication. 119 00:08:01,060 --> 00:08:04,020 Now I know how... what happened to my husband. Now I understand how it happened, 120 00:08:04,020 --> 00:08:08,700 that there's been no system-based intervention to ensure 121 00:08:09,780 --> 00:08:14,450 that lab tests are followed up on, 122 00:08:14,450 --> 00:08:17,120 that pathologies and radiology reports are followed up on. 123 00:08:17,120 --> 00:08:20,630 To know that this happens in our country, that's unacceptable. 124 00:08:23,170 --> 00:08:25,420 [siren wailing] 125 00:08:27,380 --> 00:08:28,890 [tense music] 126 00:08:30,930 --> 00:08:36,230 [narrator 2]Most of us think of a hospital as a place 127 00:08:36,230 --> 00:08:38,600 Nwhere people go after they have an accident, 128 00:08:38,600 --> 00:08:41,730 not as a place where people go to have accidents. 129 00:08:41,730 --> 00:08:45,320 However, like just about any place, there are safety hazards in a hospital. 130 00:08:46,240 --> 00:08:51,080 Some are unique to the hospital environment, and some are not. 131 00:08:51,580 --> 00:08:56,040 Generally, the hospital staff is very aware of medical safety practices, 132 00:08:56,040 --> 00:09:01,040 such as the proper handling of infectious cases, 133 00:09:01,040 --> 00:09:04,800 careful checking of patient ID before administering any medication, 134 00:09:04,800 --> 00:09:09,760 keeping things sanitary and disinfected. 135 00:09:09,760 --> 00:09:11,540 Yet, all of us at times tend tooverlook some 136 00:09:11,550 --> 00:09:13,600 potential hazards that we are around every day. 137 00:09:13,600 --> 00:09:18,310 We must try to learn to think safety in everything we do. 138 00:09:19,020 --> 00:09:23,690 But safety doesn't come just by learning a lot of rules. It comes from an attitude. 139 00:09:24,150 --> 00:09:29,030 For everyone who works in a hospital, safety has to be a full time job. 140 00:09:33,040 --> 00:09:36,340 [Albert Wu] This is a problem that's, you know, hiding in plain sight. 141 00:09:38,920 --> 00:09:42,670 And I think that no one is really surprised when they think about it for a minute. 142 00:09:42,670 --> 00:09:47,510 If we think the amount of harm that is currently existing is just fine, 143 00:09:47,510 --> 00:09:51,600 then maybe it's not a crisis, it's not a problem. If that's okay, then we're done. 144 00:09:51,600 --> 00:09:56,730 Most of us in medicine just said, "Well, that's the way it is, you know. 145 00:09:56,730 --> 00:10:00,980 Things go wrong. People make mistakes. There's nothing you can do about it." 146 00:10:00,980 --> 00:10:05,030 It's pretty obvious that safety is not number one priority in most hospitals. 147 00:10:05,030 --> 00:10:09,700 When it is, wonderful things happen. 148 00:10:09,700 --> 00:10:12,540 What is the problem you're trying to solve? 149 00:10:12,540 --> 00:10:15,670 And the answer is, for most hospital administrators, 150 00:10:15,670 --> 00:10:19,090 life is too short to get the doctors angry at you. 151 00:10:19,090 --> 00:10:22,670 Building a new cancer center, your oncologists love you, 152 00:10:22,670 --> 00:10:26,390 the other doctors love you, it brings in revenue, the community loves you. 153 00:10:26,390 --> 00:10:29,760 If you reduce medical error, you can't advertise it because 154 00:10:29,770 --> 00:10:33,390 the patients all think that everything's safe anyway. Nobody knows the problem exists. 155 00:10:33,400 --> 00:10:37,360 The doctors are angry because you start to talk about medical error. 156 00:10:37,360 --> 00:10:40,820 So, that's why you have an invisible problem. 157 00:10:40,820 --> 00:10:44,490 Every human being will make mistakes, and will... so the goal cannot be zero errors. 158 00:10:44,490 --> 00:10:49,910 Our goal needs to be zero harm, because we know errors will occur. 159 00:10:49,910 --> 00:10:53,960 So, how do we make sure those errors don't actually lead to harm and are caught early? 160 00:10:53,960 --> 00:10:59,260 10 or 15 years ago, we thought central line infections were not preventable. 161 00:10:59,260 --> 00:11:04,010 We thought that was part of kind of doing business in healthcare that, okay, 162 00:11:04,010 --> 00:11:08,060 people have central lines, occasionally they'll get infections, 163 00:11:08,060 --> 00:11:11,020 and that's just... Now we know infections can go down to zero. 164 00:11:11,020 --> 00:11:16,150 Preventing preventable harm is a skill and a commitment and a technology 165 00:11:17,780 --> 00:11:23,280 all of its own. It's not glamorous, but it's what keeps all of us safe. 166 00:11:23,280 --> 00:11:28,040 If you believe, "First, do no harm", 167 00:11:28,040 --> 00:11:31,670 there is no excuse for not investing in things which will prevent harm. 168 00:11:32,080 --> 00:11:37,050 Health care nowadays is incredibly complicated. 169 00:11:37,050 --> 00:11:40,090 A patient has literally hundreds of things done to them, 170 00:11:40,090 --> 00:11:44,720 having blood drawn for a test or getting an x-ray or whatever. 171 00:11:44,720 --> 00:11:48,810 And so, there are many, many, many opportunities for things to go wrong. 172 00:11:48,810 --> 00:11:53,230 So, even when nurses and doctors and technicians and radiologists 173 00:11:53,230 --> 00:11:57,490 are functioning at a 99% level, which is, you know, pretty good for human activity, 174 00:11:57,490 --> 00:12:02,660 that still means a lot of opportunity for things to go wrong. 175 00:12:02,660 --> 00:12:07,000 I think this is a general problem that you have when you deal with people. 176 00:12:07,000 --> 00:12:12,630 We are not built to not make mistakes. We are not built to be perfect. 177 00:12:13,170 --> 00:12:16,090 Are you going to try and change the person or 178 00:12:16,090 --> 00:12:19,130 are you going to try and change the situation? 179 00:12:19,130 --> 00:12:24,430 One way to do it is to design, say the work environment, 180 00:12:24,430 --> 00:12:30,020 in a way that would not necessarily prevent the error, 181 00:12:32,400 --> 00:12:36,860 but would assume the error. 182 00:12:36,860 --> 00:12:39,950 We have to acknowledge that to err is human, and then to figure out 183 00:12:39,950 --> 00:12:45,000 what do we do with that fact in terms of building a system that's safe for patients. 184 00:12:45,160 --> 00:12:49,830 [tense music] 185 00:12:50,290 --> 00:12:55,630 [Sue Sheridan] Between Cal's patient safety event and Pat's patient safety event, 186 00:12:55,630 --> 00:13:00,180 we had Mackenzie in the middle there. Exactly at 16 hours, just like Cal, 187 00:13:00,180 --> 00:13:04,890 she also had a very high bilirubin, which the hospital took action. 188 00:13:05,730 --> 00:13:10,770 They tested it and they treated it. I took a shower. 189 00:13:10,770 --> 00:13:14,650 And it was the first shower after delivery and I remember 190 00:13:14,650 --> 00:13:17,780 I stayed in the shower for an hour and they sent a female chaplain in, and I was crying. 191 00:13:17,780 --> 00:13:19,970 And the chaplain thought I was crying because my 192 00:13:19,970 --> 00:13:22,580 daughter was getting treated for her jaundice 193 00:13:22,620 --> 00:13:28,290 and I explained to them I was not crying because of that. I was crying because 194 00:13:28,290 --> 00:13:32,170 I witnessed what the only thing they had to do with my son, 195 00:13:32,170 --> 00:13:35,970 that it was so easy to prevent what happened to my son. 196 00:13:37,050 --> 00:13:39,810 [tense music] 197 00:13:42,350 --> 00:13:45,090 [Mackenzie Sheridan] When I got into about first grade, people started asking me, 198 00:13:45,090 --> 00:13:47,770 "What's wrong with your brother? Why, like, can't he move like the rest of us?" 199 00:13:47,770 --> 00:13:51,440 I didn't really get it, because I was never told necessarily, 200 00:13:51,440 --> 00:13:55,990 you know, your brother has cerebral palsy, your brother has kernicterus. 201 00:13:55,990 --> 00:13:59,490 You know, to me, he was just my brother. 202 00:14:05,920 --> 00:14:10,630 [ambient music] 203 00:14:40,660 --> 00:14:46,500 [Mackenzie Sheridan] Recently, I became more interested in the case, my brother's case, 204 00:14:46,500 --> 00:14:52,380 because I knew, before looking it up, I knew that he wasn't given a bilirubin test 205 00:14:52,800 --> 00:14:57,970 and because of that he got cerebral palsy and kernicterus. 206 00:14:57,970 --> 00:15:01,810 And I got frustrated and I got angry and confused and 207 00:15:01,810 --> 00:15:04,720 my mom has taught me that I can do something 208 00:15:04,720 --> 00:15:07,690 positive with that kind of anger and fervor. 209 00:15:07,690 --> 00:15:13,070 I can, you know, go out and make sure that those kind of things don't happen. 210 00:15:20,330 --> 00:15:25,420 So, I used to be a little scared hearing all of the things 211 00:15:25,420 --> 00:15:28,550 that could go wrong in the health system. 212 00:15:28,550 --> 00:15:31,550 I just learned to be cautious and to ask questions and to, you know, 213 00:15:31,550 --> 00:15:36,600 ask the doctors, "What are you doing? Have you washed your hands? Have you done this?" 214 00:15:36,600 --> 00:15:39,930 I look at doctors in a different sense than, I think, 215 00:15:39,930 --> 00:15:42,650 a lot of people do and as a child I looked at doctors differently as well. 216 00:15:42,650 --> 00:15:44,300 I know why kids would think like a doctor don't 217 00:15:44,300 --> 00:15:45,980 make mistakes, but I knew from a very young age 218 00:15:45,980 --> 00:15:50,450 that they do, and that their mistakes could cost a life. 219 00:15:50,450 --> 00:15:54,370 The first thing that we wanted was to tell somebody. 220 00:15:54,370 --> 00:15:56,700 Some kind of high authority that could tell all of the hospitals 221 00:15:56,700 --> 00:16:01,460 about what happened, so all hospitals could implement change. 222 00:16:01,460 --> 00:16:05,340 And I thought somebody was in charge of patient safety in the United States, 223 00:16:05,340 --> 00:16:10,340 and I learned that that simply does not exist. 224 00:16:12,550 --> 00:16:18,180 When people think about science in healthcare, they think about genes and cells 225 00:16:18,190 --> 00:16:23,650 and drugs and chemistry. Yeah, that's science. That's one science. 226 00:16:23,650 --> 00:16:28,360 But, there's another science, which is the science of organizing care, 227 00:16:28,360 --> 00:16:30,470 which is how to you actually get the help, what 228 00:16:30,480 --> 00:16:32,620 are the flows like, how do you do surgery. 229 00:16:32,620 --> 00:16:38,080 How do you take care of a chronic illness. There's science there too, 230 00:16:38,080 --> 00:16:42,420 and luckily this country began investing in that really in the past few decades. 231 00:16:42,420 --> 00:16:46,300 The Agency for Healthcare Research and Quality, for example, it's an American 232 00:16:46,300 --> 00:16:49,590 investment in developing the sciences for delivering better care. 233 00:16:49,590 --> 00:16:54,260 [narrator] In 2000, after speaking with leaders in healthcare, 234 00:16:54,270 --> 00:16:58,190 President Bill Clinton made a bold statement regarding 235 00:16:58,190 --> 00:17:00,980 the country's new efforts in managing medical errors. 236 00:17:00,980 --> 00:17:04,740 Just think about it, we can cut preventable medical errors in half in five years. 237 00:17:05,280 --> 00:17:11,030 [narrator] The Agency forHealthcare Research and Quality took on this task. 238 00:17:11,030 --> 00:17:16,200 Today, AHRQ remains focused on improving the quality 239 00:17:16,210 --> 00:17:20,000 and safety of healthcare for Americans. 240 00:17:20,000 --> 00:17:23,010 It does so by funding research, developing tools and training, 241 00:17:23,920 --> 00:17:27,840 and collecting measures and data on the healthcare system as a whole. 242 00:17:27,840 --> 00:17:31,640 In 2016, a report was released on the recent progress in patient safety efforts. 243 00:17:32,060 --> 00:17:37,440 The report showed that fromn2010-2015, there were 3 million 244 00:17:37,440 --> 00:17:42,030 fewer hospital-acquired conditions, showing a 21% reduction. 245 00:17:42,030 --> 00:17:47,490 125,000 deaths were prevented, saving $28 billion in healthcare costs. 246 00:17:47,530 --> 00:17:53,250 All with a budget that annually hovered between $400-$450 million. 247 00:17:53,250 --> 00:17:58,960 But it's part of a healthcare system that spends over $3 trillion, 248 00:17:58,960 --> 00:18:03,260 and has more than 5,000 hospitals, with over 800,000 physicians, 4 million nurses, 249 00:18:03,260 --> 00:18:09,600 and 330 million patients. That means the agency is working with 1/100th of a percent 250 00:18:09,600 --> 00:18:16,060 of national health spending and is tasked with improving the other 99.99%. 251 00:18:16,940 --> 00:18:21,690 It is such an underinvestment that, you know, a doubling of the amount 252 00:18:22,450 --> 00:18:27,200 for the agency would be a vast improvement, but it still is not nearly enough. 253 00:18:27,200 --> 00:18:31,910 We need this information for us to take care of our patients properly, for health plans, 254 00:18:31,910 --> 00:18:36,630 for leaders of large clinics to say, "Actually, no, I need to better understand 255 00:18:36,630 --> 00:18:41,880 the choices I make, how it impacts our ability to deliver safe care." 256 00:18:41,880 --> 00:18:46,090 It has funded some of the seminal studies that have had massive improvements 257 00:18:46,100 --> 00:18:49,770 in patient safety. So, it funded the studies that led us to create the checklists for 258 00:18:49,770 --> 00:18:54,350 central line infections. That alone has saved the American healthcare system 259 00:18:54,350 --> 00:18:59,110 hundreds of millions of dollars, if not billions of dollars, 260 00:18:59,110 --> 00:19:01,530 but more importantly, has probably saved tens of thousands of lives. 261 00:19:01,530 --> 00:19:04,410 [Ashish Jha] There are tens of thousands of Americans walking around today 262 00:19:04,410 --> 00:19:08,500 who would be dead if it had not been for some of the work that AHRQ has funded. 263 00:19:08,790 --> 00:19:14,290 It's really about how we apply the best of science 264 00:19:15,040 --> 00:19:19,760 to your individual needs and preferences. 265 00:19:19,760 --> 00:19:23,010 To some extent I do know some systems that are doing a terrific job, 266 00:19:23,010 --> 00:19:26,970 and when I learn from them about how they are doing it, a lot of them 267 00:19:26,970 --> 00:19:30,770 are using the tools and methods pioneered by AHRQ. 268 00:19:35,820 --> 00:19:40,530 [ambient music] 269 00:19:41,160 --> 00:19:43,700 Much of the work that we use to train around patient safety 270 00:19:43,700 --> 00:19:47,450 and how to make healthcare safer is actually derived from AHRQ research and tools. 271 00:19:47,450 --> 00:19:53,040 When they put out a toolkit or research tools, I know that they've been vetted 272 00:19:53,040 --> 00:19:58,090 and they've been tried and investigated and shown to be of benefit. 273 00:19:58,090 --> 00:20:02,720 So, a big problem that we face in safety in hospitals is really improving handoffs, 274 00:20:02,890 --> 00:20:07,930 which is when a patient moves from one area to another 275 00:20:07,930 --> 00:20:11,270 or when their doctors or nurses change shifts. 276 00:20:11,270 --> 00:20:12,780 Handoffs are somewhat invisible to patients, 277 00:20:12,780 --> 00:20:14,360 but they actually have a huge impact on them. 278 00:20:14,360 --> 00:20:19,150 Like, if an average patient got hospitalized tomorrow, 279 00:20:19,160 --> 00:20:21,990 they would face upwards of 15 handovers. 280 00:20:21,990 --> 00:20:24,450 And we know from AHRQ-funded research, it's got to be more than just 281 00:20:24,450 --> 00:20:28,080 a passive listening where you're like, uh-huh, okay, I got it, 282 00:20:28,080 --> 00:20:31,880 but really engage, ask questions, because often times you'll pick up things. 283 00:20:31,880 --> 00:20:36,670 Combining AHRQ TeamSTEPPS, with a standardized tool to improve handoffs 284 00:20:37,340 --> 00:20:42,140 actually led to a 30% reduction in preventable adverse events. 285 00:20:42,140 --> 00:20:46,680 We also develop our own home-grown patient safety teaching programs. 286 00:20:46,680 --> 00:20:51,310 One of my personal favorites that we've actually 287 00:20:51,310 --> 00:20:53,900 developed here is called the Room of Horrors. 288 00:20:53,900 --> 00:20:57,110 We take 10 patient safety hazards 289 00:20:57,110 --> 00:20:59,740 and we embed it into a hospital room, into a simulation. 290 00:20:59,740 --> 00:21:02,870 This is training where you're walking into a room 291 00:21:02,870 --> 00:21:05,960 and you're actually seeing with your own eyes, can you spot what's wrong? 292 00:21:05,960 --> 00:21:07,970 [Trainee 1] Ammonia. C-diff positive. So, 293 00:21:07,970 --> 00:21:10,920 probably should be some kind of like precautions. 294 00:21:11,670 --> 00:21:16,590 [Trainee 2] Yeah, he should be contacted less. 295 00:21:16,590 --> 00:21:18,430 Allergies, latex and penicillin. That's fine. 296 00:21:18,430 --> 00:21:22,060 Umm. Let's see here. Oh, those are gloves over there. 297 00:21:22,060 --> 00:21:26,390 [Trainee 2] Are these latex gloves? Uh-oh, we got latex gloves. 298 00:21:27,020 --> 00:21:31,480 So, it looks like he's got some [unintelligible] hanging, 299 00:21:31,480 --> 00:21:33,780 and he's allergic to penicillin so that's definitely not ideal. 300 00:21:33,780 --> 00:21:38,530 [Trainee 1] Yes, absolutely. Why does he have magnesium? 301 00:21:38,530 --> 00:21:41,370 I don't know. It's actually not for his name. His name is Washington, right? 302 00:21:41,370 --> 00:21:45,830 [Trainee 1] Yeah. Michael Johnson. Alright. 303 00:21:45,830 --> 00:21:47,960 Different Michael. I'm also going to put the stress ulcer. 304 00:21:47,960 --> 00:21:48,670 Okay. Good call. 305 00:21:48,670 --> 00:21:54,050 [Vinny Arora] They have 10 minutes to identify all the hazards that they can, 306 00:21:54,050 --> 00:21:57,680 and then right after, when they come out, I actually debrief with them, 307 00:21:57,680 --> 00:22:01,020 so we go over how they did, not only what they got right, 308 00:22:01,020 --> 00:22:05,230 where did they miss things, and perhaps why did they miss those things. 309 00:22:05,230 --> 00:22:08,690 If you train people this way, this is the way their brain is running in the background. 310 00:22:08,690 --> 00:22:10,950 Every time they enter a room they can automatically 311 00:22:10,960 --> 00:22:12,650 spot it from the corner of their eye. 312 00:22:12,650 --> 00:22:14,850 As an organization, we cannot improve patient 313 00:22:14,850 --> 00:22:16,990 safety unless we have front line personnel, 314 00:22:16,990 --> 00:22:22,450 including our residents and nurses and everyone else that works in healthcare 315 00:22:22,460 --> 00:22:25,540 raising their hand to say, "Hey, I saw something wrong." 316 00:22:25,540 --> 00:22:28,290 And so that's why it's really important to embed people into a clinical situation 317 00:22:28,300 --> 00:22:32,800 where they are able to recognize what types of events they should report. 318 00:22:32,800 --> 00:22:37,430 [somber music] 319 00:22:38,510 --> 00:22:44,190 [Bob Wachter] Probably the most important foundational thinker in the field of 320 00:22:44,190 --> 00:22:48,070 patient safety is a gentleman by the name of James Reason, 321 00:22:48,070 --> 00:22:51,150 who is now retired or semi-retired psychologist in Manchester, England. 322 00:22:51,150 --> 00:22:52,850 What Reason was doing was, as a psychologist, 323 00:22:52,860 --> 00:22:54,740 studying what he called organizational accidents. 324 00:22:54,740 --> 00:22:59,700 How did terrible errors and accidents and harm happen in industries, 325 00:22:59,700 --> 00:23:04,670 whether it was nuclear power or space shuttles or intelligence failures in the CIA? 326 00:23:04,670 --> 00:23:07,510 So, he studied a bunch of them, and what he 327 00:23:07,510 --> 00:23:10,670 found was the same pattern over and over again. 328 00:23:10,670 --> 00:23:12,760 What he found was if you look at it superficially, 329 00:23:12,760 --> 00:23:14,590 you would see a human being who screwed up. 330 00:23:14,590 --> 00:23:19,720 That was the superficial understanding. It was easy because it fit with 331 00:23:19,730 --> 00:23:22,600 the human model that I need to blame somebody 332 00:23:22,600 --> 00:23:25,230 and if I can just point a finger, you know, I have solved a problem. 333 00:23:25,230 --> 00:23:28,190 What was really right was that in unsafe organizations, 334 00:23:28,190 --> 00:23:32,030 these organizational accidents happen because of a long chain of events 335 00:23:32,030 --> 00:23:36,910 that allowed that human error, sometimes several human errors to cause terrible harm. 336 00:23:36,910 --> 00:23:42,120 So, he came up with a model that, to me, I remember the first time I read this, 337 00:23:42,130 --> 00:23:45,380 it's called the Swiss Cheese Model. A little lightbulb went off and I said, 338 00:23:45,380 --> 00:23:48,260 "Aha! Oh, now I get it." And now I look back on errors 339 00:23:48,260 --> 00:23:51,880 I have seen through my entire career, and now it makes sense. 340 00:23:51,890 --> 00:23:55,350 Organizations build in protections to block 341 00:23:55,350 --> 00:23:58,770 those simple human glitches from causing harm. 342 00:23:58,770 --> 00:24:01,640 The problem is, those layers of protections 343 00:24:01,650 --> 00:24:04,730 he likened to pieces of Swiss cheese, they all have holes. 344 00:24:04,730 --> 00:24:07,240 If I kind of blow something one day, I kind of forget something, 345 00:24:07,240 --> 00:24:08,580 or write something in the wrong space, most 346 00:24:08,580 --> 00:24:10,070 days the first layer of Swiss cheese blocks it. 347 00:24:10,070 --> 00:24:15,410 But, on a bad day, the first layer misses. It goes through the hole 348 00:24:15,410 --> 00:24:19,160 and it hits the second layer and the second layer blocks it. 349 00:24:19,170 --> 00:24:21,830 When we kill someone in medicine because we gave them the wrong medicine or cut off 350 00:24:21,830 --> 00:24:26,590 the wrong leg or there's a space shuttle crash or Three Mile Island and you look back, 351 00:24:26,590 --> 00:24:31,970 you realize there were a lot of layers, each one of them had a lot of holes, 352 00:24:31,970 --> 00:24:36,730 and also that particular day the karma was pretty terrible 353 00:24:36,730 --> 00:24:40,600 and it just happened to be that all of the holes aligned. 354 00:24:40,610 --> 00:24:43,360 And that's how the error made it through all of these quote "protections" 355 00:24:43,360 --> 00:24:46,900 to cause terrible harm. My instinct was no longer, "Let me figure out who screwed up." 356 00:24:46,900 --> 00:24:51,950 My instinct was now Swiss cheese. It became automatic. 357 00:24:51,950 --> 00:24:53,730 Here's a bad error, what's the Swiss cheese? 358 00:24:53,730 --> 00:24:55,580 What are the layers of protection that we had 359 00:24:55,580 --> 00:25:00,290 that failed, how do we shrink the size of the holes, and how do we create enough overlap 360 00:25:00,290 --> 00:25:03,180 in layers of cheese so an error never makes it 361 00:25:03,180 --> 00:25:06,760 through all those layers to cause terrible harm? 362 00:25:07,590 --> 00:25:12,310 [ambient music] 363 00:25:43,220 --> 00:25:48,890 [Sue Sheridan] With Pat, I actually spoke to the pathologist about why he didn't 364 00:25:48,890 --> 00:25:54,430 pick up the phone and call the neurosurgeon when they learned it was cancer, 365 00:25:54,440 --> 00:25:59,730 and it was a rare kind of cancer, and his answer was, "It's not my job." 366 00:26:05,450 --> 00:26:09,660 [Mackenzie Sheridan] The doctor told our family, you know, your dad is fine. 367 00:26:09,660 --> 00:26:12,750 He's benign, the tumor is benign, everything's great, go on and live your life. 368 00:26:12,750 --> 00:26:16,170 A few months later my dad got very sick. 369 00:26:16,170 --> 00:26:18,010 [Sue Sheridan] And I got the documents from the 370 00:26:18,010 --> 00:26:20,230 neurosurgeon and it said that the pathology 371 00:26:20,510 --> 00:26:25,380 was an atypical spindle cell neoplasm, which the doctor said was benign. 372 00:26:25,390 --> 00:26:30,350 We expected the hospital to fully describe to us what had happened, to... 373 00:26:31,390 --> 00:26:37,400 you know, take care of us, and we were discharged without any explanation. 374 00:26:38,020 --> 00:26:43,070 So, we left there with all the documents in our hands 375 00:26:43,070 --> 00:26:45,410 with absolutely no explanation that this was an error. 376 00:26:45,410 --> 00:26:48,910 [sighs] 377 00:26:50,250 --> 00:26:53,060 I think our first reaction was fear. We were scared. It scared 378 00:26:53,060 --> 00:26:56,250 us that a hospital, a well-known hospital, with professionals, 379 00:26:56,250 --> 00:27:01,300 would intentionally cover up that kind of information. 380 00:27:01,300 --> 00:27:04,840 So, the first, the first emotion was fear. 381 00:27:04,840 --> 00:27:08,600 One day, Pat woke up paralyzed from his waist down, 382 00:27:08,600 --> 00:27:12,480 and we're at home in Boise, Idaho, and we thought maybe he had a stroke. 383 00:27:12,480 --> 00:27:17,150 We learned then that his cancer had returned explosively. 384 00:27:17,530 --> 00:27:21,490 They estimated he had about 10 days to live. 385 00:27:22,410 --> 00:27:25,450 [Mackenzie Sheridan] And I remember my mom sitting Cal and I down right before 386 00:27:25,450 --> 00:27:29,830 and she said, "You know, your dad is sick, and he is going to no longer be with us." 387 00:27:32,290 --> 00:27:35,560 [Sue Sheridan] I requested a meeting with the doctor, and with the CEO, 388 00:27:37,760 --> 00:27:42,550 and with the risk manager. They agreed to it and I flew down there 389 00:27:42,720 --> 00:27:47,560 and nobody showed up, except the chaplain. 390 00:27:47,560 --> 00:27:49,200 I demanded that they implement a disclosure 391 00:27:49,200 --> 00:27:51,020 procedure that when there was an error at their 392 00:27:51,020 --> 00:27:53,760 hospital that they sit down with the family, 393 00:27:53,760 --> 00:27:57,070 which, you know, which they did not with us. 394 00:28:01,990 --> 00:28:05,830 [David Mayer] Historically, you've probably heard the term deny and defend. 395 00:28:05,830 --> 00:28:08,910 That was the model that is still existent today unfortunately at many hospitals; 396 00:28:08,910 --> 00:28:11,570 That if we cause a preventable medical harm, 397 00:28:11,570 --> 00:28:14,840 the goal has always been to shut things down, 398 00:28:15,000 --> 00:28:20,430 let the lawyers handle it, don't talk to the patients and families, 399 00:28:20,430 --> 00:28:24,510 and then it turns into a legal battle for 4, 5, 6 years where 400 00:28:24,520 --> 00:28:28,930 the hope is that the patient and family will just give up and go away 401 00:28:28,940 --> 00:28:33,400 and that's been the model. And now we've moved to more open and honest communication. 402 00:28:35,820 --> 00:28:38,150 [Heather Young] We do a simulation on 403 00:28:40,570 --> 00:28:42,370 how to tell someone that you've made an error, 404 00:28:42,370 --> 00:28:44,990 and that's a skill that's very difficult to develop, 405 00:28:45,000 --> 00:28:47,620 to do in a way that conveys that you care and that you are concerned about 406 00:28:47,620 --> 00:28:52,380 the person's safety and that you are going to do something about it 407 00:28:52,380 --> 00:28:55,590 when you may face a family member who is irate, very upset by the news. 408 00:28:55,590 --> 00:29:00,800 And you know, as a new clinician, you need to have the skills to be open and 409 00:29:00,800 --> 00:29:05,220 transparent and talk honestly and authentically with people. 410 00:29:05,230 --> 00:29:10,190 So, I'm about to go in and see a standardized patient, is what we call it. 411 00:29:10,190 --> 00:29:14,530 It's an actor that I have no idea how he's going to react. 412 00:29:14,530 --> 00:29:18,030 We're going to break him some bad news about a test result that we missed 3 months ago. 413 00:29:18,030 --> 00:29:21,740 They are told to react differently to each student. 414 00:29:21,740 --> 00:29:25,790 So I don't know what I'm going to get when I break him the news. 415 00:29:25,790 --> 00:29:28,580 He could be angry, frustrated, or he could go easy on me. I just don't know. 416 00:29:28,580 --> 00:29:34,090 One of the things that was ordered a couple weeks ago was a CT scan, uh, which, 417 00:29:35,590 --> 00:29:41,510 umm, indicated, umm, some results that could indicate colon cancer. 418 00:29:42,850 --> 00:29:47,690 [Doctor] Listen, I've got to cut this. Um, you don't want to say there was another 419 00:29:48,480 --> 00:29:54,450 test result that might indicate colon cancer at this short intro into it, right? 420 00:29:56,200 --> 00:29:59,950 - Ah? - Oh. 421 00:30:01,580 --> 00:30:04,010 I mean you went right to: "That could be colon cancer." His dad died 422 00:30:04,020 --> 00:30:06,420 of colon cancer. You could have a patient falling apart in moments. 423 00:30:06,420 --> 00:30:11,050 Do you want to look at those pearls on effective communication? 424 00:30:11,050 --> 00:30:16,550 Lay out the facts, that you know them, and say that 3 months ago on the CT... 425 00:30:16,800 --> 00:30:21,810 And then as he's like, and I know your dad passed, 426 00:30:21,810 --> 00:30:26,650 it could be a cancer, but we don't know that yet. 427 00:30:26,650 --> 00:30:30,110 You know, all that gingerly, careful stuff. 428 00:30:30,110 --> 00:30:33,240 - Hey, Walt. How's it going? - Hey, Jason. I'm alright. 429 00:30:34,030 --> 00:30:36,370 - It's good to see you again. - Thank you. 430 00:30:36,370 --> 00:30:38,160 - How was the drive in? - Uh, fine. 431 00:30:38,160 --> 00:30:40,100 Three months ago, remember you came in three months ago? 432 00:30:40,100 --> 00:30:41,250 I do. 433 00:30:43,120 --> 00:30:46,130 It showed that you had some thickening of your colonic wall 434 00:30:46,330 --> 00:30:50,260 and some enlarged mesenteric lymph nodes. 435 00:30:51,130 --> 00:30:54,970 We need to do a colonoscopy immediately. 436 00:30:54,970 --> 00:30:58,430 We want to make sure, and I'm not saying it's colon cancer, 437 00:30:58,720 --> 00:31:01,480 but we want to make sure that it's not colon cancer and rule it out. 438 00:31:01,480 --> 00:31:05,230 Why did it take 3 months to, uh, that I know this? 439 00:31:05,230 --> 00:31:09,980 That was my mistake. We were looking for structural abnormalities on your kidneys 440 00:31:09,990 --> 00:31:13,820 and I overlooked that part of the report 3 months ago. 441 00:31:13,820 --> 00:31:17,160 [sigh] I mean, I would have been upset 442 00:31:18,120 --> 00:31:22,920 hearing it first when the CT scan happened, 443 00:31:22,920 --> 00:31:26,500 but now I'm really pissed off that it's been 3 months, that it was delayed. 444 00:31:26,500 --> 00:31:31,840 Right, and, I mean, I understand that you're angry, I can see that you're frustrated 445 00:31:31,840 --> 00:31:37,260 and I can't, I can't do anything to fix that mistake 3 months ago. 446 00:31:37,270 --> 00:31:41,140 But, what I can do now is make this a priority as your primary care provider, 447 00:31:41,140 --> 00:31:46,020 and I can't even imagine how you're feeling right now with the mistake, 448 00:31:46,020 --> 00:31:48,140 but let's take it from here, and we'll figure 449 00:31:48,140 --> 00:31:50,990 this out together. I'll make this a priority, OK? 450 00:32:31,370 --> 00:32:36,330 - [Charlie] Good morning. - Good morning. Hi Charlie. 451 00:32:36,330 --> 00:32:37,870 Hi Walt, nice to meet you. 452 00:32:37,870 --> 00:32:39,710 I'm sorry. Wait, I've met you before. 453 00:32:39,710 --> 00:32:41,960 Yeah. We've known each other for years. 454 00:32:41,960 --> 00:32:45,840 [Heather Young] The closer you are to the error, the more important it is that you have 455 00:32:45,840 --> 00:32:49,130 some accountability for it, and that you communicate 456 00:32:49,140 --> 00:32:51,930 with the people who might be harmed by it. 457 00:32:51,930 --> 00:32:53,890 And so all of us need to learn the skills to be able to 458 00:32:53,890 --> 00:32:57,770 acknowledge what we've donewrong and what we're planning to do to fix it. 459 00:32:57,770 --> 00:33:01,610 [Don Berwick] We built it completely wrong. We were trained, I was trained, 460 00:33:02,150 --> 00:33:06,110 "No, you don't talk about your mistakes with a patient, 461 00:33:06,110 --> 00:33:09,200 that's liability, the lawyers will be all over us." 462 00:33:09,200 --> 00:33:12,410 This is a time for openness and honesty, and so we can learn and grow together. 463 00:33:12,410 --> 00:33:17,870 Healthcare is not like a toaster where I make it and I sell it to you, 464 00:33:17,880 --> 00:33:21,130 and you take it and plug it in. No, it's always a cooperative enterprise so that 465 00:33:21,130 --> 00:33:26,380 the patient and the family, and the doctor and the nurse, they're co-producing the care. 466 00:33:26,380 --> 00:33:30,430 And now that we're more aware of that over time, 467 00:33:30,430 --> 00:33:33,490 there's a lot of possibility for much more participation by both. 468 00:33:36,440 --> 00:33:41,570 [John Eisenberg] I recalled a woman whom I took care of. 469 00:33:41,570 --> 00:33:45,780 We had had a pap test done to screen her for cervical cancer. 470 00:33:45,780 --> 00:33:49,620 The result was suspicious, but I never knew that, 471 00:33:50,370 --> 00:33:54,500 because I never got the report back. And I didn't realize 472 00:33:54,790 --> 00:33:58,670 that I hadn't gotten the report back until she called me and asked about the report. 473 00:33:58,670 --> 00:34:03,470 I tracked it down. I found out that it was suspicious. We followed it up and fortunately 474 00:34:04,130 --> 00:34:10,010 it turned out not to be anything serious. But that was a near miss. 475 00:34:10,060 --> 00:34:14,980 It was a near miss that could have been a tragedy had she not called me. 476 00:34:16,310 --> 00:34:21,690 Senator, when I spoke at three medical school graduations last Spring, 477 00:34:21,690 --> 00:34:26,740 I asked all the students who were graduating, and I asked all of the faculty 478 00:34:26,740 --> 00:34:31,910 to raise their hands if they had ever made a mistake in taking care of a patient, 479 00:34:32,250 --> 00:34:34,790 and every single student raised his or her hand, 480 00:34:34,790 --> 00:34:37,670 every faculty member raised his or her hand. 481 00:34:37,880 --> 00:34:43,680 When I was a medical student on one of my very first rotations, 482 00:34:43,930 --> 00:34:48,470 I inadvertently, during a code, 483 00:34:48,470 --> 00:34:52,480 gave a full syringe of morphine to a patient IV and they had a respiratory arrest. 484 00:34:54,770 --> 00:35:00,820 Fortunately, the person was intubated and resuscitated and did just fine. 485 00:35:00,940 --> 00:35:05,490 That was a shocking experience, 486 00:35:05,490 --> 00:35:10,160 and made me aware at a very early point in my medical career 487 00:35:10,160 --> 00:35:15,040 that we have the potential to do things wrong and to potentially harm patients. 488 00:35:15,040 --> 00:35:19,340 No one ever heard about it besides me and that nurse. 489 00:35:19,340 --> 00:35:23,050 So, it's not clear to me that any changes were ever made 490 00:35:23,470 --> 00:35:26,300 as a result, and I don't think the patient ever heard. 491 00:35:26,310 --> 00:35:28,350 I've made medical errors; I have, uh, 492 00:35:28,350 --> 00:35:31,730 I prescribed the wrong medication on a patient. There were two patients of mine 493 00:35:31,730 --> 00:35:35,650 with very similar names and I just prescribed it on the wrong patient. 494 00:35:35,650 --> 00:35:38,900 I felt terrible. I felt incompetent. I felt a little ashamed. 495 00:35:38,900 --> 00:35:44,450 And I, my first instinct was not just to fix the problem, but then not to tell anybody. 496 00:35:44,450 --> 00:35:49,160 That's just a normal human instinct. 497 00:35:49,160 --> 00:35:52,540 It is completely understandable why people's first reaction is 498 00:35:52,540 --> 00:35:57,380 cover it up, don't talk about it. It's a very human response. 499 00:35:57,380 --> 00:36:01,180 Doesn't make it the right thing, it's actually clearly 500 00:36:01,180 --> 00:36:03,640 not the right thing, it's clearly bad to do that. 501 00:36:03,640 --> 00:36:06,430 But I think we have to begin by acknowledging that it's a very human response. 502 00:36:06,430 --> 00:36:10,230 You can feel very self-righteous. You can say, 503 00:36:10,230 --> 00:36:12,230 "Well, the patient got the wrong drug, fire the nurse. 504 00:36:12,230 --> 00:36:14,770 There's a complication of the surgery, bad surgeon." 505 00:36:14,780 --> 00:36:18,690 You're wrong. You're almost always wrong. 506 00:36:18,700 --> 00:36:21,410 It feels good to blame someone. You've got a culprit? Put them in jail, fire them. 507 00:36:21,410 --> 00:36:25,120 Many things caused it. So, who's responsible? Everybody's responsible. 508 00:36:25,120 --> 00:36:29,000 Everybody can contribute to the enterprise of closing the vulnerabilities, 509 00:36:29,000 --> 00:36:33,340 of making the whole thing less likely to go wrong. 510 00:36:33,340 --> 00:36:37,220 The most recent survey I have seen is that nearly 50% of nurses in America 511 00:36:37,220 --> 00:36:43,140 still don't feel it is safe to talk about a mistake they've made. 512 00:36:43,180 --> 00:36:47,640 That's an absolute disgrace. 513 00:36:47,640 --> 00:36:50,510 If something bad is going to happen to you when you speak up 514 00:36:51,520 --> 00:36:55,230 about something you've seen or done that could help. 515 00:36:55,240 --> 00:37:00,860 If you're going to get punished for that, why would you speak up? You don't. 516 00:37:00,870 --> 00:37:05,250 You run and hide. You lie. That's normal human behavior. 517 00:37:05,250 --> 00:37:10,000 We're not talking about bad people; we're talking about normal people become frightened. 518 00:37:10,000 --> 00:37:13,550 And so leaders, you got a choice: you can scare your workforce and give up the hope 519 00:37:13,550 --> 00:37:18,260 for improvement, or you can celebrate, invite, work with your workforce, 520 00:37:18,510 --> 00:37:23,810 and have a chance of learning together to get to a better world. 521 00:37:23,810 --> 00:37:26,560 What we have learned from other industries is that if you could change the culture 522 00:37:26,560 --> 00:37:30,770 and reward people for being open, reward people for being honest, 523 00:37:30,770 --> 00:37:34,190 reward people for coming forth and talking about their errors, 524 00:37:34,190 --> 00:37:38,110 then you being to counter that kind of normal instinct that we all have, 525 00:37:38,120 --> 00:37:42,240 and begin to create a culture of patient safety where people are much more open. 526 00:37:42,240 --> 00:37:45,910 And the system gets better because it learns from mistakes and doesn't hide them. 527 00:37:45,910 --> 00:37:50,090 And we found in the food industry they were years ahead of us. They had programs. 528 00:37:50,090 --> 00:37:54,880 For instance, Burger King had a program if the employee saw another one 529 00:37:54,880 --> 00:37:59,180 not washing their hands, they went over and they tapped them and said, "Got you", 530 00:37:59,180 --> 00:38:03,350 and then they got either two hours compensation off 531 00:38:03,350 --> 00:38:07,190 or some other reward. I mean, they're on board. 532 00:38:07,190 --> 00:38:09,040 Safety reporting is like democracy. Democracy 533 00:38:09,050 --> 00:38:11,400 isn't about having a free and fair election. 534 00:38:11,650 --> 00:38:16,410 We can do that. Democracy is about having a second free and fair election. 535 00:38:17,160 --> 00:38:22,290 The same thing is true with safety reporting. It's not about 536 00:38:22,290 --> 00:38:24,660 filing a safety report, it's about filing a second. 537 00:38:24,660 --> 00:38:27,130 And where you see an organization with a high rate of reported error, 538 00:38:27,130 --> 00:38:28,750 what that tells you is it tells you that they 539 00:38:28,750 --> 00:38:30,380 must be doing something about those reports, 540 00:38:30,380 --> 00:38:34,800 because if they are just sitting on them, people will stop reporting. 541 00:38:34,800 --> 00:38:38,300 Because even if you tell people they have to, in the end it's all voluntary. 542 00:38:38,310 --> 00:38:42,140 I mean, you can't solve it if you can't see it. We can see it. 543 00:38:42,140 --> 00:38:45,480 And more and more people are aware of it. That's the good news. 544 00:38:45,480 --> 00:38:47,980 Bad news is you're still at risk, really at risk. 545 00:38:47,980 --> 00:38:51,690 I mean we haven't pervaded healthcare with the designs and approaches and cultures 546 00:38:52,900 --> 00:38:58,120 that actually make you super safe and that's the task ahead. 547 00:38:58,120 --> 00:39:01,910 It's amazing how quickly hospitals can completely overhaul their safety 548 00:39:03,000 --> 00:39:07,500 when they know that it's important to their patients. 549 00:39:08,000 --> 00:39:10,920 Hospitals had to hear the message from their own patients 550 00:39:10,920 --> 00:39:12,640 that it matters that they wash their hands, 551 00:39:12,640 --> 00:39:14,470 it matters that they keep a safe environment, 552 00:39:14,470 --> 00:39:19,680 it matters that they put the safety and protection of their patients first 553 00:39:19,680 --> 00:39:24,980 every minute of every day. The only way they're really going to get that message 554 00:39:24,980 --> 00:39:29,610 is when the American public gets involved and pushes. 555 00:39:32,780 --> 00:39:38,410 [narrator]One way to improve the quality of hospitals in America 556 00:39:38,410 --> 00:39:41,210 is to put a microscope on the data they do actually provide. 557 00:39:41,210 --> 00:39:45,000 [narrator] Leah Binder and her team at the Leapfrog Group in Washington, DC, 558 00:39:45,920 --> 00:39:49,720 worked with leaders in patient safety to create a new way 559 00:39:49,720 --> 00:39:52,840 to rate the quality of hospitals that patients can understand. 560 00:39:52,840 --> 00:39:56,810 We worked with the foremost experts in patient safety 561 00:39:57,180 --> 00:39:58,970 and we asked them to look at all this data and 562 00:39:58,970 --> 00:40:00,640 decide which of the data is most reliable, 563 00:40:00,640 --> 00:40:05,400 which gives us the best information about the safety of a hospital, 564 00:40:05,400 --> 00:40:08,360 and then help us figure out a reliable criteria to put it all together. 565 00:40:08,360 --> 00:40:12,700 And then, we did something else. We decided to issue a letter grade. 566 00:40:12,700 --> 00:40:15,190 The letter grade would apply to each hospital on 567 00:40:15,190 --> 00:40:18,120 how safe they are relative to other hospitals. 568 00:40:18,160 --> 00:40:22,750 So, were they an A, B, C, D, or F? 569 00:40:22,750 --> 00:40:26,050 When we first did it, we got calls from some hospital CEOs who said 570 00:40:26,050 --> 00:40:30,220 to me, memorably, "I've decided I don't want a letter grade from you." 571 00:40:30,220 --> 00:40:34,470 And I said, "Well, I've decided you're getting one anyway, because you serve the public, 572 00:40:34,470 --> 00:40:39,100 and the public you serve deserves to know how you're doing." 573 00:40:39,100 --> 00:40:42,190 It's very important to do these kinds of ratings because 574 00:40:42,190 --> 00:40:43,870 who wants to work in a terrible organization? And 575 00:40:43,870 --> 00:40:45,440 so if you can make it very obvious to all the 576 00:40:45,440 --> 00:40:49,860 doctors and nurses in that hospital that this is a highly unsafe hospital, 577 00:40:49,860 --> 00:40:51,700 I think there is going to be internal pressure 578 00:40:51,700 --> 00:40:53,580 to reform and internal pressure to get better. 579 00:40:53,580 --> 00:40:57,910 But, certainly I think it's true that, like, if you're in an isolated area, 580 00:40:57,920 --> 00:41:00,790 there's one hospital in town or you could be in the middle of Chicago, 581 00:41:00,790 --> 00:41:04,090 but your insurance company covers one hospital only, 582 00:41:04,090 --> 00:41:05,890 it's going to be a challenge of choices. But 583 00:41:05,890 --> 00:41:07,880 that doesn't mean you couldn't go to your doctor 584 00:41:07,880 --> 00:41:12,180 who works in that hospital and be like, "Hey, why are you guys a D hospital?" 585 00:41:12,180 --> 00:41:15,680 And I think if consumers started talking to doctors and nurses that way, 586 00:41:15,680 --> 00:41:19,020 it would actually begin to change the conversation, where doctors would say, 587 00:41:19,020 --> 00:41:23,190 "Why do I work at a hospital that has such high infection rates?" 588 00:41:23,190 --> 00:41:26,070 Virtually every other industry in this country has their products and services 589 00:41:26,070 --> 00:41:29,740 in a transparent market, and people choose. 590 00:41:29,740 --> 00:41:31,990 So, if you're buying a car, you can look up auto reviews 591 00:41:31,990 --> 00:41:36,790 and you can compare among different cars and different features. 592 00:41:36,790 --> 00:41:39,630 In New York, which I know particularly well, 593 00:41:39,630 --> 00:41:42,210 restaurants that had, for many years, been getting public ratings 594 00:41:42,210 --> 00:41:46,010 from the health department on how safe they were; 595 00:41:46,010 --> 00:41:49,180 those were all public, but nobody paid any attention to them. 596 00:41:49,180 --> 00:41:52,470 So, the health department said, from now on you're going to get a grade 597 00:41:52,470 --> 00:41:55,640 on how safe you are and you have to post it in your window. 598 00:41:55,640 --> 00:41:57,450 So, restaurants started posting it, and within 599 00:41:57,450 --> 00:41:59,400 six months any restaurant that didn't have an A 600 00:41:59,400 --> 00:42:04,900 was either out of business or they were very quickly getting to their A. 601 00:42:04,900 --> 00:42:09,280 So we said, "Well, let's do the same thing with hospitals." 602 00:42:09,280 --> 00:42:12,200 I mean in our dream, hospitals would put their letter grade on 603 00:42:12,200 --> 00:42:16,580 you know, their front door and everyone would know that this hospital was safe or not. 604 00:42:16,580 --> 00:42:21,260 [sombr music] 605 00:42:22,340 --> 00:42:28,550 [Helen Burstin]John Eisenberg used to tell a great story of the drunk who lost his keys. 606 00:42:28,560 --> 00:42:32,640 And he's out in front of the bar in the street looking for his keys 607 00:42:33,270 --> 00:42:36,150 and some guy comes over and goes, "What are you doing?" 608 00:42:36,150 --> 00:42:38,610 He says, "I'm looking for my key." 609 00:42:38,610 --> 00:42:40,530 "Well, why are you only looking right here?" 610 00:42:40,530 --> 00:42:42,570 He said, "Well, nthat's where the lamplight is." 611 00:42:42,570 --> 00:42:44,450 [clock ticking] 612 00:42:44,450 --> 00:42:46,450 [narrator] This is known as the streetlight effect. 613 00:42:46,450 --> 00:42:49,230 Many in the patient safety field have been looking outside 614 00:42:49,870 --> 00:42:53,460 healthcare for solutions to preventable errors. 615 00:42:53,460 --> 00:42:56,080 Industries like nuclear power, aircraft carriers, and commercial aviation 616 00:42:56,090 --> 00:43:00,670 have become known as high-reliability organizations 617 00:43:00,670 --> 00:43:03,720 due to significant efforts to improve safety. 618 00:43:03,720 --> 00:43:07,470 High reliability is different in healthcare because it points directly at examples 619 00:43:08,010 --> 00:43:12,890 of very hazardous industries, organizations 620 00:43:12,890 --> 00:43:17,820 that have solved the problem of getting to zero harm that healthcare has not solved. 621 00:43:18,190 --> 00:43:23,030 Tools and methods and lessons from that work 622 00:43:23,030 --> 00:43:27,330 are very directly applicable to healthcare and we're starting to see 623 00:43:27,330 --> 00:43:31,830 healthcare organizations use them to make improvements 624 00:43:31,830 --> 00:43:35,500 at a level that we have never seen before. 625 00:43:35,500 --> 00:43:38,170 So over here we have the complex system of the modern American hospital, 626 00:43:38,170 --> 00:43:43,220 and over here we have other industries that have learned to 627 00:43:43,220 --> 00:43:48,100 simplify and deal with these complex systems. 628 00:43:48,100 --> 00:43:52,310 In the last calendar year there has been no fatality worldwide 629 00:43:52,310 --> 00:43:57,150 in commercial aviation due to an accident. 630 00:43:57,150 --> 00:44:00,700 Compare that to our business where we have 20 wrong-site surgeries every week. 631 00:44:01,110 --> 00:44:05,870 [David Mayer] Pilots make one error per hour in the cockpit every day they work 632 00:44:06,830 --> 00:44:11,830 and yet we wonder why planes aren't falling out of the sky. 633 00:44:11,830 --> 00:44:15,710 If aviation had said, "Well, you know what, to fly you 600 miles an hour it's going to 634 00:44:15,710 --> 00:44:20,680 come with some mishap. And you got to expect a plane or two to fall out of the sky," 635 00:44:20,680 --> 00:44:24,800 and thank god they didn't say that and they said, "No, we can drive it to zero. 636 00:44:24,810 --> 00:44:29,480 We can drive it down to virtually no mishap," and they have. 637 00:44:29,480 --> 00:44:31,500 The aviation industry is the safest it's ever 638 00:44:31,500 --> 00:44:33,480 been since the invention of the jet engine. 639 00:44:33,480 --> 00:44:35,680 What we're really doing when we go up in an 640 00:44:35,680 --> 00:44:38,530 airliner is pushing a tube filled with people 641 00:44:38,740 --> 00:44:44,990 through the upper atmosphere, 7 or 8 miles above the earth, at 80% the speed of sound, 642 00:44:44,990 --> 00:44:50,710 in a hostile environment with outside air pressure one-quarter that at the surface, 643 00:44:50,710 --> 00:44:55,380 and we must return it safely to the surface every time, and we do. 644 00:44:55,380 --> 00:45:01,140 In this country alone, 28,000 times a day, 10.2 million times a year. 645 00:45:01,470 --> 00:45:06,980 In a little over 100 years you've gone from quite a dangerous 646 00:45:06,980 --> 00:45:12,150 industry to the first ultra-safe mode of transport bar none. 647 00:45:12,150 --> 00:45:17,780 One of the reasons is because it is studied so well, and every single event 648 00:45:17,780 --> 00:45:23,280 is clearly understood and is made public so others can learn from them. 649 00:45:23,290 --> 00:45:27,370 I had been flying airplanes for 42 years. I had 20,000 hours in the air. 650 00:45:27,370 --> 00:45:32,960 And throughout that entire time, I had never been so challenged in an airplane 651 00:45:32,960 --> 00:45:36,340 I doubted the outcome. I never thought I would be. I was wrong. 652 00:45:36,340 --> 00:45:41,640 [narrator] In January of 2009, Captain Sullenberger's training and instincts 653 00:45:41,640 --> 00:45:46,640 saved the lives of all 155 passengers aboard US Airways Flight 1549 654 00:45:46,770 --> 00:45:52,400 after it struck a flock of geese and lost all engine power. 655 00:45:52,400 --> 00:45:56,570 The dramatic landing on thenHudson River reminded Americans 656 00:45:56,570 --> 00:45:59,530 of the importance of experience in the cockpit. 657 00:45:59,530 --> 00:46:03,440 In an industry in which we work very hard to make everything easy and routine and safe, 658 00:46:06,210 --> 00:46:11,380 100 seconds after takeoff we were suddenly confronted with an 659 00:46:11,380 --> 00:46:15,010 ultimate challenge of a lifetime, to do something we'd never done before 660 00:46:15,010 --> 00:46:17,800 and get it right the first time never having practiced it. 661 00:46:17,800 --> 00:46:20,470 In a similar fashion in medicine, there are some things that just can't be 662 00:46:20,470 --> 00:46:24,440 practiced safely any other way than in a simulation for the first time. 663 00:46:24,440 --> 00:46:28,230 And it gives you a chance to practice things over and over and over again. 664 00:46:28,230 --> 00:46:33,780 And so it's important that the simulations be done not simply individually, 665 00:46:33,780 --> 00:46:37,910 but also collectively as a whole team. 666 00:46:45,790 --> 00:46:51,840 [narrator]Flight simulators have been used to train pilots for nearly 100 years. 667 00:46:51,840 --> 00:46:56,390 And while medicine has used cadavers to train doctors for much longer, 668 00:46:56,390 --> 00:47:00,390 only recently have institutions begun using robotics 669 00:47:00,390 --> 00:47:03,730 to simulate any kind of situation a care provider may face. 670 00:47:03,730 --> 00:47:07,900 [Heather Young] Simulation is a very big part of our educational program here 671 00:47:08,980 --> 00:47:10,860 and it involves anything from patients who come 672 00:47:10,860 --> 00:47:12,900 in as actors and will work with a student, 673 00:47:12,910 --> 00:47:18,700 all the way up to very high-fidelity robots, and environments that 674 00:47:18,700 --> 00:47:22,790 are tricked out to look out exactly like a hospital operating room 675 00:47:22,790 --> 00:47:26,420 or an emergency department or hospital ward. 676 00:47:26,420 --> 00:47:29,590 The airline industry is the prototype of using simulation 677 00:47:31,720 --> 00:47:35,300 where you can practice landing in San Diego with a terrible storm or a tsunami 678 00:47:35,300 --> 00:47:40,140 or on a very calm day and you can practice 679 00:47:40,140 --> 00:47:42,480 all different kinds of failures within the airplane. 680 00:47:42,480 --> 00:47:45,900 It's newer in healthcare, but it's really something that's catching on, 681 00:47:45,900 --> 00:47:49,820 and you can really put people through the steps of handling many important situations. 682 00:47:50,610 --> 00:47:55,990 [Ian Julie] So we're going to be practicing our new simulated protocol 683 00:47:55,990 --> 00:47:59,620 for our actual sepsis patients. Sepsis care can be very, very difficult. 684 00:47:59,620 --> 00:48:04,040 We know the science behind it, we know what helps, but we don't necessarily know 685 00:48:04,040 --> 00:48:07,460 how to do it in a way that's organized and consistent. 686 00:48:07,460 --> 00:48:10,430 We'd rather practice on our friend the mannequin here who it's very hard to injure, 687 00:48:10,430 --> 00:48:15,100 rather than on real patients. That way we can standardize things within our hospital 688 00:48:15,100 --> 00:48:19,270 and give our nurses and doctors a chance to practice what it is their doing, 689 00:48:19,270 --> 00:48:23,100 before they have to do it on real patients. 690 00:48:23,110 --> 00:48:25,400 Hi Robert. My name is Emily. I'm going to be you nurse today. 691 00:48:25,400 --> 00:48:28,490 I'm here to do your morning assessment and take your vital signs. How are you feeling? 692 00:48:28,490 --> 00:48:32,360 [mannequin] I'm not feeling very well. 693 00:48:32,370 --> 00:48:34,030 [Nurse 1] You're not? What's going on? 694 00:48:34,030 --> 00:48:36,080 [mannequin] I just can't catch my breath this morning and 695 00:48:36,080 --> 00:48:38,250 I feel like my cough is worse. 696 00:48:38,250 --> 00:48:40,080 He's remaining stable. Based on the alert, um, 697 00:48:40,080 --> 00:48:42,790 and the lactic acid, I think I'm going to start some oxygen. 698 00:48:42,790 --> 00:48:46,800 [Nurse 2] Okay, are there signs or symptoms of an infection? 699 00:48:47,170 --> 00:48:51,510 [Nurse 1] Well, he's saying that he has an increased work of breathing. 700 00:48:51,510 --> 00:48:54,470 He has a white count of 16. 701 00:48:54,470 --> 00:48:56,680 [Nurse 2] OK, sound good. I'll be right over. 702 00:48:56,680 --> 00:48:58,520 - OK, thank you! - [Nurse 2] Alright. 703 00:48:58,520 --> 00:49:00,980 - [Nurse 2] Hi, Mr. Robert! - [mannequin] Hi! 704 00:49:02,020 --> 00:49:04,520 -How are you feeling? -[mannequin] I've had better days. 705 00:49:04,530 --> 00:49:07,900 - Are you short of breath? - [Mannequin] Yeah. 706 00:49:07,910 --> 00:49:10,070 [Nurse 2] OK and when did this start? 707 00:49:10,070 --> 00:49:12,120 Here you go. You've drawn cultures already, correct? 708 00:49:12,120 --> 00:49:17,410 - [Nurse 2] This is Robert Doe? - [Nurse 1] Yes, Robert Doe. 709 00:49:17,410 --> 00:49:21,170 [Ian Julie] We can make the scenario more complex, and we do on occasion. 710 00:49:21,460 --> 00:49:24,710 We could have the patient enter a state of shock, 711 00:49:24,710 --> 00:49:27,130 or not respond properly to the fluids or the antibiotics. 712 00:49:27,130 --> 00:49:30,390 So, much of what we've done is related to the need to kind of 713 00:49:30,390 --> 00:49:34,100 fulfill the recommendations that have been given. 714 00:49:34,100 --> 00:49:36,980 In addition to wanting to do what's right for the patient 715 00:49:36,980 --> 00:49:39,310 and following through on the best available scientific evidence. 716 00:49:39,310 --> 00:49:42,690 When I graduated as a nurse, the first time I ever had a chance 717 00:49:42,690 --> 00:49:46,110 to shock a person whose heart had stopped 718 00:49:46,110 --> 00:49:48,780 was in the middle of the night in a rural hospital 719 00:49:48,780 --> 00:49:51,370 and it was my first time I had ever turned on the paddles in my life. 720 00:49:51,370 --> 00:49:55,000 And someone's life depended on that. That's not acceptable. 721 00:49:55,000 --> 00:49:59,630 We want our students to practice and practice and practice how to shock people 722 00:49:59,630 --> 00:50:04,670 in a simulated situation, so that when someone is really depending on them, 723 00:50:04,670 --> 00:50:08,760 they do it right the first time. 724 00:50:08,760 --> 00:50:11,510 I shudder to remember how I was trained as a doctor to learn how to do stuff. 725 00:50:11,520 --> 00:50:16,730 Lumbar punctures, spinal taps, put IVs in, or even chest tubes. 726 00:50:16,730 --> 00:50:21,780 You practiced on the patients. I mean, that was the only option. 727 00:50:21,780 --> 00:50:25,860 Some patient, some time, was the first patient I ever put a chest tube in, 728 00:50:25,860 --> 00:50:29,620 and that person paid the price. They were paying for my tuition. 729 00:50:29,620 --> 00:50:34,160 You know, we don't do that with pilots, we put them in the simulator 730 00:50:34,160 --> 00:50:36,880 and they fly something that isn't really a plane for a while, first, 731 00:50:36,880 --> 00:50:39,800 with high fidelity. Now we know how to do that in health care. 732 00:50:39,800 --> 00:50:42,380 The growth of simulation so that the first chest tube doesn't go into a human being, 733 00:50:42,380 --> 00:50:46,090 it goes in a mannequin that looks like a human being, that's great. And I think that 734 00:50:46,090 --> 00:50:51,020 it's one of the emerging ways to help build skills 735 00:50:51,020 --> 00:50:54,980 hmm, in a work force without the patients paying the tuition. 736 00:50:54,980 --> 00:50:59,400 [narrator] Many aspects of the aviation industry have been applied to medicine, 737 00:51:01,030 --> 00:51:04,610 from checklists before an operation to monitoring physicians for fatigue. 738 00:51:04,610 --> 00:51:08,530 But there are still elements of safety in aviation that have not been explored. 739 00:51:08,540 --> 00:51:13,080 One of the most well-known improvements in airline safety is the black box. 740 00:51:13,080 --> 00:51:17,790 A surgeon in Toronto has been working with a group of designers 741 00:51:17,800 --> 00:51:21,420 to create a similar tool for the operating room. 742 00:51:21,420 --> 00:51:24,760 [Teodor Grantcharov] I want my patients to feel the same way when 743 00:51:24,760 --> 00:51:27,550 they enter the operating room as I feel when I enter a modern aircraft. 744 00:51:27,560 --> 00:51:31,350 Unless we create a system where we understand, that we tolerate, 745 00:51:31,350 --> 00:51:35,810 and we learn from our errors, we will never be able to improve. 746 00:51:35,810 --> 00:51:39,400 We've tried for many years to create something like the black box. 747 00:51:39,400 --> 00:51:42,240 Finally, in 2012 we were able to create a technology that allows us to capture 748 00:51:42,240 --> 00:51:48,160 video and audio and data from everything that's happening in an operating room. 749 00:51:48,160 --> 00:51:53,250 We've been developing and implemented a number of sensors. 750 00:51:53,250 --> 00:51:56,290 So, we know how many times a door opens and closes. 751 00:51:56,300 --> 00:51:59,090 We know how we wash our hands prior to a surgical procedure. 752 00:51:59,090 --> 00:52:02,720 And all these data feeds are combined 753 00:52:02,720 --> 00:52:05,220 and perfectly synchronized on the same platform. 754 00:52:05,220 --> 00:52:07,770 When we talk with our patients about the black box and what we are doing here, 755 00:52:07,770 --> 00:52:11,350 the first reaction, the most common reaction in 90% of the patients is, 756 00:52:11,350 --> 00:52:15,310 "I can't believe this hasn't been done before." 757 00:52:15,320 --> 00:52:17,900 From the point we started recording our surgeries, 758 00:52:17,900 --> 00:52:20,780 we had a tremendous amount of media attention. 759 00:52:20,780 --> 00:52:24,240 Everybody believed in the transparency doctor, that's what he was nicknamed. 760 00:52:26,040 --> 00:52:30,420 He doesn't have anything to hide. I'm definitely going to go to him. 761 00:52:30,420 --> 00:52:33,630 Patients need to know that when they walk into a hospital, 762 00:52:33,630 --> 00:52:36,630 everything is being done to learn from mistakes and possible risks that take place. 763 00:52:36,630 --> 00:52:41,050 This has to be common standard practice. 764 00:52:41,050 --> 00:52:43,800 We've heard for too long that healthcare is complex, that our patients are not aircraft, 765 00:52:43,810 --> 00:52:49,390 that surgeons are not pilots. I just want us to start doing something and changing it. 766 00:52:49,390 --> 00:52:54,150 We're trying to create a system that identifies 767 00:52:54,150 --> 00:52:57,400 performance deficiencies and improves safety. 768 00:52:57,400 --> 00:53:00,570 A new gadget comes out from an industry provider all the time. 769 00:53:00,570 --> 00:53:04,700 Usually it's a very emotional attachment like, "Oh, this looks sexy," 770 00:53:04,700 --> 00:53:08,000 or "I like how this handle feels when I'm using it during surgery." 771 00:53:08,000 --> 00:53:11,040 You need something deeper beyond that. 772 00:53:11,040 --> 00:53:13,340 So, this is what we call a full surgical timeline and you can see 773 00:53:13,340 --> 00:53:16,510 the entire procedure broken down from beginning to end. 774 00:53:16,510 --> 00:53:19,930 As you scroll down this timeline, you'll start to see little beeps here, 775 00:53:19,930 --> 00:53:25,350 and that's where our surgical expert analysts have coded where they saw errors. 776 00:53:25,350 --> 00:53:29,850 This screenshot establishes what one of the errors is. So in this case 777 00:53:29,860 --> 00:53:33,610 an error took place during the suturing task, and it was inadequate visualizations. 778 00:53:33,610 --> 00:53:37,530 As the surgeon was suturing with the needle and driver, he might have gone off frame, 779 00:53:37,530 --> 00:53:42,740 which is incorrect because now you have no idea where that needle is. 780 00:53:42,740 --> 00:53:46,250 You even see issues in say, leadership or communication, 781 00:53:46,250 --> 00:53:49,130 and we have a whole toolset determining exactly that. 782 00:53:49,130 --> 00:53:52,130 This points a really interesting storyline because the blue bar establishes 783 00:53:52,130 --> 00:53:53,890 that the surgical resident, the trainee under 784 00:53:53,890 --> 00:53:56,110 the main surgeon was doing the actual case, 785 00:53:56,340 --> 00:54:01,260 and then when a cluster of errors takes place, 786 00:54:01,260 --> 00:54:03,600 you can see the switch over to the main surgeon, to Dr. Grantcharov. 787 00:54:03,600 --> 00:54:06,850 This entire timeline is the data quantified. 788 00:54:06,850 --> 00:54:11,690 We're breaking down the entire set of errors into tangible areas 789 00:54:11,690 --> 00:54:15,950 to provide further education on it to essentially improve it. 790 00:54:15,950 --> 00:54:19,410 Analytics is at the heart of what the black box does, 791 00:54:19,410 --> 00:54:22,120 but we're jumping into different areas. 792 00:54:22,120 --> 00:54:24,540 Our engineers are working on tools to improve handwashing 793 00:54:24,540 --> 00:54:27,670 to essentially create a detector that lets you know, 794 00:54:27,670 --> 00:54:31,210 yes, you've spent the right amount of time and the right technique to wash your hands. 795 00:54:31,210 --> 00:54:34,510 So over here, one of our engineers, Kevin, has been working on just that. 796 00:54:34,510 --> 00:54:38,100 It's a motion sensing tool that will look at how you wash your hands 797 00:54:38,100 --> 00:54:42,100 and look at the surface plane you are working with, the amount of time 798 00:54:42,100 --> 00:54:44,890 spent on washing your hands, and give you real time feedback. 799 00:54:44,890 --> 00:54:48,360 The key here is the data. So, I can go and tell any surgeon 800 00:54:48,360 --> 00:54:51,570 you have to wash your hands this many number of times in this fashion, 801 00:54:51,570 --> 00:54:55,780 but if I have hard data showing... because of doing it this particular way 802 00:54:55,780 --> 00:55:00,240 we have reduced site infections by this much, it's irrefutable. 803 00:55:04,830 --> 00:55:09,510 [piano music] 804 00:59:07,800 --> 00:59:13,230 [Sue Sheridan] Cal has gone on to become part of a comedy community. 805 00:59:13,230 --> 00:59:17,190 He's producing. He's produced two comedy shows. 806 00:59:17,190 --> 00:59:20,190 Cal uses his comedy in really novel ways 807 00:59:20,190 --> 00:59:24,530 that helps him deal with losing a dad. 808 00:59:42,800 --> 00:59:48,180 Everybody develops their own way to deal with death or loss or grief 809 00:59:48,180 --> 00:59:52,850 and I think that comedy is Cal's, uh, his outlet. 810 01:00:04,030 --> 01:00:09,620 [Sue Sheridan] He feels like he doesn't suffer, but he sometimes struggles 811 01:00:09,620 --> 01:00:13,840 to be understood because his speech is impaired. He struggles when he rides on airplanes 812 01:00:14,130 --> 01:00:19,170 because sometimes his scooters or walkers are broken. 813 01:00:19,170 --> 01:00:22,970 He struggles in environments where it's not easy to get around. 814 01:00:23,680 --> 01:00:28,310 The first year, or year and a half, we had 183 separate medical visits 815 01:00:30,650 --> 01:00:35,440 for physical therapy, and ENT, eyes and ears and teeth, and neurology, 816 01:00:36,070 --> 01:00:40,150 and during that time, in my heart, I knew something was wrong with Cal. 817 01:00:40,160 --> 01:00:44,080 Our local doctors were really not willing to offer a diagnosis. 818 01:00:44,080 --> 01:00:48,460 We took... I took Cal out of state to a leading university where a team of specialists 819 01:00:48,750 --> 01:00:53,750 reviewed Cal's charts that I had never looked at. 820 01:00:53,750 --> 01:00:56,460 I didn't think that there was any reason for me to look at my birthing charts, 821 01:00:56,460 --> 01:00:59,680 and back then charts weren't that available to patients. 822 01:00:59,680 --> 01:01:03,540 And they showed to me a report from an MRI that they did on Cal when he was 5 days old 823 01:01:04,470 --> 01:01:10,350 that clearly shared abnormalities in his brain from his jaundice. 824 01:01:11,940 --> 01:01:16,690 And our healthcare system really didn't... 825 01:01:16,700 --> 01:01:20,660 Well, they covered up. They covered up Cal's injury and, umm... 826 01:01:24,080 --> 01:01:26,500 I wasn't empowered with information and knowledge 827 01:01:29,040 --> 01:01:31,840 to challenge some of it or ask the appropriate questions. 828 01:01:31,840 --> 01:01:34,460 You know, in healthcare they say that patients, we need to ask more questions, 829 01:01:34,460 --> 01:01:37,970 but sometimes we simply don't know what to ask. 830 01:01:39,600 --> 01:01:44,930 [Michael Millenson] Understand, before you go in for any particular procedure, 831 01:01:44,930 --> 01:01:49,900 what are the questions you need to ask to keep yourself safe? 832 01:01:49,900 --> 01:01:53,070 And if we all start asking those questions, 833 01:01:53,070 --> 01:01:55,820 then pretty soon it will become clear to any hospital 834 01:01:55,820 --> 01:01:58,660 that's not doing those things that there is pressure on them to do it. 835 01:01:58,660 --> 01:02:03,120 If you're in a hospital, by definition today, you're seeing a lot of different doctors, 836 01:02:03,120 --> 01:02:07,880 there's a lot of caregivers coming in and out of the room, 837 01:02:07,880 --> 01:02:10,630 most of whom work to communicate with each other, but sometimes they miss. 838 01:02:10,630 --> 01:02:13,960 So, if you see something that doesn't look right, or sound right, you say, 839 01:02:13,960 --> 01:02:17,840 "Whoa, wait a minute, that's not what they told me." 840 01:02:17,850 --> 01:02:20,430 Patient safety is a team sport. And one of the ways 841 01:02:20,430 --> 01:02:23,520 to really make a difference is you've got to get patients engaged. 842 01:02:23,520 --> 01:02:26,190 So, if patients begin walking into hospitals with an expectation 843 01:02:26,190 --> 01:02:30,020 that they are not going to get an infection and they start saying, 844 01:02:30,020 --> 01:02:33,030 "Hey, have you washed your hands before you come over to see me?" That's how it happens. 845 01:02:33,030 --> 01:02:36,820 If they've done it outside the room, or they've done it at the nurse's station, 846 01:02:36,820 --> 01:02:41,450 on the way into the room they're touching the door, they're touching things, 847 01:02:41,450 --> 01:02:45,000 and then they are coming in, so that doesn't count. 848 01:02:45,000 --> 01:02:47,080 It has to be before and after patient contact. 849 01:02:47,080 --> 01:02:50,040 Here's what I look for in a hospital that's really outstanding on safety; 850 01:02:50,050 --> 01:02:52,250 the sink is placed in a way that it is easy to 851 01:02:52,250 --> 01:02:54,630 walk into a room and immediately wash your hands. 852 01:02:54,630 --> 01:03:00,260 You'll see charts on patient floors, right there for anyone to see, that will show 853 01:03:00,270 --> 01:03:06,060 how they are doing on patient falls, for instance, or how they are doing on infections. 854 01:03:06,060 --> 01:03:10,980 People have an attitude about safety, you just can feel it. 855 01:03:10,990 --> 01:03:12,820 There's an apocryphal story of President Kennedy 856 01:03:12,820 --> 01:03:14,570 visiting Cape Canaveral during his presidency 857 01:03:14,570 --> 01:03:19,910 and he takes aside a custodian and says, "What's your job?" And the custodian says, 858 01:03:19,910 --> 01:03:24,670 "Mr. President, my job is to help get a man to the moon and return him to earth safely." 859 01:03:25,080 --> 01:03:29,760 Everybody has a job to do to protect patients, not just doctors. 860 01:03:31,170 --> 01:03:37,100 Every nurse, pharmacist, physician, custodian, has a role in safety. 861 01:03:37,220 --> 01:03:43,020 I think it's deeply unfair to expect patients who are sick, in the middle of an illness, 862 01:03:43,020 --> 01:03:47,270 to try and sort this out on their own. Now, it may be unfair, 863 01:03:47,280 --> 01:03:50,690 but the reality is that's where we are. 864 01:03:50,700 --> 01:03:53,360 The best thing they can do is have a family member or a friend around, because again, 865 01:03:53,360 --> 01:03:57,280 when in the middle of an illness it's very hard for you to pay attention 866 01:03:57,290 --> 01:04:00,200 to know what's going on, but your friend can, your family member can. 867 01:04:00,210 --> 01:04:03,170 If somebody says you're going to get medication X, 868 01:04:03,170 --> 01:04:06,040 is that the medication that actually showed up? 869 01:04:06,050 --> 01:04:08,800 And asking in a very friendly and respectful way, 870 01:04:08,800 --> 01:04:12,300 when a nurse comes by to hang a medication or give you a pill, 871 01:04:12,300 --> 01:04:16,100 you know, what is this? What am I getting? It's a totally reasonable question. 872 01:04:16,100 --> 01:04:20,600 Patients should feel comfortable doing it. And if you have a provider 873 01:04:20,600 --> 01:04:23,860 that responds badly to that, you should try to figure out if you can switch providers. 874 01:04:23,860 --> 01:04:28,320 My father was a doctor in a small town in Connecticut. 875 01:04:28,320 --> 01:04:32,570 For a lot of time he was the only doctor there and he was revered. 876 01:04:32,570 --> 01:04:37,620 You didn't question him. It wasn't my father's fault in any way. 877 01:04:37,620 --> 01:04:42,210 He was a proud and successful professional honored by his community. 878 01:04:42,210 --> 01:04:45,880 That's not actually adaptive if we really want care to be what it can be. 879 01:04:45,880 --> 01:04:50,550 I think, I understand the hesitation people may feel to ask the doctor, 880 01:04:50,550 --> 01:04:53,360 "What's going on here?" But that's healthy, 881 01:04:53,360 --> 01:04:56,850 that's good, and we need to train doctors to, 882 01:04:56,890 --> 01:05:02,480 not just to accept that, but to absolutely welcome it. It's better medicine. 883 01:05:02,480 --> 01:05:06,940 [tense music] 884 01:05:07,820 --> 01:05:13,580 [Sully Sullenberger] If, as reports indicate, there are as many as 440,000 preventable 885 01:05:13,580 --> 01:05:19,000 medical deaths in this country alone every year, that is the equivalent of 886 01:05:19,000 --> 01:05:23,340 7 or 8 airliners crashing every day with no survivors. 887 01:05:23,340 --> 01:05:27,760 Before the first day of that kind of carnage was complete, airplanes would be grounded, 888 01:05:27,760 --> 01:05:33,300 airlines would stop operating, airports would close, no one would fly 889 01:05:33,310 --> 01:05:38,020 until some of the fundamental issues had been resolved. 890 01:05:38,020 --> 01:05:42,230 But because aviation accidents are dramatic, 891 01:05:42,230 --> 01:05:45,610 they receive the kind of attention that they do, and the public awareness. 892 01:05:45,610 --> 01:05:47,710 Medical deaths occur singly and often behind 893 01:05:47,710 --> 01:05:50,450 the scenes, but in aggregate the harm is huge. 894 01:05:52,490 --> 01:05:57,790 We need to change the way we think about these medical deaths. 895 01:05:57,790 --> 01:06:03,170 We need to think about them not as unavoidable, but as unthinkable. 896 01:06:03,300 --> 01:06:08,640 We've got to get better at making sure whatever hospital you go into in the U.S. 897 01:06:10,180 --> 01:06:15,390 you're getting the same quality care, and we are not there. 898 01:06:15,390 --> 01:06:18,480 I mean, you're asking people to do things differently. 899 01:06:18,480 --> 01:06:20,820 You're asking doctors to think differently and work differently. 900 01:06:20,820 --> 01:06:23,320 You're asking architects to build different spaces, nurses to work 901 01:06:23,320 --> 01:06:27,410 differently in teams, patients to have a different role. 902 01:06:27,410 --> 01:06:30,160 To change patient safety, you have to change everything. If you look at preventable harm 903 01:06:30,160 --> 01:06:34,160 across American hospitals, it has gone down considerably, 904 01:06:34,160 --> 01:06:37,210 you know, saving hundreds of thousands of lives 905 01:06:37,210 --> 01:06:39,380 and billions of dollars. That doesn't mean we fixed it. 906 01:06:39,380 --> 01:06:43,590 It is quietly, slowly, but definitely becoming the professional norm 907 01:06:43,970 --> 01:06:49,760 to take certain precautions, to do things in a certain way so that patients are safe. 908 01:06:49,760 --> 01:06:52,590 Because you can be satisfied if you have very 909 01:06:52,590 --> 01:06:55,600 low expectations and the reality is that all of 910 01:06:55,600 --> 01:07:00,940 our expectations should be raised, that we all get very high quality, safe healthcare. 911 01:07:00,940 --> 01:07:05,280 We all want to do well. We all want to get better. Nobody comes to work 912 01:07:05,280 --> 01:07:10,700 to harm a patient or wanting to harm a patient or to give bad care. 913 01:07:10,700 --> 01:07:15,500 We haven't made this a public health issue where the public is really 914 01:07:15,500 --> 01:07:19,340 thinking about this, and yet when you talk to any person 915 01:07:19,340 --> 01:07:23,380 who's had a family member or themselves in healthcare, they all have a story. 916 01:07:23,380 --> 01:07:26,850 I've also talked to doctors and nurses who have committed a terrible error and they say, 917 01:07:26,850 --> 01:07:30,520 "I know I can't take that back, but what will really give that meaning is 918 01:07:30,520 --> 01:07:34,310 if I do something that makes the system safer for the next person." 919 01:07:34,310 --> 01:07:37,230 I think, in part, the job of people like me in leadership roles 920 01:07:37,230 --> 01:07:40,110 is to harness that passion, harness that energy. 921 01:07:40,110 --> 01:07:42,780 All of the rest of these guys are much more serious about medical error reduction 922 01:07:42,780 --> 01:07:46,620 than they ever were. Is it going as fast as it could? No, of course not. 923 01:07:46,620 --> 01:07:50,080 It is not "stuff happens" anymore. 924 01:07:50,080 --> 01:07:53,580 That's where we're going, and that's the good future that we're moving towards. 925 01:07:53,580 --> 01:07:57,380 It feels like we should be further along than we are, but actually 926 01:07:57,380 --> 01:08:00,630 I think we've made tremendous progress in 15 years. It is on the map. 927 01:08:00,630 --> 01:08:04,010 We have these examples in the U.S. and around the world. 928 01:08:04,010 --> 01:08:07,810 It's not any longer a question of possibility, it's a question of will. 929 01:08:14,020 --> 01:08:19,740 Many of us go kind of through a self-blame. Although we know it wasn't our fault, 930 01:08:19,740 --> 01:08:24,490 we feel like we didn't... protect our son. 931 01:08:24,780 --> 01:08:29,870 And so there was really, really significant grieving, 932 01:08:29,870 --> 01:08:34,420 so the anger at first was immeasurable. 933 01:08:34,420 --> 01:08:39,010 When we discovered Pat's error, we both felt tremendous fear. 934 01:08:39,010 --> 01:08:43,590 I think at that point it was just plain disbelief. 935 01:08:43,600 --> 01:08:47,680 He said: "Whatever you do, do not give up on patient safety." 936 01:08:47,680 --> 01:08:52,190 So, that led me onto a journey to... I wanted to make sure our healthcare system, 937 01:08:52,190 --> 01:08:55,940 our government knew what happened to Pat and Cal. 938 01:08:55,940 --> 01:08:59,360 So, it took us 8 years, but we really did make some changes in our healthcare system 939 01:08:59,360 --> 01:09:03,830 where babies being discharged would have a bilirubin test before they were discharged. 940 01:09:03,830 --> 01:09:09,250 [Mark Graber]Thanks to Sue and the work that she's done, there are now processes 941 01:09:09,250 --> 01:09:13,090 in place in every hospital to screen for that condition. 942 01:09:13,090 --> 01:09:16,420 And the odds that that's going to happen again are now approaching zero, 943 01:09:16,420 --> 01:09:20,890 and that's what we'd like to see happen throughout medicine. 944 01:09:20,890 --> 01:09:23,890 And the work that Sue has donenis our model for how to do that. 945 01:09:23,890 --> 01:09:28,230 She turned what she had gone through into empowerment and positivity, 946 01:09:28,520 --> 01:09:33,730 and if she can do it, so can I, and so can a lot of people. 947 01:09:33,730 --> 01:09:38,490 I've obviously always idolized my Mom and I understood her job very well 948 01:09:38,490 --> 01:09:43,580 and people would ask me, "What does your Mom do?" I would say, "Well, she saves lives." 949 01:09:43,580 --> 01:09:47,120 Having witnessed these tragic outcomes in our healthcare system, 950 01:09:47,120 --> 01:09:51,380 the one place that we should feel unquestionably safe. 951 01:09:51,380 --> 01:09:57,130 [Mackenzie Sheridan] And it kind of ignited a fire inside me that wanted to, you know, 952 01:09:57,140 --> 01:10:01,470 do what my Mom does, which is, you know, talk with hospitals and talk with doctors 953 01:10:01,470 --> 01:10:05,230 and figure out how we can make those kind of things not happen again. 954 01:10:05,230 --> 01:10:09,110 So, I went to Portland State University and I chose to do Public Health 955 01:10:09,110 --> 01:10:14,070 because I wanted to feel like I was making a difference and feel like I, you know, 956 01:10:14,070 --> 01:10:20,330 could help prevent things that happened to my family, happening to other people. 957 01:10:28,210 --> 01:10:33,420 I am unwilling to believe that we have done all that we can do. 958 01:10:33,420 --> 01:10:38,140 My experience with diagnostic errors and the healthcare system has been without a doubt 959 01:10:38,850 --> 01:10:44,310 the most powerfully emotional experience in my life. 960 01:10:44,480 --> 01:10:50,110 However, my family's story is also a story of awakening, 961 01:10:50,110 --> 01:10:55,490 of passion, of change, and hope for the future. 962 01:10:56,660 --> 01:11:01,450 I cannot change what happened to Cal and Pat, but I've always felt 963 01:11:02,040 --> 01:11:06,880 that I can somehow be part of it and make a difference. 964 01:11:06,880 --> 01:11:10,590 My teacher in courage, in hope, 965 01:11:11,420 --> 01:11:15,640 in determination, in passion, 966 01:11:15,640 --> 01:11:19,010 of course he's my teacher in sense of humor which he believes his mother has none of, 967 01:11:19,010 --> 01:11:24,230 but he's the reason for what's in me. 968 01:11:51,130 --> 01:11:55,970 [audience applauding] 969 01:12:04,360 --> 01:12:10,240 [Sue Sheridan] You know, when Pat was dying he said, "Never give up on patient safety." 970 01:12:10,240 --> 01:12:14,950 At that time, I did not envision my whole family being engaged. 971 01:12:14,950 --> 01:12:18,620 Before we went on stage today, I thought about Pat, my daughter in the front row, 972 01:12:18,620 --> 01:12:23,380 my son on stage. It was, umm, just surreal. 973 01:12:28,470 --> 01:12:33,300 [singing Happy Birthday] 974 01:12:42,020 --> 01:12:47,820 [Mackenzie Sheridan] On March 8th, which is the day that my dad passed away, 975 01:12:47,820 --> 01:12:51,280 we spread his ashes on Table Rock. 976 01:12:51,370 --> 01:12:54,590 Whenever we go there I always feel like a warm, just like, presence. 977 01:12:57,160 --> 01:13:01,370 It's because it's such a beautiful place, 978 01:13:01,380 --> 01:13:03,290 and it's beautiful that he's there as well. 979 01:13:03,290 --> 01:13:06,920 [Sue Sheridan] Pat will always be alive in our hearts and in our memories, 980 01:13:08,670 --> 01:13:12,640 and it was very hard for them to lose a Dad when they were only 4 and 6. 981 01:13:12,640 --> 01:13:16,220 They will continue to honor andnmiss and wonder about their Dad. 982 01:13:21,560 --> 01:13:23,400 I've always had this hope: 983 01:13:28,280 --> 01:13:31,310 I will not believe that our leadership in our country, in 984 01:13:31,320 --> 01:13:34,450 our healthcare system, will continue to think this is okay. 985 01:13:34,450 --> 01:13:36,910 Because it's not. 986 01:13:46,680 --> 01:13:51,350 [piano music] 95192

Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.