Would you like to inspect the original subtitles? These are the user uploaded subtitles that are being translated:
1
00:00:02,000 --> 00:00:07,000
Downloaded from
YTS.MX
2
00:00:08,000 --> 00:00:13,000
Official YIFY movies site:
YTS.MX
3
00:00:17,810 --> 00:00:22,520
It was devastating. It was very hard.
4
00:00:22,520 --> 00:00:25,990
It was hard for us to understand
and believe that this could happen
5
00:00:26,280 --> 00:00:31,700
in a developed country
like the United States.
6
00:00:33,080 --> 00:00:37,620
Both my husband, Pat, and
my son, Cal, experienced
7
00:00:37,620 --> 00:00:41,290
what I would say classic diagnostic errors.
8
00:00:41,290 --> 00:00:44,670
Cal suffered brain damage from his
newborn jaundice when it was misdiagnosed
9
00:00:44,670 --> 00:00:49,050
and it was never tested
or treated appropriately,
10
00:00:49,050 --> 00:00:51,350
and today he has
significant cerebral palsy.
11
00:00:51,350 --> 00:00:54,520
At about 16 hours, the nurse
charted that he was yellow,
12
00:00:55,810 --> 00:00:59,360
it was no big deal. We
were basically discharged
13
00:00:59,360 --> 00:01:02,980
with a very sick baby, but we
were told he was a well baby.
14
00:01:02,980 --> 00:01:06,070
I was familiar with jaundice
and it was communicated to us
15
00:01:06,070 --> 00:01:10,200
that it was no big deal
and not to worry about it.
16
00:01:10,200 --> 00:01:12,030
And they asked me if I
was a first-time mom.
17
00:01:12,040 --> 00:01:14,200
I said I was, and they reminded
me that first-time moms are
18
00:01:14,210 --> 00:01:19,080
often over-reactive, and they
didn't seem worried at all.
19
00:01:19,090 --> 00:01:23,710
I didn't really know at the time,
20
00:01:23,720 --> 00:01:26,970
but I learned later on that Cal
was in the process of dying.
21
00:01:26,970 --> 00:01:32,060
We actually watched our son suffer brain
damage in the hospital before our eyes,
22
00:01:33,230 --> 00:01:38,730
and... Quite honestly,
that will haunt me forever.
23
00:01:40,070 --> 00:01:44,400
And Pat, my husband, died
when he was 45 from cancer,
24
00:01:45,110 --> 00:01:49,780
a cancer that was appropriately diagnosed,
25
00:01:49,790 --> 00:01:52,870
but the pathology failed to get
communicated to the doctor or Pat.
26
00:01:52,870 --> 00:01:57,920
They did an MRI and they
discovered that there was a mass
27
00:01:57,920 --> 00:02:00,590
in his neck at the base of the
skull, and so Pat had surgery.
28
00:02:00,590 --> 00:02:04,220
Six months later, the pain
returned in Pat's neck.
29
00:02:04,220 --> 00:02:07,890
A whole series of doctors
came through asking Pat
30
00:02:07,890 --> 00:02:10,810
why he never got treatment
after his first surgery,
31
00:02:10,810 --> 00:02:14,440
and I had all the documents, and I
said, well, because it was benign.
32
00:02:14,440 --> 00:02:17,400
And then by the time the third doctor
came through, I said, wait a second,
33
00:02:17,400 --> 00:02:22,030
what was his pathology
on the first surgery?
34
00:02:22,030 --> 00:02:25,160
And the final pathology was a high-grade
malignant synovial cell sarcoma.
35
00:02:25,160 --> 00:02:27,100
And that document either
never arrived or was
36
00:02:27,100 --> 00:02:29,620
placed in his chart
without the doctor seeing.
37
00:02:30,830 --> 00:02:36,040
And I remember showing it to
Pat, and I remember Pat crying.
38
00:02:36,040 --> 00:02:41,260
You know, to think that another error
had taken place, and this time with him,
39
00:02:41,260 --> 00:02:46,090
that was difficult for us to
witness in our healthcare system.
40
00:02:46,100 --> 00:02:49,720
[somber music]
41
00:02:49,730 --> 00:02:51,980
Stories like Sue Sheridan
and what happened to her,
42
00:02:51,980 --> 00:02:55,900
where a small mistake can really be a
life-altering event that remind us the human cost
43
00:02:56,150 --> 00:03:01,490
of what we're talking about.
These are not theoretical events.
44
00:03:01,490 --> 00:03:04,280
These are not just things that happen
to other people. They happen to us.
45
00:03:04,280 --> 00:03:07,540
They happen to our families. And they
are things that we need to work on.
46
00:03:07,540 --> 00:03:11,750
[tense music]
47
00:03:12,210 --> 00:03:17,880
[narrator]In 1999, the first
significant report on medical mistakes
48
00:03:17,880 --> 00:03:22,010
was released by the Institute of
Medicine. They called it To Err Is Human.
49
00:03:22,010 --> 00:03:24,710
This report claimed that
as manyas 98,000 people
50
00:03:24,710 --> 00:03:27,850
die every yearnas a result
of medical mistakes.
51
00:03:28,980 --> 00:03:34,610
Over the next 15 years, efforts to
better understand this number increased,
52
00:03:34,900 --> 00:03:39,820
but so did the number itself.
53
00:03:41,200 --> 00:03:44,540
Recent studies have raised the projected
number of deaths to as high as 440,000. To
54
00:03:44,540 --> 00:03:48,470
put this in perspective, that's more than
the number of graves in Arlington Cemetery.
55
00:03:48,750 --> 00:03:53,960
It's the equivalent of 2-3
jumbonjets crashing every single day.
56
00:03:53,960 --> 00:03:58,880
So, where does that rank medical
mistakes on the leading causes of death?
57
00:03:58,880 --> 00:04:03,100
Number three. Right behind
cancer and heart disease.
58
00:04:04,930 --> 00:04:08,480
Now suddenly, whoa. This isn't
just some egg-headed study.
59
00:04:10,100 --> 00:04:14,860
This is a big deal. This could
be you, and they're right.
60
00:04:14,860 --> 00:04:17,610
Wait a second, you mean those hospitals,
my local hospital was killing people?
61
00:04:17,610 --> 00:04:21,870
Is that what you're really saying?
62
00:04:21,870 --> 00:04:24,160
We could prevent many, many,
many of these deaths immediately
63
00:04:24,160 --> 00:04:28,660
if we just put in the effort.
64
00:04:28,670 --> 00:04:30,830
Things are happening.
Let's take a look at this.
65
00:04:30,840 --> 00:04:33,630
I just think this is like a massive
epidemic that we have underestimated,
66
00:04:33,630 --> 00:04:37,840
and the reason is because it's happening
to people who are already sick.
67
00:04:37,840 --> 00:04:43,260
But, like, they were sick, that
doesn't mean they were going to die.
68
00:04:43,260 --> 00:04:45,850
And their death is no less of a tragedy
because they already had a medical problem.
69
00:04:45,850 --> 00:04:50,860
Every time you get on a plane, you
don't expect that plane to crash.
70
00:04:50,860 --> 00:04:53,980
And everybody who dies in a plane crash,
you say, "Well, those people were healthy.
71
00:04:53,990 --> 00:04:57,150
They were going to do fine otherwise."
I think the problem with patient safety
72
00:04:57,160 --> 00:05:00,620
is you say, "Oh, well, these
people were sick anyway."
73
00:05:00,620 --> 00:05:02,700
And I think it's a very problematic
way to look at the world.
74
00:05:02,700 --> 00:05:04,110
Maybe they didn't die,
but they spend the rest
75
00:05:04,110 --> 00:05:05,580
of their life in a
wheelchair or a nursing home
76
00:05:05,580 --> 00:05:09,460
and that accelerates their death and
obviously harms their quality of life.
77
00:05:09,460 --> 00:05:12,710
So, the numbers about
deaths are a big deal,
78
00:05:12,710 --> 00:05:16,010
but in some ways they underestimate
the overall toll of preventable harm.
79
00:05:16,010 --> 00:05:20,140
We don't have a stable, agreed
way to measure safety or injuries.
80
00:05:20,390 --> 00:05:25,390
Actually, the number you
get depends on how you look.
81
00:05:25,390 --> 00:05:28,520
One rule is the harder you
look, the more you find.
82
00:05:28,520 --> 00:05:31,110
So, when you really throw the
book at it and you do everything
83
00:05:31,110 --> 00:05:35,070
you can to look for injuries,
you're going to find a ton of them.
84
00:05:35,070 --> 00:05:38,120
When people start debating, you know,
is it 40,000 or 90,000 or 100,000?
85
00:05:38,120 --> 00:05:43,830
Uh, it's a lot. It's a ton.
And our job is to make it zero.
86
00:05:43,910 --> 00:05:48,880
This is urgent. It's a
public health emergency.
87
00:05:48,880 --> 00:05:53,220
[narrator] While the number ofdeaths
related to medical error is staggering,
88
00:05:54,510 --> 00:05:58,220
the number of patients who experience
non-fatal errors is even bigger.
89
00:05:58,220 --> 00:06:00,400
Recent studies suggest
one-third of all hospital
90
00:06:00,400 --> 00:06:02,770
admissions experience a medical mistake,
91
00:06:03,230 --> 00:06:08,690
and 1.7 million hospital-acquired
infections occur every year.
92
00:06:08,690 --> 00:06:13,450
69% of those infections
could have been prevented
93
00:06:13,950 --> 00:06:17,280
through methods that already
exist, like hand washing.
94
00:06:17,280 --> 00:06:20,910
But healthcare workers wash their
hands less than 50% of the time,
95
00:06:20,910 --> 00:06:25,210
with some research suggesting
it's as low as 30%.
96
00:06:25,210 --> 00:06:28,880
There are even more dramatic
examples. In a five-year span,
97
00:06:29,510 --> 00:06:33,970
surgeons operated on the wrong
body part over 2,000 times,
98
00:06:33,970 --> 00:06:38,640
left nearly 5,000 tools inside patients,
99
00:06:38,930 --> 00:06:42,310
and in 27 cases operated on
the wrong patient entirely.
100
00:06:42,690 --> 00:06:47,360
But diagnostic errors, like the ones that
left Cal Sheridan with cerebral palsy
101
00:06:48,360 --> 00:06:50,640
and delayed the detection ofPat Sheridan's
102
00:06:50,640 --> 00:06:53,030
cancer contribute to 1
in 10 patient deaths.
103
00:06:53,030 --> 00:06:55,950
But whether it's a diagnostic error or any
104
00:06:55,950 --> 00:06:59,200
other preventable harm,
the only way to fix it
105
00:06:59,200 --> 00:07:04,210
is to first understand what causes it.
106
00:07:04,210 --> 00:07:07,300
[Boaz Keysar] When we study communication
in my lab, we look at how people communicate
107
00:07:07,300 --> 00:07:12,180
and what are the reasons
for miscommunication.
108
00:07:12,180 --> 00:07:13,980
In very simple experiments,
when we ask people
109
00:07:13,980 --> 00:07:16,160
to communicate something to somebody else,
110
00:07:17,100 --> 00:07:22,020
about 50% of the time when they thought
the other person understood them,
111
00:07:22,600 --> 00:07:26,730
they were wrong. Now, I don't know the
extent of miscommunication in medicine,
112
00:07:26,730 --> 00:07:32,110
but I am sure it is more
than, uh, physicians think.
113
00:07:32,280 --> 00:07:38,370
Part of the problem is that when you,
when, when, say a doctor miscommunicates,
114
00:07:38,370 --> 00:07:43,000
he or she might not know. That's
the core of the problem, right?
115
00:07:43,710 --> 00:07:47,800
They might not get immediate
feedback that they miscommunicated.
116
00:07:47,800 --> 00:07:51,130
And if that happens, then that error
117
00:07:51,140 --> 00:07:56,430
could amplify without anybody
realizing that the source was
118
00:07:56,560 --> 00:08:01,060
just a minor miscommunication.
119
00:08:01,060 --> 00:08:04,020
Now I know how... what happened to my
husband. Now I understand how it happened,
120
00:08:04,020 --> 00:08:08,700
that there's been no system-based
intervention to ensure
121
00:08:09,780 --> 00:08:14,450
that lab tests are followed up on,
122
00:08:14,450 --> 00:08:17,120
that pathologies and radiology
reports are followed up on.
123
00:08:17,120 --> 00:08:20,630
To know that this happens in our
country, that's unacceptable.
124
00:08:23,170 --> 00:08:25,420
[siren wailing]
125
00:08:27,380 --> 00:08:28,890
[tense music]
126
00:08:30,930 --> 00:08:36,230
[narrator 2]Most of us think
of a hospital as a place
127
00:08:36,230 --> 00:08:38,600
Nwhere people go after
they have an accident,
128
00:08:38,600 --> 00:08:41,730
not as a place where people
go to have accidents.
129
00:08:41,730 --> 00:08:45,320
However, like just about any place,
there are safety hazards in a hospital.
130
00:08:46,240 --> 00:08:51,080
Some are unique to the hospital
environment, and some are not.
131
00:08:51,580 --> 00:08:56,040
Generally, the hospital staff is very
aware of medical safety practices,
132
00:08:56,040 --> 00:09:01,040
such as the proper handling
of infectious cases,
133
00:09:01,040 --> 00:09:04,800
careful checking of patient ID
before administering any medication,
134
00:09:04,800 --> 00:09:09,760
keeping things sanitary and disinfected.
135
00:09:09,760 --> 00:09:11,540
Yet, all of us at times
tend tooverlook some
136
00:09:11,550 --> 00:09:13,600
potential hazards that
we are around every day.
137
00:09:13,600 --> 00:09:18,310
We must try to learn to think
safety in everything we do.
138
00:09:19,020 --> 00:09:23,690
But safety doesn't come just by learning
a lot of rules. It comes from an attitude.
139
00:09:24,150 --> 00:09:29,030
For everyone who works in a hospital,
safety has to be a full time job.
140
00:09:33,040 --> 00:09:36,340
[Albert Wu] This is a problem that's,
you know, hiding in plain sight.
141
00:09:38,920 --> 00:09:42,670
And I think that no one is really surprised
when they think about it for a minute.
142
00:09:42,670 --> 00:09:47,510
If we think the amount of harm that
is currently existing is just fine,
143
00:09:47,510 --> 00:09:51,600
then maybe it's not a crisis, it's not a
problem. If that's okay, then we're done.
144
00:09:51,600 --> 00:09:56,730
Most of us in medicine just said,
"Well, that's the way it is, you know.
145
00:09:56,730 --> 00:10:00,980
Things go wrong. People make mistakes.
There's nothing you can do about it."
146
00:10:00,980 --> 00:10:05,030
It's pretty obvious that safety is not
number one priority in most hospitals.
147
00:10:05,030 --> 00:10:09,700
When it is, wonderful things happen.
148
00:10:09,700 --> 00:10:12,540
What is the problem you're trying to solve?
149
00:10:12,540 --> 00:10:15,670
And the answer is, for most
hospital administrators,
150
00:10:15,670 --> 00:10:19,090
life is too short to get
the doctors angry at you.
151
00:10:19,090 --> 00:10:22,670
Building a new cancer center,
your oncologists love you,
152
00:10:22,670 --> 00:10:26,390
the other doctors love you, it brings
in revenue, the community loves you.
153
00:10:26,390 --> 00:10:29,760
If you reduce medical error,
you can't advertise it because
154
00:10:29,770 --> 00:10:33,390
the patients all think that everything's
safe anyway. Nobody knows the problem exists.
155
00:10:33,400 --> 00:10:37,360
The doctors are angry because you
start to talk about medical error.
156
00:10:37,360 --> 00:10:40,820
So, that's why you have
an invisible problem.
157
00:10:40,820 --> 00:10:44,490
Every human being will make mistakes, and
will... so the goal cannot be zero errors.
158
00:10:44,490 --> 00:10:49,910
Our goal needs to be zero harm,
because we know errors will occur.
159
00:10:49,910 --> 00:10:53,960
So, how do we make sure those errors don't
actually lead to harm and are caught early?
160
00:10:53,960 --> 00:10:59,260
10 or 15 years ago, we thought central
line infections were not preventable.
161
00:10:59,260 --> 00:11:04,010
We thought that was part of kind of
doing business in healthcare that, okay,
162
00:11:04,010 --> 00:11:08,060
people have central lines,
occasionally they'll get infections,
163
00:11:08,060 --> 00:11:11,020
and that's just... Now we know
infections can go down to zero.
164
00:11:11,020 --> 00:11:16,150
Preventing preventable harm is a skill
and a commitment and a technology
165
00:11:17,780 --> 00:11:23,280
all of its own. It's not glamorous,
but it's what keeps all of us safe.
166
00:11:23,280 --> 00:11:28,040
If you believe, "First, do no harm",
167
00:11:28,040 --> 00:11:31,670
there is no excuse for not investing
in things which will prevent harm.
168
00:11:32,080 --> 00:11:37,050
Health care nowadays is
incredibly complicated.
169
00:11:37,050 --> 00:11:40,090
A patient has literally
hundreds of things done to them,
170
00:11:40,090 --> 00:11:44,720
having blood drawn for a test
or getting an x-ray or whatever.
171
00:11:44,720 --> 00:11:48,810
And so, there are many, many, many
opportunities for things to go wrong.
172
00:11:48,810 --> 00:11:53,230
So, even when nurses and doctors
and technicians and radiologists
173
00:11:53,230 --> 00:11:57,490
are functioning at a 99% level, which is,
you know, pretty good for human activity,
174
00:11:57,490 --> 00:12:02,660
that still means a lot of
opportunity for things to go wrong.
175
00:12:02,660 --> 00:12:07,000
I think this is a general problem that
you have when you deal with people.
176
00:12:07,000 --> 00:12:12,630
We are not built to not make mistakes.
We are not built to be perfect.
177
00:12:13,170 --> 00:12:16,090
Are you going to try
and change the person or
178
00:12:16,090 --> 00:12:19,130
are you going to try and
change the situation?
179
00:12:19,130 --> 00:12:24,430
One way to do it is to design,
say the work environment,
180
00:12:24,430 --> 00:12:30,020
in a way that would not
necessarily prevent the error,
181
00:12:32,400 --> 00:12:36,860
but would assume the error.
182
00:12:36,860 --> 00:12:39,950
We have to acknowledge that to err
is human, and then to figure out
183
00:12:39,950 --> 00:12:45,000
what do we do with that fact in terms of
building a system that's safe for patients.
184
00:12:45,160 --> 00:12:49,830
[tense music]
185
00:12:50,290 --> 00:12:55,630
[Sue Sheridan] Between Cal's patient safety
event and Pat's patient safety event,
186
00:12:55,630 --> 00:13:00,180
we had Mackenzie in the middle there.
Exactly at 16 hours, just like Cal,
187
00:13:00,180 --> 00:13:04,890
she also had a very high bilirubin,
which the hospital took action.
188
00:13:05,730 --> 00:13:10,770
They tested it and they
treated it. I took a shower.
189
00:13:10,770 --> 00:13:14,650
And it was the first shower
after delivery and I remember
190
00:13:14,650 --> 00:13:17,780
I stayed in the shower for an hour and they
sent a female chaplain in, and I was crying.
191
00:13:17,780 --> 00:13:19,970
And the chaplain thought
I was crying because my
192
00:13:19,970 --> 00:13:22,580
daughter was getting
treated for her jaundice
193
00:13:22,620 --> 00:13:28,290
and I explained to them I was not crying
because of that. I was crying because
194
00:13:28,290 --> 00:13:32,170
I witnessed what the only thing
they had to do with my son,
195
00:13:32,170 --> 00:13:35,970
that it was so easy to prevent
what happened to my son.
196
00:13:37,050 --> 00:13:39,810
[tense music]
197
00:13:42,350 --> 00:13:45,090
[Mackenzie Sheridan] When I got into about
first grade, people started asking me,
198
00:13:45,090 --> 00:13:47,770
"What's wrong with your brother? Why,
like, can't he move like the rest of us?"
199
00:13:47,770 --> 00:13:51,440
I didn't really get it, because
I was never told necessarily,
200
00:13:51,440 --> 00:13:55,990
you know, your brother has cerebral
palsy, your brother has kernicterus.
201
00:13:55,990 --> 00:13:59,490
You know, to me, he was just my brother.
202
00:14:05,920 --> 00:14:10,630
[ambient music]
203
00:14:40,660 --> 00:14:46,500
[Mackenzie Sheridan] Recently, I became more
interested in the case, my brother's case,
204
00:14:46,500 --> 00:14:52,380
because I knew, before looking it up, I
knew that he wasn't given a bilirubin test
205
00:14:52,800 --> 00:14:57,970
and because of that he got
cerebral palsy and kernicterus.
206
00:14:57,970 --> 00:15:01,810
And I got frustrated and I
got angry and confused and
207
00:15:01,810 --> 00:15:04,720
my mom has taught me
that I can do something
208
00:15:04,720 --> 00:15:07,690
positive with that kind
of anger and fervor.
209
00:15:07,690 --> 00:15:13,070
I can, you know, go out and make sure
that those kind of things don't happen.
210
00:15:20,330 --> 00:15:25,420
So, I used to be a little
scared hearing all of the things
211
00:15:25,420 --> 00:15:28,550
that could go wrong in the health system.
212
00:15:28,550 --> 00:15:31,550
I just learned to be cautious and
to ask questions and to, you know,
213
00:15:31,550 --> 00:15:36,600
ask the doctors, "What are you doing? Have
you washed your hands? Have you done this?"
214
00:15:36,600 --> 00:15:39,930
I look at doctors in a
different sense than, I think,
215
00:15:39,930 --> 00:15:42,650
a lot of people do and as a child I
looked at doctors differently as well.
216
00:15:42,650 --> 00:15:44,300
I know why kids would
think like a doctor don't
217
00:15:44,300 --> 00:15:45,980
make mistakes, but I knew
from a very young age
218
00:15:45,980 --> 00:15:50,450
that they do, and that their
mistakes could cost a life.
219
00:15:50,450 --> 00:15:54,370
The first thing that we
wanted was to tell somebody.
220
00:15:54,370 --> 00:15:56,700
Some kind of high authority that
could tell all of the hospitals
221
00:15:56,700 --> 00:16:01,460
about what happened, so all
hospitals could implement change.
222
00:16:01,460 --> 00:16:05,340
And I thought somebody was in charge
of patient safety in the United States,
223
00:16:05,340 --> 00:16:10,340
and I learned that that
simply does not exist.
224
00:16:12,550 --> 00:16:18,180
When people think about science in
healthcare, they think about genes and cells
225
00:16:18,190 --> 00:16:23,650
and drugs and chemistry. Yeah,
that's science. That's one science.
226
00:16:23,650 --> 00:16:28,360
But, there's another science, which
is the science of organizing care,
227
00:16:28,360 --> 00:16:30,470
which is how to you
actually get the help, what
228
00:16:30,480 --> 00:16:32,620
are the flows like, how do you do surgery.
229
00:16:32,620 --> 00:16:38,080
How do you take care of a chronic
illness. There's science there too,
230
00:16:38,080 --> 00:16:42,420
and luckily this country began investing
in that really in the past few decades.
231
00:16:42,420 --> 00:16:46,300
The Agency for Healthcare Research and
Quality, for example, it's an American
232
00:16:46,300 --> 00:16:49,590
investment in developing the
sciences for delivering better care.
233
00:16:49,590 --> 00:16:54,260
[narrator] In 2000, after speaking
with leaders in healthcare,
234
00:16:54,270 --> 00:16:58,190
President Bill Clinton made
a bold statement regarding
235
00:16:58,190 --> 00:17:00,980
the country's new efforts
in managing medical errors.
236
00:17:00,980 --> 00:17:04,740
Just think about it, we can cut preventable
medical errors in half in five years.
237
00:17:05,280 --> 00:17:11,030
[narrator] The Agency forHealthcare
Research and Quality took on this task.
238
00:17:11,030 --> 00:17:16,200
Today, AHRQ remains focused
on improving the quality
239
00:17:16,210 --> 00:17:20,000
and safety of healthcare for Americans.
240
00:17:20,000 --> 00:17:23,010
It does so by funding research,
developing tools and training,
241
00:17:23,920 --> 00:17:27,840
and collecting measures and data on
the healthcare system as a whole.
242
00:17:27,840 --> 00:17:31,640
In 2016, a report was released on the
recent progress in patient safety efforts.
243
00:17:32,060 --> 00:17:37,440
The report showed that
fromn2010-2015, there were 3 million
244
00:17:37,440 --> 00:17:42,030
fewer hospital-acquired conditions,
showing a 21% reduction.
245
00:17:42,030 --> 00:17:47,490
125,000 deaths were prevented, saving
$28 billion in healthcare costs.
246
00:17:47,530 --> 00:17:53,250
All with a budget that annually
hovered between $400-$450 million.
247
00:17:53,250 --> 00:17:58,960
But it's part of a healthcare
system that spends over $3 trillion,
248
00:17:58,960 --> 00:18:03,260
and has more than 5,000 hospitals, with
over 800,000 physicians, 4 million nurses,
249
00:18:03,260 --> 00:18:09,600
and 330 million patients. That means the
agency is working with 1/100th of a percent
250
00:18:09,600 --> 00:18:16,060
of national health spending and is
tasked with improving the other 99.99%.
251
00:18:16,940 --> 00:18:21,690
It is such an underinvestment that,
you know, a doubling of the amount
252
00:18:22,450 --> 00:18:27,200
for the agency would be a vast improvement,
but it still is not nearly enough.
253
00:18:27,200 --> 00:18:31,910
We need this information for us to take care
of our patients properly, for health plans,
254
00:18:31,910 --> 00:18:36,630
for leaders of large clinics to say,
"Actually, no, I need to better understand
255
00:18:36,630 --> 00:18:41,880
the choices I make, how it impacts
our ability to deliver safe care."
256
00:18:41,880 --> 00:18:46,090
It has funded some of the seminal studies
that have had massive improvements
257
00:18:46,100 --> 00:18:49,770
in patient safety. So, it funded the studies
that led us to create the checklists for
258
00:18:49,770 --> 00:18:54,350
central line infections. That alone has
saved the American healthcare system
259
00:18:54,350 --> 00:18:59,110
hundreds of millions of dollars,
if not billions of dollars,
260
00:18:59,110 --> 00:19:01,530
but more importantly, has probably
saved tens of thousands of lives.
261
00:19:01,530 --> 00:19:04,410
[Ashish Jha] There are tens of thousands
of Americans walking around today
262
00:19:04,410 --> 00:19:08,500
who would be dead if it had not been for
some of the work that AHRQ has funded.
263
00:19:08,790 --> 00:19:14,290
It's really about how we
apply the best of science
264
00:19:15,040 --> 00:19:19,760
to your individual needs and preferences.
265
00:19:19,760 --> 00:19:23,010
To some extent I do know some systems
that are doing a terrific job,
266
00:19:23,010 --> 00:19:26,970
and when I learn from them about
how they are doing it, a lot of them
267
00:19:26,970 --> 00:19:30,770
are using the tools and
methods pioneered by AHRQ.
268
00:19:35,820 --> 00:19:40,530
[ambient music]
269
00:19:41,160 --> 00:19:43,700
Much of the work that we use
to train around patient safety
270
00:19:43,700 --> 00:19:47,450
and how to make healthcare safer is actually
derived from AHRQ research and tools.
271
00:19:47,450 --> 00:19:53,040
When they put out a toolkit or research
tools, I know that they've been vetted
272
00:19:53,040 --> 00:19:58,090
and they've been tried and investigated
and shown to be of benefit.
273
00:19:58,090 --> 00:20:02,720
So, a big problem that we face in safety
in hospitals is really improving handoffs,
274
00:20:02,890 --> 00:20:07,930
which is when a patient moves
from one area to another
275
00:20:07,930 --> 00:20:11,270
or when their doctors
or nurses change shifts.
276
00:20:11,270 --> 00:20:12,780
Handoffs are somewhat
invisible to patients,
277
00:20:12,780 --> 00:20:14,360
but they actually have
a huge impact on them.
278
00:20:14,360 --> 00:20:19,150
Like, if an average patient
got hospitalized tomorrow,
279
00:20:19,160 --> 00:20:21,990
they would face upwards of 15 handovers.
280
00:20:21,990 --> 00:20:24,450
And we know from AHRQ-funded research,
it's got to be more than just
281
00:20:24,450 --> 00:20:28,080
a passive listening where you're
like, uh-huh, okay, I got it,
282
00:20:28,080 --> 00:20:31,880
but really engage, ask questions, because
often times you'll pick up things.
283
00:20:31,880 --> 00:20:36,670
Combining AHRQ TeamSTEPPS, with a
standardized tool to improve handoffs
284
00:20:37,340 --> 00:20:42,140
actually led to a 30% reduction
in preventable adverse events.
285
00:20:42,140 --> 00:20:46,680
We also develop our own home-grown
patient safety teaching programs.
286
00:20:46,680 --> 00:20:51,310
One of my personal favorites
that we've actually
287
00:20:51,310 --> 00:20:53,900
developed here is called
the Room of Horrors.
288
00:20:53,900 --> 00:20:57,110
We take 10 patient safety hazards
289
00:20:57,110 --> 00:20:59,740
and we embed it into a hospital
room, into a simulation.
290
00:20:59,740 --> 00:21:02,870
This is training where
you're walking into a room
291
00:21:02,870 --> 00:21:05,960
and you're actually seeing with your
own eyes, can you spot what's wrong?
292
00:21:05,960 --> 00:21:07,970
[Trainee 1] Ammonia. C-diff positive. So,
293
00:21:07,970 --> 00:21:10,920
probably should be some
kind of like precautions.
294
00:21:11,670 --> 00:21:16,590
[Trainee 2] Yeah, he
should be contacted less.
295
00:21:16,590 --> 00:21:18,430
Allergies, latex and
penicillin. That's fine.
296
00:21:18,430 --> 00:21:22,060
Umm. Let's see here. Oh,
those are gloves over there.
297
00:21:22,060 --> 00:21:26,390
[Trainee 2] Are these latex
gloves? Uh-oh, we got latex gloves.
298
00:21:27,020 --> 00:21:31,480
So, it looks like he's got
some [unintelligible] hanging,
299
00:21:31,480 --> 00:21:33,780
and he's allergic to penicillin
so that's definitely not ideal.
300
00:21:33,780 --> 00:21:38,530
[Trainee 1] Yes, absolutely.
Why does he have magnesium?
301
00:21:38,530 --> 00:21:41,370
I don't know. It's actually not for his
name. His name is Washington, right?
302
00:21:41,370 --> 00:21:45,830
[Trainee 1] Yeah. Michael Johnson. Alright.
303
00:21:45,830 --> 00:21:47,960
Different Michael. I'm also
going to put the stress ulcer.
304
00:21:47,960 --> 00:21:48,670
Okay. Good call.
305
00:21:48,670 --> 00:21:54,050
[Vinny Arora] They have 10 minutes to
identify all the hazards that they can,
306
00:21:54,050 --> 00:21:57,680
and then right after, when they come
out, I actually debrief with them,
307
00:21:57,680 --> 00:22:01,020
so we go over how they did,
not only what they got right,
308
00:22:01,020 --> 00:22:05,230
where did they miss things, and
perhaps why did they miss those things.
309
00:22:05,230 --> 00:22:08,690
If you train people this way, this is the
way their brain is running in the background.
310
00:22:08,690 --> 00:22:10,950
Every time they enter a
room they can automatically
311
00:22:10,960 --> 00:22:12,650
spot it from the corner of their eye.
312
00:22:12,650 --> 00:22:14,850
As an organization, we
cannot improve patient
313
00:22:14,850 --> 00:22:16,990
safety unless we have front line personnel,
314
00:22:16,990 --> 00:22:22,450
including our residents and nurses and
everyone else that works in healthcare
315
00:22:22,460 --> 00:22:25,540
raising their hand to say,
"Hey, I saw something wrong."
316
00:22:25,540 --> 00:22:28,290
And so that's why it's really important
to embed people into a clinical situation
317
00:22:28,300 --> 00:22:32,800
where they are able to recognize what
types of events they should report.
318
00:22:32,800 --> 00:22:37,430
[somber music]
319
00:22:38,510 --> 00:22:44,190
[Bob Wachter] Probably the most important
foundational thinker in the field of
320
00:22:44,190 --> 00:22:48,070
patient safety is a gentleman
by the name of James Reason,
321
00:22:48,070 --> 00:22:51,150
who is now retired or semi-retired
psychologist in Manchester, England.
322
00:22:51,150 --> 00:22:52,850
What Reason was doing
was, as a psychologist,
323
00:22:52,860 --> 00:22:54,740
studying what he called
organizational accidents.
324
00:22:54,740 --> 00:22:59,700
How did terrible errors and accidents
and harm happen in industries,
325
00:22:59,700 --> 00:23:04,670
whether it was nuclear power or space
shuttles or intelligence failures in the CIA?
326
00:23:04,670 --> 00:23:07,510
So, he studied a bunch of them, and what he
327
00:23:07,510 --> 00:23:10,670
found was the same pattern
over and over again.
328
00:23:10,670 --> 00:23:12,760
What he found was if you
look at it superficially,
329
00:23:12,760 --> 00:23:14,590
you would see a human being who screwed up.
330
00:23:14,590 --> 00:23:19,720
That was the superficial understanding.
It was easy because it fit with
331
00:23:19,730 --> 00:23:22,600
the human model that I
need to blame somebody
332
00:23:22,600 --> 00:23:25,230
and if I can just point a finger,
you know, I have solved a problem.
333
00:23:25,230 --> 00:23:28,190
What was really right was
that in unsafe organizations,
334
00:23:28,190 --> 00:23:32,030
these organizational accidents happen
because of a long chain of events
335
00:23:32,030 --> 00:23:36,910
that allowed that human error, sometimes
several human errors to cause terrible harm.
336
00:23:36,910 --> 00:23:42,120
So, he came up with a model that, to me,
I remember the first time I read this,
337
00:23:42,130 --> 00:23:45,380
it's called the Swiss Cheese Model. A
little lightbulb went off and I said,
338
00:23:45,380 --> 00:23:48,260
"Aha! Oh, now I get it." And
now I look back on errors
339
00:23:48,260 --> 00:23:51,880
I have seen through my entire
career, and now it makes sense.
340
00:23:51,890 --> 00:23:55,350
Organizations build in protections to block
341
00:23:55,350 --> 00:23:58,770
those simple human
glitches from causing harm.
342
00:23:58,770 --> 00:24:01,640
The problem is, those layers of protections
343
00:24:01,650 --> 00:24:04,730
he likened to pieces of Swiss
cheese, they all have holes.
344
00:24:04,730 --> 00:24:07,240
If I kind of blow something one
day, I kind of forget something,
345
00:24:07,240 --> 00:24:08,580
or write something in the wrong space, most
346
00:24:08,580 --> 00:24:10,070
days the first layer of
Swiss cheese blocks it.
347
00:24:10,070 --> 00:24:15,410
But, on a bad day, the first layer
misses. It goes through the hole
348
00:24:15,410 --> 00:24:19,160
and it hits the second layer
and the second layer blocks it.
349
00:24:19,170 --> 00:24:21,830
When we kill someone in medicine because
we gave them the wrong medicine or cut off
350
00:24:21,830 --> 00:24:26,590
the wrong leg or there's a space shuttle
crash or Three Mile Island and you look back,
351
00:24:26,590 --> 00:24:31,970
you realize there were a lot of layers,
each one of them had a lot of holes,
352
00:24:31,970 --> 00:24:36,730
and also that particular day
the karma was pretty terrible
353
00:24:36,730 --> 00:24:40,600
and it just happened to be
that all of the holes aligned.
354
00:24:40,610 --> 00:24:43,360
And that's how the error made it
through all of these quote "protections"
355
00:24:43,360 --> 00:24:46,900
to cause terrible harm. My instinct was no
longer, "Let me figure out who screwed up."
356
00:24:46,900 --> 00:24:51,950
My instinct was now Swiss
cheese. It became automatic.
357
00:24:51,950 --> 00:24:53,730
Here's a bad error,
what's the Swiss cheese?
358
00:24:53,730 --> 00:24:55,580
What are the layers of
protection that we had
359
00:24:55,580 --> 00:25:00,290
that failed, how do we shrink the size of the
holes, and how do we create enough overlap
360
00:25:00,290 --> 00:25:03,180
in layers of cheese so
an error never makes it
361
00:25:03,180 --> 00:25:06,760
through all those layers
to cause terrible harm?
362
00:25:07,590 --> 00:25:12,310
[ambient music]
363
00:25:43,220 --> 00:25:48,890
[Sue Sheridan] With Pat, I actually spoke
to the pathologist about why he didn't
364
00:25:48,890 --> 00:25:54,430
pick up the phone and call the neurosurgeon
when they learned it was cancer,
365
00:25:54,440 --> 00:25:59,730
and it was a rare kind of cancer, and
his answer was, "It's not my job."
366
00:26:05,450 --> 00:26:09,660
[Mackenzie Sheridan] The doctor told
our family, you know, your dad is fine.
367
00:26:09,660 --> 00:26:12,750
He's benign, the tumor is benign,
everything's great, go on and live your life.
368
00:26:12,750 --> 00:26:16,170
A few months later my dad got very sick.
369
00:26:16,170 --> 00:26:18,010
[Sue Sheridan] And I got
the documents from the
370
00:26:18,010 --> 00:26:20,230
neurosurgeon and it said that the pathology
371
00:26:20,510 --> 00:26:25,380
was an atypical spindle cell neoplasm,
which the doctor said was benign.
372
00:26:25,390 --> 00:26:30,350
We expected the hospital to fully
describe to us what had happened, to...
373
00:26:31,390 --> 00:26:37,400
you know, take care of us, and we were
discharged without any explanation.
374
00:26:38,020 --> 00:26:43,070
So, we left there with all
the documents in our hands
375
00:26:43,070 --> 00:26:45,410
with absolutely no explanation
that this was an error.
376
00:26:45,410 --> 00:26:48,910
[sighs]
377
00:26:50,250 --> 00:26:53,060
I think our first reaction was
fear. We were scared. It scared
378
00:26:53,060 --> 00:26:56,250
us that a hospital, a well-known
hospital, with professionals,
379
00:26:56,250 --> 00:27:01,300
would intentionally cover
up that kind of information.
380
00:27:01,300 --> 00:27:04,840
So, the first, the first emotion was fear.
381
00:27:04,840 --> 00:27:08,600
One day, Pat woke up
paralyzed from his waist down,
382
00:27:08,600 --> 00:27:12,480
and we're at home in Boise, Idaho,
and we thought maybe he had a stroke.
383
00:27:12,480 --> 00:27:17,150
We learned then that his cancer
had returned explosively.
384
00:27:17,530 --> 00:27:21,490
They estimated he had
about 10 days to live.
385
00:27:22,410 --> 00:27:25,450
[Mackenzie Sheridan] And I remember my
mom sitting Cal and I down right before
386
00:27:25,450 --> 00:27:29,830
and she said, "You know, your dad is sick,
and he is going to no longer be with us."
387
00:27:32,290 --> 00:27:35,560
[Sue Sheridan] I requested a meeting
with the doctor, and with the CEO,
388
00:27:37,760 --> 00:27:42,550
and with the risk manager. They
agreed to it and I flew down there
389
00:27:42,720 --> 00:27:47,560
and nobody showed up, except the chaplain.
390
00:27:47,560 --> 00:27:49,200
I demanded that they implement a disclosure
391
00:27:49,200 --> 00:27:51,020
procedure that when there
was an error at their
392
00:27:51,020 --> 00:27:53,760
hospital that they sit
down with the family,
393
00:27:53,760 --> 00:27:57,070
which, you know, which
they did not with us.
394
00:28:01,990 --> 00:28:05,830
[David Mayer] Historically, you've
probably heard the term deny and defend.
395
00:28:05,830 --> 00:28:08,910
That was the model that is still existent
today unfortunately at many hospitals;
396
00:28:08,910 --> 00:28:11,570
That if we cause a
preventable medical harm,
397
00:28:11,570 --> 00:28:14,840
the goal has always been
to shut things down,
398
00:28:15,000 --> 00:28:20,430
let the lawyers handle it, don't
talk to the patients and families,
399
00:28:20,430 --> 00:28:24,510
and then it turns into a legal
battle for 4, 5, 6 years where
400
00:28:24,520 --> 00:28:28,930
the hope is that the patient and
family will just give up and go away
401
00:28:28,940 --> 00:28:33,400
and that's been the model. And now we've
moved to more open and honest communication.
402
00:28:35,820 --> 00:28:38,150
[Heather Young] We do a simulation on
403
00:28:40,570 --> 00:28:42,370
how to tell someone that
you've made an error,
404
00:28:42,370 --> 00:28:44,990
and that's a skill that's
very difficult to develop,
405
00:28:45,000 --> 00:28:47,620
to do in a way that conveys that you
care and that you are concerned about
406
00:28:47,620 --> 00:28:52,380
the person's safety and that you
are going to do something about it
407
00:28:52,380 --> 00:28:55,590
when you may face a family member
who is irate, very upset by the news.
408
00:28:55,590 --> 00:29:00,800
And you know, as a new clinician, you
need to have the skills to be open and
409
00:29:00,800 --> 00:29:05,220
transparent and talk honestly
and authentically with people.
410
00:29:05,230 --> 00:29:10,190
So, I'm about to go in and see a
standardized patient, is what we call it.
411
00:29:10,190 --> 00:29:14,530
It's an actor that I have no
idea how he's going to react.
412
00:29:14,530 --> 00:29:18,030
We're going to break him some bad news about
a test result that we missed 3 months ago.
413
00:29:18,030 --> 00:29:21,740
They are told to react
differently to each student.
414
00:29:21,740 --> 00:29:25,790
So I don't know what I'm going
to get when I break him the news.
415
00:29:25,790 --> 00:29:28,580
He could be angry, frustrated, or he
could go easy on me. I just don't know.
416
00:29:28,580 --> 00:29:34,090
One of the things that was ordered a
couple weeks ago was a CT scan, uh, which,
417
00:29:35,590 --> 00:29:41,510
umm, indicated, umm, some results
that could indicate colon cancer.
418
00:29:42,850 --> 00:29:47,690
[Doctor] Listen, I've got to cut this. Um,
you don't want to say there was another
419
00:29:48,480 --> 00:29:54,450
test result that might indicate colon
cancer at this short intro into it, right?
420
00:29:56,200 --> 00:29:59,950
- Ah? - Oh.
421
00:30:01,580 --> 00:30:04,010
I mean you went right to: "That
could be colon cancer." His dad died
422
00:30:04,020 --> 00:30:06,420
of colon cancer. You could have a
patient falling apart in moments.
423
00:30:06,420 --> 00:30:11,050
Do you want to look at those
pearls on effective communication?
424
00:30:11,050 --> 00:30:16,550
Lay out the facts, that you know them,
and say that 3 months ago on the CT...
425
00:30:16,800 --> 00:30:21,810
And then as he's like, and
I know your dad passed,
426
00:30:21,810 --> 00:30:26,650
it could be a cancer, but
we don't know that yet.
427
00:30:26,650 --> 00:30:30,110
You know, all that gingerly, careful stuff.
428
00:30:30,110 --> 00:30:33,240
- Hey, Walt. How's it going?
- Hey, Jason. I'm alright.
429
00:30:34,030 --> 00:30:36,370
- It's good to see you again. - Thank you.
430
00:30:36,370 --> 00:30:38,160
- How was the drive in? - Uh, fine.
431
00:30:38,160 --> 00:30:40,100
Three months ago, remember
you came in three months ago?
432
00:30:40,100 --> 00:30:41,250
I do.
433
00:30:43,120 --> 00:30:46,130
It showed that you had some
thickening of your colonic wall
434
00:30:46,330 --> 00:30:50,260
and some enlarged mesenteric lymph nodes.
435
00:30:51,130 --> 00:30:54,970
We need to do a colonoscopy immediately.
436
00:30:54,970 --> 00:30:58,430
We want to make sure, and I'm
not saying it's colon cancer,
437
00:30:58,720 --> 00:31:01,480
but we want to make sure that it's
not colon cancer and rule it out.
438
00:31:01,480 --> 00:31:05,230
Why did it take 3 months
to, uh, that I know this?
439
00:31:05,230 --> 00:31:09,980
That was my mistake. We were looking for
structural abnormalities on your kidneys
440
00:31:09,990 --> 00:31:13,820
and I overlooked that part
of the report 3 months ago.
441
00:31:13,820 --> 00:31:17,160
[sigh] I mean, I would have been upset
442
00:31:18,120 --> 00:31:22,920
hearing it first when the CT scan happened,
443
00:31:22,920 --> 00:31:26,500
but now I'm really pissed off that it's
been 3 months, that it was delayed.
444
00:31:26,500 --> 00:31:31,840
Right, and, I mean, I understand that you're
angry, I can see that you're frustrated
445
00:31:31,840 --> 00:31:37,260
and I can't, I can't do anything
to fix that mistake 3 months ago.
446
00:31:37,270 --> 00:31:41,140
But, what I can do now is make this a
priority as your primary care provider,
447
00:31:41,140 --> 00:31:46,020
and I can't even imagine how you're
feeling right now with the mistake,
448
00:31:46,020 --> 00:31:48,140
but let's take it from
here, and we'll figure
449
00:31:48,140 --> 00:31:50,990
this out together. I'll
make this a priority, OK?
450
00:32:31,370 --> 00:32:36,330
- [Charlie] Good morning.
- Good morning. Hi Charlie.
451
00:32:36,330 --> 00:32:37,870
Hi Walt, nice to meet you.
452
00:32:37,870 --> 00:32:39,710
I'm sorry. Wait, I've met you before.
453
00:32:39,710 --> 00:32:41,960
Yeah. We've known each other for years.
454
00:32:41,960 --> 00:32:45,840
[Heather Young] The closer you are to the
error, the more important it is that you have
455
00:32:45,840 --> 00:32:49,130
some accountability for it,
and that you communicate
456
00:32:49,140 --> 00:32:51,930
with the people who might be harmed by it.
457
00:32:51,930 --> 00:32:53,890
And so all of us need to
learn the skills to be able to
458
00:32:53,890 --> 00:32:57,770
acknowledge what we've donewrong and
what we're planning to do to fix it.
459
00:32:57,770 --> 00:33:01,610
[Don Berwick] We built it completely
wrong. We were trained, I was trained,
460
00:33:02,150 --> 00:33:06,110
"No, you don't talk about
your mistakes with a patient,
461
00:33:06,110 --> 00:33:09,200
that's liability, the
lawyers will be all over us."
462
00:33:09,200 --> 00:33:12,410
This is a time for openness and honesty,
and so we can learn and grow together.
463
00:33:12,410 --> 00:33:17,870
Healthcare is not like a toaster
where I make it and I sell it to you,
464
00:33:17,880 --> 00:33:21,130
and you take it and plug it in. No, it's
always a cooperative enterprise so that
465
00:33:21,130 --> 00:33:26,380
the patient and the family, and the doctor
and the nurse, they're co-producing the care.
466
00:33:26,380 --> 00:33:30,430
And now that we're more
aware of that over time,
467
00:33:30,430 --> 00:33:33,490
there's a lot of possibility for
much more participation by both.
468
00:33:36,440 --> 00:33:41,570
[John Eisenberg] I recalled
a woman whom I took care of.
469
00:33:41,570 --> 00:33:45,780
We had had a pap test done to
screen her for cervical cancer.
470
00:33:45,780 --> 00:33:49,620
The result was suspicious,
but I never knew that,
471
00:33:50,370 --> 00:33:54,500
because I never got the report
back. And I didn't realize
472
00:33:54,790 --> 00:33:58,670
that I hadn't gotten the report back until
she called me and asked about the report.
473
00:33:58,670 --> 00:34:03,470
I tracked it down. I found out that it was
suspicious. We followed it up and fortunately
474
00:34:04,130 --> 00:34:10,010
it turned out not to be anything
serious. But that was a near miss.
475
00:34:10,060 --> 00:34:14,980
It was a near miss that could have
been a tragedy had she not called me.
476
00:34:16,310 --> 00:34:21,690
Senator, when I spoke at three medical
school graduations last Spring,
477
00:34:21,690 --> 00:34:26,740
I asked all the students who were
graduating, and I asked all of the faculty
478
00:34:26,740 --> 00:34:31,910
to raise their hands if they had ever made
a mistake in taking care of a patient,
479
00:34:32,250 --> 00:34:34,790
and every single student
raised his or her hand,
480
00:34:34,790 --> 00:34:37,670
every faculty member
raised his or her hand.
481
00:34:37,880 --> 00:34:43,680
When I was a medical student on
one of my very first rotations,
482
00:34:43,930 --> 00:34:48,470
I inadvertently, during a code,
483
00:34:48,470 --> 00:34:52,480
gave a full syringe of morphine to a patient
IV and they had a respiratory arrest.
484
00:34:54,770 --> 00:35:00,820
Fortunately, the person was intubated
and resuscitated and did just fine.
485
00:35:00,940 --> 00:35:05,490
That was a shocking experience,
486
00:35:05,490 --> 00:35:10,160
and made me aware at a very
early point in my medical career
487
00:35:10,160 --> 00:35:15,040
that we have the potential to do things
wrong and to potentially harm patients.
488
00:35:15,040 --> 00:35:19,340
No one ever heard about it
besides me and that nurse.
489
00:35:19,340 --> 00:35:23,050
So, it's not clear to me that
any changes were ever made
490
00:35:23,470 --> 00:35:26,300
as a result, and I don't
think the patient ever heard.
491
00:35:26,310 --> 00:35:28,350
I've made medical errors; I have, uh,
492
00:35:28,350 --> 00:35:31,730
I prescribed the wrong medication on a
patient. There were two patients of mine
493
00:35:31,730 --> 00:35:35,650
with very similar names and I just
prescribed it on the wrong patient.
494
00:35:35,650 --> 00:35:38,900
I felt terrible. I felt incompetent.
I felt a little ashamed.
495
00:35:38,900 --> 00:35:44,450
And I, my first instinct was not just to fix
the problem, but then not to tell anybody.
496
00:35:44,450 --> 00:35:49,160
That's just a normal human instinct.
497
00:35:49,160 --> 00:35:52,540
It is completely understandable
why people's first reaction is
498
00:35:52,540 --> 00:35:57,380
cover it up, don't talk about
it. It's a very human response.
499
00:35:57,380 --> 00:36:01,180
Doesn't make it the right
thing, it's actually clearly
500
00:36:01,180 --> 00:36:03,640
not the right thing, it's
clearly bad to do that.
501
00:36:03,640 --> 00:36:06,430
But I think we have to begin by acknowledging
that it's a very human response.
502
00:36:06,430 --> 00:36:10,230
You can feel very
self-righteous. You can say,
503
00:36:10,230 --> 00:36:12,230
"Well, the patient got the
wrong drug, fire the nurse.
504
00:36:12,230 --> 00:36:14,770
There's a complication of
the surgery, bad surgeon."
505
00:36:14,780 --> 00:36:18,690
You're wrong. You're almost always wrong.
506
00:36:18,700 --> 00:36:21,410
It feels good to blame someone. You've got
a culprit? Put them in jail, fire them.
507
00:36:21,410 --> 00:36:25,120
Many things caused it. So, who's
responsible? Everybody's responsible.
508
00:36:25,120 --> 00:36:29,000
Everybody can contribute to the
enterprise of closing the vulnerabilities,
509
00:36:29,000 --> 00:36:33,340
of making the whole thing
less likely to go wrong.
510
00:36:33,340 --> 00:36:37,220
The most recent survey I have seen is
that nearly 50% of nurses in America
511
00:36:37,220 --> 00:36:43,140
still don't feel it is safe to
talk about a mistake they've made.
512
00:36:43,180 --> 00:36:47,640
That's an absolute disgrace.
513
00:36:47,640 --> 00:36:50,510
If something bad is going to
happen to you when you speak up
514
00:36:51,520 --> 00:36:55,230
about something you've seen
or done that could help.
515
00:36:55,240 --> 00:37:00,860
If you're going to get punished for
that, why would you speak up? You don't.
516
00:37:00,870 --> 00:37:05,250
You run and hide. You lie.
That's normal human behavior.
517
00:37:05,250 --> 00:37:10,000
We're not talking about bad people; we're
talking about normal people become frightened.
518
00:37:10,000 --> 00:37:13,550
And so leaders, you got a choice: you can
scare your workforce and give up the hope
519
00:37:13,550 --> 00:37:18,260
for improvement, or you can celebrate,
invite, work with your workforce,
520
00:37:18,510 --> 00:37:23,810
and have a chance of learning
together to get to a better world.
521
00:37:23,810 --> 00:37:26,560
What we have learned from other industries
is that if you could change the culture
522
00:37:26,560 --> 00:37:30,770
and reward people for being open,
reward people for being honest,
523
00:37:30,770 --> 00:37:34,190
reward people for coming forth
and talking about their errors,
524
00:37:34,190 --> 00:37:38,110
then you being to counter that kind
of normal instinct that we all have,
525
00:37:38,120 --> 00:37:42,240
and begin to create a culture of patient
safety where people are much more open.
526
00:37:42,240 --> 00:37:45,910
And the system gets better because it
learns from mistakes and doesn't hide them.
527
00:37:45,910 --> 00:37:50,090
And we found in the food industry they
were years ahead of us. They had programs.
528
00:37:50,090 --> 00:37:54,880
For instance, Burger King had a
program if the employee saw another one
529
00:37:54,880 --> 00:37:59,180
not washing their hands, they went over
and they tapped them and said, "Got you",
530
00:37:59,180 --> 00:38:03,350
and then they got either
two hours compensation off
531
00:38:03,350 --> 00:38:07,190
or some other reward. I
mean, they're on board.
532
00:38:07,190 --> 00:38:09,040
Safety reporting is like
democracy. Democracy
533
00:38:09,050 --> 00:38:11,400
isn't about having a
free and fair election.
534
00:38:11,650 --> 00:38:16,410
We can do that. Democracy is about
having a second free and fair election.
535
00:38:17,160 --> 00:38:22,290
The same thing is true with
safety reporting. It's not about
536
00:38:22,290 --> 00:38:24,660
filing a safety report,
it's about filing a second.
537
00:38:24,660 --> 00:38:27,130
And where you see an organization
with a high rate of reported error,
538
00:38:27,130 --> 00:38:28,750
what that tells you is
it tells you that they
539
00:38:28,750 --> 00:38:30,380
must be doing something
about those reports,
540
00:38:30,380 --> 00:38:34,800
because if they are just sitting on
them, people will stop reporting.
541
00:38:34,800 --> 00:38:38,300
Because even if you tell people they
have to, in the end it's all voluntary.
542
00:38:38,310 --> 00:38:42,140
I mean, you can't solve it if
you can't see it. We can see it.
543
00:38:42,140 --> 00:38:45,480
And more and more people are
aware of it. That's the good news.
544
00:38:45,480 --> 00:38:47,980
Bad news is you're still
at risk, really at risk.
545
00:38:47,980 --> 00:38:51,690
I mean we haven't pervaded healthcare with
the designs and approaches and cultures
546
00:38:52,900 --> 00:38:58,120
that actually make you super
safe and that's the task ahead.
547
00:38:58,120 --> 00:39:01,910
It's amazing how quickly hospitals
can completely overhaul their safety
548
00:39:03,000 --> 00:39:07,500
when they know that it's
important to their patients.
549
00:39:08,000 --> 00:39:10,920
Hospitals had to hear the
message from their own patients
550
00:39:10,920 --> 00:39:12,640
that it matters that they wash their hands,
551
00:39:12,640 --> 00:39:14,470
it matters that they
keep a safe environment,
552
00:39:14,470 --> 00:39:19,680
it matters that they put the safety
and protection of their patients first
553
00:39:19,680 --> 00:39:24,980
every minute of every day. The only way
they're really going to get that message
554
00:39:24,980 --> 00:39:29,610
is when the American public
gets involved and pushes.
555
00:39:32,780 --> 00:39:38,410
[narrator]One way to improve the
quality of hospitals in America
556
00:39:38,410 --> 00:39:41,210
is to put a microscope on the
data they do actually provide.
557
00:39:41,210 --> 00:39:45,000
[narrator] Leah Binder and her team at
the Leapfrog Group in Washington, DC,
558
00:39:45,920 --> 00:39:49,720
worked with leaders in patient
safety to create a new way
559
00:39:49,720 --> 00:39:52,840
to rate the quality of hospitals
that patients can understand.
560
00:39:52,840 --> 00:39:56,810
We worked with the foremost
experts in patient safety
561
00:39:57,180 --> 00:39:58,970
and we asked them to
look at all this data and
562
00:39:58,970 --> 00:40:00,640
decide which of the data is most reliable,
563
00:40:00,640 --> 00:40:05,400
which gives us the best information
about the safety of a hospital,
564
00:40:05,400 --> 00:40:08,360
and then help us figure out a reliable
criteria to put it all together.
565
00:40:08,360 --> 00:40:12,700
And then, we did something else.
We decided to issue a letter grade.
566
00:40:12,700 --> 00:40:15,190
The letter grade would
apply to each hospital on
567
00:40:15,190 --> 00:40:18,120
how safe they are relative
to other hospitals.
568
00:40:18,160 --> 00:40:22,750
So, were they an A, B, C, D, or F?
569
00:40:22,750 --> 00:40:26,050
When we first did it, we got calls
from some hospital CEOs who said
570
00:40:26,050 --> 00:40:30,220
to me, memorably, "I've decided I
don't want a letter grade from you."
571
00:40:30,220 --> 00:40:34,470
And I said, "Well, I've decided you're getting
one anyway, because you serve the public,
572
00:40:34,470 --> 00:40:39,100
and the public you serve deserves
to know how you're doing."
573
00:40:39,100 --> 00:40:42,190
It's very important to do
these kinds of ratings because
574
00:40:42,190 --> 00:40:43,870
who wants to work in a
terrible organization? And
575
00:40:43,870 --> 00:40:45,440
so if you can make it
very obvious to all the
576
00:40:45,440 --> 00:40:49,860
doctors and nurses in that hospital
that this is a highly unsafe hospital,
577
00:40:49,860 --> 00:40:51,700
I think there is going
to be internal pressure
578
00:40:51,700 --> 00:40:53,580
to reform and internal
pressure to get better.
579
00:40:53,580 --> 00:40:57,910
But, certainly I think it's true that,
like, if you're in an isolated area,
580
00:40:57,920 --> 00:41:00,790
there's one hospital in town or you
could be in the middle of Chicago,
581
00:41:00,790 --> 00:41:04,090
but your insurance company
covers one hospital only,
582
00:41:04,090 --> 00:41:05,890
it's going to be a
challenge of choices. But
583
00:41:05,890 --> 00:41:07,880
that doesn't mean you
couldn't go to your doctor
584
00:41:07,880 --> 00:41:12,180
who works in that hospital and be like,
"Hey, why are you guys a D hospital?"
585
00:41:12,180 --> 00:41:15,680
And I think if consumers started
talking to doctors and nurses that way,
586
00:41:15,680 --> 00:41:19,020
it would actually begin to change the
conversation, where doctors would say,
587
00:41:19,020 --> 00:41:23,190
"Why do I work at a hospital that
has such high infection rates?"
588
00:41:23,190 --> 00:41:26,070
Virtually every other industry in this
country has their products and services
589
00:41:26,070 --> 00:41:29,740
in a transparent market, and people choose.
590
00:41:29,740 --> 00:41:31,990
So, if you're buying a car,
you can look up auto reviews
591
00:41:31,990 --> 00:41:36,790
and you can compare among different
cars and different features.
592
00:41:36,790 --> 00:41:39,630
In New York, which I
know particularly well,
593
00:41:39,630 --> 00:41:42,210
restaurants that had, for many
years, been getting public ratings
594
00:41:42,210 --> 00:41:46,010
from the health department
on how safe they were;
595
00:41:46,010 --> 00:41:49,180
those were all public, but nobody
paid any attention to them.
596
00:41:49,180 --> 00:41:52,470
So, the health department said, from
now on you're going to get a grade
597
00:41:52,470 --> 00:41:55,640
on how safe you are and you
have to post it in your window.
598
00:41:55,640 --> 00:41:57,450
So, restaurants started
posting it, and within
599
00:41:57,450 --> 00:41:59,400
six months any restaurant
that didn't have an A
600
00:41:59,400 --> 00:42:04,900
was either out of business or they
were very quickly getting to their A.
601
00:42:04,900 --> 00:42:09,280
So we said, "Well, let's do
the same thing with hospitals."
602
00:42:09,280 --> 00:42:12,200
I mean in our dream, hospitals
would put their letter grade on
603
00:42:12,200 --> 00:42:16,580
you know, their front door and everyone would
know that this hospital was safe or not.
604
00:42:16,580 --> 00:42:21,260
[sombr music]
605
00:42:22,340 --> 00:42:28,550
[Helen Burstin]John Eisenberg used to tell a
great story of the drunk who lost his keys.
606
00:42:28,560 --> 00:42:32,640
And he's out in front of the bar
in the street looking for his keys
607
00:42:33,270 --> 00:42:36,150
and some guy comes over and
goes, "What are you doing?"
608
00:42:36,150 --> 00:42:38,610
He says, "I'm looking for my key."
609
00:42:38,610 --> 00:42:40,530
"Well, why are you only
looking right here?"
610
00:42:40,530 --> 00:42:42,570
He said, "Well, nthat's
where the lamplight is."
611
00:42:42,570 --> 00:42:44,450
[clock ticking]
612
00:42:44,450 --> 00:42:46,450
[narrator] This is known
as the streetlight effect.
613
00:42:46,450 --> 00:42:49,230
Many in the patient safety
field have been looking outside
614
00:42:49,870 --> 00:42:53,460
healthcare for solutions
to preventable errors.
615
00:42:53,460 --> 00:42:56,080
Industries like nuclear power, aircraft
carriers, and commercial aviation
616
00:42:56,090 --> 00:43:00,670
have become known as
high-reliability organizations
617
00:43:00,670 --> 00:43:03,720
due to significant
efforts to improve safety.
618
00:43:03,720 --> 00:43:07,470
High reliability is different in healthcare
because it points directly at examples
619
00:43:08,010 --> 00:43:12,890
of very hazardous industries, organizations
620
00:43:12,890 --> 00:43:17,820
that have solved the problem of getting to
zero harm that healthcare has not solved.
621
00:43:18,190 --> 00:43:23,030
Tools and methods and
lessons from that work
622
00:43:23,030 --> 00:43:27,330
are very directly applicable to
healthcare and we're starting to see
623
00:43:27,330 --> 00:43:31,830
healthcare organizations use
them to make improvements
624
00:43:31,830 --> 00:43:35,500
at a level that we have never seen before.
625
00:43:35,500 --> 00:43:38,170
So over here we have the complex
system of the modern American hospital,
626
00:43:38,170 --> 00:43:43,220
and over here we have other
industries that have learned to
627
00:43:43,220 --> 00:43:48,100
simplify and deal with
these complex systems.
628
00:43:48,100 --> 00:43:52,310
In the last calendar year there
has been no fatality worldwide
629
00:43:52,310 --> 00:43:57,150
in commercial aviation due to an accident.
630
00:43:57,150 --> 00:44:00,700
Compare that to our business where we
have 20 wrong-site surgeries every week.
631
00:44:01,110 --> 00:44:05,870
[David Mayer] Pilots make one error per
hour in the cockpit every day they work
632
00:44:06,830 --> 00:44:11,830
and yet we wonder why planes
aren't falling out of the sky.
633
00:44:11,830 --> 00:44:15,710
If aviation had said, "Well, you know what,
to fly you 600 miles an hour it's going to
634
00:44:15,710 --> 00:44:20,680
come with some mishap. And you got to expect
a plane or two to fall out of the sky,"
635
00:44:20,680 --> 00:44:24,800
and thank god they didn't say that and
they said, "No, we can drive it to zero.
636
00:44:24,810 --> 00:44:29,480
We can drive it down to virtually
no mishap," and they have.
637
00:44:29,480 --> 00:44:31,500
The aviation industry
is the safest it's ever
638
00:44:31,500 --> 00:44:33,480
been since the invention of the jet engine.
639
00:44:33,480 --> 00:44:35,680
What we're really doing when we go up in an
640
00:44:35,680 --> 00:44:38,530
airliner is pushing a
tube filled with people
641
00:44:38,740 --> 00:44:44,990
through the upper atmosphere, 7 or 8 miles
above the earth, at 80% the speed of sound,
642
00:44:44,990 --> 00:44:50,710
in a hostile environment with outside air
pressure one-quarter that at the surface,
643
00:44:50,710 --> 00:44:55,380
and we must return it safely to
the surface every time, and we do.
644
00:44:55,380 --> 00:45:01,140
In this country alone, 28,000 times
a day, 10.2 million times a year.
645
00:45:01,470 --> 00:45:06,980
In a little over 100 years you've
gone from quite a dangerous
646
00:45:06,980 --> 00:45:12,150
industry to the first ultra-safe
mode of transport bar none.
647
00:45:12,150 --> 00:45:17,780
One of the reasons is because it is
studied so well, and every single event
648
00:45:17,780 --> 00:45:23,280
is clearly understood and is made
public so others can learn from them.
649
00:45:23,290 --> 00:45:27,370
I had been flying airplanes for 42
years. I had 20,000 hours in the air.
650
00:45:27,370 --> 00:45:32,960
And throughout that entire time, I had
never been so challenged in an airplane
651
00:45:32,960 --> 00:45:36,340
I doubted the outcome. I never
thought I would be. I was wrong.
652
00:45:36,340 --> 00:45:41,640
[narrator] In January of 2009, Captain
Sullenberger's training and instincts
653
00:45:41,640 --> 00:45:46,640
saved the lives of all 155 passengers
aboard US Airways Flight 1549
654
00:45:46,770 --> 00:45:52,400
after it struck a flock of
geese and lost all engine power.
655
00:45:52,400 --> 00:45:56,570
The dramatic landing on thenHudson
River reminded Americans
656
00:45:56,570 --> 00:45:59,530
of the importance of
experience in the cockpit.
657
00:45:59,530 --> 00:46:03,440
In an industry in which we work very hard to
make everything easy and routine and safe,
658
00:46:06,210 --> 00:46:11,380
100 seconds after takeoff we
were suddenly confronted with an
659
00:46:11,380 --> 00:46:15,010
ultimate challenge of a lifetime, to
do something we'd never done before
660
00:46:15,010 --> 00:46:17,800
and get it right the first
time never having practiced it.
661
00:46:17,800 --> 00:46:20,470
In a similar fashion in medicine, there
are some things that just can't be
662
00:46:20,470 --> 00:46:24,440
practiced safely any other way than
in a simulation for the first time.
663
00:46:24,440 --> 00:46:28,230
And it gives you a chance to practice
things over and over and over again.
664
00:46:28,230 --> 00:46:33,780
And so it's important that the simulations
be done not simply individually,
665
00:46:33,780 --> 00:46:37,910
but also collectively as a whole team.
666
00:46:45,790 --> 00:46:51,840
[narrator]Flight simulators have been
used to train pilots for nearly 100 years.
667
00:46:51,840 --> 00:46:56,390
And while medicine has used cadavers
to train doctors for much longer,
668
00:46:56,390 --> 00:47:00,390
only recently have institutions
begun using robotics
669
00:47:00,390 --> 00:47:03,730
to simulate any kind of situation
a care provider may face.
670
00:47:03,730 --> 00:47:07,900
[Heather Young] Simulation is a very
big part of our educational program here
671
00:47:08,980 --> 00:47:10,860
and it involves anything
from patients who come
672
00:47:10,860 --> 00:47:12,900
in as actors and will work with a student,
673
00:47:12,910 --> 00:47:18,700
all the way up to very high-fidelity
robots, and environments that
674
00:47:18,700 --> 00:47:22,790
are tricked out to look out exactly
like a hospital operating room
675
00:47:22,790 --> 00:47:26,420
or an emergency department
or hospital ward.
676
00:47:26,420 --> 00:47:29,590
The airline industry is the
prototype of using simulation
677
00:47:31,720 --> 00:47:35,300
where you can practice landing in San
Diego with a terrible storm or a tsunami
678
00:47:35,300 --> 00:47:40,140
or on a very calm day and you can practice
679
00:47:40,140 --> 00:47:42,480
all different kinds of
failures within the airplane.
680
00:47:42,480 --> 00:47:45,900
It's newer in healthcare, but it's
really something that's catching on,
681
00:47:45,900 --> 00:47:49,820
and you can really put people through the
steps of handling many important situations.
682
00:47:50,610 --> 00:47:55,990
[Ian Julie] So we're going to be
practicing our new simulated protocol
683
00:47:55,990 --> 00:47:59,620
for our actual sepsis patients. Sepsis
care can be very, very difficult.
684
00:47:59,620 --> 00:48:04,040
We know the science behind it, we know
what helps, but we don't necessarily know
685
00:48:04,040 --> 00:48:07,460
how to do it in a way that's
organized and consistent.
686
00:48:07,460 --> 00:48:10,430
We'd rather practice on our friend the
mannequin here who it's very hard to injure,
687
00:48:10,430 --> 00:48:15,100
rather than on real patients. That way we
can standardize things within our hospital
688
00:48:15,100 --> 00:48:19,270
and give our nurses and doctors a chance
to practice what it is their doing,
689
00:48:19,270 --> 00:48:23,100
before they have to do it on real patients.
690
00:48:23,110 --> 00:48:25,400
Hi Robert. My name is Emily.
I'm going to be you nurse today.
691
00:48:25,400 --> 00:48:28,490
I'm here to do your morning assessment and
take your vital signs. How are you feeling?
692
00:48:28,490 --> 00:48:32,360
[mannequin] I'm not feeling very well.
693
00:48:32,370 --> 00:48:34,030
[Nurse 1] You're not? What's going on?
694
00:48:34,030 --> 00:48:36,080
[mannequin] I just can't catch
my breath this morning and
695
00:48:36,080 --> 00:48:38,250
I feel like my cough is worse.
696
00:48:38,250 --> 00:48:40,080
He's remaining stable.
Based on the alert, um,
697
00:48:40,080 --> 00:48:42,790
and the lactic acid, I think
I'm going to start some oxygen.
698
00:48:42,790 --> 00:48:46,800
[Nurse 2] Okay, are there signs
or symptoms of an infection?
699
00:48:47,170 --> 00:48:51,510
[Nurse 1] Well, he's saying that he
has an increased work of breathing.
700
00:48:51,510 --> 00:48:54,470
He has a white count of 16.
701
00:48:54,470 --> 00:48:56,680
[Nurse 2] OK, sound
good. I'll be right over.
702
00:48:56,680 --> 00:48:58,520
- OK, thank you! - [Nurse 2] Alright.
703
00:48:58,520 --> 00:49:00,980
- [Nurse 2] Hi, Mr. Robert!
- [mannequin] Hi!
704
00:49:02,020 --> 00:49:04,520
-How are you feeling?
-[mannequin] I've had better days.
705
00:49:04,530 --> 00:49:07,900
- Are you short of breath?
- [Mannequin] Yeah.
706
00:49:07,910 --> 00:49:10,070
[Nurse 2] OK and when did this start?
707
00:49:10,070 --> 00:49:12,120
Here you go. You've drawn
cultures already, correct?
708
00:49:12,120 --> 00:49:17,410
- [Nurse 2] This is Robert Doe?
- [Nurse 1] Yes, Robert Doe.
709
00:49:17,410 --> 00:49:21,170
[Ian Julie] We can make the scenario
more complex, and we do on occasion.
710
00:49:21,460 --> 00:49:24,710
We could have the patient
enter a state of shock,
711
00:49:24,710 --> 00:49:27,130
or not respond properly to
the fluids or the antibiotics.
712
00:49:27,130 --> 00:49:30,390
So, much of what we've done is
related to the need to kind of
713
00:49:30,390 --> 00:49:34,100
fulfill the recommendations
that have been given.
714
00:49:34,100 --> 00:49:36,980
In addition to wanting to do
what's right for the patient
715
00:49:36,980 --> 00:49:39,310
and following through on the best
available scientific evidence.
716
00:49:39,310 --> 00:49:42,690
When I graduated as a nurse, the
first time I ever had a chance
717
00:49:42,690 --> 00:49:46,110
to shock a person whose heart had stopped
718
00:49:46,110 --> 00:49:48,780
was in the middle of the
night in a rural hospital
719
00:49:48,780 --> 00:49:51,370
and it was my first time I had ever
turned on the paddles in my life.
720
00:49:51,370 --> 00:49:55,000
And someone's life depended on
that. That's not acceptable.
721
00:49:55,000 --> 00:49:59,630
We want our students to practice and
practice and practice how to shock people
722
00:49:59,630 --> 00:50:04,670
in a simulated situation, so that when
someone is really depending on them,
723
00:50:04,670 --> 00:50:08,760
they do it right the first time.
724
00:50:08,760 --> 00:50:11,510
I shudder to remember how I was trained
as a doctor to learn how to do stuff.
725
00:50:11,520 --> 00:50:16,730
Lumbar punctures, spinal taps,
put IVs in, or even chest tubes.
726
00:50:16,730 --> 00:50:21,780
You practiced on the patients. I
mean, that was the only option.
727
00:50:21,780 --> 00:50:25,860
Some patient, some time, was the first
patient I ever put a chest tube in,
728
00:50:25,860 --> 00:50:29,620
and that person paid the price.
They were paying for my tuition.
729
00:50:29,620 --> 00:50:34,160
You know, we don't do that with
pilots, we put them in the simulator
730
00:50:34,160 --> 00:50:36,880
and they fly something that isn't
really a plane for a while, first,
731
00:50:36,880 --> 00:50:39,800
with high fidelity. Now we know
how to do that in health care.
732
00:50:39,800 --> 00:50:42,380
The growth of simulation so that the first
chest tube doesn't go into a human being,
733
00:50:42,380 --> 00:50:46,090
it goes in a mannequin that looks like a
human being, that's great. And I think that
734
00:50:46,090 --> 00:50:51,020
it's one of the emerging
ways to help build skills
735
00:50:51,020 --> 00:50:54,980
hmm, in a work force without
the patients paying the tuition.
736
00:50:54,980 --> 00:50:59,400
[narrator] Many aspects of the aviation
industry have been applied to medicine,
737
00:51:01,030 --> 00:51:04,610
from checklists before an operation
to monitoring physicians for fatigue.
738
00:51:04,610 --> 00:51:08,530
But there are still elements of safety
in aviation that have not been explored.
739
00:51:08,540 --> 00:51:13,080
One of the most well-known improvements
in airline safety is the black box.
740
00:51:13,080 --> 00:51:17,790
A surgeon in Toronto has been
working with a group of designers
741
00:51:17,800 --> 00:51:21,420
to create a similar tool
for the operating room.
742
00:51:21,420 --> 00:51:24,760
[Teodor Grantcharov] I want my
patients to feel the same way when
743
00:51:24,760 --> 00:51:27,550
they enter the operating room as I
feel when I enter a modern aircraft.
744
00:51:27,560 --> 00:51:31,350
Unless we create a system where
we understand, that we tolerate,
745
00:51:31,350 --> 00:51:35,810
and we learn from our errors, we
will never be able to improve.
746
00:51:35,810 --> 00:51:39,400
We've tried for many years to
create something like the black box.
747
00:51:39,400 --> 00:51:42,240
Finally, in 2012 we were able to create
a technology that allows us to capture
748
00:51:42,240 --> 00:51:48,160
video and audio and data from everything
that's happening in an operating room.
749
00:51:48,160 --> 00:51:53,250
We've been developing and
implemented a number of sensors.
750
00:51:53,250 --> 00:51:56,290
So, we know how many times
a door opens and closes.
751
00:51:56,300 --> 00:51:59,090
We know how we wash our hands
prior to a surgical procedure.
752
00:51:59,090 --> 00:52:02,720
And all these data feeds are combined
753
00:52:02,720 --> 00:52:05,220
and perfectly synchronized
on the same platform.
754
00:52:05,220 --> 00:52:07,770
When we talk with our patients about the
black box and what we are doing here,
755
00:52:07,770 --> 00:52:11,350
the first reaction, the most common
reaction in 90% of the patients is,
756
00:52:11,350 --> 00:52:15,310
"I can't believe this
hasn't been done before."
757
00:52:15,320 --> 00:52:17,900
From the point we started
recording our surgeries,
758
00:52:17,900 --> 00:52:20,780
we had a tremendous
amount of media attention.
759
00:52:20,780 --> 00:52:24,240
Everybody believed in the transparency
doctor, that's what he was nicknamed.
760
00:52:26,040 --> 00:52:30,420
He doesn't have anything to hide.
I'm definitely going to go to him.
761
00:52:30,420 --> 00:52:33,630
Patients need to know that
when they walk into a hospital,
762
00:52:33,630 --> 00:52:36,630
everything is being done to learn from
mistakes and possible risks that take place.
763
00:52:36,630 --> 00:52:41,050
This has to be common standard practice.
764
00:52:41,050 --> 00:52:43,800
We've heard for too long that healthcare is
complex, that our patients are not aircraft,
765
00:52:43,810 --> 00:52:49,390
that surgeons are not pilots. I just want
us to start doing something and changing it.
766
00:52:49,390 --> 00:52:54,150
We're trying to create
a system that identifies
767
00:52:54,150 --> 00:52:57,400
performance deficiencies
and improves safety.
768
00:52:57,400 --> 00:53:00,570
A new gadget comes out from an
industry provider all the time.
769
00:53:00,570 --> 00:53:04,700
Usually it's a very emotional
attachment like, "Oh, this looks sexy,"
770
00:53:04,700 --> 00:53:08,000
or "I like how this handle feels
when I'm using it during surgery."
771
00:53:08,000 --> 00:53:11,040
You need something deeper beyond that.
772
00:53:11,040 --> 00:53:13,340
So, this is what we call a full
surgical timeline and you can see
773
00:53:13,340 --> 00:53:16,510
the entire procedure broken
down from beginning to end.
774
00:53:16,510 --> 00:53:19,930
As you scroll down this timeline,
you'll start to see little beeps here,
775
00:53:19,930 --> 00:53:25,350
and that's where our surgical expert
analysts have coded where they saw errors.
776
00:53:25,350 --> 00:53:29,850
This screenshot establishes what one
of the errors is. So in this case
777
00:53:29,860 --> 00:53:33,610
an error took place during the suturing
task, and it was inadequate visualizations.
778
00:53:33,610 --> 00:53:37,530
As the surgeon was suturing with the needle
and driver, he might have gone off frame,
779
00:53:37,530 --> 00:53:42,740
which is incorrect because now you
have no idea where that needle is.
780
00:53:42,740 --> 00:53:46,250
You even see issues in say,
leadership or communication,
781
00:53:46,250 --> 00:53:49,130
and we have a whole toolset
determining exactly that.
782
00:53:49,130 --> 00:53:52,130
This points a really interesting
storyline because the blue bar establishes
783
00:53:52,130 --> 00:53:53,890
that the surgical
resident, the trainee under
784
00:53:53,890 --> 00:53:56,110
the main surgeon was doing the actual case,
785
00:53:56,340 --> 00:54:01,260
and then when a cluster
of errors takes place,
786
00:54:01,260 --> 00:54:03,600
you can see the switch over to the
main surgeon, to Dr. Grantcharov.
787
00:54:03,600 --> 00:54:06,850
This entire timeline
is the data quantified.
788
00:54:06,850 --> 00:54:11,690
We're breaking down the entire
set of errors into tangible areas
789
00:54:11,690 --> 00:54:15,950
to provide further education on
it to essentially improve it.
790
00:54:15,950 --> 00:54:19,410
Analytics is at the heart
of what the black box does,
791
00:54:19,410 --> 00:54:22,120
but we're jumping into different areas.
792
00:54:22,120 --> 00:54:24,540
Our engineers are working on
tools to improve handwashing
793
00:54:24,540 --> 00:54:27,670
to essentially create a
detector that lets you know,
794
00:54:27,670 --> 00:54:31,210
yes, you've spent the right amount of time
and the right technique to wash your hands.
795
00:54:31,210 --> 00:54:34,510
So over here, one of our engineers,
Kevin, has been working on just that.
796
00:54:34,510 --> 00:54:38,100
It's a motion sensing tool that
will look at how you wash your hands
797
00:54:38,100 --> 00:54:42,100
and look at the surface plane you
are working with, the amount of time
798
00:54:42,100 --> 00:54:44,890
spent on washing your hands,
and give you real time feedback.
799
00:54:44,890 --> 00:54:48,360
The key here is the data. So,
I can go and tell any surgeon
800
00:54:48,360 --> 00:54:51,570
you have to wash your hands this
many number of times in this fashion,
801
00:54:51,570 --> 00:54:55,780
but if I have hard data showing...
because of doing it this particular way
802
00:54:55,780 --> 00:55:00,240
we have reduced site infections
by this much, it's irrefutable.
803
00:55:04,830 --> 00:55:09,510
[piano music]
804
00:59:07,800 --> 00:59:13,230
[Sue Sheridan] Cal has gone on to
become part of a comedy community.
805
00:59:13,230 --> 00:59:17,190
He's producing. He's
produced two comedy shows.
806
00:59:17,190 --> 00:59:20,190
Cal uses his comedy in really novel ways
807
00:59:20,190 --> 00:59:24,530
that helps him deal with losing a dad.
808
00:59:42,800 --> 00:59:48,180
Everybody develops their own way
to deal with death or loss or grief
809
00:59:48,180 --> 00:59:52,850
and I think that comedy
is Cal's, uh, his outlet.
810
01:00:04,030 --> 01:00:09,620
[Sue Sheridan] He feels like he doesn't
suffer, but he sometimes struggles
811
01:00:09,620 --> 01:00:13,840
to be understood because his speech is impaired.
He struggles when he rides on airplanes
812
01:00:14,130 --> 01:00:19,170
because sometimes his scooters
or walkers are broken.
813
01:00:19,170 --> 01:00:22,970
He struggles in environments
where it's not easy to get around.
814
01:00:23,680 --> 01:00:28,310
The first year, or year and a half,
we had 183 separate medical visits
815
01:00:30,650 --> 01:00:35,440
for physical therapy, and ENT, eyes
and ears and teeth, and neurology,
816
01:00:36,070 --> 01:00:40,150
and during that time, in my heart,
I knew something was wrong with Cal.
817
01:00:40,160 --> 01:00:44,080
Our local doctors were really
not willing to offer a diagnosis.
818
01:00:44,080 --> 01:00:48,460
We took... I took Cal out of state to a
leading university where a team of specialists
819
01:00:48,750 --> 01:00:53,750
reviewed Cal's charts that
I had never looked at.
820
01:00:53,750 --> 01:00:56,460
I didn't think that there was any reason
for me to look at my birthing charts,
821
01:00:56,460 --> 01:00:59,680
and back then charts weren't
that available to patients.
822
01:00:59,680 --> 01:01:03,540
And they showed to me a report from an MRI
that they did on Cal when he was 5 days old
823
01:01:04,470 --> 01:01:10,350
that clearly shared abnormalities
in his brain from his jaundice.
824
01:01:11,940 --> 01:01:16,690
And our healthcare system really didn't...
825
01:01:16,700 --> 01:01:20,660
Well, they covered up. They
covered up Cal's injury and, umm...
826
01:01:24,080 --> 01:01:26,500
I wasn't empowered with
information and knowledge
827
01:01:29,040 --> 01:01:31,840
to challenge some of it or
ask the appropriate questions.
828
01:01:31,840 --> 01:01:34,460
You know, in healthcare they say that
patients, we need to ask more questions,
829
01:01:34,460 --> 01:01:37,970
but sometimes we simply
don't know what to ask.
830
01:01:39,600 --> 01:01:44,930
[Michael Millenson] Understand, before
you go in for any particular procedure,
831
01:01:44,930 --> 01:01:49,900
what are the questions you need
to ask to keep yourself safe?
832
01:01:49,900 --> 01:01:53,070
And if we all start asking those questions,
833
01:01:53,070 --> 01:01:55,820
then pretty soon it will
become clear to any hospital
834
01:01:55,820 --> 01:01:58,660
that's not doing those things that
there is pressure on them to do it.
835
01:01:58,660 --> 01:02:03,120
If you're in a hospital, by definition today,
you're seeing a lot of different doctors,
836
01:02:03,120 --> 01:02:07,880
there's a lot of caregivers
coming in and out of the room,
837
01:02:07,880 --> 01:02:10,630
most of whom work to communicate with
each other, but sometimes they miss.
838
01:02:10,630 --> 01:02:13,960
So, if you see something that doesn't
look right, or sound right, you say,
839
01:02:13,960 --> 01:02:17,840
"Whoa, wait a minute, that's
not what they told me."
840
01:02:17,850 --> 01:02:20,430
Patient safety is a team
sport. And one of the ways
841
01:02:20,430 --> 01:02:23,520
to really make a difference is
you've got to get patients engaged.
842
01:02:23,520 --> 01:02:26,190
So, if patients begin walking
into hospitals with an expectation
843
01:02:26,190 --> 01:02:30,020
that they are not going to get an
infection and they start saying,
844
01:02:30,020 --> 01:02:33,030
"Hey, have you washed your hands before you
come over to see me?" That's how it happens.
845
01:02:33,030 --> 01:02:36,820
If they've done it outside the room, or
they've done it at the nurse's station,
846
01:02:36,820 --> 01:02:41,450
on the way into the room they're touching
the door, they're touching things,
847
01:02:41,450 --> 01:02:45,000
and then they are coming
in, so that doesn't count.
848
01:02:45,000 --> 01:02:47,080
It has to be before and
after patient contact.
849
01:02:47,080 --> 01:02:50,040
Here's what I look for in a hospital
that's really outstanding on safety;
850
01:02:50,050 --> 01:02:52,250
the sink is placed in a
way that it is easy to
851
01:02:52,250 --> 01:02:54,630
walk into a room and
immediately wash your hands.
852
01:02:54,630 --> 01:03:00,260
You'll see charts on patient floors, right
there for anyone to see, that will show
853
01:03:00,270 --> 01:03:06,060
how they are doing on patient falls, for
instance, or how they are doing on infections.
854
01:03:06,060 --> 01:03:10,980
People have an attitude about
safety, you just can feel it.
855
01:03:10,990 --> 01:03:12,820
There's an apocryphal
story of President Kennedy
856
01:03:12,820 --> 01:03:14,570
visiting Cape Canaveral
during his presidency
857
01:03:14,570 --> 01:03:19,910
and he takes aside a custodian and says,
"What's your job?" And the custodian says,
858
01:03:19,910 --> 01:03:24,670
"Mr. President, my job is to help get a man
to the moon and return him to earth safely."
859
01:03:25,080 --> 01:03:29,760
Everybody has a job to do to
protect patients, not just doctors.
860
01:03:31,170 --> 01:03:37,100
Every nurse, pharmacist, physician,
custodian, has a role in safety.
861
01:03:37,220 --> 01:03:43,020
I think it's deeply unfair to expect patients
who are sick, in the middle of an illness,
862
01:03:43,020 --> 01:03:47,270
to try and sort this out on
their own. Now, it may be unfair,
863
01:03:47,280 --> 01:03:50,690
but the reality is that's where we are.
864
01:03:50,700 --> 01:03:53,360
The best thing they can do is have a family
member or a friend around, because again,
865
01:03:53,360 --> 01:03:57,280
when in the middle of an illness it's
very hard for you to pay attention
866
01:03:57,290 --> 01:04:00,200
to know what's going on, but your
friend can, your family member can.
867
01:04:00,210 --> 01:04:03,170
If somebody says you're
going to get medication X,
868
01:04:03,170 --> 01:04:06,040
is that the medication
that actually showed up?
869
01:04:06,050 --> 01:04:08,800
And asking in a very
friendly and respectful way,
870
01:04:08,800 --> 01:04:12,300
when a nurse comes by to hang a
medication or give you a pill,
871
01:04:12,300 --> 01:04:16,100
you know, what is this? What am I getting?
It's a totally reasonable question.
872
01:04:16,100 --> 01:04:20,600
Patients should feel comfortable
doing it. And if you have a provider
873
01:04:20,600 --> 01:04:23,860
that responds badly to that, you should try
to figure out if you can switch providers.
874
01:04:23,860 --> 01:04:28,320
My father was a doctor in a
small town in Connecticut.
875
01:04:28,320 --> 01:04:32,570
For a lot of time he was the only
doctor there and he was revered.
876
01:04:32,570 --> 01:04:37,620
You didn't question him. It wasn't
my father's fault in any way.
877
01:04:37,620 --> 01:04:42,210
He was a proud and successful
professional honored by his community.
878
01:04:42,210 --> 01:04:45,880
That's not actually adaptive if we
really want care to be what it can be.
879
01:04:45,880 --> 01:04:50,550
I think, I understand the hesitation
people may feel to ask the doctor,
880
01:04:50,550 --> 01:04:53,360
"What's going on here?" But that's healthy,
881
01:04:53,360 --> 01:04:56,850
that's good, and we need
to train doctors to,
882
01:04:56,890 --> 01:05:02,480
not just to accept that, but to absolutely
welcome it. It's better medicine.
883
01:05:02,480 --> 01:05:06,940
[tense music]
884
01:05:07,820 --> 01:05:13,580
[Sully Sullenberger] If, as reports indicate,
there are as many as 440,000 preventable
885
01:05:13,580 --> 01:05:19,000
medical deaths in this country alone
every year, that is the equivalent of
886
01:05:19,000 --> 01:05:23,340
7 or 8 airliners crashing
every day with no survivors.
887
01:05:23,340 --> 01:05:27,760
Before the first day of that kind of carnage
was complete, airplanes would be grounded,
888
01:05:27,760 --> 01:05:33,300
airlines would stop operating,
airports would close, no one would fly
889
01:05:33,310 --> 01:05:38,020
until some of the fundamental
issues had been resolved.
890
01:05:38,020 --> 01:05:42,230
But because aviation
accidents are dramatic,
891
01:05:42,230 --> 01:05:45,610
they receive the kind of attention
that they do, and the public awareness.
892
01:05:45,610 --> 01:05:47,710
Medical deaths occur
singly and often behind
893
01:05:47,710 --> 01:05:50,450
the scenes, but in
aggregate the harm is huge.
894
01:05:52,490 --> 01:05:57,790
We need to change the way we
think about these medical deaths.
895
01:05:57,790 --> 01:06:03,170
We need to think about them not as
unavoidable, but as unthinkable.
896
01:06:03,300 --> 01:06:08,640
We've got to get better at making sure
whatever hospital you go into in the U.S.
897
01:06:10,180 --> 01:06:15,390
you're getting the same quality
care, and we are not there.
898
01:06:15,390 --> 01:06:18,480
I mean, you're asking people
to do things differently.
899
01:06:18,480 --> 01:06:20,820
You're asking doctors to think
differently and work differently.
900
01:06:20,820 --> 01:06:23,320
You're asking architects to build
different spaces, nurses to work
901
01:06:23,320 --> 01:06:27,410
differently in teams, patients
to have a different role.
902
01:06:27,410 --> 01:06:30,160
To change patient safety, you have to change
everything. If you look at preventable harm
903
01:06:30,160 --> 01:06:34,160
across American hospitals, it
has gone down considerably,
904
01:06:34,160 --> 01:06:37,210
you know, saving hundreds
of thousands of lives
905
01:06:37,210 --> 01:06:39,380
and billions of dollars. That
doesn't mean we fixed it.
906
01:06:39,380 --> 01:06:43,590
It is quietly, slowly, but definitely
becoming the professional norm
907
01:06:43,970 --> 01:06:49,760
to take certain precautions, to do things
in a certain way so that patients are safe.
908
01:06:49,760 --> 01:06:52,590
Because you can be
satisfied if you have very
909
01:06:52,590 --> 01:06:55,600
low expectations and the
reality is that all of
910
01:06:55,600 --> 01:07:00,940
our expectations should be raised, that we
all get very high quality, safe healthcare.
911
01:07:00,940 --> 01:07:05,280
We all want to do well. We all want
to get better. Nobody comes to work
912
01:07:05,280 --> 01:07:10,700
to harm a patient or wanting to
harm a patient or to give bad care.
913
01:07:10,700 --> 01:07:15,500
We haven't made this a public health
issue where the public is really
914
01:07:15,500 --> 01:07:19,340
thinking about this, and yet
when you talk to any person
915
01:07:19,340 --> 01:07:23,380
who's had a family member or themselves
in healthcare, they all have a story.
916
01:07:23,380 --> 01:07:26,850
I've also talked to doctors and nurses who
have committed a terrible error and they say,
917
01:07:26,850 --> 01:07:30,520
"I know I can't take that back, but
what will really give that meaning is
918
01:07:30,520 --> 01:07:34,310
if I do something that makes the
system safer for the next person."
919
01:07:34,310 --> 01:07:37,230
I think, in part, the job of
people like me in leadership roles
920
01:07:37,230 --> 01:07:40,110
is to harness that passion,
harness that energy.
921
01:07:40,110 --> 01:07:42,780
All of the rest of these guys are much
more serious about medical error reduction
922
01:07:42,780 --> 01:07:46,620
than they ever were. Is it going as
fast as it could? No, of course not.
923
01:07:46,620 --> 01:07:50,080
It is not "stuff happens" anymore.
924
01:07:50,080 --> 01:07:53,580
That's where we're going, and that's the
good future that we're moving towards.
925
01:07:53,580 --> 01:07:57,380
It feels like we should be further
along than we are, but actually
926
01:07:57,380 --> 01:08:00,630
I think we've made tremendous progress
in 15 years. It is on the map.
927
01:08:00,630 --> 01:08:04,010
We have these examples in the
U.S. and around the world.
928
01:08:04,010 --> 01:08:07,810
It's not any longer a question of
possibility, it's a question of will.
929
01:08:14,020 --> 01:08:19,740
Many of us go kind of through a self-blame.
Although we know it wasn't our fault,
930
01:08:19,740 --> 01:08:24,490
we feel like we didn't... protect our son.
931
01:08:24,780 --> 01:08:29,870
And so there was really,
really significant grieving,
932
01:08:29,870 --> 01:08:34,420
so the anger at first was immeasurable.
933
01:08:34,420 --> 01:08:39,010
When we discovered Pat's error,
we both felt tremendous fear.
934
01:08:39,010 --> 01:08:43,590
I think at that point it
was just plain disbelief.
935
01:08:43,600 --> 01:08:47,680
He said: "Whatever you do, do
not give up on patient safety."
936
01:08:47,680 --> 01:08:52,190
So, that led me onto a journey to... I
wanted to make sure our healthcare system,
937
01:08:52,190 --> 01:08:55,940
our government knew what
happened to Pat and Cal.
938
01:08:55,940 --> 01:08:59,360
So, it took us 8 years, but we really did
make some changes in our healthcare system
939
01:08:59,360 --> 01:09:03,830
where babies being discharged would have a
bilirubin test before they were discharged.
940
01:09:03,830 --> 01:09:09,250
[Mark Graber]Thanks to Sue and the work
that she's done, there are now processes
941
01:09:09,250 --> 01:09:13,090
in place in every hospital
to screen for that condition.
942
01:09:13,090 --> 01:09:16,420
And the odds that that's going to
happen again are now approaching zero,
943
01:09:16,420 --> 01:09:20,890
and that's what we'd like to
see happen throughout medicine.
944
01:09:20,890 --> 01:09:23,890
And the work that Sue has donenis
our model for how to do that.
945
01:09:23,890 --> 01:09:28,230
She turned what she had gone through
into empowerment and positivity,
946
01:09:28,520 --> 01:09:33,730
and if she can do it, so can
I, and so can a lot of people.
947
01:09:33,730 --> 01:09:38,490
I've obviously always idolized my Mom
and I understood her job very well
948
01:09:38,490 --> 01:09:43,580
and people would ask me, "What does your Mom
do?" I would say, "Well, she saves lives."
949
01:09:43,580 --> 01:09:47,120
Having witnessed these tragic
outcomes in our healthcare system,
950
01:09:47,120 --> 01:09:51,380
the one place that we should
feel unquestionably safe.
951
01:09:51,380 --> 01:09:57,130
[Mackenzie Sheridan] And it kind of ignited
a fire inside me that wanted to, you know,
952
01:09:57,140 --> 01:10:01,470
do what my Mom does, which is, you know,
talk with hospitals and talk with doctors
953
01:10:01,470 --> 01:10:05,230
and figure out how we can make those
kind of things not happen again.
954
01:10:05,230 --> 01:10:09,110
So, I went to Portland State University
and I chose to do Public Health
955
01:10:09,110 --> 01:10:14,070
because I wanted to feel like I was making
a difference and feel like I, you know,
956
01:10:14,070 --> 01:10:20,330
could help prevent things that happened
to my family, happening to other people.
957
01:10:28,210 --> 01:10:33,420
I am unwilling to believe that
we have done all that we can do.
958
01:10:33,420 --> 01:10:38,140
My experience with diagnostic errors and the
healthcare system has been without a doubt
959
01:10:38,850 --> 01:10:44,310
the most powerfully emotional
experience in my life.
960
01:10:44,480 --> 01:10:50,110
However, my family's story
is also a story of awakening,
961
01:10:50,110 --> 01:10:55,490
of passion, of change,
and hope for the future.
962
01:10:56,660 --> 01:11:01,450
I cannot change what happened to
Cal and Pat, but I've always felt
963
01:11:02,040 --> 01:11:06,880
that I can somehow be part
of it and make a difference.
964
01:11:06,880 --> 01:11:10,590
My teacher in courage, in hope,
965
01:11:11,420 --> 01:11:15,640
in determination, in passion,
966
01:11:15,640 --> 01:11:19,010
of course he's my teacher in sense of humor
which he believes his mother has none of,
967
01:11:19,010 --> 01:11:24,230
but he's the reason for what's in me.
968
01:11:51,130 --> 01:11:55,970
[audience applauding]
969
01:12:04,360 --> 01:12:10,240
[Sue Sheridan] You know, when Pat was dying
he said, "Never give up on patient safety."
970
01:12:10,240 --> 01:12:14,950
At that time, I did not envision
my whole family being engaged.
971
01:12:14,950 --> 01:12:18,620
Before we went on stage today, I thought
about Pat, my daughter in the front row,
972
01:12:18,620 --> 01:12:23,380
my son on stage. It was, umm, just surreal.
973
01:12:28,470 --> 01:12:33,300
[singing Happy Birthday]
974
01:12:42,020 --> 01:12:47,820
[Mackenzie Sheridan] On March 8th, which
is the day that my dad passed away,
975
01:12:47,820 --> 01:12:51,280
we spread his ashes on Table Rock.
976
01:12:51,370 --> 01:12:54,590
Whenever we go there I always feel
like a warm, just like, presence.
977
01:12:57,160 --> 01:13:01,370
It's because it's such a beautiful place,
978
01:13:01,380 --> 01:13:03,290
and it's beautiful that he's there as well.
979
01:13:03,290 --> 01:13:06,920
[Sue Sheridan] Pat will always be alive
in our hearts and in our memories,
980
01:13:08,670 --> 01:13:12,640
and it was very hard for them to lose
a Dad when they were only 4 and 6.
981
01:13:12,640 --> 01:13:16,220
They will continue to honor
andnmiss and wonder about their Dad.
982
01:13:21,560 --> 01:13:23,400
I've always had this hope:
983
01:13:28,280 --> 01:13:31,310
I will not believe that our
leadership in our country, in
984
01:13:31,320 --> 01:13:34,450
our healthcare system, will
continue to think this is okay.
985
01:13:34,450 --> 01:13:36,910
Because it's not.
986
01:13:46,680 --> 01:13:51,350
[piano music]
95192
Can't find what you're looking for?
Get subtitles in any language from opensubtitles.com, and translate them here.