Insulin and Oral Hypoglycemic Toxicity
Download
Report
Transcript Insulin and Oral Hypoglycemic Toxicity
Diagnostic Testing
Ethan Cowan, MD, MS
Department of Emergency Medicine
Jacobi Medical Center
Department of Epidemiology and Population Health
Albert Einstein College of Medicine
The Provider Dilemma
A
26 year old pregnant female presents after
twisting her ankle. She has no abdominal or
urinary complaints. The nurse sends a UA
and uricult dipslide prior to you seeing the
patient. What should you do with the
results of these tests?
The Provider Dilemma
Should a provider give
antibiotics if either one
or both of these tests
come back positive?
Why Order a Diagnostic Test?
When the diagnosis is
uncertain
Incorrect diagnosis
leads to clinically
significant morbidity
or mortality
Diagnostic test result
changes management
Test is cost effective
Clinician Thought Process
Clinician derives patient
prior prob. of disease:
H & P
Literature
Experience
“Index of Suspicion”
0% - 100%
“Low, Med., High”
Threshold Approach to
Diagnostic Testing
Probability of Disease
0%
100%
Testing Zone
P(-)
P(+)
P < P(-)
Dx testing & therapy not indicated
P(-) < P < P(+) Dx testing needed prior to therapy
P > P(+)
Only intervention needed
Pauker and Kassirer, 1980, Gallagher, 1998
Threshold Approach to
Diagnostic Testing
Probability of Disease
0%
100%
Testing Zone
P(-)
P(+)
Width of testing zone depends on:
Test properties
Risk of excess morbidity/mortality attributable to the test
Risk/benefit ratio of available therapies for the Dx
Pauker and Kassirer, 1980, Gallagher, 1998
Test Characteristics
Reliability
Inter observer
Intra observer
Correlation
B&A Plot
Simple Agreement
Kappa Statistics
Validity
Sensitivity
Specificity
NPV
PPV
ROC Curves
Reliability
The extent to which
results obtained with a
test are reproducible.
Reliability
Not Reliable
Reliable
Intra rater reliability
Extent
to which a
measure produces
the same result at
different times for
the same subjects
Inter rater reliability
Extent to which a
measure produces the
same result on each
subject regardless of
who makes the
observation
Correlation (r)
For continuous data
r=1
perfect
r=0
none
O1
O1 = O2
O2
Bland & Altman, 1986
Correlation (r)
Measures relation
strength, not
O1
agreement
Problem: even near
perfect correlation
may indicate
significant differences
between observations
r = 0.8
O1 = O2
O2
Bland & Altman, 1986
Bland & Altman Plot
O1 – O 2
For continuous data
Plot of observation
differences versus the
means
Data that are evenly
distributed around 0
and are within 2 STDs
exhibit good
agreement
10
0
-10
[O1 + O2] / 2
Bland & Altman, 1986
Simple Agreement
Rater 1
+
total
Extent
Rater 2
+
a
b
c
d
a+c b+d
total
a+b
c+d
N
to which two or more raters agree on the
classifications of all subjects
% of concordance in the 2 x 2 table (a + d) / N
Not ideal, subjects may fall on diagonal by chance
Kappa
Rater 1
+
total
The
Rater 2
+
a
b
c
d
a+c b+d
total
a+b
c+d
N
proportion of the best possible improvement in
agreement beyond chance obtained by the observers
K = (pa – p0)/(1-p0)
Pa = (a+d)/N (prop. of subjects along the main diagonal)
Po = [(a + b)(a+c) + (c+d)(b+d)]/N2 (expected prop.)
Interpreting Kappa Values
K=1
K > 0.80
0.60 < K < 0.80
0.40 < K < 0.60
0 < K < 0.40
K=0
K<0
Perfect
Excellent
Good
Fair
Poor
Chance (pa = p0)
Less than chance
Weighted Kappa
Rater 1
1
2
.
.
C
total
Rater 2
1
2
n11
n12
n21
n22
.
.
.
.
nC1
nC2
n.1
n.2
...
...
...
...
...
...
...
C
n1C
n2C
.
.
nCC
n.C
total
n1.
n2.
.
.
nC.
N
Used for more than 2 observers or categories
Perfect agreement on the main diagonal weighted
more than partial agreement off of it.
Validity
The degree to which a
test correctly diagnoses
people as having or not
having a condition
Internal Validity
External Validity
Validity
Valid, not reliable
Reliable and Valid
Internal Validity
Performance Characteristics
Sensitivity
Specificity
NPV
PPV
ROC Curves
2 x 2 Table
Disease Status
Test
Result
+
total
cases
noncases
TP
FN
FP
TN
cases
noncases
TP = True Positives
FP = False Positives
total
positives
negatives
N
TN = True Negatives
FN = False Negatives
Gold Standard
Definitive
test used
to identify cases
Example:
traditional agar
culture
The dipstick and
dipslide are
measured against
the gold standard
Sensitivity (SN)
Disease Status
Test
Result
+
total
cases
noncases
TP
FN
FP
TN
cases
noncases
total
positives
negatives
N
Probability
of correctly identifying a true case
TP/(TP + FN) = TP/ cases
High SN, Negative test result rules out Dx (SnNout)
Sackett & Straus, 1998
Specificity (SP)
Disease Status
Test
Result
+
total
cases
noncases
TP
FN
FP
TN
cases
noncases
total
positives
negatives
N
Probability
of correctly identifying a true noncase
TN/(TN + FP) = TN/ noncases
High SP, Positive test result rules in Dx (SpPin)
Sackett & Straus, 1998
Problems with
Sensitivity and Specificity
Remain
constant over patient populations
But, SN and SP convey how likely a test
result is positive or negative given the
patient does or does not have disease
Paradoxical inversion of clinical logic
Prior knowledge of disease status obviates
need of the diagnostic test
Gallagher, 1998
Positive Predictive Value (PPV)
Disease Status
Test
Result
+
total
cases
noncases
TP
FN
FP
TN
cases
noncases
total
positives
negatives
N
Probability
that a labeled (+) is a true case
TP/(TP + FP) = TP/ total positives
High SP corresponds to very high PPV (SpPin)
Sackett & Straus, 1998
Negative Predictive Value (NPV)
Disease Status
Test
Result
+
total
cases
noncases
TP
FN
FP
TN
cases
noncases
total
positives
negatives
N
Probability
that a labeled (-) is a true noncase
TN/(TN + FN) = TP/ total negatives
High SN corresponds to very high NPV (SnNout)
Sackett & Straus, 1998
Predictive Value Problems
Vulnerable
to Disease Prevalence (P) Shifts
Do not remain constant over patient populations
As P
PPV
NPV
As P
PPV
NPV
Gallagher, 1998
Flipping a Coin to Dx AMI for
People with Chest Pain
ED AMI Prevalence 6%
AMI
No AMI
Heads (+) 3
47
50
Tails (-) 3
47
50
6
94
100
SN = 3 / 6 = 50%
SP = 47 / 94 = 50%
PPV= 3 / 50 = 6%
NPV = 47 / 50 = 94%
Worster, 2002
Flipping a Coin to Dx AMI for
People with Chest Pain
CCU AMI Prevalence 90%
AMI
No AMI
Heads (+) 45
5
50
Tails (-) 45
5
50
10
100
90
SN = 45 / 90 = 50%
SP = 5 / 10 = 50%
PPV= 45 / 50 = 90%
NPV = 5 / 50 = 10%
Worster, 2002
Receiver Operator Curve
1.0
Sensitivity
(TPR)
0.0
0.0 1-Specificity (FPR) 1.0
Allows consideration of test performance across a
range of threshold values
Well suited for continuous variable Dx Tests
Receiver Operator Curve
Avoids
the “single
cutoff trap”
Sepsis
No Effect Effect
WBC Count
Gallagher, 1998
Area Under the Curve (θ)
1.0
Sensitivity
(TPR)
0.0
0.0 1-Specificity (FPR) 1.0
Measure
of test accuracy
(θ) 0.5 – 0.7 no to low discriminatory power
(θ) 0.7 – 0.9 moderate discriminatory power
(θ) > 0.9
high discriminatory power
Gryzybowski, 1997
Problem with ROC curves
Same
problems as SN and SP “Reverse
Logic”
Mainly used to describe Dx test
performance
Appendicitis Example
Study design:
Prospective cohort
Gold standard:
Pathology report from
appendectomy or CT
finding (negatives)
Diagnostic Test:
Total WBC
Physical Exam
+
OR
+
Appy
CT Scan
-
No Appy
Cardall, 2004
Appendicitis Example
WBC
Appy
Not Appy Total
> 10,000
66
89
155
< 10,000
21
98
119
Total
87
187
274
SN 76% (65%-84%)
SP 52% (45%-60%)
PPV 42% (35%-51%)
NPV 82% (74%-89%)
Cardall, 2004
Appendicitis Example
Patient WBC:
13,000
Management:
Get CT with PO & IV
Contrast
Physical Exam
+
OR
+
Appy
CT Scan
-
No Appy
Cardall, 2004
Abdominal CT
Follow UP
CT
result: acute
appendicitis
Patient taken to
OR for
appendectomy
But, was WBC necessary?
Answer given in talk on Likelihood Ratios