evidence-based practice for physical therapists

Download Report

Transcript evidence-based practice for physical therapists

THE BASICS: EVIDENCE-BASED
PRACTICE FOR PHYSICAL
THERAPISTS AND PHYSICAL
THERAPIST ASSISTANTS,
ONLINE TRAINING MODULE
1
Jessica Lambeth, MPT
[email protected]
Module Purpose
The purpose of this online training module is
to share the basics of evidence-based practice
(EBP). This module focuses on general concepts
of EBP, with clinical scenarios related to schoolbased physical therapy.
2
Module Purpose, continued
The main emphasis of this module is not to
promote certain treatments or tests and measures
or “tell you” what the recent literature reveals about
pediatric physical therapy practices. As will be
discussed in the coming slides, the “research
answer” alone is not the correct answer. In
addition, EBP emphasizes the necessity of learning
to perform searches and evaluate the material
independently, related to clinicians’ and patients’
current circumstances. While initially time
consuming, the results are worthwhile.
3
Module Purpose, continued
For those without a recent EBP background
or training, this module will not result in immediate
efficiency in literature searches, statistical
interpretation, etc. Hopefully it will, however,
encourage more questions and motivation for EPB.
It will also enable the NC DPI to focus educational
sessions on meeting needs in the area of EBP and
to gauge the current knowledge of school-based
PTs and PTAs. Your feedback and comments will
help to plan future learning opportunities and
resources.
4
Pretest
Please stop here and complete the pretest if
you have not already done so. Please also keep
track of your time, as directed in the instructions.
Also, “I don’t know” answers have been
included to ensure that we receive appropriate
feedback. If you really do not know the answer, in
the pretest or the posttest, please indicate that in
your response. This will greatly assist in providing
necessary learning opportunities in the future, as
well as help us to evaluate the effectiveness of the
online module format and material.
5
References:
The content of this EBP module is largely based
on curriculum from courses within the transitional
Doctor of Physical Therapy program at Rocky
Mountain University of Health Professions in Provo,
Utah. (web address: www.rmuohp.edu)
6
References, continued:
Much of the material within this online training module can be
found in the following texts:
Guyatt G, Rennie D. Users' Guides to the Medical Literature- Essentials of
Evidence-Based Clinical Practice. Chicago: AMA Press; 2002.
Jaeschke R, Singer J, Guyatt GH. Measurement of health status:
ascertaining the minimal clinically important difference. Control Clin Trials
1989;10:407-15.
Portney LG, Watkins MP. Foundations of Clinical Research: Applications to
Practice. 2nd ed. Upper Saddle River, NJ: Prentice-Hall Inc, 2000.
Rothstein JM. Autonomous Practice or Autonomous Ignorance? (Editorial)
Physical Therapy 81(10), October 2001.
7
References, continued:
Much of the material within this online training module can be
found in the following texts:
Straus SE, Richardson WS, Glasziou P, Haynes RB. Evidence-based
Medicine: How to Practice and Teach EBM. 3rd ed. Edinburgh: Churchill
Livingstone; 2005.
All clipart came from Microsoft Office:
http://office.microsoft.com/en-us/clipart/default.aspx.
Accessed 05/09/2009.
8
Module Outline
1) What is EBP? Slides 9-22
2) Statistics Review: Basic Research, Slides 23-115
3) How to Search for EBP, Slides 116-152
4) How to Interpret Research Related to EBP, Slides
153-174
9
Section 1 Outline
What is Evidence-Based Practice?
• Evidence-Based Practice: General Introduction,
10-12
• Survey Results (NC school-based PTs view of
EBP),13-18
• Guiding Steps to Practice EBP,19-20
• Two Fundamental Principles of EBP, 21
• Best Research Evidence, 22
10
Evidence-Based Practice:
General Introduction
EBP is the integration of the best research evidence,
clinical expertise, and the patient’s values and
circumstances
• Best Research Evidence: valid and clinically
relevant research with a focus on patient-centered
clinical research
• Clinical Expertise: use of clinical skills and
experiences
• Patient’s Values and Circumstances: the
patient’s unique preferences, concerns, and
expectations in his or her setting
11
(Straus et al, 2005)
Evidence-Based Practice:
General Introduction
EBP is Not:
• Focused only on research studies
• Only to be used or understood by professionals
who routinely participate in research studies
• A discouragement from trying new treatment
 There may be little or no research on a
particular topic, or studies with small sample
sizes may have lacked the power to
demonstrate statistical significance (as later
explained in the statistics section)
12
Evidence-Based Practice:
General Introduction
“Because RCTs are so difficult, we will always have
areas that lack evidence, we will need to find other
credible research approaches to supply evidence.
Keep in mind that an absence of evidence is
different from negative evidence. An absence of
evidence is not an excuse to ignore the growing
body of data available to guide practice.”
(RCT = randomized clinical or controlled trials)
13
(Rothstein 2001)
Survey Results
Survey responses from North Carolina SchoolBased PTs/PTAs about EBP:
• Survey distributed at the NC Exceptional
Children’s Conference, Nov 2008
• 41 Respondents
(www.med.unc.edu/ahs/physical/schoolbasedpt for
detailed results)
• 73% had participated in a conference or course
on the use of EBP in the last 5 years
• Of those that participated in a course,
respondent data suggests the course changed
their view of EBP, but their use and practice
14
of EBP was changed to a lesser degree.
Survey Results, Continued…
• Highest frequency response as to why we should
use EBP  positive impact on our clinical practice
• 39 of 41 respondents agreed that EBP is relevant
and necessary for PTs in the school system (2 left
that question blank)
 Highest frequency responses as to why it is
relevant and necessary: A focus on EBP results in
improved clinical practice and provides validation
and justification for our role as school-based
PTs/PTAs
15
Survey Results, Continued…
• When respondents were asked if they were
comfortable searching for and using EBP:
• Yes: 17 (41%),
• No: 18 (44%),
• No Response: 6 (15%)
• Internet was the most frequently
used source for searching EBP,
but continuing education courses
ranked highest for preference.
16
Survey Results, Continued…
• A majority of the respondents were comfortable:
 Determining the level of evidence and
interpreting the authors’ conclusions
• A majority of the respondents were not comfortable:
 Interpreting statistics, even though statistical
knowledge is often helpful to evaluate conclusions
drawn by the author
17
Survey Results, Continued…
The primary barriers to searching and using EBP, as
listed by NC school PTs/PTAs:
1. Time
2. Access
3. Uncertain how to search for EBP
Factors that enhance the search and use of EBP,
as listed by NC school PTs/PTAs listed:
1. Additional time
2. Education on the appropriate use of EBP
3. Education in how to access EBP resources
18
Survey Results, Continued…
Questions generated from the survey:
• How to efficiently and effectively increase the
knowledge and practice of EBP by NC school
PTs/PTAs?
• How to address barriers to using and searching
for EBP?
• How to enhance use and access of EBP in
school-based practice?
19
Guiding Steps to Practice EBP
1. Analyze what we know and what we do not know,
in relation to improving our clinical practice. Form
answerable questions to address any gaps in our
knowledge.
2. Search for and find the best research evidence to
address our questions.
3. Critically appraise the information, based on its
validity, impact or size of effect, and applicability.
20
(Straus et al, 2005)
Guiding Steps to Practice EBP
4. Integrate information gathered from the best
research evidence with clinical expertise and the
patient’s values and circumstances
5. Evaluate the effectiveness of any intervention
taken based on steps 1-4, and the effectiveness
and efficiency of the process
21
(Straus et al, 2005)
Two Fundamental Principles of EBP
1. “Evidence alone is never sufficient to make a
clinical decision” (page 8)
Consider risks and benefits, costs, inconvenience,
alternative treatment strategies, patient
preferences/values and circumstances.
2. “EBM posits a hierarchy of evidence to guide
clinical decision making” (page 8)
Not all research is equal in terms of relevance and
statistical support, however, that does not mean
lower level evidence is not worthwhile.
(Guyatt
and Rennie use the term Evidence-Based
22
Medicine, EBM)
(Guyatt and Rennie, 2002)
Best Research Evidence
• The three sections that follow will focus on the
“best research evidence” branch of Straus’ three
components of EBM. (Straus et al, 2005)
• Best research evidence is not more important than
the other two branches; it is prominent in this module
because knowledge concerning clinical expertise and
patient values and expectations will vary from
situation to situation.
• Section 2 (Statistics Review, Basic Research) will
provide the background information necessary to
perform effective searches and interpret the best
23
research evidence (Sections 3 and 4).
Section 2 Outline
Statistics Review, Basic Research
•
•
•
•
Types of Research, 25-33
Hierarchy of Evidence, 34-35
Variables, 36-44
Measurement Validity
• Types, 45-48
• Statistics, 49-67
• Sensitivity and Specificity
• Positive and Negative Predictive Value
• Positive and Negative Likelihood
• Receiver Operating Characteristic
(ROC) Curve
• Responsiveness to Change
•
Effect Size versus MCID
24
Section 2 Outline, continued…
Statistics Review, Basic Research, continued
•
•
•
Measurement Reliability, 68-72
Descriptive Statistics,73-83
• Frequency and Shape of Distribution
• Central Tendency Measures
• Measures of Variability
Inferential Statistics, 84-115
• Probability
• Sampling Error
• Confidence Intervals
• Hypothesis Testing
• Power
25
Types of Research
Nonexperimental (Observational)
• Descriptive or Exploratory
• No control or manipulation of variables
• Examines populations and relationships
Experimental
• Researcher manipulates/controls variable(s)
• Comparison of interventions or groups,
examines cause and effect
26
(Portney and Watkins, 2000)
Types of Research
Portney and Watkins suggest viewing various designs
as a continuum of research with a descriptive,
exploratory, or experimental purpose.
Certain designs may include elements of more than
one classification.
27
(Portney and Watkins,2000)
Descriptive Research
Examples:
• Case Study: Description of one or more patients,
may document unusual conditions or response to
intervention
• Developmental Research: Examines patterns of
growth and change, or documents natural history
of a disease or condition
• Normative Research: Establishes typical values
for specific variables
28
(Portney and Watkins, 2000)
Descriptive Research, continued…
Examples:
• Qualitative Research: Collection of subjective,
narrative information (rather than quantitative,
numerical data) in an effort to better understand
experiences
• Evaluation Research: Assessment of a program
or policy, often by the collection of descriptive
information through surveys or questionnaires
29
(Portney and Watkins, 2000)
Exploratory Research
Examples:
• Correlational Methods: Examines relationships
between areas of interest, may be used to predict
or suggest, but cannot offer cause and effect
• Cohort and Case-Control Studies: Used often in
epidemiological research to describe and predict
risks for certain conditions
• Methodological Studies: Used to evaluate the
validity and reliability of measuring instruments
• Historical Research: Use of archives or other
records to reconstruct the past to generate
questions or suggest relationships of historical
30
interest
(Portney and Watkins, 2000)
Experimental Research
Example
• Randomized Clinical or Controlled Trial (RCT):
In general, a clinical treatment, or experimental
condition, is compared to a control condition, often
a placebo but in some cases an alternative
treatment, where subjects are randomly assigned
to a group.
31
(Portney and Watkins, 2000)
Experimental Research, continued…
Examples:
• Single-Subject Design: Variation of RCT, study
of an individual over time with repeated
measurement and determined design phases
(Portney and Watkins, 2000)
In an N=1 RCT, a single individual receives
alternating treatment and placebo or alternative
treatment, with the patient and the assessor
blinded to intervention allocation. Objective or
subjective measures are then recorded during the
allocation periods. (Guyatt and Rennie, 2002)
32
Experimental Research, continued…
Examples:
• Sequential Clinical Trial: Variation of RCT,
technique that allows for the continuous analysis of
data as it becomes available, does not require a
fixed sample
• Quasi-Experimental Research: Comparative
research in which subjects cannot be randomly
assigned to a group, or control groups cannot be
used. Lower level of evidence than RCTs.
33
(Portney and Watkins 2000)
Experimental Research, continued
Examples:
• Systematic Review: Combination of several
studies with the same or similar variables, in which
the studies are summarized and analyzed (Guyatt
and Rennie, 2002)
• Meta-analysis: Statistical combination of the data
from several studies with the same or similar
variables, to determine an overall outcome (Portney
and Watkins, 2000; Guyatt and Rennie, 2002)
34
Hierarchy of Evidence for Treatment Decisions:
Greatest (Top) to Least (Bottom)
1.
2.
3.
4.
N of 1 randomized controlled trial
Systematic review of randomized trials*
Single randomized trial
Systematic review of observational studies
addressing patient-important outcomes
5. Single observational study addressing patientimportant outcomes
6. Physiological studies (studies of blood pressure,
cardiac output, exercise capacity, bone density,
and so forth)
7. Unsystematic clinical observations
35
*A meta-analysis is often considered higher than a
systematic review
(Guyatt and Rennie, 2002)
Hierarchy of Evidence
Ideally, evidence from individual studies would
be compiled or synthesized into systematic reviews,
with that information succinctly consolidated into
easily and quickly read synopses. All relevant
information would be integrated and linked to a
specific patient’s circumstance. The medical search
literature is still far from this, but working towards that
goal. Efforts include clinical prediction guidelines and
APTA’s emphasis on EBP.
36
(Straus et al, 2005)
Variables
Variables: Characteristic that can be manipulated or
observed
• Types of Variables
• Independent or Dependent
• Measurement Scales/Levels
Classification is useful for communication, so that
readers are aware of the author’s hypothesis of what
situation or intervention (independent variable) will
predict or cause a given outcome (dependent
variable)
37
(Portney and Watkins, 2000)
Variables: Independent or Dependent
• Independent Variable: A variable that is
manipulated or controlled by the researcher,
presumed to cause or determine another
(dependent) variable
• Dependent Variable: A response variable that is
assumed to depend on or be caused by another
(independent) variable
38
(Portney and Watkins, 2000)
Variables: Measurement Scales
• Useful to convey information to the reader about
the type of variables observed
• Necessary to determine what statistical analysis
approach should be used to examine relationships
between variables
• From lowest to highest level of measurement,
the scales are nominal, ordinal, interval, and ratio
39
(Portney and Watkins, 2000)
Variables: Measurement Scales
Nominal Scales (Classification Scale)
• Data, with no quantitative value, are organized
into categories
• Categorizes are based on some criterion
• Categories are mutually exclusive and
exhaustive (each piece of data will be assigned to
only one category)
• Only permissible mathematical operation is
counting (such as the number of items within each
category)
• Examples: Gender, Blood Type, Side of
40
Hemiplegic Involvement
(Portney and Watkins, 2000)
Variables: Measurement Scales
Ordinal Scales
• Data are organized into categories, which are
rank-ordered on the basis of a defined
characteristic or property
• Categories exhibit a “greater than-less than”
relationship with each other and intervals between
categories may not be consistent and may not be
known
41
(Portney and Watkins, 2000)
Variables: Measurement Scales
Ordinal Scales, continued
• If categories are labeled with a numerical value,
the number does not represent a quantity, but only
a relative position within a distribution (for
example, manual muscle test grades of 0-5)
• Not appropriate to use arithmetic operations
• Examples: Pain Scales, Reported Sensation,
Military Rank, Amount of Assistance Required
(Independent, Minimal…)
42
(Portney and Watkins, 2000)
Variables: Measurement Scales
Interval Scales
• Data are organized into categories, which are
rank-ordered with known and equal intervals
between units of measurement
• Not related to a true zero
• Data can be added or subtracted, but actual
quantities and ratios cannot be interpreted, due to
lack of a true zero
• Examples: Intelligence testing scores,
temperature in degrees centigrade or Fahrenheit,
calendar years in AD or BC
43
(Portney and Watkins, 2000)
Variables: Measurement Scales
Ratio Scales
• Interval score with an absolute zero point (so
negative numbers are not possible)
• All mathematical and statistical operations are
permissible
• Examples: time, distance, age, weight
44
(Portney and Watkins, 2000)
Variables: Clinical Example
A study investigates how a strengthening
program impacts a child’s ability to independently
walk. In this case, the strengthening program is the
independent variable and the ability to independently
walk is the dependent variable. Amount of
assistance required (if ranked maximal, moderate,
minimal, independently, not based on weight put on a
crutch or other quantitative testing) would be an
example of ordinal data.
Studies often have more than one independent or
dependent variable
45
Measurement Validity
• Measurement Validity examines the “extent to
which an instrument measures what it is intended to
measure” (Portney and Watkins, 2000)
• For example, how accurate
is a test or instrument at
discriminating, evaluating,
or predicting certain items?
46
Measurement Validity
Validity of Diagnostic Tests
• Based on the ability for a test to accurately
determine the presence or absence of a condition
• Compare the test’s results to known results,
such as a gold standard.
• For example, a test determining balance
difficulties likely to result in falls could be compared
against the number of falls an individual actually
experiences within a certain time frame. A clinical
test for a torn ACL could be compared against an
MRI.
47
(Portney and Watkins, 2000)
Measurement Validity: Types
• Face Validity: Examines if an instrument appears to
measure what it is supposed to measure (weakest
form of measurement validity)
• Content Validity: Examines if the items within an
instrument adequately comprise the entire content of
a given domain reported to be measured by the
instrument
• Construct Validity: Examines if an instrument can
measure an abstract concept
48
(Portney and Watkins, 2000)
Measurement Validity: Types
• Criterion-related Validity: Examines if the outcomes
of the instrument can be used as a substitute
measure for an established gold standard test.
 Concurrent Validity: Examination of Criterionrelated Validity, when the instrument being examined
and the gold standard are compared at the same time
 Predictive Validity: Examination of Criterion-related
Validity, when the outcome of the instrument being
examined can be used to predict a future outcome
determined by a gold standard
49
(Portney and Watkins, 2000)
Measurement Validity: Statistics
Ways to Evaluate Usefulness of Clinical Screening or
Diagnostic Tools
• Sensitivity and Specificity
• Positive and Negative Predictive Value
• Positive and Negative Likelihood Ratios
• Receiver Operating Characteristic (ROC) Curve
The above mentioned statistical procedures are often
used when researchers are introducing (and validating)
the test. Hopefully the values from these operations can
be found tool’s testing manual or in articles evaluating the
tool’s validity within certain populations or settings.
50
Measurement Validity: Statistics
Diagnostic Reference Table
Condition
Present
Absent
Test Result
Positive
Negative
True Positive
a
False Positive
b
False Negative
c
True Negative
d
(Guyatt and Rennie, 2002; Portney and Watkins, 2000;
Straus et al, 2005)
51
Measurement Validity: Statistics
Table Labels:
(a) True Positive: The subject matter has the
condition, and the test accurately identifies the
presence of the condition
(d) True Negative: The subject matter does not have
the condition, and the test accurately identifies the
absence of the condition
(b) False Positive: The subject matter does not have
the condition, and the test incorrectly identifies the
presence of the condition
(c) False Negative: The subject matter has the
condition, and the test incorrectly identifies the
52
absence of the condition
(Portney and Watkins, 2000)
Measurement Validity: Statistics
• Positive test result = the test identifies the condition
as being present;
• Negative result = the test identifies the condition as
being absent
(This may or may not be accurate when compared to
the gold standard).
The test’s sensitivity and specificity provide
information about the accuracy of the test.
53
(Portney and Watkins, 2000)
Measurement Validity: Statistics
Sensitivity
• The ability of the test to obtain a positive test when
the condition is present; the ability to detect a true
positive (a)
• a/(a + c) The proportion that test positive out of
those with the condition
54
(Portney and Watkins, 2000)
Measurement Validity: Statistics
Specificity
• The ability of the test to obtain a negative result
when the condition is absent, the ability to detect a
true negative (d)
• d/(b + d) The proportion that test negative out of
those without the condition
Sensitivity and Specificity are often provided in terms
of percents, from 0% to 100% (low to high)
55
(Portney and Watkins, 2000)
Measurement Validity: Statistics
Helpful Hints to remember
Sensitivity and Specificity
• Sensitivity: SnNout: When a test has a high
sensitivity (Sn), a negative result (N), rules out (out)
the diagnosis.
• Specificity: SpPin: When a test has a high
specificity (Sp), a positive result (P), rules in (in) the
diagnosis
56
(Straus et al, 2005)
Measurement Validity: Statistics
Clinical Example
Example:
You’re choosing between two tests that screen
participation in school based on physical abilities.
A positive result means that the student’s physical
abilities impact his or her participation.
57
Measurement Validity: Statistics
Clinical Example
One test has a high sensitivity, but a low specificity. A
high sensitivity means that a negative test will
effectively rule out students whose physical abilities
do not impact participation.
However, with a low specificity, there may be many
false positives, meaning students may test “positive”
for the condition when, in fact, their abilities do not
impact participation.
58
Measurement Validity: Statistics
Clinical Example
Example:
You’re choosing between two tests that evaluate
participation in school based on physical abilities.
A positive result means that the student’s physical
abilities impact his or her participation.
59
Measurement Validity: Statistics
Clinical Example
The other test has a low sensitivity, but a high
specificity. A high specificity means that a positive
result will effectively rule in the condition.
However, with a low sensitivity, there may be many
false negatives, meaning that students may test
“negative” for the condition, when, in fact, their
abilities do impact their participation.
60
Measurement Validity: Statistics
Predictive Values
• Provided in terms of percentages, 0% to 100%,
low to high
• Positive Predictive Value (PV+)
• Probability that a person with a positive test
actually has the condition
• a/(a + b)
• High PV+ desired for screening tools, to
prevent excessive unnecessary future testing
• Negative Predictive Value (PV-)
• Probability that a person with a negative
test does not have the condition
61
• d/(c + d)
(Portney and Watkins, 2000)
Measurement Validity: Statistics
Likelihood Ratios
• Calculated from the Diagnostic Reference Table
• Requires prior calculation of the pretest
probability of the condition in question
• Easy to use when familiar with the concept, but
requires the use of a probability guide chart or a
nomogram (chart that contains pretest probability
and likelihood ratios, with a ruler connecting those
two points to determine posttest probabilities)
62
(Guyatt and Rennie, 2002)
Measurement Validity: Statistics
Likelihood Ratios, continued
• Positive and negative likelihood ratios are
calculated
• Determines the usefulness of a diagnostic test.
If a positive or negative result will change the
posttest probability of having a condition to alter
the clinician and patient’s course of action, it will
be useful. If the likelihood ratios of the test do not
result in a substantial change of knowledge, the
test most likely will not be useful.
63
(Guyatt and Rennie, 2002)
Measurement Validity: Statistics
Receiver Operating Characteristic (ROC) Curve
• Uses sensitivity and specificity information to
find the probability of correctly choosing between
presence or absence of the condition
For example, a test with an area under the ROC
curve of 0.80, would result in the correct
determination of presence or absence of a
condition 80% of the time.
64
(Portney and Watkins, 2000)
Measurement Validity: Statistics
Responsiveness to chance statistics evaluate a
measurement tool’s ability to detect change over time
• For example, will administration of the test pre
and post intervention reflect a change in status, if
one actually occurred?
• Evaluated by examining the change in scores in
a pretest-posttest design, or using effect size
65
(Portney and Watkins, 2000)
Measurement Validity: Statistics
Effect Size
• Effect size (ES) is a measure of the amount of
difference.
For example, experimental group A increased their
score on a coordination measure by an average of
15 points, while experimental group B increased
their score an average of 8 points. The ES
between groups would be 7 points, considering
the groups were homogeneous at the start.
66
(Portney and Watkins, 2000)
Measurement Validity: Statistics
Effect Size, continued…
• An effect size index is a converted effect size, a
standardized measure of change, so that change
scores across different outcome measures can be
compared.
• ES is often displayed as a correlation
coefficient, r
Portney and Watkins note that considerations of ES
vary based on the interpreting clinician, but review
Cohen’s suggestions of scores <0.4 as small
(treatment had a small effect), 0.5 as moderate,
67
and >0.8 as large
(Portney and Watkins, 2000)
Measurement Validity: Statistics
Effect Size versus Minimal Clinically Important
Difference
• In addition to numerical data revealed and
statistical significance, the clinician should also
consider what amount of change is clinically
meaningful, such as, how great a gain in strength
or endurance will result in a change in function?
This is often referred to as the minimal clinically
important difference (MCID).
68
(Jaeschke et al 1989)
Measurement Reliability: Statistics
Reliability examines a measurement’s consistency
and freedom from error
• Can be thought of as reproducibility or
dependability
• Estimate of how observed scores vary from the
actual scores
69
(Portney and Watkins, 2000)
Measurement Reliability: Statistics
Reliability Coefficient
• Ratio of reliability (many different types with
various symbol representation)
• Range between 0.00 to 1.00;
• 0.00 = no reliability;
• 1.00 = perfect reliability
• Reflection of variance, a measure of the
differences among scores within a sample
70
(Portney and Watkins, 2000)
Measurement Reliability: Statistics
Correlation
• Comparison of the degree of association
between two variables or sets of data
• Used as a basis for many reliability coefficients
71
(Portney and Watkins, 2000)
Measurement Reliability: Statistics
Test-Retest Reliability
Examines the consistency of the results of
repeated test administration
• Traditional Analysis
• Pearson product-moment coefficient of
correlation (for interval or ratio data)
• Spearman rho (for ordinal data)
• Current, sometimes considered preferred,
Analysis
• Intraclass correlation coefficient
72
(Portney and Watkins, 2000)
Measurement Reliability: Statistics
Rater Reliability
• Intrarater reliability
• Reliability of data collection from one
individual over two or more trials
• Interrater reliability
• Reliability of data collection between two or
more raters
73
(Portney and Watkins, 2000)
Descriptive Statistics
Descriptive statistics are used to describe sample
characteristics.
• A sample is a subset of a population chosen for
study. Since data often cannot be collected from an
entire population, the data chosen from a selected
sample is intended to be representative or an
estimate of the population data.
• Distribution: Total set of scores (from a sample) for
a particular variable, given the symbol, n
74
(Portney and Watkins, 2000)
Descriptive Statistics
Frequency and Shapes of Distribution
• Frequency distribution: The number of times
each value, from the variable data, occurred
• Drawing a graph of frequency distributions can
result in shapes that characterize the distributions
• Some graphs are asymmetrical, others are
symmetrical
• A symmetrical graph with a bell-shaped
distribution is referred to as a normal
distribution.
• A skewed distribution presents asymmetrically
75
(Portney and Watkins, 2000)
Descriptive Statistics
Normal Distributions, when graphed according
to frequency, present in the shape of a bell with the
majority of scores falling in the middle and
progressively fewer scores at either end. It has
special properties in statistics.
76
(Portney and Watkins, 2000)
Descriptive Statistics:
Central Tendency Measures
Used to quantitatively summarize a group’s
characteristics.
• Mode: The score that occurs most frequently
• Median: The middle score in numerically
ordered group of data. If there are an even
number of scores, the median is the midpoint
between the two middle scores
• Mean: The sum of a set of scores divided by the
number of scores, n. Often referred to as the
“average” of a data set.
77
(Portney and Watkins, 2000)
Descriptive Statistics:
Measures of Variability
Variability is the dispersion of scores.
Variability is affected (qualified) by five
characteristics:
• Range
• Percentiles
• Variance
• Standard deviation
• Coefficient of variation
78
(Portney and Watkins, 2000)
Descriptive Statistics:
Measures of Variability
Variability, continued…
Range: Difference between the highest and lowest
scores in a distribution
Percentiles: Used to describe a score’s position
within a distribution, distribution data is often divided
into quartiles, or four equal parts
Variance: (Too in-depth to describe the statistical
background for this module purpose). Reflects
variation within a set of scores, in square units.
Symbolized in sample data by s
79
(Portney and Watkins, 2000)
Descriptive Statistics:
Measures of Variability
Variability, continued
Standard Deviation: Representative of the variability
of scores surrounding the mean, in original units of
measurement. Square root of variance.
• The larger the standard deviation, the more
spread out the variable’s scores are around a mean.
• For example, data set (A) 8,9,10,11,12 and the
data set (B) 4,5,10,15,16 both have a mean of 10.
However the standard deviation for set A is 1.58
while the standard deviation of set B is 5.52.
• Symbolized in sample data by s2.
80
(Portney and Watkins, 2000)
Descriptive Statistics:
Measures of Variability
Variability, continued…
Coefficient of Variation: The ratio of the standard
deviation to the mean.
81
(Portney and Watkins, 2000)
Descriptive Statistics:
Distributions
Normal Versus Skewed (Non-normal) Distribution
• These theoretical shapes of distribution help
determine what statistical formulas or measures
should be used
• The characteristics of normally distributed data are
constant and predictable. For statistical purposes, the
normally distributed curve is often divided into
proportional areas, each equal to one standard
deviation.
• Data should be examined for “goodness-of-fit” to
see if the sample approximates the normal
distribution.
(Portney
82
and Watkins, 2000)
Descriptive Statistics:
Distributions
Normal Distribution Statistics
• 1st standard deviation on either side of the average
contains 34.13% of the data
• total of 68.62% of the data will be between +1
and -1 standard deviation
• Between 1st and 2nd standard deviation contains
13.59% of the data
• total of 95.45% of the data will be between +2
and -2 standard deviations
[(13.59 times 2) + (34.13 times 2)]
83
(Portney and Watkins, 2000)
Descriptive Statistics:
Distributions
Normal Distribution Statistics, continued
Between 2nd and 3rd standard deviation contains
2.14% of the data
• total of 99.73% of the data will be between +3
and -3 standard deviations
[(13.59 times 2) + (34.13 times 2) + (2.14 times
2)]
84
(Portney
and Watkins, 2000)
Inferential Statistics
Estimate population characteristics from sample data
• Used often when testing theories about the effects
of experimental treatments
• Requires that assumptions are made about how
well the sample represents the larger population
• Assumptions are based on the statistical concepts
of probability and sampling error
• It is important the sample be representative of the
population, so that the results of interventions on
samples can be applied to the entire population of
individuals with those characteristics.
85
(Portney
and Watkins, 2000)
Inferential Statistics
Probability
Probability
• The likelihood that an event will occur, given all
the possible outcomes. Used often in prediction.
Given the symbol p
• Probability may range from p = 1 (certain the
event will occur, 100% probability) to p = 0 (certain
that the event will not occur, 0% probability)
86
(Portney and Watkins, 2000)
Inferential Statistics
Probability
Probability, continued…
Reflective of should happen in the long run, not
necessarily what will happen on a given trial.
For example, if a treatment has a 60% chance of
success, then 60% of people will likely be
successfully treated. That does not mean the
treatment will be 60% successful in an individual,
likely it will either be a unsuccessful or successful
for that individual.
87
(Portney and Watkins, 2000)
Inferential Statistics: Probability
Clinical Example
Probability statistics can be applied to the distribution
of scores
Example, for a normally distributed data set:
Average long jump for a certain group of children is
35 inches with a standard deviation of 4 inches.
Suppose you want to know the probability that a child will
jump within one standard deviation (from 31 to 39
inches)? You know that within one standard deviation on
either side of the mean is 68.2%, so that is the probability
that a child will jump within one standard deviation of the
mean.
88
Inferential Statistics: Probability
Clinical Example
Example, continued…
If you wanted to know the probability that a
child will jump more than one standard deviation
above the mean (greater than 39 inches), you can
refer to the data and calculate 15.86%.
Charts and graphs are available to calculate the data
in between the standard deviations.
89
Inferential Statistics:
Sampling Error
Sampling Error
The difference between sample values and
population values
• The lower the sampling error, the greater the
confidence that the sample values are similar to
the population values.
• To estimate sampling error, the standard error of
the mean statistic is used (too complex to explain
statistical basis in this format)
• The larger the sample size, n, the smaller the
standard error of the mean
90
(Portney and Watkins, 2000)
Inferential Statistics
Confidence Intervals
Confidence Interval (CI)
• Range of scores with specific boundaries, or
confidence limits, that should contain the
population mean
• The boundaries of CIs are based on the sample
mean and its standard error
• Degree of confidence is expressed as a
percentage
• Often, researchers use 95% as a boundary,
which is just slightly less than 2 standard
deviations on either side of the mean
91
(Portney and Watkins, 2000)
Inferential Statistics, Confidence Interval
Clinical Example
A physical therapy treatment program resulted in
the ability for 40 children with a certain disorder to walk
an additional 8 independent steps, on average, within a
certain set of parameters.
The 95% CI for this data was ± 2 steps.
Therefore, we can be 95% certain that the population
mean, or average for all children with this disorder, is
between 6 and 10 extra independent steps.
Said another way, if there were an additional 100 children
with the same condition, 95 of them would likely have an
average that was between an additional 6 to 10
92
additional independent steps following the physical
therapy treatment.
Inferential Statistics
Hypothesis Testing
Hypothesis Testing
Used to decide if effects from an intervention are due
to chance or the result of the intervention
Results in one of two decisions: To accept or
reject the null hypothesis
93
(Portney and Watkins, 2000)
Inferential Statistics
Hypothesis Testing
Hypothesis Testing, continued…
Statistical Hypothesis (also known as the null
hypothesis):
Any observed difference (as in pretreatment to posttreatment or treatment compared to a placebo) is due
to chance.
When the null hypothesis is rejected, the researcher
concludes that the effect of treatment is not likely due
to chance.
94
(Portney and Watkins, 2000)
Inferential Statistics
Hypothesis Testing
Hypothesis Testing, continued…
Alternative Hypothesis: Any observed difference is
not due to chance.
Often the researcher is trying to support the
alternative hypothesis, as when trying to prove that
one particular treatment is better.
Sometimes, however, the researcher may be trying
to prove that certain interventions are equal.
95
(Portney and Watkins 2000)
Inferential Statistics
Hypothesis Testing, Errors
Errors in Hypothesis Testing
• Decision to accept or reject the null hypothesis
is based on the results of the statistical
procedures on collected data from samples.
• Decisions are based on sample data, so it is
possible that the results obtained are not accurate
of population data.
• There is a chance for error, that the researcher
may incorrectly accept or reject the null
hypothesis
96
(Portney and Watkins 2000)
Inferential Statistics
Hypothesis Testing, Errors
Type I error (α): Rejecting the null hypothesis when it
is true (for example, deciding that the difference seen
between a treatment group and a control group is
due to the effect of the treatment, when, in fact, the
difference is due to chance).
•A commonly used standard is α= 0.05, or the
researchers accept a 5% chance of making a
Type I error
Statistical tests completed with the sample data are
used to calculate p, the probability that an observed
difference did occur by chance.
97
(Portney and Watkins, 2000)
Inferential Statistics
Hypothesis Testing, p and α
Hypothesis Testing: Relationship between p and α
If p is greater than the chosen α, then the
researchers chose not to reject the null hypothesis
For example, in a placebo versus treatment study,
the researchers cannot conclude that the
experimental treatment had a different effect then the
placebo.
98
(Portney and Watkins, 2000)
Inferential Statistics
Hypothesis Testing, p and α
Hypothesis Testing: Relationship between p and α,
continued…
If p is less than the chosen α, then the researchers
chose to reject the null hypothesis
For example, in a placebo versus treatment study,
the researchers conclude that the experimental
treatment had a different effect then the placebo.
99
(Portney and Watkins, 2000)
Inferential Statistics
Hypothesis Testing, p and α
Hypothesis Testing: Relationship between p and α,
continued…
Confidence intervals surrounding the p value can be
calculated, hopefully these are included in data
analysis section of the research study.
The CIs between two groups should not (rare
exceptions) overlap, if a statistically significant
difference is found.
100
(Portney and Watkins, 2000)
Inferential Statistics
Hypothesis, CIs versus MCID
Hypothesis Testing : CIs (slide 90) versus MCID
(slide 67)
When the null hypothesis is rejected, the researchers
conclude that the experimental treatment (versus a
placebo, for example) had a statistically significant
effect.
The clinician should examine the effect size and the
MCID to ensure that the change is clinically and
functionally relevant.
101
(Guyatt and Rennie, 2002)
Inferential Statistics
Hypothesis, CIs versus MCID
When a null hypothesis is not rejected, the
researchers may conclude that the experimental
treatment did not have an effect.
However, the researchers should pay close attention
to the confidence intervals.
• If the CI does not include the MCID, then the trial
is most likely negative.
• If the CI includes the MCID, then the possibility
that the experimental treatment may have a
positive effect cannot be ruled out. The
researchers may wish to run a power analysis,
102
explained in later slides.
(Guyatt and Rennie, 2002)
Inferential Statistics, Hypothesis Testing
Clinical Example
Children with similar abilities/diagnosis
(homogenous sample) are randomly assigned to two
different groups:
• Group A receives a physical therapy designed to
improve gross motor skills, and
• Group B completes typical daily activities (but don’t
worry, due to ethical concerns children in Group B will
receive the same treatment as those in Group A at the
end of the study period).
103
Inferential Statistics, Hypothesis Testing
Clinical Example 1, continued…
The outcome measure will be a tool that tests gross
motor activities with a final, numerical outcome
score.
The authors hypothesis that the experimental group
will show statistically significant gains (experimental
hypothesis) and that the null hypothesis (there is no
difference between groups at the end of the
intervention period) will be rejected.
104
Inferential Statistics, Hypothesis Testing
Clinical Example 1, continued…
Initially, group A and B have similar average pre-test
scores (if not, that can be statistically corrected).
Now suppose that Group A (experimental group,
receiving additional PT services) increases an
average of 9 points from pretest to posttest, with a CI
of ±1, 8 to 10.
Group B increases an average of 0.5 points, with a
CI of ±2, -1.5 to 3.5.
105
Inferential Statistics, Hypothesis Testing
Clinical Example 1, continued…
These confidence intervals do not overlap, which
corresponds with the statistical analysis performed
that p<0.05 (the predetermined α), and there is a
statistically significant difference between groups.
The null hypothesis is rejected.
106
Inferential Statistics, Hypothesis Testing
Clinical Example 1, continued…
Prior to the experiment, the authors agreed that the
MCID was 6.
Both the statistically determined (p<0.05 ) and
clinically determined criteria in support of the
experimental group’s treatment were met (ES of
experimental group,9, was greater than the MCID,
6).
107
Inferential Statistics, Hypothesis Testing
Clinical Example 2
Initially, group A and B have similar average pre-test
scores (if not, that can be statistically corrected).
Now, change the scenario from previous slides.
Suppose that Group A (experimental group,
receiving additional PT services) increases an
average of 5 points from pretest to posttest, with a CI
of ±2, 3 to 7.
Group B increases an average of 0.5 point,s CI ±2, 1.5 to 3.5.
108
Inferential Statistics, Hypothesis Testing
Clinical Example 2, continued…
These confidence intervals overlap, and the
statistical analysis performed revealed that p>0.05
(the predetermined α). There is not a statistically
significant difference between groups.
The null hypothesis is not rejected.
109
Inferential Statistics, Hypothesis Testing
Clinical Example 2, continued…
However, prior to the experiment, the authors agreed
that the MCID was 6.
Since the CI range of Group A includes that MCID,
the beneficial effects of the treatment from Group A
over Group B cannot be ruled out.
The researchers may want to run a statistical power
analysis, to determine if the study was
underpowered, and therefore unlikely to show a
difference of the treatment.
110
Inferential Statistics
Hypothesis Testing, Errors
Errors in Hypothesis Testing, Type II
Type II error (β): Not rejecting the null hypothesis
when it is false (for example, determining that
differences are due to chance, when they are, in fact,
due to the experimental treatment)
111
(Portney and Watkins, 2000)
Inferential Statistics
Hypothesis Testing, Errors
Errors in Hypothesis Testing, Type II
The complement of Type II error is the statistical
power of the test (1-β)
• Power: the probability that a test will lead to
rejection of the null hypothesis, or the
probability of obtaining statistical significance if
the differences are not due to chance
• Many researchers use a standard of β=0.20,
or a power of 80%, as reasonable protection
against Type II error
112
(Portney and Watkins,2000)
Inferential Statistics: Power
Determinants of Statistical Power
Even though the researchers may not reject the null
hypothesis, it does not always mean that an
experimental treatment is not
effective.
The power of the study may have
been too low or small to detect a
significant difference.
113
(Guyatt and Rennie 2002, Portney and Watkins, 2000)
Inferential Statistics: Power
Determinants of Statistical Power
• Levels set of α and β
• Variance
• Sample Size (n)
• Effect Size (Difference between two treatments
or variables, Treatments with large changes or
correlations are more likely to produce significant
outcomes)
Increases of effect size, sample size, and alpha all
increase power, while decreases in variance
increases power.
114
(Portney and Watkins, 2000)
Inferential Statistics: Tests
Parametric Vs. Nonparametric Tests
Parametric statistics are statistics used to estimate
population parameters
• Primary assumptions of parametric tests:
• Samples randomly drawn from populations
with normal distributions
• Variances in samples are roughly equal
• Data are measured on interval or ratio scales
115
(Portney and Watkins, 2000)
Inferential Statistics
Tests
Parametric Vs. Nonparametric Tests
Nonparametric tests are generally not as powerful,
and researchers may choose to use parametric tests,
despite not meeting all generally held assumptions
(such as use of parametric test in a study with ordinal
data)
116
(Portney and Watkins, 2000)
Section 3 Outline
How to Search for EBP
• Formation of Clinical Questions used to search
for EBP, 117-121
• Evidence Search: Sources of Information, 122138
• Internet/World Wide Web
• Textbooks
• Specific Journal Subscriptions
• Internet Sources for Medical Information
• The Guide
• Search Strategies, 139-152
117
Formation of Clinical Questions
Used to Search for EBP
Background Questions: general knowledge about a
condition/area of interest
Foreground Questions: specific knowledge used to
inform clinical decisions or actions. Often what
researches use when investigating a particular
treatment, prognosis, outcome measure, etc.
• Clinicians require the use of background and
foreground knowledge, with proportions that may
vary over time and depend on knowledge of a
particular situation.
118
(Straus et al, 2005)
Formation of Clinical Questions
Used to Search for EBP
Background Questions
Usually consist of two components:
• A question root (who, what, where, when,
why, how) with a verb
• A condition, test, treatment, or other health
care concern
Example: “What is juvenile rheumatoid arthritis?” or
“What causes cerebral palsy?
119
(Straus et al, 2005)
Formation of Clinical Questions
Used to Search for EBP
Foreground Questions
Usually consist of four components (PICO)
• Patient and/or problem
• Intervention (or exposure)
• Comparison, if relevant
• Clinical Outcomes, including time if relevant
120
(Straus et al, 2005)
Formation of Clinical Questions
Used to Search for EBP
Foreground Questions, continued…
• Example, “For teenage girls with juvenile
rheumatoid arthritis, is treatment X or treatment Y
more effective in decreasing pain in joints of the
hand and wrist?” or “For children with spastic
diplegic cerebral palsy age 3-6 years old, which
outcome measure (can compare A or B or search
in general) most accurately assesses functional
mobility?”
121
Formation of Clinical Questions
Used to Search for EBP
Question Type: Categorization is useful for writing
the question and statistical analysis
• Therapy: Evaluates the effects of various
treatments or interventions
• Harm (not frequently investigate in PT literature):
Evaluates the effects of various treatments or
modalities on function, morbidity, mortality
• Diagnosis: Evaluates a tool or test’s ability to
distinguish among certain conditions
• Prognosis: Evaluates the course of a certain
condition
122
(Guyatt
and Rennie, 2002)
Evidence Search: Sources of Information
Where to find the research evidence, once the
foreground question is developed?
•
•
•
•
Internet/World Wide Web
Textbooks
Specific Journal Subscriptions
Internet Sources for Medical Information
123
Evidence Search: Sources of Information
Internet/World Wide Web:
This category is related to general search engines,
not specific medical information journal search
sources or health-related texts and journals available
on the web.
• Provides a rapid and abundant source of
information
• Be careful, not all sites and sources of
information meet EBP guidelines
124
(Guyatt and Rennie, 2002)
Evidence Search: Sources of Information
Internet/World Wide Web, continued…
Consider the reputability of the source and the
information found.
Examples:
• MD Consult (fee),
• Google Scholar (free for the search)
• Some articles found may be free, others may
have a fee.
• I personally have found Google Scholar very
helpful. Its own search page is displayed when
you enter “Google Scholar” into Google,
125
(Guyatt and Rennie, 2002)
Evidence Search: Sources of Information
Textbooks:
• Often more useful for answering general
background questions
• Books in print (hard-copy) may exclude valuable
recent information, due to the time taken to
compile, edit, and publish texts
• If utilizing a text, use one that is updated
frequently and well referenced, so you can access
more details if necessary
126
(Guyatt and Rennie, 2002)
Evidence Search: Sources of Information
Textbooks, continued…
Many texts that are internet-based are becoming
more evidence-based and may be an increasing
source of information for foreground questions in the
future.
Examples of internet-based texts (Guyatt and Rennie
recommend UpToDate and Clinical Evidence as
becoming more evidence-based):
• emedicine (free)
• UpToDate (fee)
• Clinical Evidence (fee)
127
(Guyatt and Rennie, 2002)
Evidence Search: Sources of Information
Specific Journal Subscriptions:
• Browsing a specific journal may be beneficial if
specific to your field of practice
• However, it mostly likely will leave out many
applicable articles published in other journals
• Browsing through full text journals to find a
specific article with appropriate quality and
relevance often takes considerable time
128
(Straus et al, 2005)
Evidence Search: Sources of Information
Commonly Used Internet Sources for Medical
Information: Prefiltered Resources, Medline, PEDro,
EBSCOhost/CINHAL
These internet-accessible search engines allow you
enter search terms to find applicable research, most
often related to published journals.
Most cites will provide an abstract (general summary)
of any article found for free, but often a fee is
charged for full text access unless you have access
to university subscriptions. Many PT programs offer
access to their resources for their Clinical
129
Instructors.
Evidence Search: Sources of Information
Prefiltered Resources:
Contain only those studies, most often systematic
reviews, considered high methodological quality.
• Benefits: Easy to search small database with
high quality studies. Drawbacks: not
comprehensive
• Examples: Best Evidence, Cochrane Library
(both require a subscription fee)
130
(Guyatt and Rennie, 2002)
Evidence Search: Sources of Information
Medline:
• Online search engine with comprehensive
coverage of medical journals, can be intimidating
due to large size and knowledge necessary to
perform an efficient search.
• Unfiltered, containing information from
prefiltered sources (such as the Cochrane Library)
plus additional expansive coverage from a wide
variety of types of studies and journal publications
131
(Guyatt and Rennie 2002)
Evidence Search: Sources of Information
Medline, continued …
• Medline, the search engine, not necessarily the
articles within Medline, is available free to the
public through PubMed
(www.ncbi.nlm.nih.gov/PubMed).
• Commercial vendor access (most often through
university or health science center subscription):
OVID, Knowledge Finder, Silver Platter
• Articles through PubMed Central,
http://www.pubmedcentral.nih.gov, are free.
132
(Guyatt and Rennie, 2002)
Evidence Search: Sources of Information
PEDro (Physiotherapy Evidence Database):
Per the website, “[PEDro] has been developed to
give rapid access to bibliographic details and
abstracts of randomized controlled trials, systematic
reviews and evidence-based clinical practice
guidelines in physiotherapy.”
(http://www.pedro.org.au/index.html)
• Not-for-profit organization supported by many
national physiotherapy associations, including the
APTA
133
Evidence Search: Sources of Information
PEDro (Physiotherapy Evidence Database):
• Helpful Aspects of PEDro: systematic reviews
are listed first, followed by articles ranked
according to EBM hierarchy (with the scores
explained based on the studies methodological
characteristics)
• Free access at:
http://www.pedro.org.au/index.html, no direct fulltext articles. When applicable, the website
provides potential links to full-text articles
134
Evidence Search: Sources of Information
EBSCOhost:
• Provider of CINAHL® (Cumulative Index to
Nursing and Allied Health Literature) – CINAHL ®
also available through several other sources
• Fee for use (Often as a university or health
science subscription): http://www.ebscohost.com
• Similar search strategies to other medical
information sources
135
Evidence Search: Sources of Information
APTA’s Open Door:
• Available with APTA membership (login to
www.apta.org, look under areas of interest,
research subheading)
• Access to research journal collections including
ProQuest, Medline, Cochrane Library, CINAHL®
(mostly contains bibliographic records, not full text
journals)
136
Evidence Search: Sources of Information
APTA’s Open Door, continued…
• Suggestions, with links, for free full text
(including BioMed Central, Directory of Open
Access Journals, Public Library of Science,
PubMed Central, and more)
• Tutorials and searching tips for EBP, along with
guidelines for finding full-text articles
• Information on PT journals and samplings of
current research in specific fields
137
Evidence Search: Sources of Information
APTA’s Hooked on Evidence:
• Available with APTA membership (login in to
www.apta.org, look under areas of interest,
research subheading)
• Database of article extractions relevant to
physical therapists: Peer reviewed information
about the methodological quality and level
evidence of the included articles
138
Evidence Search: Sources of Information
The Guide to Physical Therapist Practice (Text):
• Provides PTs with comprehensive descriptions
of scope of practice
• Details preferred practice patterns
• Indications for specific tests and measures and
interventions
APTA. Guide to Physical Therapist Practice. 2nd ed. American
Physical Therapy Association; 2001.
139
Search Strategies
Searching for the Evidence within
Medical Information Sources:
It is beyond the scope of this module to provide indepth training on searching for literature.
There are many online training courses, courses
offered at health science center libraries, and thick
packets of information on such searches.
140
Search Strategies
Searching for the Evidence within Medical
Information Sources, continued…
It is not included in the time estimate or quiz
information, but if interested in performing more
searches and you would like greater instruction,
consider a course or an online tutorial. Many on the
internet are free, as the one by PubMed,
http://www.nlm.nih.gov/bsd/disted/pubmed.html
Efficiency in searching requires time spent learning,
experimenting, and practicing!
141
Search Strategies
PubMed Search
(http://www.ncbi.nlm.nih.gov/sites/entrez)
As a brief, non-inclusive overview, the following
slides provide some basic information on searches.
PubMed Database was chosen as an example since
it is a free and comprehensive site. This is not
intended to result in the ability to perform efficient
searches, but rather to give an example, the very tip
of the iceberg, of the variety of functions that can be
performed on such a database.
142
Search Strategies
PubMed Search:
The first step in searching is to form your foreground
question and consider which information is most
important to search
• Consider the specificity of your search. There is
less information on pediatrics than adults, less
information on rare disorders than common ones,
etc.
• If you have too many results, try to be more
specific. If too few, be more general.
143
Search Strategies
PubMed Search
• Go to the PubMed website
(http://www.ncbi.nlm.nih.gov/sites/entrez) and try
entering a basic, general term of interest into the
open field: Search, PubMed for (open field – enter
term here)
• Consider the results of your search: Was there
too little or too much information, was it relevant to
what your interest, do any of the articles appear
helpful to you?
144
Search Strategies
PubMed Search
• Limits: You can limit the responses in several
categories. On the results page or on the search
page, there is a row of tabs close to the top of the
page. Click on “Limits,” the tab farthest to the left.
This allows the searcher to limit the results in
certain categories such as English language, for
studies specifically on children, type of study, etc.
145
Search Strategies
PubMed Search
Basic operators commonly included in search terms:
• * (truncation symbol) : Retrieves various
spellings; Example: the term strength* will
retrieve articles related to strength,
strengthening, strengthen, etc
• OR : Expands retrieval of various topics;
Example: Spina Bifida OR Meningocele will
return articles listed under either heading
146
Search Strategies
PubMed Search
Basic operators commonly included in search terms,
continued…
• AND : Limits search; Example: Cerebral
Palsy AND Strength* will only return articles
that are related to both terms (for example,
many of those retrieved articles will have a
major or minor focus on strengthening
programs in individuals with CP)
147
Search Strategies
PubMed Search
Basic operators commonly included in search terms,
continued…
• [tw] (text word) : Retrieves articles with a
specific word in the title or abstract; Example:
[tw]orthotic will retrieve articles with orthotic in
the title or possibly the abstract. Be careful
though, it might leave out articles with terms
such as AFOs, DAFOs, orthoses, etc. That is
when MeSH headings (slide 149) are useful.
148
Search Strategies
PubMed Search
Basic operators commonly included in search terms,
continued…
There are many additional terms/commands
that can be added to a basic search, which again is
beyond the scope of this module. This is only to
make you aware that terms can be added to
designate a type of study (such as treatment,
prognosis, diagnosis, etc), methodological qualities
of the study (specificity, sensitivity, validity, reliability,
responsiveness to change, etc), and many more. 149
Search Strategies
PubMed Search, MeSH
MeSH (Medical Subject Heading) terms provide a
way to search for a specific topic that may use
different terminology for the same concept. Tutorials
are available on PubMed.
• MeSH database: Try this search! On the main
PubMed page, look at the blue column on the left.
Under “PubMed Services” click on “MeSH
Database”
150
Search Strategies
PubMed Search, MeSH
At the top of the page, try a search in the open field.
“Search MeSH for” (open field – enter topic of your
choice)
• Checking on the numbered box with your
subheading will allow the search to include all related
“Entry Terms” listed below the topic.
•To search for this in PubMed, click on the tab towards
the top of the page next to “Display” currently with
“Summary” or “Full” in the field.
•Scroll down. “PubMed Links” will include a general,
more comprehensive search, while “PubMed –
151
Major Topic Links” will include articles more
specific to your topic.
Search Strategies
PubMed Search, Clinical Queries
Clinical Queries database: Try this search! On the
main PubMed page, look at the blue column on the
left. Under “PubMed Services” click on “Clinical
Queries” Database.
This page provides specialized searches for
clinicians, the search can be performed for a specific
topic, related to a type of study. For example, the
searcher could investigate cerebral palsy, related to
etiology, diagnosis, therapy, prognosis, or clinical
152
prediction guides.
Search Strategies
PubMed Search, Clinical Queries
Clinical Queries database, continued…
It is also a place to search for systematic reviews
(Don’t be disappointed if you cannot find very many
for rare conditions, especially in pediatrics. More
studies are being done, but there have not been as
many performed yet in those fields)
You can perform a narrow, specific search or a
board sensitive search.
153
Pick out a topic, category, and scope, and try one!
Section 4 Outline
How to Interpret Research Related to EBP
• Interpretation of Evidence, 154
• For therapy/intervention, 155-166
• For diagnosis, 167-174
154
Interpretation of Evidence
• Consider the type of study and where the study
falls on the hierarchy of evidence.
• When evaluating the study, ask three questions:
Are the Results Valid, What Are the Results, How
Can I Apply the Results to Patient Care?
• Don’t forget to incorporate clinical expertise and
patient values and preferences after evaluating any
research material.
• As examples, the following slides will detail
interpretation of evidence for therapy and diagnosis
studies.
155
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Validity
Are the results valid? (Slides 148 to 156)
• Did the experimental (or treatment) groups
begin the study with a similar prognosis?
If not, the end results may be due to differences in
the groups versus differences due to intervention
• Were patients randomized into the groups?
If the participants were not randomly assigned to a
group, this decreases the validity of the study, as
one group’s conditions may have impacted the
156
outcomes.
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Validity
Was randomization concealed (blinded or masked)?
If randomization was not concealed, this decreases
the validity of the study. Those enrolling the
participants may choose, unintentionally or
intentionally, certain participants for certain groups if
randomization is not concealed.
157
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Validity
Were patients analyzed in the groups to which they
were randomized?
It may be tempting to move patients between groups.
For example, if a patient in an experimental group did
not participate in the strengthening activity or take
the medication prescribed for that group, it may be
tempting to move that person to the control group.
However, any cross-over or dropping out decreases
the validity of the results.
158
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Validity
Were patients analyzed in the groups to which they
were randomized?
Reasons for non-compliance or failure to finish
treatment may be related to prognosis, and switching
groups negates the unbiased comparison of
randomization.
Researchers are encouraged to follow an “intentionto-treat” analysis where all participants are analyzed
in groups to which they were originally assigned. If
drop-outs occurred, statistical methods are present to
159
compensate.
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Validity
Were patients in the treatment group and control
groups similar with respect to known prognostic
factors?
If not, this decreases the validity of the study, as
groups’ prognostic differences may impact the
outcome.
Statistical testing can determine if groups are similar,
and can also compensate for potential differences in
final analyses. The magnitude of initial differences
should be examined.
160
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Validity
Did experimental and control groups retain a similar
prognosis after the study started?
Were patients aware of group allocation? If so, this
decreases the validity of the study. Often in PT
research, it is difficult to blind a participant to the fact
that they received a certain treatment.
However, patients’ perceptions of received treatment
may impact results and responses.
161
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Validity
Were clinicians aware of group allocation?
If so, this decreases the validity of the study. The
researchers may treat patients in different groups
differently, aside from the treatment, such as
spending more time with a certain participants.
This may bias the results.
162
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Validity
Were outcome assessors aware of group allocation?
If so, this decreases the validity of the study.
If the researcher knows the group assignment of the
enrolled participant, it may unintentionally impact the
results, especially if the outcome measures are more
subjective.
163
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Validity
Was follow-up complete?
If not, this decreases the validity of the study.
When there is loss of follow-up, the researcher may
consider a “worse-case” scenario, such as the
experimental treatment that was supposed to
improve function decreased function while those in
the control group improved.
Statistical analysis can determine how results may
have been different.
164
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Results
What are the results?
• How large was the treatment effect? Usually
involves examination of statistical analysis. Was
there a statistically significant difference? How
large was that difference or effect size?
• How precise was the estimate of the treatment
effect? What do the statistics, such as confidence
intervals and minimal clinically important
differences, reveal about the treatment effect?
(Refer to statistics knowledge for these questions)
165
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Application
How can I apply the results to patient care?
• Were the study patients similar to my patient?
Would your patient have qualified to be included in
the study? Can you extrapolate any useful
information if the patient would not have qualified?
• Were all clinically important outcomes
considered? Are the outcomes of interest areas of
importance in your practice?
166
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Therapy/Intervention: Application
Are the likely treatment benefits worth the potential
harm and cost?
The clinician and patient would need to contemplate
the impact of any potentially adverse effects, as well
as the amount of time, effort, and additional costs of
the treatment.
167
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Diagnosis
This information does not apply only to a medical
diagnoses, it can also be used with physical therapy
tests and measures that evaluate a multitude of
conditions.
For example, tools that evaluate: risk for falls,
developmentally appropriate gross motor skills,
sensory processing abilities impact on performance
168
Interpretation of Evidence
For Diagnosis: Validity
Are the Results Valid? (slides 168-171)
Did clinicians face diagnostic uncertainty? If not, this
decreases the validity of the results. The tool is only
helpful if it can distinguish among those with
uncertainty in diagnosis.
For example, if testing a tool determining fall risk, it
would not be beneficial to study the tool with young
adults with no history of falls and no risks for falls.
The tool would need to be investigated in individuals
with the potential to be at risk for falls.
169
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Diagnosis: Validity
Was there a blind comparison with an independent
gold standard applied similarly to the treatment group
and the control group?
If not, this decreases the validity of the results.
When the comparison is not blinded, there is a
potential for bias, especially if the tools contain more
subjective material. The gold standard, also known
as a reference standard, should be separate from the
test being evaluated.
170
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Diagnosis: Validity
Was there a blind comparison with an independent
gold standard applied similarly to the treatment group
and the control group? continued…
If two different test batteries are being compared, for
example, they should not contain the same items.
Finally, it is important to consider the reputability of
the gold standard. In some cases, there might not be
a universally accepted gold standard of comparison.
The clinician should consider if the reference
standard is acceptable.
171
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Diagnosis: Validity
Did the results of the test being evaluated influence
the decision to perform the reference standard?
If so, this decreases the validity of the results. The
reference standard should be applied in every case,
regardless of the results of the test being evaluated.
172
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Diagnosis: Results
What are the results?
Consider the statistical and clinical implications
What likelihood ratios were associated with the range
of possible test results?
The researcher could also examine other related
statistics, such as sensitivity and specificity, predictive
values, likelihood ratios, and the ROC curve. The
researcher can evaluate the psychometric properties
of the diagnostic test: its validity, reliability, and
responsiveness to change (as discussed in
173
previous slides).
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Diagnosis: Application
How can I apply the results to patient care?
Will the reproducibility of the test result and its
interpretation be satisfactory in my clinical setting? It
is important that the test produce reliable results that
can be effectively interpreted.
Are the results applicable to the patient in my
practice? How similar are the patients in the study to
your patient?
174
(Guyatt and Rennie, 2002)
Interpretation of Evidence
For Diagnosis: Application
How can I apply the results to patient care?
Will the results change my management strategy?
Will your treatment of the patient change based on
the test results?
Will patients be better off as a result of the test?
What are the potential risks versus benefits of the
test?
175
(Guyatt and Rennie, 2002)
Thank You!
Please mark your time of completion for this module
and complete the posttest (same as the pretest, just
print another copy).
You are free also to look back to the slides to answer
questions in the posttest.
However, if you are unable to answer any question,
even upon review of the PowerPoint module, “I don’t
know” answers have been included to ensure that
we receive appropriate feedback.
176
Thank You!
Please also complete the comments section
for this online module. This will greatly assist in
providing necessary learning opportunities in the
future, as well as help us to evaluate the
effectiveness of the online module format and
material.
We appreciate your time and hope you found this
training worthwhile.
177
Thank You!
A special thanks to the following:
Laurie Ray, MPT
Physical Therapy Consultant for NC DPI
Project Mentor
Jane Sweeney, PT, PhD, PCS, FAPTA
Rocky Mountain University of Health Professions
Graduate Program Consultant
Jean Bryan Coe, PhD, DPT
Project Consultant, Survey Data
Stephen Allison, PT, PhD
Project Consultant, Resources for Survey Data
178