Child and Adolescent Psychopathology

Download Report

Transcript Child and Adolescent Psychopathology

Chapter 2:
Strategies for Evidence-Based
Assessment of Children and
Adolescents
Erica A. Youngstrom
Thomas A. Frazier
“Three P’s” of Assessment
 Predict
 Does the assessment method predict important criteria?
 Prescribe
 Does the assessment method prescribe specific
treatments?
 Process
 Does the assessment method inform our understanding
of processes in developmental psychopathology?
Factors Contributing to
Stagnation Within the Field
 Lack of a clear linkage between measurement tools
and clinical practice
 Erosion of service reimbursement and the challenge to
demonstrate cost effectiveness of psychological
assessments
 Current assessment devices are selected primarily
based on convention rather than on any clear sense
that they are the best tools for specific purposes
 Nosological systems were initially developed for adults
and then just adapted for use with adolescents and
children
Prediction
 Which assessment tools should be used to
measure the development of psychopathology?
 How can users compare the different tests that are
available and make an informed choice about
them?
 Prediction can mean demonstrating concurrent
correlations, as well as showing associations with
criteria that are separated by time.
Concurrent Criterion Validity
 What does it mean?
 Correlating assessment tools with other meaningful criteria
 An iterative process involving elaboration and consolidation
 Elaboration  involves expanding the network of correlations to
include not just validators, but also correlates in terms of typical
treatment response, outcome, quality of life, and functioning
 Crucial to contextualizing the construct
 Consolidation  involves clarifying when different measures are
assessing the same underlying latent variables, as well as
creating models about relationships among these underlying
processes
Statistical Methods: Correlations
 Correlations
 A measure of effect size that can be compared across
studies and across measures within a field
 Has several drawbacks:
• As an index of prognosis, is difficult to apply to individual cases
• Can consider only one correlate at a time
• Can also be unduly influenced by extreme scores, especially when
you have a small sample size
 May be more useful to directly compare validity
correlations for different measures
Statistical Methods: Regression
 Several advantages over bivariate correlational analyses:
 Preserve the actual units of measurement in the unstandardized
regression weights
 Provide formal tests of whether combinations of predictors provide
incremental improvement over a single predictor
 Make predictions about an individual’s score on the criterion variable
 Create a framework where it is possible to test statistical mediation or
moderation of relationships
 Some limitations:
 Lack of sufficient published statistics
 Computational burden on the clinician
 Regression weights are dependent on the specific combination of
instruments used
 One possible solution:
 Publish more complete details of regression results
Additional Statistical Methods
 Raw Scores
 Working in raw units would indicate the actual scores that
would be predicted on the dependent variable, which
brings the focus back to how clinicians would actually use
the assessment
 Incremental Validity
 An additional variable provides a statistically significant
improvement in prediction of the criterion
 Comparing Observed Versus Predicted
Performance
 Most commonly used to compare cognitive ability and
academic function
Statistical Methods for
Consolidation of Scores
 Three methods for determining the number of
underlying dimensions underlying a battery
1. Partial Analysis
2. Scree Test
3. Minimum Average Partials
 All three methods converge on more parsimonious
factor structures, which has several advantages:
 Fewer scales to interpret  reduces Type I error risk
 Larger number of items per factor  greater internal
consistency, smaller standard errors of measurement,
and more accurate description of the individual
Grouping Individuals: “QMethods”
 Factor Analysis
 Aggregate people with similar profiles of scores across
variables, rather than the more widespread approach of
grouping variables with similar scores across people
 Cluster Analysis
 Latent Class Analysis
 Cluster cases on the basis of observed categorical
indicator variables
 Mixture Modeling
 Takes into account residual relationships among
indicators
Advantages of Q-Methods
 Allow the data to dictate how many core profiles
emerge, and what the relative prevalence is of
each profile
 Provide a parsimonious response to the problem of
comorbidity because most core profiles involve
elevations on multiple scales
 Multivariate profiles can incorporate information
from multiple informants
Event History Methods and
Individual Trajectories
 Event History Analysis
 Methods for looking at the length of time until a particular event
of interest happens
 Survival Analysis and Cox Regression
 Uniquely situated to evaluate onset, cessation, relapse, and
recovery
 Individual Trajectories
 Hierarchical Linear Models
• Repeated measures are “nested” within the individual participant
 Growth Curve Models
• Observed variables are treated as indicators of change for indirectly
observed latent variables
 Growth Mixture Models or Latent Class Growth Models
• Unique trajectories of indicators are empirically defined
Improving Predictive Value
of Assessments
 Publish studies that consolidate existing measures
into more parsimonious dimensions
 Conduct studies that elaborate the connections
between measures and constructs across
development
 Directly compare the predictive value of multiple
tests under the same conditions
 Supplement or supplant correlational analyses with
multivariate regression, reported in enough detail
to allow application to individual cases
Prescription
 Sometimes referred to as “treatment matching”
between a diagnosis and an intervention strategy
 Clinical diagnosis is “imperfectly reliable”
 We can never be 100% certain that any child has a given
disorder
 Once our estimate of the probability exceeds a certain
point, we can be confident to proceed with treatment
Primary, Secondary, and Tertiary
Intervention
 Primary Intervention
 “Universal prevention” that treats everyone, regardless of risk
 Most rational approach when treatment costs and risks are low, and
benefits clearly outweigh risk and cost
 Example: fluoridating drinking water to prevent dental problems
 Secondary Intervention
 Focus on those who are at risk of developing a condition, but have
not yet developed the condition
 Appropriate for cases where an individual falls between the
assessment and treatment thresholds
 Tertiary Intervention
 Most expensive and may also have the most serious risk of sideeffects
 Assessment question is not if there is a problem, but rather what the
specific nature of the problem is
Using Assessment to Aid in
Diagnosis
 Sensitivity: The percentage of cases with the diagnosis
that would be classified correctly by the test
 Specificity: Quantifies the percentage of cases without
the diagnosis that would be classified correctly by the
test
 Receiver Operating Characteristics (ROC) analysis
 Method for quantifying the relationship between the sensitivity
and specificity of an assessment tool for a particular diagnosis
 Sensitivity is plotted as a function of specificity
Nomogram
 Used to simplify the process of determining a posterior probability
(probability of a diagnosis combined with test results estimates a new
probability)
 Advantages
 Makes it possible to apply test results to a specific individual, directly estimating
the posterior probability
 Substantial improvements in the accuracy and precision of test interpretations
 Reducing the influence of cognitive heuristics on the interpretation of test results
 Flexibility in choice of starting point
 Elimination of computation
 Facilitation of discussion between families and practitioners
 Limitations
 Unfamiliar to most researchers
 Few researchers are publishing the likelihood ratios directly in their research
reports
Assessments as Aid in
Treatment Selection
 Absolute Risk Reduction (ARR)
 Subtracts the rate of a negative categorical event in the treatment
group from the rate in the comparison group
 Number Needed to Treat (NNT)
 1 divided by the ARR
 Number Needed to Harm (NNH)
 1 divided by the rate of adverse events in the group receiving the
treatment versus the rate of the same adverse events in the
comparison group
 Likelihood of Help Versus Harm (LHH)
 The ratio of the NNT and NNH, with both being adjusted for the
clinician’s judgment of the patient’s risk compared to the risk of the
average patient in studies providing the values of NNT and NNH
 Limitation of these parameters  all assume dichotomous
outcomes
Process
 Clinical Outcomes
 Quantifies the degree of change occurring as the result of
treatment
 Encompasses measurement of variables that are
informative about the mechanisms for growth and
change
 Assessment can also be used to detect variables
that will change the prognosis or response to
treatment
Treatment Outcome Evaluation
 Gives clinicians access to feedback about change
during the course of treatment, which in turn:
 Produces lower rates of dropout
 Leads to more efficient resource allocation across
caseloads
 Results in better outcomes
Measuring Treatment Outcome
 Repeated, Brief Measures
 Example: Longitudinal Interview Follow-up Evaluation
(LIFE): Ratings on a fixed scale that are repeated each
week
 Advantages: Low burden on the respondent, sensitivity to
trends of change over treatment
 Disadvantage: Resulting data are difficult to analyze
using conventional statistical methods
 Behavior Checklists and Questionnaires
 Example: Child Depression Inventory
 Advantage: Issue of converting outcomes to effect sizes
Clinically Significant Change
 Measured in two ways:
1. Reliable change
• Is the amount of change shown by an individual case large enough
to reflect “true” change, in the classical theory sense rather than
being attributable to the unreliability of the measure?
• Reliable Change Index (RCI) is expressed as a z-score with the
standard error of the difference acting as the denominator
2. Change that moves the patient’s score below a
predefined normative threshold
Mediators and Moderators
 Mediators
 The intermediate variables that provide mechanisms of
change
 Measurement error, potential suppression effects, and
low statistical power complicate the assessment of
mediational models
 Moderators
 Also known as an “interaction effect” because moderators
change the relationship between another pair of variables
 Diathesis-Stress Model is an example
Adherence and Fidelity
 Both influence the effectiveness of an intervention
 Measuring adherence
 Rate of completion of homework assignments
 Attendance of scheduled sessions
 Measuring fidelity
 Therapist progress notes, checklists, or rating scales
completed by the therapist
Reappraisal of the “Three Ps” –
Evaluating Assessments
 Practitioners should critically read the literature and
look for the best measures suited to their evaluation
 Conduct studies that directly compare assessment
methods on these three dimensions
 Conduct more meta-analytic studies
 If new measures are developed, adaptive testing using
Item Response Theory to quantify item properties will
make it possible to obtain increased precision and a
better range of scores without increasing participant
burden