Detection of Effort and Malingering
Download
Report
Transcript Detection of Effort and Malingering
Detection of Effort and
Malingering: State of the Art
Jason Gravano, M.S.
6/30/14
Speaking euphemistically?
• ‘‘cognitive effort,’’ ‘‘mental effort,’’ ‘‘insufficient effort,’’ ‘‘poor effort,’’
‘‘invalid effort,’’ or even ‘‘faked effort.’’
• response bias, disingenuous, dissimulation, non-credible, malingered,
or non- or sub-optimal effort
Bigler, 2012
Outline
• Introduction to Symptom Validity testing
• Definitional Issues
• Assumptions
• Measures of cognitive “Effort”
• Measures of validity on self report
• “Diagnosing” malingering
Validity of the Evaluation: the crux of the
issue?
• We do not measure cognition or symptoms directly, rather the behavioral
output of cognition (test scores, performances, self reports etc.)
• NP tests require the participant's best effort/cooperation for our inferences
about their cognition to be valid
• Likewise, self-report of symptoms may not be taken at face value
• Thus, in instances where the evaluation informs clinical or legal decision
making (arguably, all cases), validity checks become necessary
Bigler, 2012
Validity of the Evaluation: the crux of the
issue?
• Especially because research has shown repeatedly that experienced
experts are inaccurate in identifying valid versus invalid ability
performances from mere observation of behavior or test scores
• E.g.:
• Ekman, O’Sullivan, & Frank, 1999
• Faust, 1995
• Faust, Hart, Guilmette, & Arkes, 1988
AACN 2010
Symptom Validity Testing:
• Methods and instruments designed to detect exaggeration or
fabrication of neuropsychological dysfunction, as well as somatic and
psychological symptom complaints.
AACN 2010
Detection
• Symptom validity tests
• “effort measures”
• Any measure that is normally performed (and in some cases, perfectly performed) by a
wide range of patients who have bona fide neurologic, psychiatric, or developmental
problems
• Have to try to perform poorly?
• Performance validity/Response validity
• Negative response bias
AACN 2010
Keep it consistent, people.
• Invalid performances/report should manifest as inconsistencies:
• In ability
•
•
•
•
•
•
Consistency within a measure (hard vs. easy)
Consistency within a cognitive domain
Consistency within a battery
Consistency between two or more testing sessions
Consistency of NP results with real-world performances
Consistency with expected profile of disorder (& severity)
• In self report of sys
• More on this later…
“Effort” Measures
…now with more political correctness!
Effort measures
• Typically rely on extremely low ceilings
• Measures that are so easy, “anyone can do them”
• Inference: If they don’t do well, they must not be “trying”
• If performance is poor, raises questions about the validity of your
examination
• Can be used as supporting evidence for malingering, especially when
performance is below chance
Effort measures
• Embedded measures (usually found after publication to have utility)
• Examples: CVLTii forced choice; Reliable Digit Span; WMS Faces; WCST “other
categories”; TMT time.
• PROS- Face valid, don’t take extra time, relevant to the validity of the specific
measure, hard(er) to coach
• Stand alone measures (designed purposefully for eval of effort)
• Examples: TOMM (Test of Memory Malingering); Word Memory test; Portland
Digit)
• PROS- Created for this reason, likely better research base
AACN 2010
• Embedded measures can take advantage of floor effects:
• Use an alpha of .05 from a distribution of Severe TBI on a case of Mild
TBI
• Stand-alone cognitive effort tests
• Forced-choice
• Significantly below chance is direct evidence of deliberate underperformance
• “invalid performance can be identified using thresholds that are well
above a level that is significantly below chance”
• Non-forced-choice
• evaluate random responding
• unrealistically slow or erroneous responding
• inconsistency of response patterns
AACN 2010
• Bimodal distribution?
• Cut Scores=/= chance performance
• Based on validation studies
Bigler, 2012
Effort testing studies
• Rogers (2008)
• (a) simulation;
• (b) criterion groups (‘‘known’’ groups);
• (c) differential prevalence designs
AACN 2010
Differential prevalence design
• Assumes that response bias varies as a function of context (forensic,
clinical, incentives offered, etc.)
• Useful in initial validation studies
• Lacks objective accuracy standards (base rates are not 100%)
AACN 2010
Simulation
• Random assignment of non-clinical participants to
• simulate or fake
• Control
• Often first, proof of concept studies
• Lacks generalizability
AACN 2010
Criterion “known groups”
• Apriori criteria (often an “established” or “gold standard” measure +
Slick criteria for malingering) to define groups
• Criteria chosen to have low false positive rate
• Occasionally take out individuals with “borderline” performance on
the gold standard test……. So how does it perform across the board?
AACN 2010
Issues Bigler (2012) raises about SVT research:
• Most SVT studies do not have IRB approval
• No Systematic Study Lesion/Localization Studies of SVT performance
• Implication: there may be some lesions that disrupt SVT performance
• Neurogenic drive & motivation problems?
• Psychological/Psychiatric factors may influence SVT performance
• Cut scores will misclassify
• Is type 1 or type 2 error worse?
• Is the entire eval invalid?
Bigler, 2012
Resources for effort measures & embedded
validity measures
• Sweet, 2009
• Boone, 2007;
• Hom & Denney, 2002; Larrabee, 2007
• Vickery, Berry, Inman, Harris, & Orey, 2001
• Babikian, Boone, Lu, & Arnold, 2006;
• Heinly, Greve, Bianchini, Love, & Brennan, 2005;
• Iverson & Tulsky, 2003
• + the other readings from this week
AACN 2010
Consensus Recs: Effort
• Use tools designed (or re-designed) to detect sub-optimal effort
• Use both stand alone and embedded measures
• Use multiple measures, across the eval to account for varying effort
• Encourage optimal effort (through rapport)?
• Make sure the normative samples are appropriate
• Not all effort measures are created equal. Do some homework to see
which ones perform best with your population.
Symptom Reporting
Self-Report
• Disorder-specific inventories; symptom inventories
• E.g., ptsd, depression, etc.
• When evaluating the validity of self-report, if possible, clinicians
include measures that possess an internal means of assessing
response bias.
• In general, disorder-specific inventories and symptom checklists that
do not contain effective means for determining response bias and
possible response invalidity should not be used in isolation.
AACN 2010
Self-Report
• General personality inventories
•
•
•
•
Strong validity research & design
Keep up with research, appropriate populations
Don’t “interpret” clinical scales when validity is at issue
Be conservative in assuming intentionality (without supporting evidence)
AACN 2010
Review: Validity Scales on the MMPI-2
• Traditional validity scales
•
•
•
•
L – “Lie” present favorably
F- “Faking good or Faking Bad”
K – “defensiveness”
Fb- Infrequent items at the end of the test; Random responding at the end of the
exam?
• VRIN- Consistency
• TRIN- Consistency
• Other
•
•
•
•
Infrequency Psychopathology scale (Fp; Arbisi & Ben-Porath, 1995)
F-K (Gough, 1950)
Obvious minus Subtle Scale (O-S; Greene, 1991; Wiener, 1948)
Dissimulation scale (Gough,c1954, 1957) for the MMPI-2 (Ds2; Dsr2)
MMPI-2
• Other, newer scales that may be important:
• FBS-r on MMPI-2-rf
• RBS- Response Bias Scale- predict insufficient effort on neuropsychological
evaluation (Gervais et al., 2007a)
• HHI- Henry Heilbronner Index- exaggeration of somatic or
‘‘pseudoneurological’’ symptoms (Henry et al., 2006, 2008).
Butcher 2003 on FBS
• Originally termed “Fake Bad Scale”
• The FBS was developed by identifying items rationally on a content basis utilizing unpublished frequency
counts of malingerers’ MMPI responses and the authors’ subjective observations of personal injury (LeesHaley et al., 1991).
• Item pool contains 43 items that include somatic symptoms, unusual beliefs, and deviant attitudes.
• Scores for the FBS were then compared between
•
•
•
•
•
20 personal injury claimants who appeared “notably credible,”
A group of 25 personal injury claimants who appeared clearly to be malingering
16 medical outpatients asked to simulate emotional distress caused by a motor vehicle accident,
15 medical outpatients asked to simulate emotional distress caused by a toxic exposure,
36 medical outpatients asked to simulate emotional distress caused by stress on the job (Lees-Haley et al., 1991).
FBS
• The cumulative FBS literature suggests that the scale continues to
differentiate groups (“malingerers” or “over reporters” vs. valid) as
well as, and under certain conditions superior to, other MMPI-2
validity scales (including all of the F-family scales).
• In particular, two factors that are particularly relevant to practicing
neuropsychologists, effort status and TBI, substantially moderated
FBS magnitudes.
Nelson, 2010
FBS
• Early criticisms of FBS
• The basis for assigning individuals to the notably credible group and the
malingered group was not explained.
• Original norming was questionable (private practice)
• No T scores, just mean data
• Cutoffs were not intended for mental health arenas
• FBS may not measure a single dimension
• Items contained on the FBS are generally not those that are rare or
infrequently endorsed
• Low cross over with other validity scales
• FBS does correlate with psychopathology…
• Gender specific cutoffs?
Butcher 2003
Psychological/Somatic
• Inference: Symptom exaggeration, symptom promotion, fabrication
• Evaluation:
• In depth diagnostic eval
• Onset, course, symptom picture, comorbidities, treatment efforts, response
to treatment
• Inconsistent, contradictory data noted & Documented
AACN 2010
Onset
• What is the typical age of onset
• When was the first symptom (e.g., after the arrest?)
• Personal history (abuse, neglect, etc.)
• Family history
AACN 2010
Symptom Presentation
• Keep an eye out for atypical symptoms
• Layman’s understanding of the disorder
• Atypical combo of symptoms
• **Distinct clinical presentations and/or prevalence of
psychopathology may differ among some cultures and ethnic groups
(Draguns & Tanaka-Matsumi, 2003).
AACN 2010
Course
• Consistency of symptoms
• Expected trajectory
AACN 2010
Response to Treatment
• Have they responded to treatment? Why or why not? When? Prior to
litigation/compensation?
AACN 2010
Consensus Recs: Symptom Reporting
• Use self reports with validity measurements
• Make sure the normative samples are appropriate
• Evaluate ability and disorder response bias separately
Malingering
Malingering
• In considering the diagnosis of malingering, the clinician is explicitly
making a determination of intent: more specifically, a determination
of intentionally exaggerated symptoms and/or intentionally
diminished capability with the goal of obtaining an external reward.
• To do this:
•
•
•
•
•
the context of the evaluation
overall presentation of the examinee
background information
history information gathered during interview
observations, neuropsychological tests, and measures of response bias
AACN 2010
Effort Malingering
• Failed Effort measure =/= malingering, just as passing one =/= not malingering
• Assessing effort is one dimension of the evaluation of Malingering
• equating ‘‘poor effort’’ with malingering is an oversimplification
• It takes effort to fake!
• The process of detecting malingering is one in which consideration is given to
multiple dimensions of behavior that differentiate malingering from other
entities
•
•
•
•
factitious disorder
conversion disorder
cogniform disorder
somatoform disorder.
AACN 2010
Diagnostic classifications for Malingering
• DSM
• NP operationalized systems
• Bianchini et al., 2005
• Slick, Sherman, & Iverson, 1999
• Reliable classification indices; operationalized
• Guidelines recommend incorporation of multiple sources of
data/information:
• AACN Practice Guidelines for Neuropsychological Assessment and Consultation; AACN, 2007
• Specialty Guidelines for Forensic Psychologists; Committee on Ethical Guidelines for Forensic
Psychologists, 1991
AACN 2010
“Diagnosing” malingering
• Context of the evaluation
• Clinical vs. forensic
• Gain, secondary gain, external gain, financial gain
• ***MUST HAVE SOMETHING TO GAIN***
• Necessary, but far from sufficient
• Examples: financial reward, compensated time away from work,
avoidance of military duty, relief from legal consequences, and
obtaining medications/narcotics
AACN 2010
Assessing intent?
• The best way to assess intent is by ruling out other possible conditions
(e.g., psychological, neurological, developmental) that might otherwise
explain the suspicious behavioral presentation and by requiring the
presence of multiple improbable performances and/or atypical
symptomatic complaints.
• The differential diagnosis of malingering is a clinical process that:
•
•
•
•
(1) requires careful analysis on the part of the examiner,
(2) is based on objective criteria;
(3) incorporates indicators that have established classification accuracy,
(4) combines clinical judgment with the results of scientifically validated measures in
this process.
AACN 2010
The possibility of malingering
• (1) disparity between real-world observations and either test
performance or self-report,
• (2) inconsistency between type or severity of injury and test
performances,
• (3) inconsistency between an individual’s behavior when he/she is
aware of being evaluated versus when not aware of not being
evaluated, and
• (4) inconsistency across serial testings that cannot be explained by an
underlying neurological process or known psychiatric condition.
AACN 2010
The assessment
•
•
•
•
(1) review of records;
(2) psychosocial history obtained by interviewing;
(3) observations of the claimant’s behavior during the assessment period;
(4) consideration of information from collateral sources, such as significant
others, employers, etc., when available and appropriate;
• (5) formal psychological/neuropsychological testing;
• (6) response validity assessment procedures; and
• (7) surveillance video/audio, when available.
• ‘‘Does what I am learning about this claimant make sense in light of the
putative claim, the diagnosis, history and totality of the presentation?’’
AACN 2010
• Unlike psychometric indicators of invalid performance, objective
standards for the evaluation of inconsistencies between the clinical
presentation and evidence of capacities outside the clinical setting
may not exist.
• In the absence of such standards, a determination that an
inconsistency indicates invalidity should be cautious and conservative
and in some cases may be left to the trier-of-fact.
AACN 2010
Consensus Recs: Malingering
• Consider & investigate “real world” performance
• Be aware of potential litigation cases
• Be aware of discrepancy between test results and known-group
performances
• Evidence for/against malingering should be considered incrementally
Consensus recs: Practice
• List SVIs in reports, but be cautious of test security
• Consider serial evaluations to discriminate unrealistic performances
(e.g., lack of practice effects)
• Use collateral informants (when appropriate)
• Don’t rely on the manuals and cut scores: use up-to-date research
AACN 2010
Questions for the Class!
• If you see invalid performance in testing, how should you address it?
• Could we use and interpret effort measures in looking at intentional
disorders (e.g., subcortical disorders)?
• What do you do if you suspect there is genuine neurological disorder
& malingering?
Consensus recs: Research
• How to interpret multiple effort measurements
• Populations that “fail” effort measures despite effort
• Deliberate intent to feign
• Future ability tests should have validity indicators
• Pediatric samples?
• Rule out or rule in?
AACN 2010
The end