Transcript classmar2

IS 4800 Empirical Research Methods
for Information Science
Class Notes March 2, 2012
Instructor: Prof. Carole Hafner, 446 WVH
[email protected] Tel: 617-373-5116
Course Web site: www.ccs.neu.edu/course/is4800sp12/
Outline
Finish discusion of usability testing
Hypothesis testing review
Sampling, Power and Effect Size
Chi square – review and SPSS application
Correlation – review and SPSS application
Begin t-test if time permits
UI/Usabililty evaluation
•
•
•
•
What are the three approaches ??
What are the advantages and disadvantages of each?
Explain a usability experiment that is within-subjects
Explain a usability experiment that is betweensubjects
• What are the advantages and disadvantages of each ?
What is a Usability Experiment?
Usability testing in a controlled environment
•There is a test set of users
•They perform pre-specified tasks
•Data is collected (usually quantitative and qualitative)
•Take mean and/or median value of quantitative attributes
•Compare to goal or another system
Contrasted with “expert review” and “field study” evaluation
methodologies
The growth of usability groups and usability laboratories
Usability Experiment
•
•
•
•
•
•
Defining the variables to collect ?
Techniques for data collection ?
Descriptive statistics to use
Potential for inferential statistics
Basis for correlational vs experimental claims
Reliability and validity
Experimental factors
Subjects
representative
sufficient sample
Variables
independent variable (IV)
characteristic changed to produce different conditions.
e.g. interface style, number of menu items.
dependent variable (DV)
characteristics measured in the experiment
e.g. time taken, number of errors.
Experimental factors (cont.)
Hypothesis
prediction of outcome framed in terms of IV and DV
null hypothesis: states no difference between conditions
aim is to disprove this.
Experimental design
within groups design
each subject performs experiment under each condition.
transfer of learning possible
less costly and less likely to suffer from user variation.
between groups design
each subject performs under only one condition
no transfer of learning
more users required
variation can bias results.
Summative Analysis
What to measure? (and it’s relationship to a usability goal)
Total task time
User “think time” (dead time??)
Time spent not moving toward goal
Ratio of successful actions/errors
Commands used/not used
frequency of user expression of:
confusion, frustration, satisfaction
frequency of reference to manuals/help system
percent of time such reference provided the needed answer
Measuring User Performance
Measuring learnability
Time to complete a set of tasks
Learnability/efficiency trade-off
Measuring efficiency
Time to complete a set of tasks
How to define and locate “experienced” users
Measuring memorability
The most difficult, since “casual” users are hard
to find for experiments
Memory quizzes may be misleading
Measuring User Performance (cont.)
Measuring user satisfaction
Likert scale (agree or disagree)
Semantic differential scale
Physiological measure of stress
Measuring errors
Classification of minor v. serious
Reliability and Validity
Reliability means repeatability. Statistical significance
is a measure of reliability
Validity means will the results transfer into a real-life
situation. It depends on matching the users, task,
environment
Reliability - difficult to achieve because of high
variability in individual user performance
Formative Evaluation
What is a Usability Problem??
Unclear - the planned method for using the system is not
readily understood or remembered (info. design level)
Error-prone - the design leads users to stray from the
correct operation of the system (any design level)
Mechanism overhead - the mechanism design creates awkward
work flow patterns that slow down or distract users.
Environment clash - the design of the system does not
fit well with the users’ overall work processes. (any design level)
Ex: incomplete transaction cannot be saved
Qualitative methods for collecting usability
problems
Thinking aloud studies
Difficult to conduct
Experimenter prompting, non-directive
Alternatives: constructive interaction, coaching
method, retrospective testing
Output: notes on what users did and expressed: goals,
confusions or misunderstandings, errors, reactions expressed
Questionnaires
Should be usability-tested beforehand
Focus groups, interviews
Observational Methods - Think Aloud
user observed performing task
user asked to describe what he is doing and why, what he thinks is
happening etc.
Advantages
simplicity - requires little expertise
can provide useful insight
can show how system is actually use
Disadvantages
subjective
selective
act of describing may alter task performance
Observational Methods - Cooperative evaluation
variation on think aloud
user collaborates in evaluation
both user and evaluator can ask each other questions throughout
Additional advantages
less constrained and easier to use
user is encouraged to criticize system
clarification possible
Observational Methods - Protocol analysis
paper and pencil
cheap, limited to writing speed
audio
good for think aloud, diffcult to match with other protocols
video
accurate and realistic, needs special equipment, obtrusive
computer logging
automatic and unobtrusive, large amounts of data difficult to analyze
user notebooks
coarse and subjective, useful insights, good for longitudinal studies
Mixed use in practice.
Transcription of audio and video difficult and requires skill.
Some automatic support tools available
Query Techniques - Interviews
analyst questions user on one to one basis
usually based on prepared questions
informal, subjective and relatively cheap
Advantages
can be varied to suit context
issues can be explored more fully
can elicit user views and identify unanticipated problems
Disadvantages
very subjective
time consuming
Query Techniques - Questionnaires
Set of fixed questions given to users
Advantages
quick and reaches large user group
can be analyzed more rigorously
Disadvantages
less flexible
less probing
Laboratory studies: Pros and Cons
Advantages:
specialist equipment available
uninterrupted environment
Disadvantages:
lack of context
difficult to observe several users cooperating
Appropriate
if actual system location is dangerous or impractical for
to allow controlled manipulation of use.
Steps in a usability experiment
1. The planning phase
2. The execution phase
3. Data collection techniques
4. Data analysis
The planning phase (your proposal)
Who, what, where, when and how much?
•Who are test users, and how will they be recruited?
•Who are the experimenters?
•When, where, and how long will the test take?
•What equipment/software is needed?
•How much will the experiment cost? <not required>
Prepare detailed test protocol
*What test tasks? (written task sheets)
*What user aids? (written manual)
*What data collected? (include questionnaire)
How will results be analyzed/evaluated?
Pilot test protocol with a few users <one user>
Execution Phase: Designing Test Tasks
Tasks:
Are representative
Cover most important parts of UI
Don’t take too long to complete
Goal or result oriented (possibly with scenario)
Not frivolous or humorous (unless part of product goal)
First task should build confidence
Last task should create a sense of accomplishment
Detailed Test Protocol
What tasks?
Criteria for completion?
User aids
What will users be asked to do (thinking aloud
studies)?
Interaction with experimenter
What data will be collected?
All materials to be given to users as part of the test,
including detailed description of the tasks.
Execution phase
Prepare environment, materials, software
Introduction should include:
purpose (evaluating software)
voluntary and confidential
explain all procedures
recording
question-handling
invite questions
During experiment
give user written task description(s), one at a time
only one experimenter should talk
De-briefing
Execution phase: ethics of human
experimentation applied to usability testing
Users feel exposed using unfamiliar tools and making errors
Guidelines:
•Re-assure that individual results not revealed
•Re-assure that user can stop any time
•Provide comfortable environment
•Don’t laugh or refer to users as subjects or guinea pigs
•Don’t volunteer help, but don’t allow user to struggle too
long
•In de-briefing
•answer all questions
•reveal any deception
•thanks for helping
Data collection - usability labs and equipment
Pad and paper the only absolutely necessary data
collection tool!
Observation areas (for other experimenters, developers,
customer reps, etc.) - should be shown to users
Videotape (may be overrated) - users must sign a
release
Video display capture
Portable usability labs
Usability kiosks
Analysis of data
Before you start to do any statistics:
look at data
save original data
Choice of statistical technique depends on
type of data
information required
Type of data
discrete - finite number of values
continuous - any value
Testing usability in the field (6 things you can
do)
1. Direct observation in actual use discover new
uses take notes, don’t help, chat later
2. Logging actual use objective, not intrusive great
for identifying errors which features are/are not
used privacy concerns
Testing Usability in the Field (cont.)
3. Questionnaires and interviews with real users
ask users to recall critical incidents
questionnaires must be short and easy to return
4. Focus groups
6-9 users
skilled moderator with pre-planned script
computer conferencing??
5 On-line direct feedback mechanisms
initiated by users
may signal change in user needs
trust but verify
6. Bulletin boards and user groups
Field Studies: Pros and Cons
Advantages:
natural environment
context retained (though observation may alter it)
longitudinal studies possible
Disadvantages:
distractions
noise
Appropriate
for “beta testing”
where context is crucial for longitudinal studies
Statistical Thinking (samples and populations)
• H1: Research Hypothesis:
– Population 1 is different than Population 2
• H0: Null Hypothesis:
– No difference between Pop 1 and Pop 2
• State test criteria (a, tails)
• Compute p(observed difference|H0)
– ‘p’ = probability observed difference is due to random
variation
• If p < alpha then reject H0 => accept H1
– alpha typically set to 0.05 for most work
– p is called the “level of significance” (actual)
– alpha is called the criterion
31
Relationship between alpha, beta,
and power.
“The Truth”
Decide to Reject H0
& accept H1
Do not Reject H0
& do not accept H1
H1 True
H1 False
Correct
p = power
Type I err
p = alpha
Type II err Correct
p = beta
p = 1-alpha
32
Relationship Between Population
and Samples When a Treatment
Had No Effect
Population

Sample 1
M1
Sample 2
M2
33
Relationship Between Population
and Samples When a Treatment
Had An Effect
Control
group
population
c
Treatment
group
population
t
Treatment
group
sample
Control
group
sample
Mc
Mt
34
Some Basic Concepts
• Sampling Distribution
– The distribution of every possible sample taken from a
population (with size n)
• Sampling Error
– The difference between a sample mean and the population
mean: M - μ
– The standard error of the mean is a measure of sampling
error (std dev of distribution of means)  M
35
Some Basic Concepts
• Degrees of Freedom
– The number of scores in sample with a known mean that are
free to vary and is defined as n-1
– Used to find the appropriate tabled critical value of a statistic
• Parametric vs. Nonparametric Statistics
– Parametric statistics make assumptions about the nature of
an underlying population
– Nonparametric statistics make no assumptions about the
nature of an underlying population
36
Sampling
Mean?
Variance?
Population

Sample of size N
Mean values from all possible
samples of size N
aka “distribution of means”
2
X

M=
SD 2 =
2
(
X

M
)

N
MM = 
 M2 =
N
2
N
ZM = ( M -  ) /  M
Estimating the Population Variance
S2 is an estimate of σ2
S2 = SS/(N-1) for one sample (take sq root for S)
For two independent samples – “pooled estimate”:
S2 = df1/dfTotal * S12 + df2/dfTotal * S22
dfTotal = df1 + df2 = (N1 -1) + (N2 – 1)
From this calculate variance of sample means: S2M = S2/N
needed to compute t statistic
Z tests and t-tests
t is like Z:
Z=M-μ/ 
M
t=M–0/ S
M
We use a stricter criterion (t) instead of Z
because S is based on an estimate of the
M
population variance while  Mis based on a
known population variance.
T-test with paired samples
Given info about
population of change
scores and the
sample size we will
be using (N)
We can compute the
distribution of means
?
=0
S2 est 2 from sample = SS/df
Now, given a
particular sample of
change scores of
size N
S2M = S2/N
We compute its mean
and finally determine
the probability that this
mean occurred by
chance
t=
M
SM
df = N-1
t test for independent samples
Given two
samples
Estimate population
variances
(assume same)
Estimate variances
of distributions
of means
Estimate variance
of differences
between means
(mean = 0)
This is now your
comparison distribution
t test for independent samples, continued
Distribution of differences
between means
This is your
comparison distribution
NOT normal, is a ‘t’
distribution
Shape changes depending on
df
df = (N1 – 1) + (N2 – 1)
Compute t = (M1-M2)/SDifference
Determine if beyond cutoff score
for test parameters (df,sig, tails)
from lookup table.
Effect size
• The amount of change in the DVs seen.
• Can have statistically significant test but small
effect size.
43
Power Analysis
• Power
– Increases with effect size
– Increases with sample size
– Decreases with alpha
• Should determine number of subjects you need ahead
of time by doing a ‘power analysis’
• Standard procedure:
– Fix alpha and beta (power)
– Estimate effect size from prior studies
• Categorize based on Table 13-8 in Aron (sm/med/lg)
– Determine number of subjects you need
– For Chi-square, see Table 13-10 in Aron reading
44
• X^2 tests
– For nominal measures
– Can apply to a single measure (goodness of fit)
• Correlation tests
– For two numeric measures
• t-test for independent means
– For categorical IV, numeric DV
45
Categorial Examples
• Observational study/descriptive claim
– Do NU students prefer Coke or Pepsi?
• Study with correlational claim
– Is there a difference between males and females in
Coke or Pepsi preference?
• Experimental Study with causal claim
– Does exposure to advertising affect Coke or Pepsi
preference? (students assigned to treatments)
Understanding numeric measures
• Sources of variance
– IV
– Other uncontrolled factors (“error variance”)
• If (many) independent, random variables with the
same distribution are added, the result approximately
a normal curve
– The Central Limit Theorem
47
The most important parts of the
normal curve (for testing)
5%
Z=1.65 48
The most important parts of the
normal curve (for testing)
2.5%
Z=-1.96
2.5%
Z=1.9649
Hypothesis testing – one tailed
• Hypothesis: sample (of 1) will be significantly greater
than known population distribution
– Population completely known (not an estimate)
• Example – WizziWord experiment:
–
–
–
–
H1:  WizziWord > Word
a = 0.05 (one-tailed)
Population (Word users):  Word =150, =25
What level of performance do we need to see before we can
accept H1?
50
Hypothesis testing – two tailed
• Hypothesis: sample (of 1) will be significantly
different from known population distribution
• Example – WizziWord experiment:
– H1:  WizziWord   Word
– a = 0.05 (two-tailed)
– Population (Word users):  Word =150, =25
– What level of performance do we need to see before we can
accept H1?
51
Standard testing criteria for
experiments
• a = 0.05
• Two-tailed
52
Don’t try this at home
• You would never do a study this way.
• Why?
– Can’t control extraneous variables through
randomization.
– Usually don’t know population statistics.
– Can’t generalize from an individual.
53
Sampling
• Sometimes you really can measure the entire
population (e.g., workgroup, company), but this is
rare…
• More typical: “Convenience sample”
– Cases are selected only on the basis of feasibility or ease of
data collection.
• Assumed ideal: Random sample
– e.g., random digit dialing (approx)
54
Hypothesis testing with a sample wrt distribution of
means
Given info about
population and the
sample size we will
be using (N)
Now, given a
particular sample
of size N
We can compute the
distribution of means
and finally determine
the probability that
this mean occurred
by chance
We compute its mean
55
Sampling
Mean?
Variance?
Population

X

M=
Sample of size N
Mean values from all
possible samples of size N
aka “distribution of means”
2
SD 2 =
2
(
X

M
)

N
N

NOTE: This is
a normal curve
56
 =
2
M
2
N
t-statistics,
t-distributions &
t-tests
57
Single sample t-test
• What if you know comparison pop’s mean but
not stddev?
– Estimate population variance from sample
variance
• Estimate of S^2 = SS/(N-1)
• S^2M = S^2/N
– Comparison is now a t-test, t=(M-u)/SM
– df=N-1
58
t-test for dependent means
aka “paired sample t-test”
59
t-test for dependent means
When to use
• One factor, two-level, within-subjects/repeated
measures design
-or-
• One factor, two-level, between-subjects, matched pair
design
• In general, a bivariate categorical IV and numeric
DV when the DV scores are correlated.
• Assumes
– Population distribution of individual scores is normal
60
Wanted: a statistic for differences
between paired individuals
• In a repeated-measures or matched-pair design,
you directly compare one subject with
him/herself or another specific subject (not
groups to groups).
• So, start with a sample of change (difference)
scores:
Sample 1 = Mary’s wpm using Wizziword –
Mary’s wpm using Word
61
Hypothesis testing with paired samples
Given info about
population of change
scores and the
sample size we will
be using (N)
We can compute the
distribution of means
?
=0
est 2 from sample
Now, given a
particular sample of
change scores of
size N
We compute its mean
and finally determine
the probability that this
mean occurred by
chance
t=
M
62
SM
df = N-1
SPSS
64
Analyze/Compare Means/Paired Sample t-test
65
Results
Paired Samples Test
Paired Differences
Std. Error
Mean
Std. Deviation Mean
r 1 Condition1 - Condition2 -168.000
199.332
63.034
95% Confidence
Interval of the
Difference
Lower
Upper
-310.594 -25.406
paired t(9)=2.665, p<.05
t
-2.665
df
Sig. (2-tailed)
9
.026
Between-Subjects Design
• Have two experimental conditions (treatments, levels,
groups)
• Randomly assign subjects to conditions (why?)
• Measure numeric outcome in each group
• Each group is a sample from a population
• Big question: are the populations the same (null
hypothesis) or are they significantly different?
– What statistic tests this?
68
t-test for independent means
• Tests association between binomial IV and
numeric DV.
• Examples:
– WizziWord vs. Word => wpm
– Small vs. Large Monitors => wpd
– Wait time sign vs. none => satisfaction
69
t-test for independent means
• Two samples
• No other information about comparison
distribution
70
Solution – take two samples, gathered at
same time
Intervention
Control
The big question: which is correct?
H1
Intervention
Control
H0
Intervention
Control
71
Wanted: a statistic to measure how similar
two samples are
(of numeric measures)
• “t score for the difference between two means”
M1  M 2
t=
S?
• If samples are identical, t=0
• As samples become more different, t increases.
• What is the comparison distribution?
– Want to compute probability of getting a particular t score IF the
samples actually came from the same distribution (what is the t score
for this case?).
72
Why t?
• In this situation, you do not know the population
parameters; they must be estimated from the samples.
• When you have to estimate a comparison population’s
variance, the resulting distribution is not normal – it is
a “t distribution”.
• The particular kind of t distribution we are using in
this case is called a “distribution of the difference of
means”.
73
All things t
• t distribution shape is parameterized by
“degrees of freedom”
• For a distribution of the difference of means,
df = df1  df 2 = ( N1 1)  ( N2 1)
74
Only remaining loophole
M1  M 2
t=
S?
75
Assumptions for t
– Scores are sampled randomly from the population
– The sampling distribution of means is normal
– Variances of the two populations (whether they are
the same or different) are the same.
• Typical assumption.
76
Finally – the t test for independent samples
S? = Sdifference
Est of Mean
Pop1
Dist of
Means 1
Pooled est of common
variance
Est of Mean
Dist of
Difference
of Means
Pop2
Dist of
Means 2
This is now your
comparison distribution
Reporting results
• Significant results
t(df)=tscore, p<sig
e.g., t(38)=4.72, p<.05
• Non-significant results
e.g., t(38)=4.72, n.s.
78
SPSS
79
SPSS
80
SPSS
Equal variances
assumed
t(10)=3.796, p<.05
Sidebar: Control groups
• To demonstrate a cause and effect hypothesis, an experiment
must show that a phenomenon occurs after a certain treatment
is given to a subject, and that the phenomenon does not occur
in the absence of the treatment.
• A controlled experiment (“experimental design”) generally
compares the results obtained from an experimental sample
against a control sample, which is identical to the
experimental sample except for the one aspect whose effect is
being tested.
• You must carefully select your control group in order to
demonstrate that only the IV of interest is changing between
groups.
82
Sidebar: Control groups
•
•
•
•
Standard-of-care control (new vs. old)
Non-intervention control
“A vs. B” design (shootout)
“A vs. A+B” design (e.g., S-O-C vs. S-O-C+intervention)
• Problem: the “intervention” may cause more than just the desired
effect.
– Example: giving more attention to intervention Ss in educational
intervention
• Some solutions:
– Attention control
– Placebo control
– Wait list control (also addresses measurement issues)
83
Sidebar: Control groups
Related concepts
• Blind test – S does not know group
• Double blind test – neither S nor experimenter know
• Manipulation check
– Test performed just to see if your manipulation is working. Necessary if
immediate effect of manipulation is not obvious.
– “Positive control” test for intervention effect
– “Negative control” test for lack of intervention effect
– Example:
• Student Center Sign: ask students if they saw & read the sign
84
Relationship Between Population
and Samples When a Treatment
Had No Effect
Population

Sample 1
M1
Sample 2
M2
85
Relationship Between Population
and Samples When a Treatment
Had An Effect
Control
group
population
c
Treatment
group
population
t
Treatment
group
sample
Control
group
sample
Mc
Mt
86
Some Basic Concepts
• Sampling Distribution
– The distribution of every possible sample taken from a
population
• Sampling Error
– The difference between a sample mean and the population
mean
– The standard error of the mean is a measure of sampling
error (std dev of distribution of means)
87
Some Basic Concepts
• Degrees of Freedom
– The number of scores in sample with a known mean that are
free to vary and is defined as n-1
– Used to find the appropriate tabled critical value of a statistic
• Parametric vs. Nonparametric Statistics
– Parametric statistics make assumptions about the nature of
an underlying population
– Nonparametric statistics make no assumptions about the
nature of an underlying population
88
Parametric Statistics
• Assumptions
– Scores are sampled randomly from the population
– The sampling distribution of the mean is normal
– Within-groups variances are homogeneous
• Two-Sample Tests
– t test for independent samples used when subjects were
randomly assigned to your two groups
– t test for dependent samples (aka “paired-sample t-test”)
used when samples are not independent (e.g., repeated
measure)
89
Finally – the t test for independent samples
Given two
samples
Estimate population
variances
(assume same)
Estimate variances
of distributions
of means
Estimate variance
of differences
between means
(mean = 0)
This is now your
comparison distribution
Finally – the t test for independent samples, continued
Distribution of differences
between means
This is your
comparison distribution
NOT normal, is a ‘t’
distribution
Shape changes depending on
df
df = (N1 – 1) + (N2 – 1)
Compute t = (M1-M2)/SDifference
Determine if beyond cutoff score
for test parameters (df,sig, tails)
from lookup table.