Transcript a 1
Statistical &
Uncertainty Analysis
UC REU Programs
Instructor: Lilit Yeghiazarian, PhD
Environmental Engineering Program
1
Instructor
Dr. Lilit Yeghiazarian
Environmental Engineering Program
Office: 746 Engineering Research Center (ERC)
Email: [email protected]
Phone: 513-556-3623
2
Textbooks
Applied Numerical Methods with MATLAB
for Engineers and Scientists, 3rd edition,
S.C. Chapra, McGraw-Hill Companies,
Inc., 2012
An Introduction to Error Analysis:
The Study of Uncertainties in Physical
Measurements, 2nd editions, J.R. Taylor,
University Science Books, Sausalito, CA
3
Outline for today
Error
numerical error
data uncertainty in measurement error
Statistics & Curve Fitting
mean
standard deviation
linear regression
t-test
ANOVA
4
Types of Error
General Error (cannot blame computer) :
Blunders
Formulation or model error
incomplete mathematical model
Data uncertainty
human error
limited to significant figures in physical measurements
Numerical Error:
Round-off error (due to computer approximations)
Truncation error (due to mathematical approximations)
5
Gare Montparnasse, Paris, 1895
Accident causes:
Faulty brake
Engine driver trying to make up lost time
6
Accuracy and Precision
Errors associated with both
calculations &
measurements can be
characterized in terms of:
•
Accuracy: how closely a
computed/measured value agrees
with true value
•
Precision: how closely individual
computed/measured values agree
with each other
(a) inaccurate and imprecise
(b) accurate and imprecise
(c) inaccurate and precise
(d) accurate and precise
Note: Inaccuracy = bias
Imprecision = uncertainty
Figure 4.1, Chapra
7
Accuracy and Precision
• Inaccuracy: systemic
deviation from truth
• Imprecision: magnitude
of scatter
(a) inaccurate and imprecise
(b) accurate and imprecise
(c) inaccurate and precise
(d) accurate and precise
Note: Inaccuracy = bias
Imprecision = uncertainty
Figure 4.1, Chapra
8
Error, Accuracy and Precision
In this class we refer to Error as collective
term to represent both inaccuracy and
imprecision of our predictions
9
Round-off Errors
Occur because digital computers have a limited
ability to represent numbers
Digital computers have size & precision limits on
their ability to represent numbers
Some numerical manipulations highly sensitive
to round-off errors arising from
mathematical
considerations and/or
performance of arithmetic operations on computers
10
Computer Representation of Numbers
Numerical round-off errors are directly related to
way numbers are stored in computer
The
fundamental unit whereby information is represented
is a word
A
word consists of a string of binary digits or bits
Numbers
are stored in one or more words, e.g.,
-173 could look like this in binary on a 16-bit computer:
off “0”
on “1”
(10101101)2=27+25+23+22+20=17310
11
As good as it gets on our PCs …
-1.797693134862316 x 10308
1.797693134862316 x 10308
0
- Overflow
15 significant
figures
Underflow
15 significant
figures
Overflow +
“Hole”
on either
side of
zero
-2.225073858507201 x 10-308
2.225073858507201 x 10-308
For 64-bit, IEEE double precision format systems 12
Implications of Finite Number of bits (1)
Range
Finite
range of numbers a computer can represent
Overflow
error – bigger than computer can handle
For double precision (MATLAB and Excel): >1.7977 x 10308
Underflow
error – smaller than computer can handle
For double precision (MATLAB and Excel): <2.2251 x 10-308
set format long and use realmax and realmin
in MATLAB to test your computer for range
Can
13
Implications of Finite Number of bits (2)
Precision
Some
numbers cannot be expressed with a finite number
of significant figures, e.g., π, e, √7
14
Round-Off Error and
Common Arithmetic Operations
Addition
Mantissa of number with smaller exponent is modified so both
are the same and decimal points are aligned
Result is chopped
Example: hypothetical 4-digit mantissa & 1-digit exponent computer
1.557 + 0.04341 = 0.1557 x 101 + 0.004341 x 101 (so they have same exponent)
= 0.160041 x 101 = 0.1600 x 101 (because of 4-digit mantissa)
Subtraction
Similar to addition, but sign of subtrahend is reversed
Severe loss of significance during subtraction of nearly equal
numbers → one of the biggest sources of round-off error in
numerical methods – subtractive cancellation
15
Round-Off Error and
Large Computations
Even though an individual round-off error could
be small, the cumulative effect over the course
of a large computation can be significant!!
Large
numbers of computations
Computations interdependent
Later calculations depend on results of earlier ones
16
Particular Problems Arising from
Round-Off Error (1)
Adding a small and a large number
Common problem in summing infinite series (like the Taylor
series) where initial terms are large compared to the later terms
Mitigate by summing in the reverse order so each new term is
comparable in size to the accumulated sum (add small
numbers first)
Subtractive cancellation
Round-off error induced from subtracting two nearly equal
floating-point numbers
Example: finding roots of a quadratic equation or parabola
Mitigate by using alternative formulation of model to minimize
problem
17
Particular Problems Arising from
Round-Off Error (2)
Smearing
Occurs when individual terms in a summation are > summation
itself (positive and negative numbers in summation)
Really a form of subtractive cancellation – mitigate by using
alternative formulation of model to minimize problem
Inner Products
Common problem in solution of simultaneous linear algebraic
equations
Use double precision to mitigate problem (MATLAB does this
automatically)
n
x y
i 1
i
i
x1 y1 x2 y2 ... xn yn
18
Truncation Errors
Occur when exact mathematical formulations
are represented by approximations
Example:
Taylor series
19
Taylor series widely used to express
functions in an approximate fashion
Taylor’s Theorem:
Any smooth function can be approximated as a polynomial
Taylor series expansions
where h = xi+1 - xi
0th
f ( xi 1 ) f ( xi )
1st
f ( xi 1 ) f ( xi ) f ' ( xi )( h)
2nd
f " ( xi )
f ( xi 1 ) f ( xi ) f ( xi )( h)
( h) 2
2!
3rd
f " ( xi )
f (3) ( xi )
2
f ( xi 1 ) f ( xi ) f ( xi )( h)
( h)
( h) 3
2!
3!
4th
f " ( xi )
f (3) ( xi )
f ( 4) ( xi )
2
3
f ( xi 1 ) f ( xi ) f ( xi )( h)
( h)
( h)
( h) 4
2!
3!
4!
'
'
'
20
Each term adds more information:
e.g., f(x) = - 0.1x4 - 0.15x3 - 0.5x2 - 0.25x + 1.2 at x = 1
= 1.2
≈ 1.2 – 0.25(1) = 0.95
Figure 4.6, Chapra, p. 93
≈ 1.2 – 0.25(1) –(1.0/(1*2))*12 = 0.45
= 1.2 – 0.25(1) – (1.0/(1*2))*12
– (0.9/(1*2*3))*13 = 0.3
21
Total Numerical Error
Sum of
round-off error and
truncation error
As step size ↓,
# computations
round-off error (e.g.
due to subtractive cancellation
or large numbers of
computations)
truncation error ↓
Point of diminishing returns is when
round-off error begins to negate
benefits of step-size reduction
Figure 4.10, Chapra, p. 104
Trade-off here
22
Control of Numerical Errors
Experience and judgment of engineer
Practical programming guidelines:
Avoid subtracting two nearly equal numbers
Sort the numbers and work with the smallest numbers first
Use theoretical formulations to predict total numerical
errors when possible (small-scale tasks)
Check results by substituting back in original model and
see if it actually makes sense
Perform numerical experiments to increase awareness
Change step size or method to cross-check
Have two independent groups perform same calculations
23
Measurements & Uncertainty
24
Errors as Uncertainties
Error in scientific measurement means the
inevitable uncertainty that accompanies all
measurements
As such, errors are not mistakes, you cannot
eliminate them by being very careful
The best we can hope to do is to ensure that
errors are as small as reasonably possible
In this section, words error and uncertainty are
used interchangeably
25
Inevitability of Uncertainty
Carpenter wants to measure the height of
doorway before installing a door
First rough measurement: 210 cm
If pressed, the carpenter might admit that the
height in anywhere between 205 & 215 cm
For a more precise measurement, he uses a
tape measure: 211.3 cm
How can he be sure it’s not 211.3001 cm?
Use a more precise tape?
26
Measuring Length with Ruler
27
Measuring Length with Ruler
28
Measuring Length with Ruler
Note: markings are 1 mm apart
Best estimate of length = 82.5 mm
Probable range: 82 to 83 mm
We have measured the length to the nearest millimeter
29
How To Report & Use Uncertainties
Best estimate ± uncertainty
In general, the result of any measurement
of quantity x is stated as
(measured value of x) = xbest ± Δx
Δx is called uncertainty, or error, or margin
of error
Δx is always positive
30
Basic Rules About Uncertainty
Δx cannot be known/stated with too much
precision; it cannot conceivably be known to 4
significant figures
Rule for stating uncertainties: Experimental
uncertainties should almost always be rounded
to one significant figure
Example: if some calculation yields Δx=0.02385,
it should be rounded to Δx=0.02
31
Basic Rules About Uncertainty
Rule for stating answers: The last significant
figure in any stated answer should usually be of
the same order of magnitude (in the same
decimal position) as the uncertainty
Examples:
The answer 92.81 with uncertainty 0.3 should be
rounded as 92.8 ± 0.3
If the uncertainty is 3, then the answer should be
rounded as 93 ± 3
If the uncertainty is 30, then the answer should be
rounded as 90 ± 30
32
Propagation Of Uncertainty
33
Statistics & Curve Fitting
34
Curve Fitting
Could plot points and
sketch a curve that
visually conforms to the
data
Three different ways
shown:
a.
Least-squares regression for
data with scatter (covered)
b.
Linear interpolation for
precise data
c.
Curvilinear interpolation for
precise data
Figure PT4.1, Chapra
Curve Fitting and Engineering Practice
Two typical applications when fitting experimental
data:
Trend analysis – use pattern of data to make
predictions:
Imprecise or “noisy” data → regression (least-squares)
Precise data → interpolation (interpolating polynomials)
Hypothesis testing – compare existing
mathematical model with measured data
Determine unknown model coefficient values … or …
Compare predicted values with observed values to test
model adequacy
You’ve Got a Problem … especially if you are this guy
Wind tunnel data relating force
of air resistance (F) to wind
velocity (v) for our friend the
bungee jumper
Figure 13.1, Chapra
The data can be used to
discover the relationship and
find a drag coefficient (cd), i.e.,
As F , v
How to fit the “best” line or curve
to these data?
Data is not smooth, especially
at higher v’s
If F = 0 at v = 0, then the
relationship may not be linear
Figure 13.2, Chapra
Before We Can Discuss Regression
Techniques … We Need To Review
basic
terminology
descriptive statistics
for talking about sets of data
Data from TABLE 13.3
Basic Terminology
Maximum? 6.775
Minimum? 6.395
Range?
6.775 - 6.395 = 0.380
Individual data points, yi
y1 = 6.395
y2 = 6.435
↓
y24 = 6.775
Number of observations?
Degrees of freedom?
Residual?
yi y
n = 24
n – 1 = 23
Use Descriptive Statistics To
Characterize Data Sets:
Location of center of distribution of the data
Arithmetic
mean
Median (midpoint of data, or 50th percentile)
Mode (value that occurs most frequently)
Degree of spread of the data set
Standard
deviation
Variance
Coefficient
of variation (c.v.)
Data from TABLE 13.3
Arithmetic Mean
y
y
i
n
158.4
y
6.6
24
Data from TABLE 13.3
Standard Deviation
St
sy
n 1
2
(
y
y
)
i
n 1
St: total sum of squares of residuals
between data points and mean
0.217000
sy
0.097133
24 1
Data from TABLE 13.3
Variance
s y2
2
(
y
y
)
i
St
n 1
n 1
2
2
y
(
y
)
i i /n
n 1
0.217000
s
0.009435
24 1
2
y
Coefficient of
Variation (c.v.)
c.v. = standard deviation / mean
Normalized measure of spread
c.v.
sy
y
100%
0.097133
c.v.
100% 1.47%
6.6
Data from TABLE 13.3
Data from TABLE 13.3
Histogram of data
Figure 12.4, Chapra
For a large set of data,
histogram can be
approximated by
a smooth, symmetric
bell-shaped curve
→ normal distribution
Confidence Intervals
If a data set is normally distributed, ~68% of the total
measurements will fall within the range defined by
y s y to y s y
Similarly, ~95% of the total measurements will be
encompassed by the range
y 2s y to y 2s y
Descriptive Statistics in MATLAB
>>% s holds data from Table 13.2
>>s=[6.395;6.435;6.485;…;6.775]
>>mean(s), median(s),
>>min(s),
mode(s)
max(s)
ans = 6.6
ans = 6.395
ans = 6.61
ans = 6.775
ans = 6.555
>>var(s),
>>range=max(s)std(s)
min(s)
ans =
range = 0.38
0.0094348
ans = 0.097133 >>[n,x]=hist(s)
n =
1
1
3
1
4
3
5
2
2
2
x = 6.414 6.452 6.49 6.528 6.566 6.604 6.642 6.68 6.718 6.756
n is the number of elements in each bin; x is a vector specifying the midpoint of each bin
Back to the Bungee Jumper Wind Tunnel
Data … is the mean a good fit to the data?
Figure 12.1, Chapra
velocity, m/s
10
20
30
40
50
60
70
80
mean:
Force, N
25
70
380
550
610
1220
830
1450
642
Figure 13.8a, Chapra
not very!!! distribution
of residuals is large
Figure 13.2, Chapra
Curve Fitting Techniques
Least-squaresFigure
regression
12.8, Chapra, p. 209
Figure 13.8b, Chapra
Linear
Polynomial
General
linear least-squares
Nonlinear
Interpolation (not covered)
Polynomial
Splines
Can reduce the distribution
of the residuals if use curvefitting techniques such as
linear least-squares regression
Linear Least-Squares Regression
Linear least-squares regression, the simplest
example of a least-squares approximation is fitting
a straight line to a set of paired observations:
(x1, y1), (x2, y2), …, (xn, yn)
Mathematical expression for a straight line:
y = a0+ a1x + e
error or residual
intercept
slope
Least-Squares Regression:
Important Statistical Assumptions
Each x has a fixed value; it is not random and it
is known without error, this means that
x values must be error-free
regression of y versus x is not
the same as the
regression of x versus y
The y values are independent random variables
and all have the same variance
The y values for a given x must be normally
distributed
Residuals in Linear Regression
Regression line is a
measure of central
tendency for paired
observations (i.e.,
data points)
Residuals (ei) in
linear regression
represent the
vertical distance
between a data
point and the
regression line
Figure 13.7, Chapra
How to Get the “Best” Fit:
Minimize the sum of the squares of the residuals
between the measured y and the y calculated with the (linear)
model:
n
S r ei2
i 1
n
( yi ,measured yi ,model ) 2
i 1
n
( yi a0 a1 xi ) 2
i 1
Yields a unique line for a given dataset
How do we compute the best a0 and a1?
n
S r ( yi a0 a1 xi ) 2
i 1
One way is to use optimization techniques since
looking for a minimum (more common for nonlinear
case)… or …
Another way is to solve the normal equations for a0
and a1 according to the derivation in the next few
slides
Derivation of Normal Equations Used to
Solve for a0 and a1
n
S r ( yi a0 a1 xi ) 2
i 1
First, differentiate the sum of
the squares of the residuals
with respect to each unknown
coefficient
¶Sr
= -2å (yi - a0 - a1 xi )
¶a0
¶Sr
= -2å[(yi - a0 - a1 xi )]xi
¶a1
Derivation of Normal Equations Used to
Solve for a0 and a1 - continued
Set derivatives = 0
Will result in a
minimum Sr
Can be expressed as
S r
2 ( yi a0 a1 xi )
a0
0
S r
2 [( yi a0 a1 xi )] xi 0
a1
0 yi a0 a1 xi
0 yi xi a0 xi a1 xi2
Derivation of Normal Equations Used to
Solve for a0 and a1 - continued
Realizing ∑a0 = na0, we can express these
equations as a set of two simultaneous linear
equations with two unknowns (a0 and a1) called
the normal equations:
Normal Equations:
0 yi a0 a1 xi
na0 xi a1 yi
0 yi xi a0 xi a1 xi2
xi a0 xi2 a1 xi yi
Derivation of Normal Equations Used to
Solve for a0 and a1 - continued
na0 xi a1 yi
n
xi
xi a0 xi2 a1 xi yi
Finally, can solve
these normal
equations for a0
and a1
a1
x a y
x a x y
i
2
i
i
0
i
1
n xi yi xi yi
n x xi
2
i
2
and
i
a0 y a1 x
Improvement Due to Linear Regression
Figure 12.8, Chapra, p. 209
mean of the dependent
variable
sy/x
sy
best-fit line
Figure 13.8, Chapra
Spread of data around the mean
of the dependent variable
Spread of data around the best-fit line
The reduction in spread in going from (a) to (b), indicated by bellshape curves, represents improvement due to linear regression
Improvement Due to Linear Regression
Figure 12.8, Chapra, p. 209
sy/x
mean of the dependent
variable
Total sum of squares around the
mean of dependent variable
n
S t ( yi y )
i 1
2
sy
best-fit line
Sum of squares of residuals
around the best-fit
regression line
n
S r ( yi a0 a1 xi )
2
i 1
Coefficient of determination quantifies improvement or error
reduction due to describing data in terms of a straight line rather
than an average value
St S r
r
St
2
“Goodness” of Fit
St - Sr quantifies error reduction
due to using line instead of
mean
Normalize by St because scale
sensitive → r 2 = coefficient of
determination
small residual errors
r2 → 1
St S r
r
St
2
Used for comparison of several
regressions
Value of zero represents no
improvement
Value of 1 is a perfect fit, the line
explains 100% of data variability
large residual errors
r2 << 1
Figure 12.9, Chapra
Linearization of Nonlinear
Relationships
What to do when relationship
is nonlinear?
One option is polynomial
regression
Another option is to linearize
the data using transformation
techniques
Exponential
Power
Saturation-growth-rate
Figure 14.1, Chapra
Linearization Transformation Examples
Exponential model:
y 1e
1 x
Figure 13.11, Chapra
Used to characterize quantities
that increase (+β1) or decrease
(-β1) at a rate directly proportional
to their own magnitude, e.g.,
population growth or radioactive
decay
Take ln of both sides to linearize
data:
ln y ln 1 1x
Linearization Transformation Examples
Power model:
y 2 x
2
Figure 13.11, Chapra
Widely applicable in all fields
of engineering
Take log of both sides to
linearize data:
log y 2 log x log 2
Linearization Transformation Examples
Saturation-growth-rate model
x
y 3
3 x
Figure 13.11, Chapra
Used to characterize
population growth under
limiting conditions or enzyme
kinetics
To linearize, invert equation to
give:
1 3 1 1
y 3 x 3
Linear Regression with MATLAB
Polyfit can be used to determine the slope
and y-intercept as follows:
>>x=[10 20 30 40 50 60 70 80];
>>y=[25 70 380 550 610 1220 830 1450];
>>a=polyfit(x,y,1)
use 1 for linear (1st order)
a = 19.4702 -234.2857
Polyval can be used to compute a value using
the coefficients as follows:
>>y = polyval(a,45)
y = 641.8750
Polynomial Regression
Extend the linear leastsquares procedure to fit
data to a higher order
polynomial
For a quadratic (2nd order
polynomial), will have a
system of 3 normal
equations to solve
instead of 2 as for linear
For higher mth-order
polynomials, will have a
system of m+1 normal
equations to solve
Data not suited for linear least-squares
regression
Figure 14.1, Chapra
Nonlinear Regression (not covered)
Cases in engineering where nonlinear models –
models that have a nonlinear dependence on
their parameters – must be fit to data
For example,
f ( x) a0 (1 e a1x ) eerror
Like linear models in that we still minimize the
sum of the squares of the residuals
Most convenient to do this with optimization
More Statistics: Comparing 2 Means
depends on mean and
amount of variability
can tell there is a
difference when
variability is low
use t-test to do this
mathematically
http://www.socialresearchmethods.net/kb/stat_t.php
69
The t-test Is A Ratio Of “Signal To Noise”
Remember, variance (var) is just the
square of the standard deviation
Standard Error of difference between means
http://www.socialresearchmethods.net/kb/stat_t.php
70
How It Works ... Once You Have A t-value
Need to set a risk level (called the alpha level) – a
typical value is 0.05 which means that five times out of
a hundred you would find a statistically significant
difference between the means even if there was none
(i.e., by "chance").
Degrees of freedom (df) for the test = sum of # data
points in both groups (N) minus 2.
Given the alpha level, the df, and the t-value, you can
use a t-test table or computer program to determine
whether the t-value is large enough to be significant.
If calculated t is larger than t (alpha, N-2) in table, you can
conclude that the difference between the means for
the two groups is different (even given the variability).
http://www.socialresearchmethods.net/kb/stat_t.php
71
What About More Than 2 Sets Of Data?
ANOVA = Analysis of Variance commonly used for more
than 2, if ...
k samples are random and independently selected
treatment responses are normally distributed
treatment responses have equal variances
ANOVA compares variation between groups of data
to variation within groups, i.e.,
variation between groups
F =
variation within groups
Source: http://www.uwlax.edu/faculty/baggett/Math_145/HANDOUTS/anova.pdf
72
Steps for ANOVA
Define k population or treatment means being compared in context
of problem
Set up hypotheses to be tested, i.e.,
H0: all means are equal
Ha: not all means are equal (no claim as to which ones not)
Choose risk level, alpha (=0.05 is a typical level)
Check if assumptions reasonable (previous slide)
Calculate the test statistic ... pretty involved ...see next page!
Note: usually use a computer program, but is helpful to know what
computer is doing ... can do simple problems by hand
Source: http://www.uwlax.edu/faculty/baggett/Math_145/HANDOUTS/anova.pdf
73
ANOVA calculations
Collect this info from data set:
Source: http://www.uwlax.edu/faculty/baggett/Math_145/HANDOUTS/anova.pdf
74
ANOVA calculations
Fill out a table like this to compute the F ratio
statistic:
Source: http://www.uwlax.edu/faculty/baggett/Math_145/HANDOUTS/anova.pdf
75
ANOVA calculations
Now what do we do with F statistic ?
Compare it to an F distribution like we did with the t-test
This time we need
to look up F (1- alpha)(k-1, N-k)
This time we need to compare
alpha
df of numerator (k-1)
df of denominator (N-k)
F ≥ F (1- alpha)(k-1, N-k)
If yes, then more evidence against H0, reject H0
Source: http://www.uwlax.edu/faculty/baggett/Math_145/HANDOUTS/anova.pdf
76
The End
77