• Describes techniques to fit curves ( curve fitting ) to discrete data to obtain intermediate estimates.
• There are two general approaches two curve fitting:
– Data exhibit a significant degree of scatter . The strategy is to derive a single curve that represents the general trend of the data.
– Data is very precise.
The strategy is to pass a curve or a series of curves through each of the points.
• In engineering two types of applications are encountered:
– Trend analysis. Predicting values of dependent variable, may include extrapolation beyond data points or interpolation between data points.
– Hypothesis testing. Comparing existing mathematical model with measured data.
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
1
Figure PT5.1
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
2
Simple Statistics/
• In course of engineering study, if several measurements are made of a particular quantity, additional insight can be gained by summarizing the data in one or more well chosen statistics that convey as much information as possible about specific characteristics of the data set.
• These descriptive statistics are most often selected to represent
– The location of the center of the distribution of the data,
– The degree of spread of the data.
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
3
• Arithmetic mean . The sum of the individual data points (yi) divided by the number of points (n).
i y
y n
1 , , n i
•
Standard deviation . The most common measure of a spread for a sample.
S y
S t
n
S
t y i
1
(
y )
2 or
S y
2
y i
2 n
1 y i
2
/ n by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
4
• Variance . Representation of spread by the square of the standard deviation.
S y
2
( y i n
1 y )
2
Degrees of freedom
•
Coefficient of variation . Has the utility to quantify the spread of data.
c .
v .
S y
100 % y by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
5
Figure PT5.2
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
6
Figure PT5.3
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
7
• Fitting a straight line to a set of paired observations: (x
1
, y
1
), (x
2
, y
2
),…,(x n
, y n
).
y=a
0
+a
1 x+e a
1
slope a
0
intercept eerror, or residual, between the model and the observations by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
8
Criteria for a “Best” Fit/
• Minimize the sum of the residual errors for all available data: i n
1 e i
i n
1
( y i
a o
a
1 x i
) n = total number of points
• However, this is an inadequate criterion, so is the sum of the absolute values i n
1 e i
i n
1 y i
a
0
a
1 x i by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
9
Figure 17.2
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
10
• Best strategy is to minimize the sum of the squares of the residuals between the measured y and the y calculated with the linear model:
S r
i n
1 e i
2 i n
1
( y i
, measured
y i
, model )
2 i n
1
( y i
a
0
a
1 x i
)
2
• Yields a unique line for a given set of data.
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
11
Least-Squares Fit of a Straight Line/
0
0
S
a o r
S
a
1 r
2
( y i
2
( y i y i
y i x i
a
0
a a o
0 a o x i
a
1 x i a
1 x i a
1
x i
)
) x i a
1 x i
2
0
na
0 a
0
na
0 x i
a
1
y i
0
Normal equations, can be solved simultaneously a
1
n
x i y i
n
x i
2
x i
2 y i
Mean values a
y
a
1 x
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
12
Figure 17.3
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
13
Figure 17.4
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
14
Figure 17.5
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
15
“Goodness” of our fit/
If
• Total sum of the squares around the mean for the dependent variable, y, is S t
• Sum of the squares of residuals around the regression line is S r
•
S t
-S r quantifies the improvement or error reduction due to describing data in terms of a straight line rather than as an average value.
r
2
S t
S r
S t r 2 -coefficient of determination by Lale Yurttas, Texas
A&M University
Sqrt(r 2 ) – correlation coefficient
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
16
• For a perfect fit
S r
=0 and r=r 2 =1, signifying that the line explains 100 percent of the variability of the data.
• For r=r 2 =0, S r
=S t
, the fit represents no improvement.
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
17
• Some engineering data is poorly represented by a straight line. For these cases a curve is better suited to fit the data. The least squares method can readily be extended to fit the data to higher order polynomials ( Sec. 17.2
).
by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
18
y
a
0 z
0
a
1 z
1
a
2 z
2
a m z m
e z
Y
0
, z
1
,
, z m are m
matrix of the
1 basis functions calculated values of the basis functions
at the measured observed unknown residuals
valued
values of coefficien
the ts of the independen dependent variable t variable
S r
i n
1
y i m
0 j a j z ji
2 Minimized by taking its partial derivative w.r.t. each of the coefficients and setting the resulting equation equal to zero by Lale Yurttas, Texas
A&M University
Chapter 17
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
19