Sprague & Geers. The Sprague & Geers metric (S&G) is a simple integral comparison metric that includes both magnitude and phase errors, which are calculated independently and then combined to give a comprehensive error (6). The error calculation can be biased towards either the measured time history or the computed (simulation) time history, with the former being most common. Given two time histories of equal length, measured m(t) and computed c(t), the following time integrals are defined: t2 I mm (t 2 t1 ) 1 m 2 (t )dt [3] t1 I cc (t 2 t1 ) 1 t2 c 2 [4] (t )dt t1 I mc (t 2 t1 ) 1 t2 m(t ) c(t )dt [5] t1 The magnitude error is then defined as: M SG I cc / I mm 1 [6] The phase error is defined as: PSG 1 cos 1 ( I mc / I mm I cc ) [7] The comprehensive error is defined as: 2 2 CSG M SG PSG [8] Due to the relative simplicity of the error metric, it can be implemented into a spreadsheet program with little loss of accuracy (7). The integrals can be approximated by summations using the trapezoidal method. b ba N [9] f ( t ) dt [ f (ti ) f (ti1 )] a 2 N i 1 ba where N is the number of intervals such that t . Because equations [6] and [7] N use ratios of the integrals, the coefficients cancel leaving, for example; N I cc I mm [c(t ) c(t i 1 N i i 1 [m(t ) m(t i 1 i )] 2 [10] i 1 )] 2 An advantage of the Sprague & Geers metric is the independent treatment of magnitude and phase errors, i.e. the magnitude component is insensitive to phase errors and vice versa. This allows for an accurate representation of magnitude only and phase only errors. Additionally, this independence can help in determining the cause of an error. Weighted Integrated Factor. The Weighted Integrated Factor method (WIFac) has been proposed in the automotive literature, specifically for ATD model evaluation (4). This metric is associative and calculates the agreement between two signals. This is advantageous in situations where the experimental data is known to have significant error. Shown in equation [11], WIFac is weighted by the absolute maximum of the function values. This weighting has the effect of scaling local differences to increase the effect of errors at high magnitude function values (4). A modified version of this metric is examined in this paper, where the one minus portion has been removed to provide a computation of error instead of agreement. This metric can also be implemented in a spreadsheet program with the inclusion of a small delta function in the denominator in order to avoid divide by zero errors. Below is new WIFac (1-max/max previously did not have squared functions in the bottom max). 2 max( 0, f (t ) g (t )) max f (t ) , g (t ) 1 max( f (t ) 2 , g (t ) 2 ) dt Crit 1 2 2 max f (t ) , g (t ) dt 2 2 [11] In addition to the philosophical reasons for the use of an associative metric, there are also practical purposes. When utilizing an optimization routine to determine the ideal model parameters needed to achieve agreement, such as in ATD model development, associative metrics tend to produce computed results that do not under predict the test results (4). This metric has been combined with scalar errors from the peak and time of peak within the Quick Rating scheme from Madymo using a simple averaging routine. Global Evaluation Method. The Global Evaluation Method (GEM) has also been utilized in the automotive community to evaluate ATD models (8). This metric is part of a larger software program, called ModEval, which is designed to evaluate an entire ATD model, factoring in multiple data channels and multiple types of evaluation methods. The GEM is intended to be used on primary interest channels, such as head x and z acceleration in a typical forward crash without yaw. In general, to evaluate the test and model, the test curve is parameterized to define important points and shapes, and the model curve is analyzed to search for the defined parameters. A score is determined for each parameter, factoring in the magnitude and phasing of the parameter, and then an overall score is computed for the two data sets. These steps are completed by a standalone executable. Unlike the previous two metrics presented, the GEM can not readably be coded into a spreadsheet program. The parameters evaluated in the GEM are separated into three groups; points (minima and maxima), shapes (rising/falling edges and local extrema), and global. For the point group, one or two overall maxima and minima are chosen in the test data. Once found in the model data, equations [12] and [13] are used to calculate a single point score, which is the average of the time score (eqn. 12) and value score (eqn. 13). If a point is not found in the simulation data, the time and value scores are both set to zero. TimeTest _ Extremum TimeModel _ Extremum 100% Time _ ScoreOverall _ Extremum 1 Re ference _ Period ValueTest _ Extremum Value Model _ Extremum 100% Value _ ScoreOverall _ Extremum 1 Value Test _ Extremum [12] [13] The value score is essentially a relative error, with a one minus portion to calculate agreement. The reference period is 40% of the total evaluation time, chosen based on intuitive judgment for automotive crash events (8). The reference time allows the time score to be independent of the time of occurrence within the data. For example, a 5 ms difference in occurrence produces a different relative error if it occurs at 50 ms (10% relative error) or 100 ms (5% relative error), while using a reference time of 100 ms, which is 40% of the typical 250 ms test duration, produces the same error independent of when the error occurs. Setting a reference time may allow for a more logical understanding of an error, since the peak should occur in a region of interest and should not depend on the specific time. For the shape characteristics, the input curves are filtered with a low pass Butterworth filter with a corner frequency of 40 Hz in order to extract the general shape of the channels. One or two rising/falling edges are selected, along with one or two local extrema if appropriate. Equations [14] and [15] are used to determine the shape score for edges (taking the average of the time and value scores). Equations 12 and 13 are used for local extrema. Once again, if a feature is not found in the simulation data, the time and value scores are set to zero. TimeTest _ Edge TimeModel _ Edge 100% [14] Time _ Score Edge 1 Re ference _ Period min SlopeTest , Slope Model 100% [15] Value _ Score Edge 1 max SlopeTest , Slope Model The global group consists of two factors, the normalized cross-correlation function (NCCF) and a shape corridor. The NCCF determines a general time shift (phasing) of the two input curves, and the max of the NCCF provides the value. For the shape corridor, the input curves are filtered as in the shape group, and a corridor is defined based on the test data surrounded by a square rectangle with dimensions of +/5%. The time and value scores from the NCCF and shape corridor are combined to give the global score. The final GEM score is computed by summing the point score (twice) plus the shape and global scores, divided by four. Like WIFac, the GEM produces a score reflecting the agreement between test and simulation. For the purposes of this paper, the error between the two curves was calculated by subtracting the agreement score from one. By definition, the GEM score combines the error at the peak with the error associated with the time to peak and curve shape, with additional weight given to the peak value. The usefulness of this combination will be discussed in a later section. Normalized Integral Square Error (NISE). Originally developed to evaluate acceleration responses from side impact test dummies (9) and used extensively in automotive biomechanical research, the NISE method evaluates the amplitude, phase, and curve shape errors of time histories. At the heart of the method, values are calculated with and without a phase shift to allow for an independent calculation of phase error and amplitude error. Four basic values are defined (equations 16-19), similar to the S&G metric. 1 N R xx (0) X i X i [16] N i 1 1 N R yy (0) Yi Yi [17] N i 1 1 N R xy (0) X i Yi [18] N i 1 1 N n R xy ( ) [19] X iYi n N n i 1 Where n = τ / ∆t, X = test data, Y = simulation data, and N = number of points in the data set. Rxy (τ) is calculated for all time leads/lags, with the maximum value being used in the subsequent equations. Next the phase, amplitude, and shape scores are calculated. 2 R xy ( ) max 2 R xy (0) NISE phase [20] R xx (0) R yy (0) NISE amplitude Pxy ( ) 2 R xy ( ) max R xx (0) R yy (0) NISE shape 1 Pxy ( ) [21] [22] where Pxy is the correlation coefficient, defined as: Pxy ( ) R xy ( ) max) R xx (0) R yy (0) [23] A total NISE score can also be computed using equation 24, which is effectively the summation of the three components. NISE 1 2 R xy (0) R xx (0) R yy (0) [24] Because the total score is a simple summation, if there are sign differences between any of the three components, the total score can be lower than one of the components (suggesting less error than one might expect). Like S&G and WIFac, this metric can be implemented into a spreadsheet program with little difficulty. 4. Spit HH, Hovenga PE, Happee RM, Uijldert M, Dalenoort AM. Improved Prediction of Hybrid-III Injury Values using Advanced Multibody Techniques and Objective Rating. SAE 2005-01-1307. 2005. 5. Russell DM. Error Measures for Comparing Transient Data: Part II: Error Measures Case Study. Proceedings of the 68th Shock and Vibration Symposium. 185-198. 1997. 6. Sprague MA and Geers TL. A Sprectral-Element Method for Modeling Cavitation in Transient Fluid-Structure Interaction. International Journal for Numerical Methods in Engineering. 60 (15), 2467-2499. 2004. 7. Schwer LE. Validation Metrics for Response Histories: A Review with Case Studies. -UPDATE 8. Jacob C, Charras F, Trosseille X, et al. Mathematical Models Integral Rating. International Journal of Crashworthiness. 5(4), 417-731. 2000. 9. Donnelly B, Morgan R, Eppinger R. Durability, Repeatability and Reproducibility of the NHTSA Side Impact Dummy. Proceedings of the 27th Stapp Car Crash Conference, SAE Paper No. 831624. 299-310. 1983. New references for WIFac 1. Twisk D., Spit H.H., Beebe M., Depinet P. "Effect of Dummy Repeatability on Numerical Model Accuracy". SAE paper 2007-01-1173. 2007. 2. D.Twisk, P.A. Ritmeijer. "A Software Method for Demonstrating Validation of Computer Dummy Models Used In the Evaluation of Aircraft Seating Systems". SAE paper 2007-01-3925. 2007.