1 GPS RECEIVER CALIBRATION – WHY IS IT NECESSARY? Dr. Nadya Goldovsky and Dr. Ilya Kuselman National Physical Laboratory of Israel, Danciger “A’ Bldg., Givat Ram, Jerusalem 91904 Tel.: 02-6536534, fax: 02-6520797, e-mail: nadya.go@moital.gov.il Abstract Measurement uncertainty of the majority of commercial GPS receivers varies from 10-11 to 10-13 by the frequency scale, and from 100 ns to 50 ns by the time scale, being dependent on the receiver design. The main sources of uncertainty in GPS measurements are the GPS receiver position error, the orbital error, the satellite and receiver clock errors, the ionospheric and the tropospheric delays, the receiver internal delay, the satellite antenna and cable delay, the receiver noise, and the multipath error. The frequency uncertainty for a GPS receiver is larger than that for Cs-standard by 2-3 orders within a short-time interval (1 – 1000 s), and by one order within a long-term interval of about one week. A GPS receiver’s time scale bias from the Universal Coordinated Time (UTC) can be in order of microseconds. Not every GPS receiver is suitable for use as a traceable primary frequency and/or time standard. Therefore, GPS receiver calibration against the primary time and frequency standard (the Cs-atomic clock traceable to UTC) is of great importance and implies the calibration of both output frequencies and time scale. The frequencies calibration is performed through comparison of one of the outputs (1 pps, 10 MHz, and/or 5 MHz) with the corresponding reference frequency of the caesium atomic standard. The user’s GPS receiver time scale calibration is fulfilled through comparison between 1 pps signals of the user’s GPS receiver and those of the Cs-atomic standard. Keywords: Global Positioning System (GPS); GPS receiver calibration; Time and frequency measurement 2 Introduction Global Positioning System (GPS) is well known as a tool for both positioning and for high accuracy time and frequency transfer. Hundreds of companies sell GPS products, and many of manufacturers advertise their units as being time and frequency standards. The characteristics of a GPS receiver depend on the applications for which it was designed. There are two main types of GPS receivers: commercial GPS receivers intended for navigation (for example, in commercial and military aircraft, marine and automobile navigation) [1], and GPS receivers used in time and frequency metrology [2]. Commercial GPS receivers intended for navigation provide X,Y,Z coordinates of their location. Receivers used in cars are generally intended for driver navigation or for sending information on the location of a car to an emergency response center in case of accident. Receivers intended for commercial airliners have usually a data port for integration into the air navigation systems. The required uncertainty of such receivers is of the order of 1 – 100 m [1]. GPS timing receivers provide a 1 pulse per second (1 pps) electrical output. Most receivers use the coarse acquisition (C/A) code broadcast on the frequency of 1.57542 GHz (L1) as their time and frequency reference. The 1 pps output can provide time synchronization at locations anywhere on earth, or provide the timing reference for either a telecommunications network or for an Internet timeserver. Another type of GPS receivers known as GPS disciplined oscillators [3], provide standard frequencies of 5 MHz and/or 10 MHz, and sometimes produce frequencies of 1.544 MHz or 2.048 MHz used in telecommunications. Digital telecommunication networks that carry voice, data, and facsimile information require a reference frequency source with an uncertainty of 10-11 or less [2,4] for ensuring quality-of-service to the clients. Two other types of GPS receivers used for more precise measurements are commonview GPS receivers and carrier-phase GPS receivers [2]. A common-view GPS receiver is actually integrated system that combines a standard GPS timing receiver with 3 measurement hardware and software. These hardware and software allow the system to make measurements from individual satellites and to store the measurement results that can be processed later. Carrier-phase GPS receivers are designed for geodetic and surveying applications. Usually more expensive than the traditional time and frequency receivers, they track and measure the precision (P) code broadcast on two carrier frequencies L1 and L2 (the last one is at 1.2276 GHz). Their potential positioning performance is exceptional, the positioning uncertainty being within centimeters or even millimeters ranges [2]. Common-view and carrier-phase measurements require more effort, including postprocessing of the measurement data. For this reasons, they are usually used in international time and frequency comparisons between National Metrology Institutes (NMI) [2]. Although GPS broadcasts are continuously monitored by National Institute of Standards and Technology (NIST), USA, not all GPS receivers are suitable for use as traceable primary frequency standards. The importance of the GPS receiver calibration, as well as the calibration method used in the National Physical Laboratory of Israel (INPL) are discussed in the present paper. The importance of GPS receiver calibration While GPSs differ significantly, there are two common key factors that contribute to a GPS receiver performance, namely, the quality of the receiver’s internal oscillator and the quality of the software algorithm that processes data obtained from satellites [3]. The majority of GPS receivers have frequency short-term stability of about 10-10 – 10-11 [2]. Models with the best short-term stability discipline an oven-controlled quartz oscillator or a rubidium oscillator to the GPS signal. However, even GPS receivers of this type have a short-term stability that is lower than the Cs-standard stability. Such standard is maintained, for example, at INPL [5-8]. 4 Many receivers do not discipline an oscillator at all. Instead, they divide the output of a small temperature-controlled crystal oscillator down to 1pps, and then synchronize 1 pps with GPS signal. As a result, the short-term stability of these models is poor (the frequency uncertainty is about 10-9 – 10-10 over the averaging time about 1 –1000 s [2,3]). Most GPS receivers automatically select satellites used for the timing solution. However, software algorithms used to select satellites are different, and each receiver has its own thresholds at which it chooses to keep or to omit a satellite. Some algorithms limit the timing solution to just one, or to a few satellites. Others can use as many as 12 satellites in a solution. For this reason, two GPS receivers can yield different results even when connected to the same antenna at the same location. Since each GPS satellite is visible at a given location for a limited time, all GPS receivers must add and remove satellites from the group used to obtain time and frequency information. Often, adding and/or removing a satellite from the timing solution causes an instantaneous frequency change. Various receivers handle GPS broadcast errors differently. Some receivers have built-in software designed to remove “bad” data, other have no such software. But all GPS receivers might bias their time scale or even fail under certain conditions, namely, strong ionospheric disturbances (for example, ionospheric scintillations). Different types of GPS receivers are applied in different types of GPS measurements. These measurements in the field of time and frequency metrology can be divided into four general categories: one-way, common-view, carrier-phase, and two-way satellite time and frequency transfer (TWSTFT) [3]. The majority of user GPS measurements are one-way measurements. The one-way GPS technique uses the signals, obtained from a GPS receiver, as the reference for a calibration. The purpose of the measurement is either to synchronize an on-time pulse, or to calibrate a frequency source. The GPS signal is used in real time, and no post processing of the measurement results is required. 5 The main error sources in one-way GPS measurements are expressed by the basic equation for pseudorange [9]: p = ρ+dρ+c(dt-dT)+dIon+dTrop+ερ+ρmult , (1) where p is the measured pseudorange; ρ is the calculated geometric range between satellite and receiver ( i.e. [(XS-XR)2+(YS-YR)2+(ZS-ZR)2]1/2 ); XS,YS,ZS and XR,YR,ZR are the coordinates of the satellite and the receiver, respectively; dρ is the orbital error; dt and dT are the satellite and the receiver clock errors, respectively; d Ion is the ionospheric delay; dTrop is the tropospheric delay; ερ is the receiver code noise; ρmult is the multipath error [10]. Common-view, carrier-phase, and TWSTFT measurements are more precise measurements and require post-processing of the measurement data. The collected data are analyzed and processed using precise satellite orbit information and detailed models of the ionosphere and troposphere. These types of GPS measurements are usually used for international time and frequency comparisons between NMI. In common-view time transfer the time difference between two clocks is determined by simultaneous observation of the same satellite at the same time in two different laboratories. The measurement results are then exchanged and the difference between two time scales is obtained. This technique gives improved performance over the oneway technique. It eliminates satellite clock errors, and reduces orbital errors, ionospheric and tropospheric errors [3]. By forming the double and triple difference [3] using two and more satellite observations, the receiver clock errors also can be eliminated. The accuracy of common-view time transfer is typically in the 1 to 10 ns range. The GPS common-view technique has been used for many years by the Bureau International des Poids et Mesures (BIPM) as one of its main techniques for international time comparisons. The TWSTFT technique eliminates nearly all of the propagation delay errors because the signal path symmetry for both stations [3]. Recently BIPM has started using twoway technique as a primary time transfer technique. The uncertainty of TWSTFT over a 6 24 hours period is better than 1 ns and has been be as low as 100 picoseconds in some systems [3]. The traceability requirements for time and frequency measurement According to [11], the metrological traceability is “the property of a measurement result relating the result to a stated metrological reference through an unbroken chain of calibrations of a measuring system or comparisons each contributing to the stated measurement uncertainty”. The traceability chain for measurements in field of time and frequency is shown in Fig.1 [4]. The traceability chain is beginning from a measurement instrument, for example, a time interval counter. Link A is the link between the measurement instrument and the user’s GPS receiver. Measurement uncertainty sources in link A are: user’s GPS receiver uncertainty, multipath error [10], satellite clock error, uncertainty of the time delay in atmosphere, errors in satellites orbit calculation, uncertainty of calibration procedures, and human errors. The uncertainty of the user’s GPS receiver is caused by receiver’s noise, the internal oscillator’s instability, software error, uncertainty of time delay in antenna, in antenna’s cable, and finally GPS receiver internal delay. GPS receiver calibration is carried out in order to determine: 1) frequency deviation from its nominal value and uncertainty of this deviation; 2) the GPS receiver second bias from the SI second; 3) the GPS receiver time scale deviation from the Universal Coordinated Time. The deviation of the GPS receiver time scale from UTC can be relatively large (several microseconds) and is dependent on the time delays in the GPS receiver, in GPS’s antenna and in antenna’s cable; on the error of atmospheric delay calculation, and other software errors. Therefore, the calibration of GPS receiver against the time and frequency standard (the Cs-atomic clock traceable to UTC and SI second) is of great importance. It should be noted that UTC is a “paper” theoretic time scale calculated by BIPM in Paris. The time signals received by user’s GPS are referenced to real time scale called GPS time. The GPS system includes a constellation of 24 satellites [2]. The GPS satellites are controlled and operated by the United States Department of Defense. Each satellite 7 carries either rubidium or caesium oscillators, or a combination of both. The on-board oscillators provide the reference GPS time for transmission. The relation between GPS time and UTC involves a variable number of seconds as a consequence of the leap seconds and is as follows: [UTC – GPS time] = -13 s + C0 (2) where C0 is the correction given in the BIPM monthly Circular-T at 0h UTC. The C0 values vary in the range of ± 20 ns, and their expanded uncertainty is in the range of ±10 ns [12]. Link B connects the user’s GPS receiver to the broadcast signal monitored by NMI. Measurement uncertainty in this link is caused by the signal path from the GPS satellite to GPS receiver’s antenna. This uncertainty consists of tropospheric delay error, ionospheric delay error, orbital errors, and multipath errors. There are several agencies that provide calculation of satellite orbit, ionospheric delay, and tropospheric delay, for example, the Center for Orbit Determination in Europe and the International GPS Service. Among the error sources mentioned above, the major uncertainty is caused by tropospheric and ionospheric delay, which can reach to several hundreds of ns for low elevation angle satellites. Link C is the control, or monitoring, link between an NMI (for example, INPL) and the broadcast service. Link C connects the NMI to GPS satellite system. This link is established through the NMI’s measurements of the GPS signals. Link D is the chain between the UTC scale and the local Coordinated Time Scale UTC(NMI) maintained by the NMI. UTC is calculated by averaging of data collected from more than 200 atomic clocks located in about 50 national laboratories spread worldwide and regularly reported to BIPM. The uncertainty of link D can be evaluated using BIPM’s Circular T. The SI second is defined as the duration of 9,192,631,770 periods of the radiation corresponding to the transition between the two hyperfine levels of the ground state of the caesium 133 atom. Frequency (expressed in hertz) is derived from a second and 8 equals to the number of events over a 1-second interval. Thus, GPS receiver cannot be used as primary frequency and time standard, it can be used as secondary frequency and time standard only. As for every traceability chain, the measurement uncertainty in the chain shown in Fig. 1 is increasing in the direction opposite to traceability, i.e. from link D to link A. User’s GPS receiver frequency calibration at INPL Results of measurements and calibrations in field of time and frequency in Israel are traceable to the National Standard of Time and Frequency and the UTC(INPL) scale maintained by INPL, according to the law on the National Israel Time Scale since 1992. Once traceability to INPL is established, it is accepted as well in other countries that signed the mutual recognition arrangement [13]. In order to show traceability of a GPS receiver to UTC(INPL), an unbroken chain must be established between the measurements made by GPS receiver at the user’s site and the UTC(INPL) time scale. The block diagram of the GPS receiver calibration system used at INPL is shown in Fig. 2. The calibration system comprises: Cs-Standard traceable to UTC via GPS commonview link; the unit under test (UUT), for example, user’s GPS receiver; a time interval counter; a pulse distribution amplifier; a 1 pps generator (optionally, when UUT does not have its own 1 pps output); a personal computer (PC1) for GPS time transfer data processing; a INPL GPS receiver applied for common-view measurements; a satellite antenna; a microprocessor with time interval counter; a monitor; a printer; an ionospheric calibration receiver (ICR) with a satellite antenna; a commutation switch box; a personal computer (PC2) for calibration data processing; a high frequency (5 MHz) signal distribution amplifier [14, 15]. The GPS receiver calibration implies the calibration of output frequencies and the calibration of time scale. The frequencies calibration is performed through comparison of one of the outputs (1 pps, 10 MHz and/or 5 MHz) with the corresponding reference frequency of the INPL caesium atomic standard. The time interval counter measures the 9 time differences between calibrated and reference signals. UTC(INPL) calculation is developed in [5-8]. The frequency calibration procedure and the procedure for comparisons between UTC(INPL) and UTC are described in detail in [14, 15]. Frequency standards and oscillators stability is described by the fractional frequency deviation and Allan variance [16-18]. The fractional frequency deviation is: y=(fUUT–fREF)/fREF, (3) where fUUT is the calibrated frequency of the UUT; fREF is the reference frequency of the Cs-Standard. Allan variance σy2(τ) is: N-1 N-1 σy2(τ)=(1/2) · (N–1)-1·∑(yk+1-yk)2=(1/2) · (N-1)-1τ-2 ∑(xk–2xk+1+xk+2)2, k=1 (4) k=1 where N is the number of measurements; x is the difference between signals of the UUT and the reference clock; τ is the averaging time (the time between two successive x measurements). The square root from the variance σy2(τ) is the Allan deviation σy(τ). The y of the calibrated GPS receiver over all the measurement time is defined as the slope of the linear regression fit for the time comparison results between UUT (a GPS receiver) and the reference Cs-Standard. The number of measurements N of time differences x between the UUT and the reference Cs-Standard is about 20 000 – 32 000. To determine the metrological characteristics of the calibrated GPS receiver relatively to UTC, the reference frequency yREF and its uncertainty UREF should be taken into account. Thus, the corrected value of the UUT frequency yUUT-UTC and its uncertainty UUUT-UTC are calculated by the following equations: y[UUT-UTC] = y[UUT-INPL] + yREF , (5) 10 where y[UUT-INPL] is the fractional frequency deviation of the UUT relatively to INPL reference frequency; yREF = y[INPL-UTC] is the fractional frequency deviation of the INPL reference frequency relatively to UTC frequency, computed from the last BIPM Circular T. The expanded uncertainty of the calibrated GPS receiver frequency is: UUUT-UTC = k ( u2UUT-INPL + u2REF )1/2 , (6) where k is the coverage factor (k=2 for 95% confidence level in case the number of degrees of freedom N-1 is larger than 20 000); uUUT-INPL is the standard uncertainty of the calibrated GPS receiver frequency relatively to the INPL reference; uREF is the standard uncertainty of the INPL reference frequency yREF relatively to UTC, computed from BIPM Circular T for the minimal τ. The expanded uncertainty of the INPL reference frequency is: UREF = k σy(τ) / √ N (7) The expanded uncertainty of the INPL caesium atomic standard is about 10-14 over one week averaging, while the uncertainty of user’s GPS receivers relatively to INPL frequency reference uUUT-INPL is about 10-13 – 10-11. User’s GPS receiver time scale calibration at INPL The calibration of user’s GPS receiver time scale is performed at INPL according to the block diagram shown in Fig. 2. The measurements are performed simultaneously in two parallel channels. In the first channel, the user’s GPS receiver time scale is compared with the UTC(INPL). For this purpose, the 1-second signals (1 pulse per second) of the user’s GPS receiver and the INPL Cs-Standard are compared by the time interval counter. The obtained difference between UTC(INPL) and UUT time scales is D1 = UTC(INPL) – UUT time. Connecting cables delay is subtracted from the difference in the D1 value calculation. In the second channel, the international comparisons between the local and GPS time scales are performed by the INPL GPS receiver operating in the common-view mode. The obtained difference between UTC(INPL) and GPS time scales is D2 = UTC(INPL) – GPS time. These comparisons are performed continuously, and their results are 11 reported to BIPM once a week. The time delays in the INPL GPS receiver, its antenna, antenna’s cables, and connecting cables are known with the uncertainty of ± 1 ns. Moreover, the corrections for ionospheric and tropospheric delays are also taken into account at the D2 value calculation [15]. The obtained values of D1 contain the time delay in the user’s GPS receiver, in its satellite antenna, in antenna’s cable, as well as the bias caused by user’s software errors. These factors lead to a deviation of the user’s GPS time scale (UUT time) from the universal GPS time D3 = GPS time – UUT time. Determination of the D3 value can be made by subtracting the difference D2 from the difference D1: D3 = D1 – D2 (8) For this purpose, D1 and D2 should be obtained for the same real time, in other words, the measurements in both channels should be performed simultaneously from the same beginning time point and for the same calibration time. Then the D1 and D2 values obtained are fitted by two linear regressions. The D3 value is calculated as the mean difference between these two linear regressions. An example of such fitting is shown in Fig. 3. Here the mean D3 value is 1.27 microsecond. The user’s GPS receiver time scale deviation from UTC is calculated by adding [UTC – GPS time] correction from Circular T to D3 value: UTC – UUT time = D3 + [UTC – GPS time] (9) The uncertainty of the user’s GPS receiver time scale deviation from the GPS time scale is: UD3 = k (u2D1 + u2D2 + u2C + u2L)1/2 , (10) where uD1 is the standard uncertainty of linear regression fit for D1; uD2 is the standard uncertainty of linear regression fit for D2; uC is the standard uncertainty of the connecting cable delay; uL is the standard uncertainty of the INPL GPS receiver’s delay, antenna’s delay, and antenna’s cable delay; k=2 is the coverage factor. 12 As a rule, the dominant contribution of the expanded uncertainty (10) is caused by a residual deviation of the user’s GPS receiver scale from a straight line, which is uD1 > 20 ns. For comparison, in the case in Fig. 3, uD2 is about 5 ns, uC and uL are about 1 ns each. Therefore, UD3 = 41 ns here. Conclusions Due to performance differences, two user’s GPS receivers can yield different frequency results even when connected to the same antenna at the same location. The frequency uncertainty for user’s GPS receiver is larger than that for Cs-standard by 2-3 orders for a short-time interval (1 – 1000 s), and by one order for a long-term interval of about two weeks. The user’s GPS receiver time scale bias from the Universal Coordinated Time can be in order of microseconds. It depends on the time delay in GPS receiver, satellite antenna, antenna’s cable, ionospheric correction error, and software errors. Commercial GPS receiver calibration service provided by INPL ensure measurement uncertainty of about 10-14 in the frequency scale, of about 5 ns in the time scale, as well as traceability to the Universal Coordinated Time. References [1] http://www.colorado.edu/geography/gcraft/notes/gps/gps_f.html Cited 27.12.04 [2] M.A.Lombardi, L.M.Nelson, A.N.Novick, V.S.Zhang, Time and Frequency Measurements Using the Global Positioning System, Cal. Lab. (Int. Jour. of Metrology), July-September 2001, pp. 26-33. [3] http://www.boulder.nist.gov/timefreq/service/gpscal.htm Cited 27.12.04 [4] M.A.Lombardi, Traceability in Time and Frequency Metrology, Cal. Lab. (Int. Jour. of Metrology), September-October 1999, pp. 33-40. [5] A.Lepek, A.Shenhar and D.W. Allan, UTC(INPL) – a virtual time scale, Metrologia, 1995/96, 32, pp. 245-252. [6] A.Lepek, A.Shenhar, M.Bezalel, The UTC(INPL) time scale, Proceedings of Ninth International Conference of the ISQA, 1992, pp. 502-506. 13 [7] A.Lepek, Atomic scale based on optimum linear prediction, IEEE Frequency Control Symposium, 1997, pp.382-387. [8] A.Lepek, Clock prediction and characterization, Metrologia, 1997, 34, pp.379-386. [9] http://www.geomatics.ucalgary.co/links/GradTheses.html Cited 27.12.04 [10] J.Ray, Mitigation of GPS code and carrier phase multipath effects using a multiantenna system, Ph.D. Thesis, Department of Geomatics Engineering, University of Calgary, Canada, 2000. [11] International Vocabulary of Basic and General Terms in Metrology, ISO, Geneva, 2004. [12] http://www.bipm.fr Cited 27.12.04 [13] Mutual Recognition of National Measurement Standards and of Calibration and Measurement Certificates Issued by National Metrology Institutes, BIPM, Paris, 1999. [14] N.Goldovsky, M.Luria, Frequency Standard and Oscillators Calibration at the National Physical Laboratory of Israel, Proceedings of the 14-th International Conference on the Israel Society for Quality, Jerusalem, November 2002, pp. 394399. [15] N.Goldovsky, M.Luria, Ionospheric delay contribution to the uncertainty of time and frequency measurement by one-way satellite time transfer method, Measurement, 35, 2004, pp. 353-362. [16] D.W.Allan, Statistics of Atomic Frequency Standards, Proc. IEEE, vol. 54, February 1966, pp. 221-230. [17] D.W.Allan, The Measurement of Frequency and Frequency Stability of Precision Oscillators, NBS Tech. Note 669, July 1975. [18] F.L.Walls and D.W.Allan, Measurements of Frequency Stability, Proc. of the IEEE, vol. 74, №1, January 1986, pp. 162-168. 14 UTC and SI Second (BIPM) D UTC(NMI) C Broadcast GPS signal monitored by NMI B User’s GPS receiver A Measurement instrument Fig. 1. The traceability chain for time and frequency measurements. 15 Monitor Printer Commutator Distrib Ampl. Microprocessor 5 MHz GPS Pulse Distrib Ampl. 1 pps start 1 pps stop 5 MHz 5 MHz 1 pps 5 MHz ICR INPL GPS Receiver 1 pps Cs-Standard PC1 5 MHz ICR 1 pps 5 MHz Time Interval Counter PC2 start UUT GPS Receiver stop 1 pps Generator (Optionally) Fig. 2. Block diagram of GPS receiver calibration system. 16 12.000 11.800 Microsecond 11.600 11.400 D1 = UTC (INPL) - UUT time 11.200 D2 = UTC (INPL) - GPS time 11.000 10.800 10.600 10.400 0 1 2 3 4 5 6 Days Fig. 3. The time difference D3 between the user’s GPS receiver and the GPS time scales.