I. INTRODUCTION Multistatic Radar Imaging of Moving Targets LING WANG, Member, IEEE MARGARET CHENEY, Member, IEEE Rensselaer Polytechnic Institute BRETT BORDEN Naval Postgraduate School We develop a linearized imaging theory that combines the spatial, temporal, and spectral aspects of scattered waves. We consider the case of fixed sensors and a general distribution of objects, each undergoing linear motion; thus the theory deals with imaging distributions in phase space. We derive a model for the data that is appropriate for narrowband waveforms in the case when the targets are moving slowly relative to the speed of light. From this model, we develop a phase-space imaging formula that can be interpreted in terms of filtered backprojection or matched filtering. For this imaging approach, we derive the corresponding phase-space point-spread function (PSF). We show plots of the phase-space point-spread function for various geometries. We also show that in special cases, the theory reduces to: 1) range-Doppler The use of radar for detection and imaging of moving targets is a topic of great interest [1—29]. It is well known that radar signals have two important attributes, namely the time delay, which provides information on the target range, and the Doppler shift, which can be used to infer target down-range velocity. Classical “radar ambiguity” theory [30—32] shows that the transmitted waveform determines the accuracy to which target range and velocity can be obtained from a backscattered radar signal. Another well-developed theory for radar imaging is that of synthetic-aperture radar (SAR) imaging. SAR combines high-range-resolution measurements from a variety of locations to produce high-resolution radar images [21, 33—35]. Such images, however, suffer when there is unknown relative motion between the target and radar platform; research addressing this includes [3], [7]—[22]. This research generally does not consider the possibility of using waveforms other than high-range-resolution ones. When high-Doppler-resolution waveforms are transmitted from a moving platform, the Doppler-shifted returns can also be used to form images of a stationary scene [36]. These three theories are depicted on the coordinate planes in Fig. 1. None of these theories address the possibility of forming high-resolution moving-target images in the case when different waveforms are transmitted from different spatial locations and received at spatially distributed receivers. Such a theory is needed to address questions such as: imaging, 2) inverse synthetic aperture radar (ISAR), 3) synthetic aperture radar (SAR), 4) Doppler SAR, and 5) tomography of moving targets. Manuscript received November 25, 2008; revised October 15, 2009; released for publication November 11, 2010. IEEE Log No. T-AES/48/1/943610. Refereeing of this contribution was handled by M. Rangaswamy. This work was supported by the Air Force Office of Scientific Research under Agreements FA9550-06-1-0017 and FA9550-09-1-0013. Ling Wang’s stay at Rensselaer was supported by the China Scholarship Council. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of the Air Force Research Laboratory or the U.S. Government. Authors’ addresses: L. Wang, Dept. of Information and Communication Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China; M. Cheney, Dept. of Mathematical Sciences, Rensselaer Polytechnic Institute, Amos Eaton Hall, 110 Eighth St., Troy, NY 12180, E-mail: (cheney@rpi.edu); B. Borden, Physics Dept., Naval Postgraduate School, Monterey, CA 93943. c 2012 IEEE 0018-9251/12/$26.00 ° 230 Fig. 1. Notional diagram showing how our work (depicted by question mark) relates to standard radar imaging methods. In a multistatic system, which transmitters should transmit which waveforms? How many transmitters are needed, and where should they be positioned? How can data from such a system be used to form an image of unknown moving targets? What is the resolution (in position and in velocity) of such a system? IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS VOL. 48, NO. 1 JANUARY 2012 TABLE I Table of Notation Some work has been done to develop such a theory: Ambiguity theory for bistatic systems has been developed in [34], [35], [37], [38]; such systems allow for estimation of only one component of the velocity vector. For multistatic systems, the work [23]—[29] developed methods for moving-target detection. Theory for use of a multistatic system for imaging of a stationary scene is well known; see e.g. [39]—[42]. Multistatic imaging of moving targets (phase-space imaging) was developed in [43] for the case of fixed transmitters and receivers (see Fig. 2). This theory combines spatial, temporal, and spectral attributes of radar data; in particular the theory exploits the actual Doppler shift from moving targets. In appropriate special cases, this theory reduces to the standard imaging methods shown in Fig. 1. The work [43] is extended in the present paper, which sets forth the basic ideas in the special cases that are of interest to radar-based imaging, and undertakes an initial numerical exploration of the properties of the imaging system. In particular, we show that under some circumstances, this approach allows for both (spatial) image formation and also estimation of the full vector velocities of multiple targets. In Section II we develop a physics-based mathematical model that incorporates not only the waveform and wave propagation effects due to moving scatterers, but also the effects of spatial diversity of transmitters and receivers. In Section III we show how a matched-filtering technique produces images that depend on target velocity, and we relate the resulting point-spread function (PSF) to the classical radar ambiguity function. We show four-dimensional plots of the PSF for three different geometries. In Section IV we note that the general theory of Section III reduces to familiar results, including classical ambiguity theory, inverse SAR (ISAR), SAR, and Doppler SAR. Finally, in Section V we list some conclusions. Table I gives a list of notations used throughout the paper. II. MODEL FOR DATA We model wave propagation and scattering by the scalar wave equation for the wavefield Ã(t, x) due to a source waveform s̃(t, x): [r2 ¡ c¡2 (t, x)@t2 ]Ã(t, x) = s̃(t, x): (1) For example, a single isotropic source located at y transmitting the waveform s(t) starting at time t = ¡Ty could be modeled as s̃(t, x) = ±(x ¡ y)sy (t + Ty ), where the subscript y reminds us that transmitters at different locations could transmit different waveforms. For simplicity, in this discussion we consider only Symbol Designation x t y sy (t) Position (three-dimensional vector) Time (scalar) Transmitter position (three-dimensional vector) Waveform transmitted by a transmitter located at y ¡Ty Starting time of the waveform sy !y Carrier frequency of the transmitted waveform sy (t) ky = !y =c Wavenumber of the transmitted waveform sy ay (t) Slowly varying part of sy v Velocity (three-dimensional vector) of a moving scatterer Distribution of scatters with position x and velocity v at time t = 0 Doppler scale factor for a scatter at position x moving with velocity v. Wavefield at time t and position x Free-space field at x generated by a source at y Receiver position (three-dimensional vector) x ¡ z, three-dimensional vector from z to x qv (x) ®x,v Ã(t, x) à in (t, x, y) z Rx,z ¹z,v à sc (t, z, y) 1 + R̂x,z ¢ v=c ®x,v Scattered field at the receiver location z due to the field transmitted from position y Doppler scale factor !y ¯x,v (Angular) Doppler shift p Position (three-dimensional vector) in the reconstructed scene Velocity (three-dimensional vector) of object in reconstructed scene Reconstruction of reflectivity of objects moving with velocity u Geometry-dependent weighting function for improving the image quality Point spread function (PSF) Radar ambiguity function for the waveform used by the transmitter located at y Angle of reconstructed position vector p Angle of reconstructed velocity u u Iu J K Ay μ Á localized isotropic sources; the work can easily be extended to more realistic antenna models [44]. A single scatterer moving at velocity v corresponds to an index-of-refraction distribution n2 (x ¡ vt): c¡2 (t, x) = c0¡2 [1 + n2 (x ¡ vt)] (2) where c0 denotes the speed of light in vacuum. We write qv (x ¡ vt) = c0¡2 n2 (x ¡ vt). To model multiple moving scatterers, we let qv (x ¡ vt)d3 x d3 v be the corresponding quantity for the scatterers in the volume d3 x d3 v centered at (x, v). Thus qv is the distribution in phase space, at time t = 0, of scatterers moving with velocity v. Consequently, the scatterers at x give rise to Z (3) c¡2 (t, x) = c0¡2 + qv (x ¡ vt)d3 v: We note that the physical interpretation of qv involves a choice of time origin. The choice that is particularly appropriate, in view of our assumption about linear target velocities, is a time during which WANG, ET AL.: MULTISTATIC RADAR IMAGING OF MOVING TARGETS 231 Fig. 2. Sensing geometry. Antennas are fixed and illuminate a scene in which there are multiple moving targets. the wave is interacting with targets of interest. This implies that the activation of the antenna at y takes place at a negative time which we denote by ¡Ty ; as mentioned above, we write s̃(t, x) = sy (t + Ty )±(x ¡ y). The wave equation corresponding to (3) is then ¸ · Z 2 ¡2 2 3 2 r ¡ c0 @t ¡ qv (x ¡ vt)d v @t Ã(t, x) = sy (t + Ty )±(x ¡ y): (4) From this point on, we drop the subscript on c0 , so that c denotes the speed of light in vacuum. In the absence of scatterers, the field from the antenna is sy (t + Ty ¡ jx ¡ yj=c) à in (t, x, y) = ¡ : (5) 4¼jx ¡ yj (Details are given in the Appendix.) We write à = à in + à sc and neglect multiple scattering (i.e., use the weak-scatterer model) to obtain the expression for the scattered field à sc : Z Z ±(t ¡ t0 ¡ jz ¡ xj=c) à sc (t, z, y) = qv (x ¡ vt0 )d3 v @t20 4¼jz ¡ xj sy (t0 + Ty ¡ jx ¡ yj=c) 3 £ d x dt0 : 4¼jx ¡ yj (6) (For details of the derivation, see the Appendix.) Many radar systems use a narrowband waveform, which in our case is defined by sy (t) = ay (t)e¡i!y t (7) where ay (t) is slowly varying, as a function of t, in comparison with exp(¡i!y t), where !y is the carrier frequency for the transmitter at position y. For such waveforms, the second time derivative of sy obeys s̈y (t) ¼ ¡!y2 sy (t). (For a justification of this and the other approximations made here, see the Appendix; for a derivation without this approximation, see [43].) 232 In (6), we make the change of variables x 7! x0 = x ¡ vt0 , (i.e., change the frame of reference to one in which the scatterer qv is fixed) to obtain Z Z 0 0 0 0 à sc (t, z, y) = ¡ ±(t ¡ t ¡ jx + vt ¡ zj=c) 4¼jx0 + vt0 ¡ zj qv (x ) 4¼jx0 + vt0 ¡ yj £ !y2 sy (t0 + Ty ¡ jx0 + vt0 ¡ yj=c)d 3 x0 d 3 v dt0 : (8) The physical interpretation of (8) is as follows: The wave that emanates from y at time ¡Ty encounters a target at time t0 ; this target, during the time interval [0, t0 ], has moved from x0 to x0 + vt0 ; the wave scatters with strength qv (x0 ), and then propagates from position x0 + vt0 to z, arriving at time t. Henceforth we drop the primes on x. Next we assume that the target is slowly moving. More specifically, we assume that jvjt and kjvj2 t2 are much less than jx ¡ zj and jx ¡ yj, where k = !max =c, with !max the (effective) maximum angular frequency of all the signals sy . In this case, expanding jz ¡ (x + vt0 )j around t0 = 0 yields jz ¡ (x + vt0 )j ¼ Rx,z + R̂x,z ¢ vt0 (9) where Rx,z = x ¡ z, Rx,z = jRx,z j, and R̂x,z = Rx,z =Rx,z . We can carry out the t0 integration in (8) as follows: we make the change of variables t00 = t ¡ t0 ¡ (Rx,z + R̂x,z ¢ vt0 )=c (10) which has the corresponding Jacobian ¯ 0¯ ¯ @t ¯ 1 1 ¯ ¯ : ¯ @t00 ¯ = j@t00 =@t0 j = 1 + R̂ ¢ v=c x,z We denote the denominator of this Jacobian as ¹z,v . The argument of the delta function in (8) contributes only when t00 = 0; (10) with t00 = 0 can be solved for t0 to yield t ¡ Rx,z =c : (11) t0 = 1 + R̂x,z ¢ v=c IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS VOL. 48, NO. 1 JANUARY 2012 The argument of sy in (8) is then jx + vt0 ¡ yj c à ! Rx,y R̂x,y ¢ v 0 0 ¼ t + Ty ¡ + t c c t0 + Ty ¡ = μ ¶ Rx,y Rx,z t¡ ¡ + Ty (12) c c ¢ v=c 1 ¡ R̂x,y ¢ v=c 1 + R̂x,z where we have used (9) and (11). The quantity ®x,v = 1 ¡ R̂x,y ¢ v=c 1 + R̂x,z ¢ v=c ¼ 1 ¡ (R̂x,y + R̂x,z ) ¢ v=c (13) is the Doppler scale factor. If we write the Doppler scale factor (13) as ®x,v ¼ 1 + ¯x,v , where ¯x,v = ¡(R̂x,y + R̂x,z ) ¢ v=c (14) then the quantity !y ¯x,v is the (angular) Doppler shift. We note that the Doppler shift depends on the bistatic bisector vector R̂x,y + R̂x,z (see Fig. 3). With (12) and the notation of (13), (8) becomes à sc (t, z, y) = ¡!y2 Z sy [®x,v (t ¡ Rx,z =c) ¡ Rx,y =c + Ty ] (4¼)2 Rx,z Rx,y ¹z,v qv (x)d3 x d 3 v: (15) We use the narrowband assumption (7) to write the numerator of (15) as sy [®x,v (t ¡ Rx,z =c) ¡ Rx,y=c + Ty ] ¼ ay (t ¡ Rx,z =c ¡ Rx,y =c + Ty ) £ expf¡i!y [(1 + ¯x,v )(t ¡ Rx,z =c) ¡ Rx,y =c + Ty ]g (16) where we have used the fact that ay is slowly varying to replace ®x,v by 1. With (16), (15) becomes Z !y2 ei'x,v e¡i!y ®x,v t à sc (t, z, y) ¼ ¡ (4¼)2 Rx,z Rx,y ¹z,v £ ay (t + Ty ¡ (Rx,z + Rx,y )=c)qv (x)d3 x d3 v: (17) Here we have collected the time-independent terms in the exponent into the quantity 'x,v = ky [Rx,y ¡ Ty + ®˜ x,v Rx,z ] (18) where ky = !y =c. We recognize (17) as the superposition of time-delayed and Doppler-shifted copies of the transmitted waveform. The key feature of this model is that it correctly incorporates the positions of the transmitters and receivers, the transmitted waveform, and the target position and velocity. In particular, the Fig. 3. Bistatic range (sum of distances Rx,z and Rx,y ), and bistatic bisector R̂x,y + R̂x,z . time delay and Doppler shift depend correctly on the target and sensor positions and on the target velocity. III. IMAGING We now address the question of extracting information from the scattered waveform described by (15). Our image formation algorithm depends on the data we have available. In general, we form a (coherent) image as a filtered adjoint or weighted matched filter [44], but the variables we integrate over depend on whether we have multiple transmitters, multiple receivers, and/or multiple frequencies. The number of variables we integrate over, in turn, determines how many image dimensions we can hope to reconstruct. If our data depend on two variables, for example, then we expect to be able to reconstruct a two-dimensional scene; in this case the scene is commonly assumed to lie on a known surface. If, in addition, we want to reconstruct two-dimensional velocities, then we are seeking a four-dimensional phase-space image and thus we expect to need data depending on at least four variables. More generally, the problem of determining a distribution of positions in space and the corresponding three-dimensional velocities means that we seek to form an image in a six-dimensional phase space. A. Imaging Formula We form an image Iu (p) of the objects with velocity u that, at time t = 0, were located at position p. Thus Iu (p) is constructed to be an approximation to qu (p). We form an image by matched filtering and summing over all transmitters and receivers. From (17) with qv (x) = ±(x ¡ p)±(v ¡ u), we see that the predicted signal from a single point target at the phase-space position (p, u) (i.e., a target at position p traveling with velocity u) is ¡ WANG, ET AL.: MULTISTATIC RADAR IMAGING OF MOVING TARGETS !y2 ei'p,u e¡i!y ®p,u t a (t + Ty ¡ (Rp,z + Rp,y )=c): (4¼)2 Rp,z Rp,y ¹z,u y (19) 233 We denote the transmitters by y1 , y2 , : : : , yM , and the receivers by z1 , z2 , : : : , zN . To obtain a phase-space image Iu (p), we apply the (weighted) matched filter corresponding to the phase-space location (p, u) to the received signal à sc : N X M Z X In the second exponential of (22) we use (18) to write 'x,v ¡ 'p,u = kym [Rx,ym ¡ Tym + (1 + ¯x,v )Rx,zn ] ¡ kym [Rp,ym ¡ Tym + (1 + ¯p,u )Rp,zn ] = kym [¡c¢¿ym ,zn + ¯x,v Rx,zn ¡ ¯p,u Rp,zn ] e¡i'p,u ei!ym ®p,u t Rp,zn Rp,ym ¹zn ,u Iu (p) = (4¼)2 n=1 m=1 (23) where £ a¤ym (t + Tym ¡ (Rp,zn + Rp,ym )=c)à sc (t, zn , ym )Jn,m dt c¢¿ym ,zn = [Rp,zn + Rp,ym ] ¡ [Rx,zn + Rx,ym ] (24) (20) where Jn,m is a geometry-dependent weighting function that can be inserted to improve the image [44]. The weights Rp,z Rp,y ¹z,u are included so that they will cancel the corresponding factors in the data when p = x and u = v. In allowing the imaging operation (20) to depend on the transmitter ym , we are implicitly assuming that, at the receiver located at zn , we can identify the part of the signal à sc (t, zn , ym ) that is due to the transmitter at ym . In the case of multiple transmitters, this identification can be accomplished, for example, by having different transmitters operate at different frequencies, or possibly by quasi-orthogonal pulse-coding schemes. We are also assuming a coherent system, i.e., that a common time clock is available to all sensors, so that we are able to form the image (20) with the correct phase relationships. The operation (20) amounts to geometry-corrected and phase-corrected matched filtering with a time-delayed, Doppler-shifted version of the transmitted waveform, where the Doppler shift is defined in terms of (14). is the difference in bistatic ranges for the points x and p. If in (22) we make the change of variables t 7! t0 = t + Tym ¡ (Rx,zn + Rx,ym )=c and use the notation (26), then we obtain B. Image Analysis We note that the bistatic ambiguity function of (25) depends on the difference of bistatic ranges (24) and on the difference between the velocity projections onto the bistatic bisectors ¯p,u ¡ ¯x,v = (R̂x,ym + R̂x,zn ) ¢ v=c ¡ (R̂p,ym + R̂p,zn ) ¢ u=c. For the case of a single transmitter and single receiver (a case for which J1,1 = 1) and a point target (which enables us to choose Tym so that the exponential in the third line of (25) vanishes), and leaving aside the magnitude and phase corrections, (25) reduces to a coordinate-independent version of the ambiguity function of [38]. In order to characterize the behavior of this imaging system, we substitute (17) into (20), obtaining Z K(p, u; x, v)qv (x)d3 x d3 v Iu (p) = (21) where K, the PSF is K(p, u; x, v) = ¡ N X M Z X !y2m a¤ym n=1 m=1 £ (t + Tym ¡ (Rp,zn + Rp,ym )=c)ei!ym [¯p,u ¡¯x,v ]t £ aym (t + Tym ¡ (Rx,zn + Rx,ym )=c)ei['x,v ¡'p,u ] £ Rp,zn Rp,ym ¹zn ,u J dt: Rx,zn Rx,ym ¹zn ,v n,m (22) The PSF characterizes the behavior of the imaging system in the sense that it contains all the information about how the true phase-space reflectivity distribution qv (x) is related to the image Iu (p). If qv (x) is a delta-like point target at (x, v), then the PSF K(p, u; x, v) is the phase-space image of that target. 234 K(p, u; x, v) = ¡ M N X X n=1 m=1 !y2m Aym (!ym [¯p,u ¡ ¯x,v ], ¢¿ym ,zn ) £ expf¡ikym ¯p,u [Rx,zn ¡ Rp,zn ]g £ expfi!ym [¯p,u ¡ ¯x,v ][Rx,ym =c ¡ Tym ]g £ Rp,zn Rp,ym ¹zn ,u J Rx,zn Rx,ym ¹zn ,v n,m where ˜ ¿ ) = e¡i!y ¿ Ay (!, Z (25) ˜ dt a¤y (t ¡ ¿ )ay (t)ei!t (26) is the radar ambiguity function for the waveform ay . (Note that this definition includes a phase.) Equation (25) says the following: The multistatic phase-space PSF is a weighted coherent sum of radar ambiguity functions evaluated at appropriate bistatic ranges and bistatic velocities. C. Examples of the Point-Spread Function The PSF contains all the information about the performance of the imaging system. Unfortunately it is difficult to visualize this PSF because it depends on so many variables. In the case when the positions and velocities are restricted to a known plane, the PSF is a function of four variables. We would like to know whether we can find both the position and velocity of moving targets. Ideally, the PSF is delta-like, and so we can obtain IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS VOL. 48, NO. 1 JANUARY 2012 Fig. 4. This shows how Figs. 5—7 display the four-dimensional PSF (27). both position and velocity. If, however, the PSF is ridge-like, then there will be uncertainty in some directions or in some combination of positions and velocities. In order to look for possible ridge-like behavior, we write the PSF as K(p, u;x, v) = K(jpj(cos μ, sin μ), juj(cos Á, sin Á), x, v): (27) We plot the PSF for a fixed target position x and target velocity v. We then sample μ and Á at intervals of ¼=4, and for each choice of μ and Á, we plot jpj versus juj. This process results in 9 £ 9 = 81 plots of jpj versus juj. Finally, to show the entire four-dimensional space at a glance, we display all the 81 plots simultaneously on a grid, arranged as shown in Fig. 4. 1) Simulation Parameters: Our strategy in the simulations is to use a delta-like ambiguity function, and investigate the effect of geometry on the overall PSF. In all cases, we use a transmit time of Ty = 0. a) Waveform: In the simulations, we use a high-Doppler-resolution, linearly-frequency-modulated (LFM) pulse train. The ambiguity function of this pulse train has a “bed-of-nails” appearance [32], with a delta-like central peak that has both velocity resolution and range resolution that are good enough for many applications. We assume that the extraneous peaks of the ambiguity function occur in a region of space and velocity where no targets are present. In our simulations, we choose the duration of each pulse to be Tp = 10 £ 10¡6 s, the pulse period to be TR = 10¡4 s, and take N = 50 pulses in the pulse train so that the duration of the entire pulse train is T = NTR = 5 £ 10¡3 s. The bandwidth B of the pulse train is chosen to be B = 3 £ 106 Hz. The FM rate is B=Tp . We take the center frequency to be 30 GHz, which corresponds to a wavelength of ¸ = 0:01 m. The ambiguity function of this pulse train has the following properties. ¡1 1) The Doppler resolution is 1=T = 200 s , which in the ordinary monostatic case would correspond to down-range velocity resolution of ¢V = ¸=(2T) = 1 m/s. 2) The first ambiguous Doppler value is 1=TR = 104 s¡1 , which would correspond to a monostatic velocity ambiguity of ¢Vmax = ¸=(2TR ) = 50 m/s. 3) The delay resolution is 1=B ¼ :3 £ 10¡6 s, which would correspond to a monostatic range resolution of ¢r = c=(2B) = 50 m. 4) The first ambiguous delay is TR = 10¡4 s, which corresponds to an ambiguous (monostatic) range of ¢rmax = cTR =2 = 1:5 £ 104 m. This would be the maximum measurable unique range. b) Area of interest: The area of interest is a circular region with radius of 1000 m; points within this region differ by no more than 2000 m, which is well within the unambiguous range of 15000 m. We assume that there are no scatterers outside the region of interest. The location of every point in the region is denoted by the vector p = jpj(cos μ, sin μ). The directions μ are sampled at intervals of ¼=4, while the lengths r are sampled at intervals of 25 m. c) Velocities of Interest: The velocities are written u = juj(cos Á, sin Á). We consider velocity magnitudes in the interval [0, 30] m/s. The magnitudes s are sampled at intervals of 0.5 m/s and the directions Á are sampled at intervals of ¼=4. 2) Examples: a) Single transmitter, two receivers: The PSF for a single transmitter and two receivers is shown in Fig. 5. Here the transmitter is located at y = (0, ¡10000) m, the receivers are located at z1 = (10000, 0) m and z2 = (¡10000, 0) m, the scene of interest is a disk centered at the origin with radius 1000 m, the target location is 225(cos 45± , sin 45± ) m, and the target velocity is 20(cos 0, sin 0) m/s. We see that for this geometry, the PSF is indeed ridge-like. b) Circular geometry: Fig. 6 shows the PSF for a circular arrangement of 8 transmitters and 10 receivers. The transmitters are equally spaced around a circle of radius 10000 m; the receivers are equally spaced around a circle of radius 9000 m. The scene of interest has radius 1000 m. The true target location is 225(cos 180± , sin 180± ) m and the true velocity is 20(cos 180± , sin 180± ) m/s. For this geometry, there appear to be no ambiguities. c) Linear array: Fig. 7 shows the PSF for a linear array of 11 transmitters and a single receiver. In this case, the transmitters are equally spaced along the line ¡5000 m · x · 5000 m, y = 10000 m; the receiver is located at (0, 10000) m; the true target location is 225(cos 135± , sin 135± ) m and the target velocity is 20(cos 45± , sin 45± ) m/s. IV. REDUCTION TO FAMILIAR SPECIAL CASES The formula (25) reduces correctly to a variety of special cases. More details can be found in [43]. A. Range-Doppler Imaging In classical range-Doppler imaging, the range and velocity of a moving target are estimated from measurements made by a single transmitter and WANG, ET AL.: MULTISTATIC RADAR IMAGING OF MOVING TARGETS 235 Fig. 5. This shows geometry (not to scale) for one transmitter and two receivers, together with combined PSF. coincident receiver. Thus we take z = y, J = 1, and remove the integrals in (25) to show that the PSF is simply proportional to the classical ambiguity function. 1) Range-Doppler Imaging of a Rotating Target: If the target is rotating about a point, then we can use the connection between position and velocity to form a range-Doppler target image. For the case of turntable geometry, where for a point (x1 , x2 , 0), the _ , with μ_ range is x1 and the down-range velocity is μx 2 the target rotation rate, (25) reduces to the classical à ! _ !0 μ(p (p1 ¡ x1 ) 2 ¡ x2 ) KRD (p, u; x, v) / Ay 2 ,2 c c (28) where !0 is the center frequency. B. Inverse Synthetic-Aperture Radar ISAR systems rely on the target’s rotational motion to establish a pseudoarray of distributed receivers, from which range measurements are made. To show how ISAR follows from the results of this paper, we fix the coordinate system to the rotating target. Succeeding pulses of a pulse train 236 are transmitted from and received at positions z = y that rotate around the target as ŷ = ŷ(μ). The pulses typically used in ISAR are high range-resolution (HRR) ones, so that each pulse yields a ridge-like ambiguity function Ay (º, ¿ ) = Āy (¿ ), with Āy (¿ ) sharply peaked around ¿ = 0. Using this ambiguity function in (25) shows that the imaging process corresponds to backprojection [43]. C. Synthetic-Aperture Radar SAR also uses HRR pulses, transmitted and received at locations along the flight path y = y(³), where ³ denotes the flight path parameter. For a stationary scene, u = v = 0, which implies that ¯u = ¯v = 0. In this case, (25) corresponds to backprojection over circles [45]. See Fig. 8. D. Doppler SAR Doppler SAR uses a high Doppler-resolution waveform, such as an approximation to the ideal single-frequency waveform s(t) = exp(¡i!0 t) transmitted from locations along the flight path y = y(³). If we transform to the sensor frame of reference, IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS VOL. 48, NO. 1 JANUARY 2012 Fig. 6. This shows geometry (not to scale) for 8 transmitters and 10 receivers arranged in a circle and corresponding combined PSF. then the entire scene is moving with velocity u = v = _ ¡y(³) (see Fig. 9). For a high Doppler-resolution waveform, Ay (º, ¿ ) = Āy (º), where Āy (º) is sharply peaked around º = 0. Using this waveform in (25) shows that the imaging process corresponds to backprojection over hyperbolas, as detailed in [36]. E. SAR for Other Ridge-like Ambiguity Functions In the above scenario, SAR reconstruction can be performed with a sensor moving along the path y = y(³) while transmitting other waveforms. In particular, waveforms (such as chirps) can be used which produce a ridge-like ambiguity function along any line º = ´¿ in the delay-Doppler plane [32]. Then (25) corresponds to backprojection over the planar curves Lp (³): _ + ´Rp,y(³) : Lp (³) = ¡!0 R̂p,y(³) ¢ y(³) (29) The curve Lp (³) is the intersection of the imaging plane with a surface generated by rotating a limaçon of Pascal (see Fig. 10) about the flight velocity vector. We note that for a slowly-moving sensor and a side-looking beam pattern, the curve obtained from the limaçon is almost indistinguishable from a circle. F. Moving Target Tomography Moving target tomography has been proposed in [23], which models the signal using a simplified version of (17). For this simplified model, our imaging formula (20) reduces to the approach advocated in [23], namely matched-filter processing with a different filter for each location p and for each possible velocity u. V. CONCLUSIONS AND FUTURE WORK We have developed a linearized imaging theory that combines the spatial, temporal, and spectral aspects of scattered waves. This imaging theory is based on the general (linearized) expression we derived for waves scattered from moving objects, which we model in terms of a distribution in phase space. We form a phase-space image as a weighted matched filter. WANG, ET AL.: MULTISTATIC RADAR IMAGING OF MOVING TARGETS 237 Fig. 7. This shows geometry (not to scale) for linear array of 11 transmitters and single receiver and corresponding combined PSF. Fig. 8. SAR scenario: Stationary scene and sensor that (in start-stop approximation) does not move as it transmits and receives each pulse. Between pulses, sensor moves to different location along flight path. The theory allows for activation of multiple transmitters at different times, but the theory is simpler when they are all activated so that the waves arrive at the target at roughly the same time. 238 The general theory in this paper reduces to familiar results in a variety of special cases. We leave for the future further analysis of the PSF. We also leave for the future the case in which the sensors are moving IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS VOL. 48, NO. 1 JANUARY 2012 Fig. 9. For Doppler SAR, we use frame of reference in which sensor is fixed and plane below is moving with uniform velocity. Fig. 10. A limaçon of Pascal. Origin is located at transmitter position y. independently, and the problem of combining these ideas with tracking techniques. APPENDIX A. Details of the Derivation of (5) and (6) 1) The Transmitted Field: We consider the transmitted field à in in the absence of any scatterers; à in satisifies the version of (1) in which c is simply the constant background speed c0 , namely [r2 ¡ c0¡2 @t2 ]à in (t, x, y) = sy (t + Ty )±(x ¡ y): (30) Henceforth we drop the subscript on c. We use the Green’s function g, which satisfies WANG, ET AL.: MULTISTATIC RADAR IMAGING OF MOVING TARGETS [r2 ¡ c¡2 @t2 ]g(t, x) = ¡±(t)±(x) 239 and is given by g(t, x) = ±(t ¡ jxj=c) : 4¼jxj (31) For the validity of (16), we use the Taylor expansion of ay about the time (t ¡ Rx,z =c) ¡ tx , where tx = Rx,y =c ¡ Ty : ay [®x,v (t ¡ Rx,z =c) ¡ tx ] We use (31) to solve (30), obtaining Z à in (t, x, y) = ¡ g(t ¡ t0 , jx ¡ yj)sy (t0 + Ty )±(x ¡ y)dt0 dy sy (t + Ty ¡ jx ¡ yj=c) =¡ : 4¼jx ¡ yj (32) 2) The Scattered Field: We write à = à in + Ã̃ sc , which converts (1) into Z (33) [r2 ¡ c¡2 @t2 ]Ã̃ sc = qv (x ¡ vt)d3 v @t2 Ã: The map q 7! Ã̃ sc can be linearized by replacing the full field à on the right side of (33) by à in . (This is the Born or single-scattering approximation.) The resulting differential equation we solve with the help of the Green’s function; the result is Z à sc (t, z) = ¡ g(t ¡ t0 , jz ¡ xj) Z £ qv (x ¡ vt0 )d3 v @t20 à in (t0 , x)dt0 d3 x (34) which is (6). B. Validity of Approximations used to Obtain the Signal Model (17) = ay [(t ¡ Rx,z =c) ¡ tx ] + a_ y [(t ¡ Rx,z =c) ¡ tx ] £ (®x,v ¡ 1)(t ¡ Rx,z =c): Since ®x,v ¡ 1 = O(v=c), we see that (16) is valid when a_ y is bounded and jv=cj ¿ 1. REFERENCES [1] [2] [3] [4] [5] [6] 1) Narrowband Waveforms: For a narrowband waveform sy (t, y) = ay (t, y)e¡i!y t (35) [7] the second time derivative of sy is s̈y (t) = ¡!y2 sy (t) ¡ 2i!y a_ y e¡i!y t + äy e¡i!y t à ! 2ia_ y äy 2 ¡i!y t = ¡!y e + 2 : ay ¡ !y !y [8] (36) We see that the approximation s̈y (t) ¼ ¡!y2 sy (t) is valid when a_ y =!y and äy =!y2 are negligible compared with ay . 2) Slow-Mover Approximation: We expand R(t0 ) = jz ¡ (x + vt0 )j around t0 = 0, yielding 0 _ + O((t0 )2 ): R(t ) = R(0) + R(0)t 0 [9] [10] (37) Dividing by c, we obtain 0 ¯ ¯2 1 ¯ v 0¯ t v 0 jz ¡ (x + vt )j Rx,z B¯c ¯ C = + R̂x,z ¢ t + O @ A: c c c Rx,z =c 0 [11] (38) 0 Thus we see that (9) is valid when jv=cjt ¿ 1; this is also the same condition needed for validity of (12) and (13). 240 (39) [12] Skolnik, M. I. Radar Handbook (3rd ed.). New York: McGraw-Hill Professional, 2008. Brennan, L. E. and Reed, I. S. Theory of adaptive radar. IEEE Transactions on Aerospace and Electronic Systems, AES-9 (Mar. 1973), 237—252. Raney, R. K. Synthetic aperture imaging radar and moving targets. IEEE Transactions on Aerospace and Electronic Systems, AES-7, 3 (May 1971), 499—505. Ward, J. Space-time adaptive processing for airborne radar. MIT Lincoln Laboratory, Lexington, MA, Technical Report 1015, 1994. Klemm, R. Principles of Space-Time Adaptive Processing. London: IEE Publishing, 2002. Cooper, J. Scattering of electromagnetic fields by a moving boundary: The one-dimensional case. IEEE Transactions on Antennas and Propagation, AP-28, 6 (1980), 791—795. Werness, S., et al. Moving target imaging for SAR data. IEEE Transactions on Aerospace and Electronic Systems, 26, 1 (Jan. 1990), 57—66. Yang, H. and Soumekh, M. Blind-velocity SAR/ISAR imaging of a moving target in a stationary background. IEEE Transactions on Image Processing, 2, 1 (Jan. 1993), 80—95. Barbarossa, S. Detection and imaging of moving objects with synthetic aperture radar–Part 1: Optimal detection and parameter estimation theory. Proceedings of the IEE-F, Radar and Signal Processing, 139 (Feb. 1992), 79—88. Barbarossa, S. Detection and imaging of moving objects with synthetic aperture radar–Part 2: Joint time-frequency analysis by Wigner-Ville distribution. Proceedings of the IEE-F, Radar and Signal Processing, 139 (Feb. 1992), 89—97. Friedlander, B. and Porat, B. VSAR: A high resolution radar system for detection of moving targets. IEE Proceedings on Radar, Sonar and Navigation, 144 (1997), 205—218. Perry, R. P., Dipietro, R. C., and Fante, R. L. SAR imaging of moving targets. IEEE Transactions on Aerospace and Electronic Systems, 35, 1 (Jan. 1999), 188—200. IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS VOL. 48, NO. 1 JANUARY 2012 [13] Jao, J. K. Theory of synthetic aperture radar imaging of a moving target. IEEE Transactions on Geoscience and Remote Sensing, 39, 9 (Sept. 2001), 1984—1992. [28] Landi, L. and Adve, R. S. Time-orthogonal-waveform-space-time adaptive processing for distributed aperture radars. In Proceedings of the 2007 International Waveform Diversity and Design Conference, 2007, 13—17. [14] Fienup, J. R. Detecting moving targets in SAR imagery by focusing. IEEE Transactions on Aerospace and Electronic Systems, 37, 3 (July 2001), 794—809. [29] Capraro, G. T., et al. Waveform diversity in multistatic radar. In Proceedings of the International Waveform Diversity and Design Conference, Lihue, HI, Jan. 2006. [15] Dias, J. M. B. and Marques, P. A. C. Multiple moving target detection and trajectory estimation using a single SAR sensor. IEEE Transactions on Aerospace and Electronic Systems, 39, 2 (Apr. 2003), 604—624. [30] Cook, C. E. and Bernfeld, M. Radar Signals. New York: Academic, 1965. [31] Woodward, P. M. Probability and Information Theory, with Applications to Radar. New York: McGraw-Hill, 1953. [32] Levanon, N. Radar Principles. Hoboken, NJ: Wiley, 1998. [33] Franceschetti, G. and Lanari, R. Synthetic Aperture Radar Processing. Boca Raton, FL: CRC Press, 1999. [16] Kircht, M. Detection and imaging of arbitrarily moving targets with single-channel SAR. IEE Proceedings on Radar, Sonar and Navigation, 150, 1 (Feb. 2003), 7—11. [17] Pettersson, M. I. Detection of moving targets in wideband SAR. IEEE Transactions on Aerospace and Electronic Systems, 40 (2004), 780—796. [34] [18] Minardi, M. J., Gorham, L. A., and Zelnio, E. G. Ground moving target detection and tracking based on generalized SAR processing and change detection. Proceedings of SPIE, vol. 5808, 2005, 156—165. Willis, N. J. Bistatic Radar. Norwood, MA: Artech House, 1991. [35] Barbarossa, S. and Farina, A. Space-time-frequency processing of synthetic aperture radar signals. IEEE Transactions on Aerospace and Electronic Systems, 30, 2 (Apr. 1998), 341—358. Willis, N. J. Bistatic radar. In M. I. Skolnik, Radar Handbook (2nd ed.), New York: McGraw-Hill, 1990. [36] [20] Soumekh, M. Fourier Array Imaging. Upper Saddle River, NJ: Prentice-Hall, 1994. Borden, B. and Cheney, M. Synthetic-aperture imaging from high-Doppler-resolution measurements. Inverse Problems, 21 (2005), 1—11. [37] [21] Soumekh, M. Synthetic Aperture Radar Signal Processing with MATLAB Algorithms. Hoboken, NJ: Wiley-Interscience, 1999. Willis, N. J. and Griffiths, H. D. Advances in Bistatic Radar. Raleigh, NC: SciTech Publishing, 2007. [38] Tsao, T., et al. Ambiguity function for a bistatic radar. IEEE Transactions on Aerospace and Electronic Systems, 33 (1997), 1041—1051. [22] Stuff, M., et al. Imaging moving objects in 3d from single aperture synthetic aperture data. In Proceedings of the IEEE Radar Conference, 2004, 94—98. [39] Colton, D. and Kress, R. Inverse Acoustic and Electromagnetic Scattering Theory. New York: Springer, 1992. [40] Devaney, A. J. Inversion formula for inverse scattering within the Born approximation. Optics Letters, 7 (1982), 111—112. [41] Devaney, A. J. A filtered backpropagation algorithm for diffraction tomography. Ultrasonic Imaging, 4 (1982), 336—350. [42] Varsolt, T., Yazici, B., and Cheney, M. Wide-band pulse-echo imaging with distributed apertures in multi-path environments. Inverse Problems, 24, 4 (Aug. 2008). [43] Cheney, M. and Borden, B. Imaging moving targets from scattered waves. Inverse Problems, 24 (2008), 1—22. [44] Nolan, C. J. and Cheney, M. Synthetic aperture inversion. Inverse Problems, 18 (2002), 221—236. [45] Nolan, C. J. and Cheney, M. Synthetic aperture inversion for arbitrary flight paths and non-flat topography. IEEE Transactions on Image Processing, 12 (2003), 1035—1043. [19] [23] [24] [25] [26] [27] Himed, B., et al. Tomography of moving targets (TMT). Sensors, Systems, and Next-Generation Satellites, Proceedings of SPIE, vol. 4540, 2001, 608—619. Bradaric, I., et al. Multistatic radar systems signal processing. In Proceedings of the IEEE Radar Conference, 2006, 106—113. Bradaric, I., Capraro, G. T., and Wicks, M. C. Waveform diversity for different multistatic radar configurations. In Proceedings of the Asilomar Conference, 2007. Adve, R. S., et al. Adaptive processing for distributed aperture radars. In Proceedings of 1st Annual IEE Waveform Diversity Conference, Edinburgh, Nov. 2004. Adve, R. S., et al. Waveform-space-time adaptive processing for distributed aperture radars. In 2005 IEEE International Radar Conference Record, 2005, 93—97. WANG, ET AL.: MULTISTATIC RADAR IMAGING OF MOVING TARGETS 241 Ling Wang received the B.S. degree in electrical engineering, and the M.S. and Ph.D. degrees in information acquirement and processing from Nanjing University of Aeronautics and Astronautics, Nanjing, China, in 2000, 2003, and 2006, respectively. In 2003, she joined Nanjing University of Aeronautics and Astronautics, where she is currently an associate professor in the Department of Information and Communication Engineering. She was a postdoctoral research associate in the Department of Mathematical Sciences and Department of Electrical, Computer, and Systems Engineering, Rensselaer Polytechnic Institute, Troy, NY, from February 2008 to May 2009. Her main research interest is radar imaging. Margaret Cheney received the B.A. in mathematics and physics from Oberlin College, Oberlin, OH, in 1976 and the Ph.D. degree in mathematics from Indiana University, Bloomington, in 1982. She held positions at Stanford University and Duke University before moving to Rensselaer Polytechnic Institute, where she is a professor of mathematics. Most of her work has been in inverse problems, including those that arise in quantum mechanics, acoustics, and electromagnetic theory. She has been working specifically in radar imaging since 2001. Brett Borden received the B.S. degree in physics from the University of Wisconsin—Madison in 1978 and the Ph.D. degree in physics from the University of Texas at Austin in 1986. From 1980 to 2002 he worked in the Research Department of the Naval Air Warfare Center Weapons Division in China Lake, CA. In 2002, he moved to the Physics Department at The Naval Postgraduate School. 242 IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS VOL. 48, NO. 1 JANUARY 2012