JOINTLY ESTIMATING 3D TARGET SHAPE AND MOTION FROM RADAR DATA By Heather Palmeri A Thesis Submitted to the Graduate Faculty of Rensselaer Polytechnic Institute in Partial Fulfillment of the Requirements for the Degree of DOCTOR OF PHILOSOPHY Major Subject: MATHEMATICS Approved by the Examining Committee: Margaret Cheney, Thesis Adviser Joseph Mayhan, Member William Siegmann, Member David Isaacson, Member Rensselaer Polytechnic Institute Troy, New York November 2012 (For Graduation December 2012) c Copyright 2012 by Heather Palmeri All Rights Reserved ii CONTENTS LIST OF TABLES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi LIST OF FIGURES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii ACKNOWLEDGMENT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi ABSTRACT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii 1. Introduction and Historical Review . . . . . . . . . . . . . . . . . . . . . . 1 1.1 Inverse Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.2 Radar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.3 Synthetic Aperture Radar (SAR) . . . . . . . . . . . . . . . . . . . . 2 1.3.1 . . . . . . . . . . . . . . . 2 1.4 Inverse Synthetic Aperture Radar (ISAR) . . . . . . . . . . . . . . . 3 1.5 Interferometric Synthetic Aperture Radar (IFSAR) . . . . . . . . . . 3 1.6 Scatterer Correlation . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.7 Scatterer Correlation Techniques for Image Processing . . . . . . . . 4 1.8 3D Image Processing Techniques for Target Characterization . . . . . 5 2. Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 Stripmap SAR and Spotlight SAR 2.1 Electromagnetic Wave Propagation . . . . . . . . . . . . . . . . . . . 8 2.2 Scatterers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 2.3 Radar Observables Background . . . . . . . . . . . . . . . . . . . . . 11 2.3.1 Obtaining Raw Data . . . . . . . . . . . . . . . . . . . . . . . 11 2.3.2 Sequential Block Processing . . . . . . . . . . . . . . . . . . . 13 2.4 Sample Target . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 2.5 Assumptions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 2.6 Scatterer Correlation . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 2.7 The Shape and Motion Estimation Problem . . . . . . . . . . . . . . 22 2.8 Euler’s Dynamical Equations of Torque-Free Motion . . . . . . . . . . 23 2.8.1 2.9 Special Case: Turntable Geometry . . . . . . . . . . . . . . . 25 Simulating Truth Radar Data . . . . . . . . . . . . . . . . . . . . . . 26 2.10 Resolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 iii 2.11 Matrix Decompositions . . . . . . . . . . . . . . . . . . . . . . . . . . 27 2.11.1 The Singular Value Decomposition (SVD) . . . . . . . . . . . 27 2.11.2 The Reduced QR Factorization . . . . . . . . . . . . . . . . . 28 2.12 Basic Problem Formulation . . . . . . . . . . . . . . . . . . . . . . . 29 3. Phase-Enhanced 3D Snapshot Imaging Algorithm . . . . . . . . . . . . . . 31 3.1 2D Snapshot Imaging Equations . . . . . . . . . . . . . . . . . . . . . 32 3.2 3D Snapshot Imaging Equations . . . . . . . . . . . . . . . . . . . . . 32 3.3 Phase Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 3.3.1 Derivation from Snapshot Equations . . . . . . . . . . . . . . 33 3.3.2 Derivation From Range-Doppler Data Block . . . . . . . . . . 35 3.4 Phase-Enhanced 3D Snapshot Imaging Equations . . . . . . . . . . . 37 3.5 Augmenting Additional Roll Cuts . . . . . . . . . . . . . . . . . . . . 39 3.6 Phase-Enhanced 3D Snapshot Imaging Considerations . . . . . . . . . 40 3.6.1 Change in Roll . . . . . . . . . . . . . . . . . . . . . . . . . . 40 3.6.2 Coordinate System . . . . . . . . . . . . . . . . . . . . . . . . 41 3.6.3 Implementation . . . . . . . . . . . . . . . . . . . . . . . . . . 42 3.6.4 Resolving Phase Ambiguities . . . . . . . . . . . . . . . . . . . 43 3.7 Summing Target Space Images . . . . . . . . . . . . . . . . . . . . . . 45 3.8 Sector Processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 4. SVD Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49 4.1 Derivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49 4.2 Methodology Summary . . . . . . . . . . . . . . . . . . . . . . . . . . 56 4.3 Geometric Interpretation . . . . . . . . . . . . . . . . . . . . . . . . . 57 5. Phase-Enhanced SVD Method . . . . . . . . . . . . . . . . . . . . . . . . . 58 5.1 Scaling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 5.2 Sample Motion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61 5.3 Imperfect Correlation . . . . . . . . . . . . . . . . . . . . . . . . . . . 62 5.4 5.3.1 Simulating Imperfect Correlation . . . . . . . . . . . . . . . . 62 5.3.2 Effect of Imperfect Correlation . . . . . . . . . . . . . . . . . . 65 Added Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68 5.4.1 Simulating Added Noise . . . . . . . . . . . . . . . . . . . . . 68 5.4.1.1 The Percentage Method. . . . . . . . . . . . . . . . . 68 iv 5.4.1.2 5.4.2 The Extraneous Points Method. . . . . . . . . . . . . 70 Effect of Added Noise . . . . . . . . . . . . . . . . . . . . . . 72 5.4.2.1 The Percentage Method. . . . . . . . . . . . . . . . . 72 5.4.2.2 The Extraneous Points Method. . . . . . . . . . . . . 73 6. Correlation Algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 6.1 6.2 6.3 6.4 Iterative Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 6.1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 6.1.2 Large Number of Scatterers Without Noise . . . . . . . . . . . 77 6.1.3 Large Number of Scatterers With Noise . . . . . . . . . . . . . 85 6.1.4 Large Number of Scatterers With Phase Ambiguities Without Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 6.1.5 Large Number of Scatterers With Phase Ambiguities, Noise, and Filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 Nearest Neighbor and Slope Algorithm . . . . . . . . . . . . . . . . . 94 6.2.1 Small Number of Scatterers Without Noise . . . . . . . . . . . 94 6.2.2 Small Number of Scatterers With Noise and Phase Ambiguities 97 Proof of Theorem Justifying Iterative Algorithm . . . . . . . . . . . . 106 6.3.1 Case Two Scatterers Switch . . . . . . . . . . . . . . . . . . . 109 6.3.2 Case More than Two Scatterers Switch . . . . . . . . . . . . . 124 6.3.3 Case Scatterers Switch at Multiple Aspects . . . . . . . . . . . 133 Comparison of Algorithms . . . . . . . . . . . . . . . . . . . . . . . . 135 7. Discussion and Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . 138 7.1 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138 7.2 Main Contributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140 7.3 Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140 LITERATURE CITED . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142 v LIST OF TABLES 4.1 SVD Method Variables and Their Dimensions . . . . . . . . . . . . . . 56 5.1 3D Imaging Method Differences . . . . . . . . . . . . . . . . . . . . . . 58 vi LIST OF FIGURES 1.1 2D and 3D Target Shape and Motion Estimation Techniques Overview . 7 2.1 Transmitted and Received Pulses . . . . . . . . . . . . . . . . . . . . . 9 2.2 Coordinate System Diagram . . . . . . . . . . . . . . . . . . . . . . . . 11 2.3 Turntable Geometry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 2.4 RTI Plot for Simulated Data . . . . . . . . . . . . . . . . . . . . . . . . 14 2.5 Fixed Aspect Data Block . . . . . . . . . . . . . . . . . . . . . . . . . . 15 2.6 Extracted Ranges and Range-Rates . . . . . . . . . . . . . . . . . . . . 16 2.7 Sinusoidal Nature of Range and Range-Rate . . . . . . . . . . . . . . . 16 2.8 Sample RV Target Geometry . . . . . . . . . . . . . . . . . . . . . . . . 17 2.9 Simulated Sample RV Target Points . . . . . . . . . . . . . . . . . . . . 17 2.10 Scatterer Correlation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 2.11 Examples of Uncorrelated and Correlated Data Matrices . . . . . . . . . 21 2.12 Relationship Between Motion, Shape, and Radar Observables . . . . . . 22 2.13 Target Centered Coordinate System . . . . . . . . . . . . . . . . . . . . 24 2.14 Phase-Enhanced 3D Snapshot Imaging method . . . . . . . . . . . . . . 29 2.15 SVD Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 2.16 Phase-Enhanced SVD Method . . . . . . . . . . . . . . . . . . . . . . . 30 3.1 Localized Orthogonal Coordinate System for 3D Phase-Enhanced Processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 3.2 Measurement Collection Method . . . . . . . . . . . . . . . . . . . . . . 40 3.3 Apparent Motion of Target Between Two Roll Cuts . . . . . . . . . . . 41 3.4 a) Reconstructed Positions, b) ±2π Phase Ambiguities c) Target Space Filtered Image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 3.5 Phase Resolved Target Points, Three 2D Views . . . . . . . . . . . . . . 44 3.6 Target Space Reconstruction, 3D & Three 2D Views . . . . . . . . . . . 45 vii 3.7 Sector Processing, Pink Areas Correspond to Data Used . . . . . . . . . 47 3.8 Comparison: Sector Processing for Various Amounts of Data . . . . . . 48 3.9 Target Space Reconstruction Using Sector Processing . . . . . . . . . . 48 5.1 Shape and Motion Solutions . . . . . . . . . . . . . . . . . . . . . . . . 61 5.2 Varying Levels of Scatterer Correlation . . . . . . . . . . . . . . . . . . 62 5.3 Uncorrelated Extracted Radar Observables, p = 80 . . . . . . . . . . . 63 5.4 Uncorrelated Extracted Radar Observables . . . . . . . . . . . . . . . . 63 5.5 Uncorrelated Truth Radar Observables with Random Jumps . . . . . . 64 5.6 Shape and Motion Estimates for Varying Correlation Percentage Levels 5.7 Singular Value Distributions for Varying Correlation Percentage Levels . 66 5.8 Shape and Motion Estimates for Varying Correlation With Jumps . . . 67 5.9 Singular Value Distributions for Correlation With Varying Numbers of Jumps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67 5.10 Truth Radar Observables with p = 15 . . . . . . . . . . . . . . . . . . . 69 5.11 Truth Radar Observables with Extraneous Points . . . . . . . . . . . . 71 5.12 Shape and Motion Estimates for Varying Percentage Levels of White Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72 5.13 Singular Value Distributions for Varying Noise Percentages . . . . . . . 73 5.14 Shape and Motion Estimates for Varying Levels of Extraneous Point Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74 5.15 Singular Value Distributions for Varying Percentages of Extraneous Points 75 6.1 13 Scatterer Data Uncorrelated with Jumps . . . . . . . . . . . . . . . . 77 6.2 2 and 3 Jump Cases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81 6.3 Matrix C of All Possible Correlations . . . . . . . . . . . . . . . . . . . 82 6.4 Singular Value Distributions from 13 Scatterer Data . . . . . . . . . . . 83 6.5 13 Scatterer Data Uncorrelated with Jumps then Correlated with Iterative Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84 6.6 13 Scatterer Data Uncorrelated with Jumps and Extraneous Points . . . 85 viii 65 6.7 Singular Value Distributions, Noisy Data . . . . . . . . . . . . . . . . . 86 6.8 Data Uncorrelated with Jumps and Noise then Correlated with Iterative Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87 6.9 Uncorrelated Data With Phase Ambiguities . . . . . . . . . . . . . . . . 88 6.10 Delta Phase after Phase Constraint . . . . . . . . . . . . . . . . . . . . 89 6.11 Singular Value Distributions from 13 Scatterer Data With Phase Ambiguities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 6.12 Delta Phase after Iterative Algorithm Using Phase Constraint 6.13 Uncorrelated Data With Phase Ambiguities and Noise . . . . . . . . . . 91 6.14 Delta Phase With Noise after Phase Constraint . . . . . . . . . . . . . . 91 6.15 Singular Value Distributions from 13 Scatterer Data With Phase Ambiguities and Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92 6.16 Delta Phase with Noise after Iterative Algorithm Using Phase Constraint and Filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93 6.17 Radar Observables Correlated with Iterative Algorithm . . . . . . . . . 93 6.18 Uncorrelated Range Data . . . . . . . . . . . . . . . . . . . . . . . . . . 95 6.19 Range Data After Step 1 . . . . . . . . . . . . . . . . . . . . . . . . . . 96 6.20 Range Data After Step 2 . . . . . . . . . . . . . . . . . . . . . . . . . . 96 6.21 Uncorrelated Range Data with Extraneous Points . . . . . . . . . . . . 97 6.22 Range Data After Step 1 . . . . . . . . . . . . . . . . . . . . . . . . . . 98 6.23 Identifying an Intersection . . . . . . . . . . . . . . . . . . . . . . . . . 99 6.24 Analyzing an Intersection . . . . . . . . . . . . . . . . . . . . . . . . . . 100 6.25 Range Data After Step 3 . . . . . . . . . . . . . . . . . . . . . . . . . . 102 6.26 Range Data After Step 4 . . . . . . . . . . . . . . . . . . . . . . . . . . 103 6.27 Delta Phase Data Before Step 5 . . . . . . . . . . . . . . . . . . . . . . 104 6.28 Delta Phase Data After Step 5 . . . . . . . . . . . . . . . . . . . . . . . 104 6.29 Track Displaced Due to Phase Ambiguity . . . . . . . . . . . . . . . . . 105 6.30 Radar Observables Correlated Through Algorithm . . . . . . . . . . . . 105 ix . . . . . 90 6.31 Singular Value Distribution for 4 Scatterer Target . . . . . . . . . . . . 136 6.32 Singular Value Distribution for 5 Scatterer Target . . . . . . . . . . . . 137 x ACKNOWLEDGMENT First and foremost, I’d like to thank my advisor, Dr. Margaret Cheney, for all her help, and for getting me interested in the field of radar to begin with. Along those lines, I’d like to thank the RPI School of Science Accelerated B.S./Ph.D. Program. It was through this program that I was encouraged to start research during my sophomore year of college. I have Dr. Cheney and the program to thank for me being given the opportunity to dive in to research and summer internships in my field much earlier than many students are often able to. I’d like to thank Dr. William Siegmann, for recognizing my potential and accepting me in to this program to begin with. Dr. Cheney was wonderful in encouraging me to participate in summer internships and conferences. I’d like to thank Kevin Magde, my supervisor from my first summer internship at the Air Force Research Laboratory in Rome, NY. This was my first hands-on experience with radar, and was a great learning experience for someone starting out in the field. As a mathematics major, I often was asked ”What do you DO with math, teach?” and wasn’t quite sure how to answer that question. My first internship is what first gave me a real answer to that question. I was able to take radar measurements using a static range radar chamber at their facility and use mathematics to get information about targets out of this data. This was my first real-world experience in mathematics and inspired me to continue in the field. I’d like to thank the people I’ve worked with at MIT Lincoln Laboratory. I was able to spend three summer internships and two winter breaks working there. From where I started in group 34, I’d like to thank my supervisor, Allen Dors, and other members of the group, including Doug Koltenuk and Tim Bond. I’d also like to thank my officemates and fellow interns in the group, Andy O’Donnell and Kendra Kumley, for helping to make my experience at the lab an enjoyable one. I switched over to group 36 at MITLL in order to work with Dr. Joseph Mayhan on a project I thought would be interesting, which ended up leading in to my thesis topic. I’d like to thank Joe first of all for agreeing to take me on as an xi intern, then for his help and support throughout the rest of my internships, and then agreeing to be the external member of my doctoral committee. I’d like to thank Keh-Ping Dunn for his support and help in the group as well, and for taking time out of his busy schedule to meet with me and Joe often. Various other members in the group were very helpful in explaining concepts to me, including Noah Kubow, Shirin Kubat, and my officemates Elena Rossi and Su-May Hsu. My fellow interns in this group, specifically Jess Stigile and Adam Attarian, were great as well. As a graduate student at RPI, I was lucky enough to be a member of a great research group of students also working with Dr. Cheney. Through working with them, we formed some great friendships. A shout out is due to the original radar girls - Kaitlyn Voccola, Analee Miranda, and Tegan Webster. We were later joined by Tom Zugibe and Scott Altrichter, who weren’t exactly radar GIRLS, but we let that slide. Our radar group was lucky enough to join forces with some of the other students in RPI’s Inverse Problems Center (IPRPI) in sharing an office, so a shout out is due to them as well - Ashley Baer, Jessica Jones, and Joe Rosenthal. You guys are awesome! I’d like to thank my boyfriend, Ricky Grebe, for being there for me during pretty much my entire time in college. From the get-go with my research, he’s been supportive and even willing to listen to and critique my presentations. I’d like to thank my family for their love and support during my time in college. Even though I only moved about twenty-five minutes away from home to go to college, I was often still pretty unavailable since I was so busy with my schoolwork. They were always very understanding and supportive. My parents were always ready to help me move from apartment to apartment (eight moves in total during college) and didn’t get TOO annoyed with me for always leaving various assortments of furniture, boxes, and other belongings around the house when I couldn’t take them with me. Lastly, I’d like to thank my doctoral committee as a whole; Dr. Margaret Cheney, Dr. William Siegmann, Dr. David Isaacson, and Dr. Joseph Mayhan. They have always been very helpful and encouraging. I’ve been so lucky to have the support of all those listed above in my time as a student at RPI. Thanks to all of you. xii ABSTRACT A novel hybrid 3D radar imaging technique is presented that jointly estimates both target shape and motion using range, range-rate, and phase. This work expands on research done by the author as an intern at MIT Lincoln Laboratory. It builds on and combines the work of two papers: Phase-Enhanced 3D Snapshot ISAR Imaging and Interferometric SAR (Joseph Mayhan) and Shape and Motion Reconstruction from 3D-to-1D Orthographically Projected Data via Object-Image Relations (Matthew Ferrara). The second paper is a modification to work first presented in Derivation and Estimation of Euclidean Invariants of Far Field Range Data (Mark Stuff). The phase-enhanced 3D snapshot imaging algorithm solves for shape using known motion and uncorrelated range, range-rate, and phase data. The second method uses an SVD to jointly solve for shape and motion using correlated range data. Key features from each of these methods are incorporated in to the novel hybrid phase-enhanced 3D SVD method. Two algorithms are presented that eliminate the need for scatterer correlation so that the hybrid method can be used on uncorrelated radar data. One algorithm, applicable to targets with a small number of scatterers, methodically determines the optimal correlation for a set of data using continuity and slope conditions. This algorithm can be used in the presence of noise and phase ambiguities. The other algorithm, applicable to targets with a large number of scatterers, iterates on an optimally chosen set of possible correlations and chooses the “best” one based on a condition on the resulting singular values. This algorithm can also be used in the presence of noise and phase ambiguities. A mathematical proof is presented to show that a matrix of radar observables data is uncorrelated if and only if it has more than three nonzero singular values. This proof justifies the use of the iterative algorithm. xiii CHAPTER 1 Introduction and Historical Review 1.1 Inverse Problems Radar imaging a subset of a broader category of problems known as inverse problems. A forward problem can loosely be thought of as finding the solution to an equation. An inverse problem can be thought of as the “inverse” to this problem, i.e., given a solution, find the equation it resulted from. This process is in general not unique, as multiple equations can often yield the same solution [1]. Many problems in mathematics and science fall in to this category. Some examples are geophysics, medical imaging, and ocean acoustics. In general, these problems involve collecting data and using it to determine something about the source the data came from. 1.2 Radar The application of radar investigated here is the identification of flying objects. This is important in many areas. For example, the Federal Aviation Administration (FAA) uses radar for airport surveillance and air traffic control. Radar is of critical importance in missile defense applications. It isn’t enough to just know a target is there; oftentimes radar must be used to discriminate between actual missiles, decoys, and debris, all in real time. This is not a simple problem. There are many advantages to using radar over other imaging modalities. For one, radars can be used day or night, since they have their own form of illumination. They can image through scenes obscured by clouds and rain, since radio waves are only slightly attenuated by these conditions [2]. The term ‘radar’ is actually an acronym. It stands for ‘RAdio Detection And Ranging’, and dates back to 1940. In general, radar works by transmitting radio waves that travel to a target and reflect back. They are transmitted in a sequence of pulses at regularly spaced intervals in time. From the resulting time-delayed received pulses, certain properties of the target can be determined, such as range, 1 2 altitude, size, material properties, and more [3] [2]. A radar system includes a transmitter, a receiver, an antenna (with a pointing or steering mechanism), and a signal processing and display unit. Oftentimes the transmitter and receiver are at the same location, or “colocated”. The most easily determined property is the range of the target, since it is simply a scaled version of the time delay between the transmission of a pulse and its return after reflection off an object [2]. 1.3 Synthetic Aperture Radar (SAR) Synthetic Aperture Radar (SAR) was developed in 1951 by Carl Wiley of the Goodyear Aircraft Corporation. SAR systems involve a moving antenna and a stationary target. These systems are desirable because they form high-resolution images, and have been used to image terrain, including other planets and their moons. The movement of the antenna allows the synthetic aperture of the system to be greater than its actual aperture, which is how higher resolution is achieved. There are two main varieties of SAR, spotlight SAR and stripmap SAR [3] [2]. 1.3.1 Stripmap SAR and Spotlight SAR In stripmap SAR, the antenna is in a fixed position on the radar platform. This results in the radar viewing a strip of terrain parallel to the path of motion. In spotlight SAR, as the radar platform moves, the radar stares at one specific location. In this way, the same target is viewed from a different direction at each point in the flight path [2]. There are three main differences between stripmap SAR and spotlight SAR. • Spotlight SAR yields finer azimuth resolution than stripmap, using the same physical antenna. • Spotlight SAR allows a scene to be imaged at multiple viewing angles during a single pass. • Spotlight SAR efficiently images multiple smaller scenes, while stripmap SAR is more useful for a long strip of terrain. [2] 3 1.4 Inverse Synthetic Aperture Radar (ISAR) Inverse Synthetic Aperture Radar (ISAR) systems involve a stationary antenna and a moving target. The movement of the target, instead of the movement of the antenna, is then what creates the synthetic aperture for high-resolution imaging. Both SAR and ISAR use the similar underlying theory, since both involve the same relative motion between the radar and the target [4]. ISAR processing is very useful for identifying and classifying targets. The resulting images reveal the dominant scattering centers on the target. They are displayed in range-Doppler, or range and cross-range, coordinates. Range is the axis parallel to the direction of propagation from the radar toward the target, also known as slant range, and cross-range is the direction perpendicular to range [4]. Oftentimes, the radar observables range and range-rate are used for image processing. Range-rate is a scaled version of Doppler. It is interesting to note the similarity between ISAR and spotlight SAR. By focusing on one specific target from many aspects in spotlight SAR, a geometry very similar to ISAR results. 1.5 Interferometric Synthetic Aperture Radar (IFSAR) A SAR image inherently contains more information than an optical image. An optical image shows the amplitude of reflected light at each location. A SAR image shows the amplitude of reflected radio waves at each location, but additionally stores phase information at each location. Like optical images, SAR images are twodimensional. Oftentimes it is desirable to obtain a three-dimensional image in order to gain more information about the target [4]. Interferometric Synthetic Aperture Radar (IFSAR) compares images of a scene from different elevations and compares phase to extract information about the third dimension and hence form a three-dimensional image. This was first demonstrated by Graham in 1974. This method is widely used to make elevation maps of terrain. It can be done using two different setups; an airborne radar platform carrying a two-radar system, or a single antenna radar platform making repeat passes over the same area [4]. 4 1.6 Scatterer Correlation Scatterer correlation, or signal isolation, is the process of separating the signals that arise from each of the geometrically distinct locations, or scatterers, on the target. If a set of data is organized such that information from separate scatterers is distinguishable, we call it correlated. Otherwise, it is uncorrelated. Some imaging schemes require that the radar observables data be correlated, and some do not. In general, actual radar data will always be uncorrelated. For a more detailed explanation of scatterer correlation, see Section 2.6. 1.7 Scatterer Correlation Techniques for Image Processing The process of scatterer correlation is complicated. We discuss the history of various methods. • Prominent Point Methods: In the 1980s, W. Carrara, S. Werness, and others were able to successfully isolate a few of the strongest signals from localized regions of targets. This was done using interactive image editing tools and required an enormous amount of persistence to execute [2]. • Fully Automated Approach: In the 1990s, a fully automated approach to scatterer correlation was developed. The method assigned a likelihood ratio score to points in the range data and an optimization algorithm was then used to estimate the correlation. These estimates were then iteratively updated using least squares fits to a linearized model. This method worked well in all cases except when scatterers’ range tracks intersected [5]. • Three-Dimensional Dynamic Signal Analysis: In 1999, Mark Stuff published an approach that improves on the fully automated approach. The method exploits the fact that if scatterers intersect in range, they cannot simultaneously intersect in velocity, the derivative of range. The method uses a three-dimensional complex time-range-velocity (CTRV) distribution that “separates” the data so that no scatterers’ paths intersect in three-dimensional space. A dynamic programming algorithm is then used to extract the correctly correlated signals [5]. 5 1.8 3D Image Processing Techniques for Target Characterization Most conventional radar imaging schemes are two-dimensional. This is because radar data is generally taken in two dimensions, so much more work must be done to gain a third dimension. We review some of the general methods currently used to construct 3D images from wideband radar data: • Backprojection: At each aspect angle, backproject the data to all possible target locations that are the correct distance from the radar, where the distance is calculated using the recorded time delay. Coherently summing all contributions results in an image with the target appearing at the correct location. This is a commonly used method [3]. • Microlocal Analysis: The key idea is that singularities in the data lead to singularities in the resulting image. Microlocal analysis uses the concept of a wavefront set to help describe singularities and decipher which target characteristics they resulted from. This is a relatively new and unexplored area [6]. • 3D Snapshot Imaging: Extract radar observables (range, range-rate, amplitude, phase) for each observation, use known motion to solve general [Motion]x[Target]=[Radar Observables] equation for target scatterer locations. This method can solve for 2D target shape at one roll angle given one snapshot, or 3D target shape at one roll angle given two or more snapshots and scatterer correlation [7]. • Phase-Enhanced 3D Snapshot Imaging: This method is very similar to 3D snapshot imaging, but uses data from two snapshots taken at different roll angles. The phase difference between the snapshots is used as a third independent variable in order to solve for 3D target shape, given a priori motion information, without the need for scatterer correlation. This is a new method that is not yet widely used [8]. 6 • SVD of Correlated Observation Data: After performing centroid removal (essentially normalizing all target ranges), perform an SVD of the correlated range data matrix. This requires making the rigid-body assumption, that is, the assumption that nothing on the target itself is moving. After some manipulation, the matrices resulting from the SVD can be used to arrive at both shape and motion estimates. This method is not yet widely used. This was done using both range and range-rate by Tomasi and Kanade in 1991 [9]. It was done separately for only range by Stuff for his PhD thesis in 2002 [10]. Stuff’s method was modified by Ferrara in 2009 [11]. Of these methods, the two focused on here are the Phase-Enhanced 3D Snapshot Imaging Method and the SVD Method. It is important to notice some of the key differences between them. The SVD method requires scatterer correlation and solves for both shape and motion, while the phase-enhanced 3D snapshot imaging method does not require scatterer correlation, but solves for only shape. An overview of how these methods are related is shown in Figure 1.1. We next give much more detail on these methods, then explore their relation to each other and how they can be combined. 7 Figure 1.1: 2D and 3D Target Shape and Motion Estimation Techniques Overview CHAPTER 2 Background 2.1 Electromagnetic Wave Propagation We begin with Maxwell’s equations in the time domain. ∇·D=ρ (2.1) ∇·B=0 (2.2) ∇×E =− ∂B ∂t ∇×H=J+ (2.3) ∂D ∂t (2.4) Here D is the electric displacement field, B is the magnetic induction field, E is the electric field, H is the magnetic intensity or magnetic field, J is the current density, and ρ is the charge density. Radar waves are electromagnetic waves and can be assumed to propagate in free space. The following relations and simplifications can be made in free space. B = µ0 H D = 0 E ρ=0 J=0 By using the above relations, taking the curl of (2.3), and plugging the result in to (2.4), we arrive at ∇ × ∇ × E = −µ0 0 ∂ 2E . ∂t2 Then since ∇ × (∇ × E) = ∇(∇ · E) − ∇2 E and (2.1) in free space becomes ∇ · E = 0, we arrive at ∇2 E = µ0 0 8 ∂ 2E . ∂t2 9 So in free space, Maxwell’s equations reduce to the scalar wave equation for each component of the electric field vector E [3]. In a radar system, a pulse is transmitted at time t = 0. The pulse travels to the target, bounces off of it, then returns to the radar at time t = τ . The quantity τ is known as the time delay. An illustration of this is given in Figure 2.1. Figure 2.1: Transmitted and Received Pulses Assume that the target is a distance R from the transmitter. Then by time t = τ , the pulse has traveled a distance 2R. The pulse is estimated to travel at the speed of light, c, so the time delay τ can be calculated by τ = 2R/c. Here we assume the radar is stationary relative to the target, but in general, either the radar or the target can be moving, so we let R = R(s) to show this distance can change. If the transmitted signal is assumed to be a delta function, the received signal can be thought of as a time-delayed delta function. The equation of the received signal is then given by E(t, s) = Dδ(t − 2R(s)/c), where D measures the signal’s amplitude, and we have plugged in for the time delay τ . Notice that the received signal is given in two different time scales. The time variable t is on the order of the speed of light, describing the fast transmission of electromagnetic waves. This is known as fast time. The time variable s describes the relatively slow movement between the radar and target. This is known as slow time. Fourier transforming the received signal from time to frequency gives 10 Z E(ω, s) = Dδ(t − 2R(s)/c)e−iωt dω, which, using a well-known property of the delta function, simplifies to E(ω, s) = Dei2ωR(s)/c . (2.5) As can be seen, the received signal takes the form of a complex exponential in the frequency domain. 2.2 Scatterers A scatterer (or point scatterer or scattering center) is a distinct point on a target, such as a corner of a fin or the tip of a cone. A commonly used way to model targets to be imaged is the point scatterer model. Targets are modeled as a collection of discrete point scatterers. There are two main types of scatterers; fixed scatterers and slipping scatterers. Fixed scatterers are distinct points on the target that stay at fixed locations. As a radar views the target from different aspects, fixed scatterers appear to move since their distance from the radar is changing. On the other hand, slipping scatterers always appear to be at the same distance to the radar. A common example of a slipping scatterer is a sphere. When a sphere is viewed by a radar, the point on the sphere closest to the radar gives the biggest return. This is true no matter what aspect the sphere is viewed at. This scatterer located closest to the radar is known as a slipping scatterer. Its distance to the radar appears to remain constant, while its location on the target “slips”. Another example of a slipping scatterer a groove on a cone-shaped target. Slipping scatters are extremely difficult to image, and most imaging schemes that even attempt to take them in to account impose a filter on the data before processing to get rid of them. Due to this, the only scatterers considered here will be fixed scatterers. 11 2.3 2.3.1 Radar Observables Background Obtaining Raw Data We begin with a target centered coordinate system. The origin of the co- ordinate system is located at the center of gravity of the target. The unit look vector from the origin to the radar is denoted by k̂(s). The radar views the target at L different aspects. At the mth aspect, we have k̂(s) = k̂m . There are N distinct three-dimensional scatterers on the target, and the nth scatterer is denoted by rn = (xn , yn , zn ). The distance from the radar to each scatterer is given by R, where R is the sum of an overall range offset R0 and a normalized range Rmn such that k̂m · rn = Rmn . Figure 2.2: Coordinate System Diagram A typical way to collect radar data is in turntable geometry, as depicted in Figure 2.3. The target is located at a fixed origin and the radar moves in a counterclockwise fashion around the target. At each aspect, the radar sends and receives a sequence of electromagnetic pulses at increasing frequencies. A radar taking data at increasing but distinct frequencies, such as this, is often referred to as a stepped frequency system. The time it takes the radar to move around the target, stopping at each aspect, is on a different time scale than the time it takes for the radar to send and receive pulses at each aspect. These differing time scales are the same as the ones introduced in Section 2.1. The motion of the radar around the target is often described as slow 12 time, while fast time refers to the time taken to send and receive pulses. This results in radar data being taken in two variables, frequency and slow time. Figure 2.3: Turntable Geometry Another common way to think of ”turntable” geometry is to imagine that the target is mounted on a turntable in front of a stationary radar. Data taken in static ranges is often taken in this geometry, so it is also referred to as static range geometry. As mentioned in Section 2.1 (and is easily deduced from Figure 2.2), the radar receives the time-delayed pulse at time t = 2R(s)/c. Since the received signal includes a return from each discrete scatterer, the received signal is given by summing 2.5 over each scatterer, E(ω, s) = X Dn ei2ωRn (s)/c . n This expression gives the scattered field data E(ω, s) in frequency and slow time. Plugging in R(s) = k̂(s) · r gives E(ω, s) = X Dn ei2ωk̂(s)·rn /c , (2.6) n where the nth scatterer has complex amplitude Dn and is located at position rn on the target. Since aspect is a function of slow time, k̂(s) = k̂m denotes the radar line-of- 13 sight unit vector at the mth aspect. The frequency ω is given by ω = 2πc , λ where λ is the wavelength corresponding to frequency ω. Then the scattered field data can also be expressed as E(λ, m) = X 4π Dn ei λ k̂m rn . (2.7) n This expression gives the scattered field data E(λ, m) in wavelength and aspect. This is consistent with the Geometric Theory of Diffraction (GTD) applicable to high frequency regions. It is important to note that the target motion associated with this data is necessarily two-dimensional. This data is taken with the target at a fixed “roll” angle. A change in roll occurs when the target is rotated in the y − z plane. 2.3.2 Sequential Block Processing An image of the raw data E(ω, s) does not visually convey useful information. In order to extract useful information from this data, it can first be processed as follows. Some background for the following was found in [12]. Fourier transforming the received signal (2.6) from frequency ω to fast time t gives E(t, s) = X = X = X Z ei2ωk̂(s)·rn /c e−iωt dω Z e−iω(t−2k̂(s)·rn /c) dω Dn n Dn n Dn δ(t − 2k̂(s) · rn /c). (2.8) n This gives the received signal in the form of a time-delayed delta function, as is expected from the development in Section 2.1. We note that the introduction of the delta function assumes the signal model is valid for −∞ < ω < ∞, even though data is actually taken over only a selected frequency band. Two commonly known radar observables are range and range-rate. These are the observables that will be used here. Consequently, it will be useful to express the received signal above in terms of range, R, instead of time, t. We use the relation 14 t = 2R/c to obtain the expression E(R, s) = X n 2 Dn δ( (R − k̂(s) · rn )). c (2.9) Here we note the abuse of notation in again using E. The received signal is now given in range and slow time. Plotting E(R, s) for all R and s gives a Range-Time Intensity (RTI) plot, as shown in Figure 2.4. Figure 2.4: RTI Plot for Simulated Data Inverse Fourier transforming E(R, s) transforms slow time to Doppler frequency, ν. This is done for each fixed slow time value, s0 . Since it wouldn’t be useful to Fourier transform data at just one point, a “block” of data centered at each fixed slow time value, s0 , is used for each transform. The block is formed by multiplying the data by a“window”, W , that zeros out all the data not included in the block. The length of the window can be thought of as similar to the length of time the shutter is open on a camera. If it’s too long, the image will be blurred. The images here were made with a window length of 64 pulses. Inverse Fourier transforming the windowed data results in I(s0 , R, ν) = X n Z Dn 2 δ( (R − k̂(s) · rn ))Ws0 (s)eisν ds, c (2.10) so that the data are now in fast time and Doppler frequency. Again, it will be more useful to represent data in range-rate. Doppler frequency, ν, is related to range-rate, 15 Ṙ, by ν = 4π Ṙ, λ so we have I(s0 , R, Ṙ) = X n Z Dn 4π 2 δ( (R − k̂(s) · rn ))Ws0 (s)eis λ Ṙ ds. c (2.11) This is depicted for s0 = 45◦ in Figure 2.5. Figure 2.5: Fixed Aspect Data Block Since each (R, Ṙ) image is formed using a fixed aspect, we end up with L total images, often referred to as data blocks. Viewing these images in succession gives a “movie” of the target’s motion throughout the data acquisition period. The peaks of each image correspond to the radar observables to be extracted. Each peak has coordinates in the (R, Ṙ) plane, as well as a phase and amplitude at that location. These observables are recorded at each aspect and can be used for imaging. The process of extracting these observables from sequential images is known as sequential block processing. Figure 2.6 shows the extracted ranges and range-rates for each successive block. The radar begins viewing the target from the back end, as oriented in the first target image. As time (and aspect angle) increases, the target is viewed from all angles. The tumbling target in the figure displays how the target is oriented as time increases. 16 Figure 2.6: Extracted Ranges and Range-Rates Note the distinct sinusoidal nature of the ranges and range-rates. As the radar moves around the target in a circle, its coordinates, and therefore range to the target, can be described in terms of sines and cosines. Figure 2.7 illustrates this. Since range-rate is the time derivative of range, it follows that range-rate is also sinusoidal. Figure 2.7: Sinusoidal Nature of Range and Range-Rate 17 2.4 Sample Target Figure 2.8 shows the sample target geometry used for further simulations. The target has a cone body (grey) and three fins (blue). The fins are located 120◦ apart on the back end of the target. As mentioned previously, discrete points on targets are known as scatterers. Each of the red points in Figure 2.8 represents a scatterer on the sample target. The target has a total of thirteen fixed scatterers; four on each of the three fins and one on the cone nose. Figure 2.9 shows the three-dimensional simulated points used to model the target. Each point has a three-dimensional set of coordinates, (x, y, z). Figure 2.8: Sample RV Target Geometry Figure 2.9: Simulated Sample RV Target Points 18 2.5 Assumptions Many simplifications are made here. • Constant Amplitude For simplicity, each of the scatterers in Figure 2.9 were given equal constant amplitudes in the simulations. In reality, not all scatterers have the same amplitude, and their amplitude can vary depending on which direction they are viewed from. Variations in amplitude would show up in RTIs such as the one in Figure 2.4. • No Speculars Specular scattering occurs when a flat surface of a target is viewed at an angle exactly perpendicular to the surface, and a much higher return is generated. A useful real-life example is when some people’s foreheads show up as completely white in pictures when they are exactly perpendicular to the camera. A more realistic target model would include this effect, but this effect is ignored here. • Scatterer Resolution Notice in Figure 2.5 that the many scatterers on the back end of the target seem blurred together. This is also seen by how close together the range tracks are in Figure 2.4. This is due to the resolution of the Fourier transform. If there are many scatterers close together, they will not all be resolved, so the extracted ranges and range-rates will include some errors due to this. To help minimize that effect here, we use the simpler target shown in Figure 2.8. • No Shadowing Shadowing, or disappearing or occluded scatterers, occurs when some scatterers on a target cannot be seen because they are being blocked by other scatterers on the target. For example, when viewing the sample target looking directly at the back end of the cone, it would be impossible to see the nose of the cone. In Figure 2.4, the track on the top left for 0 < t < π 2 is the nose scatterer. It wouldn’t show up if shadowing was incorporated in to the target 19 model. Shadowing would also result in not every point in Figure 2.5 being visible. The effect of shadowing is not considered here. • Polarization Polarization refers to the geometry in which electromagnetic waves are transmitted and received. They can be transmitted and received either horizontally or vertically. Letting H denote horizontal and V denote vertical, this gives four transmit-receive combinations; HH, HV, VH, and VV. No specified polarization is included in the simulation, although it would be more realistic to include this. 20 2.6 Scatterer Correlation Scatterer correlation was briefly introduced in Section 1.6. Here we present a more detailed explanation of scatterer correlation. The range vs. range-rate images (data blocks) in Figure 2.10 display the aforementioned target at two aspect angles 30◦ apart. Scatterer correlation is the process of determining which scatterers in the second image correspond with which scatterers in the first image. For example, if we consider only scatterers A, B, and C in the first image, there are 3! = 6 possible ways these scatterers could be rearranged to get scatterers A, B, and C in the second image. From these images, it is fairly clear that A → A, B → B, and C → C. This problem gets very complicated very quickly, however. Including all thirteen scatterers on the target gives over six billion possible correlations. This problem is made much more difficult in the presence of noise and shadowing. Figure 2.10: Scatterer Correlation Figure 2.11 gives examples of uncorrelated and correlated data matrices. When data is correlated, data from specific scatterers is always entered in the same order in each row. This is not the case for uncorrelated data. Here, the letter in each matrix entry is used to indicate that the data located at that entry is from the scatterer corresponding to that letter, and has the form shown in (2.5). Unfortunately, radar data in general isn’t correlated. This is because as a radar interrogates a target and records data, it has no way of “knowing” which 21 Figure 2.11: Examples of Uncorrelated and Correlated Data Matrices data taken at a given aspect corresponds to data taken at a previous aspect. As the target is viewed at different orientations, its scatterers are viewed in different orientations as well. These ever-changing orientations dictate how the radar data is recorded. After the raw radar data is processed as discussed previously in this chapter, we arrive at Range vs. Range-Rate images. A peak-finding algorithm is used to identify all peaks in each of these images that are above a certain threshold. A peak is defined as a point in the image that is greater than all eight points surrounding it (right, left, top, bottom, top right, top left, bottom right, bottom left). In general, more points will be classified as peaks than there are scatterers on the target. Therefore, an input to the algorithm is the number of peaks to be extracted, p. In order to ensure that the p peaks chosen are most likely to be those resulting from scatterers on the target, the peaks are sorted by amplitude and the p points of highest amplitude are chosen. Their range, range-rate, and phase are recorded in the data matrices. Based on how the data is chosen to be recorded, it is easy to see that there is no guarantee the data will ever be recorded in the same order from one aspect to the next. Therefore, the data will in general be uncorrelated. However, data will remain correlated for a few aspects at a time. This is because data is recorded at every 0.25◦ . The images don’t change much over this small of a change in aspect. Therefore, data is recorded in the same order over at least a few aspects at a time, until a noticeable change occurs in the orientation of the target. 22 2.7 The Shape and Motion Estimation Problem Shape and motion estimation poses a nonlinear optimization problem which is a function of several variables including target geometry (size and shape), body center of rotation, and Euler dynamical variables, spin and precession, as explained in Section 2.8. Figure 2.12 displays the relationship between shape, motion, and radar observables. Many imaging algorithms assume either shape or motion is known in order to solve for the other. The joint estimation problem, which involves solving for both shape and motion, is a difficult and nontrivial problem. Figure 2.12: Relationship Between Motion, Shape, and Radar Observables Notice that the size of the shape matrix will never change. It will always have dimensions 3 × N , since each of the N scatterers has coordinates in three dimensions. However, the matrices of motion and radar observables will increase in size as the target is viewed at increasing numbers of aspects, since both have dimension 3L. Here, L is the number of aspects. Data taken at L aspects for all three radar observables results in a total of 3L rows of data in these matrices. 23 2.8 Euler’s Dynamical Equations of Torque-Free Motion The motion considered in this work will be torque-free motion. We give some background on what this means. The following can be found in [13]. The Euler equations are equations that describe the rotation of a rigid body. It is assumed that the reference frame is centered inside the body. The equations are Ix ω̇x + (Iz − Iy )ωy ωz = Nx Iy ω̇y + (Ix − Iz )ωz ωx = Ny Iz ω̇z + (Iy − Ix )ωx ωy = Nz . Here, the Ii are the three principal moments of inertia, ωi are the three components of the angular velocity, ω̇i are the three components of the angular acceleration, and Ni are the applied torques. Torque-free motion occurs when each Ni = 0. This yields Euler’s Dynamical Equations of Torque-Free Motion, Ix ω̇x = (Iy − Iz )ωy ωz Iy ω̇y = (Iz − Ix )ωz ωx (2.12) Iz ω̇z = (Ix − Iy )ωx ωy . The quantities ωx , ωy , and ωz depend on the Euler angles φ, ψ, and θ, ωx = ψ̇ sin θ sin φ + θ̇ cos φ ωy = ψ̇ sin θ cos φ − θ̇ sin φ ωz = φ̇ + ψ̇ cos θ. Here, ψ is the spin angle, θ is the precession cone angle, and φ is the precession rotation angle. The following is adapted from [8]. For our application, we use a target-centered coordinate system as shown in Figure 2.13. The origin is located at the center of the target. The vectors x, y, and z originate at the center of the target and are chosen to 24 Figure 2.13: Target Centered Coordinate System align with a dimension of the target. Standard spherical coordinates θ, φ characterize the look angle to the radar, which is determined by the unit vector k̂. Here, κ is the localized aspect angle from the radar to the target angular momentum vector, θp is the precession cone angle relative to the target angular momentum vector, φp is the precession rotation angle, and ψ is the spin angle. In general, for a target undergoing torque-free motion, θ(t) = θ(κ(t), θp , φp ) and φ(t) = φ(κ(t), θp , φp , ψ(t)) are complicated functions determined by Euler’s dynamical equations of motion. The aspect angle θ(t) is dependent on the precession variables, θ(t) = cos−1 (cos κ cos θp + sin κ sin θp sin φp ). (2.13) The roll angle φ(t) is dependent on the spin and precession variables, −1 φ(t) = ψ − tan cos κ sin θp − sin κ cos θp sin φp sin κ cos θp . (2.14) The target coordinates (and, hence, the image reference frame) are defined relative to the origin of the system shown in Figure 2.13. For a wideband radar operating in the higher frequency bands, the total scattering from the target can be decomposed into scattering from discrete scattering centers, rn , n = 1, . . . , N , as discussed in Section 2.3.1. Assume we can isolate a specific target scatterer. In a 25 target-centered coordinate system, this position is fixed, and the look angle to the radar, and its time derivative, are given by k̂ · rn ≡ Rn (2.15) ˙ k̂ · rn ≡ Ṙn (2.16) where (Rn , Ṙn ) are the range and range-rate radar observables corresponding to rn ˙ and k̂ and k̂ are functions of θ, θ̇, φ, φ̇ given by k̂ = sin θ cos φ x̂ + sin θ sin φ ŷ + cos θ ẑ (2.17) ˙ k̂ = θ̇θ̂ + φ̇ sin θ φ̂ (2.18) where θ̂ = cos θ cos φ x̂ + cos θ sin φ ŷ − sin θ ẑ φ̂ = − sin φ x̂ + cos φ ŷ. Equation (2.15) was previously shown geometrically in Figure 2.2. Equations (2.15) and (2.16) form a nonlinear set of equations characterizing the behavior of θ(t) and φ(t) and each scatterer rn . In [8], physical constraints are imposed on the motion θ(t) and φ(t) by constraining them to that body motion following from Euler’s dynamical equations of torque-free motion. 2.8.1 Special Case: Turntable Geometry In the turntable geometry/static range measurement case depicted in Figure 2.3, the radar is kept stationary while the target rotates a full 360◦ , which is computationally equivalent to the radar rotating as the target is kept stationary. The radar views the target in the plane θ = θ0 at a constant roll, while φ varies. For example, if θ0 = 90◦ , (2.17) and (2.18) are simplified to k̂ = cos φ x̂ + sin φ ŷ (2.19) ˙ k̂ = φ̇ (− sin φ x̂ + cos φ ŷ), (2.20) 26 where φ̇ is constant. This motion is purely two-dimensional. 2.9 Simulating Truth Radar Data In order to test the imaging algorithms developed in this work, we must simu- late data to test them on. At each aspect, the instantaneous motion is defined as in (2.17) and (2.18). For each of the L aspect angles and F frequency values, we use (2.7) to determine the exact value of the scattered field at that location. We end up with an L × F matrix of raw radar data. Note that in the case that we are simulating truth radar data resulting from turntable geometry, we use (2.19) and (2.20) instead of (2.17) and (2.18) to plug in to (2.7). The data used here was simulated to be X-band data, as is typical for radars imaging targets of this size. Specifically, F = 81 frequencies equally spaced between 8.2 and 12.2 GHz were used, making the center frequency 10.2 GHz. We assume that data is taken at 360◦ about the target, taken every 0.25◦ . This results in a total of N = 1441 aspects. The one extra aspect results from data being taken both at 0◦ and 360◦ . 2.10 Resolution The Fourier transforms here were computed using MATLAB’s Fast Fourier Transform (FFT). For computation speed, all Fourier processing here was computed using an FFT of size 28 = 256. A power of two was chosen because the FFT performs more efficiently when its size is a power of two. To calculate the wavelength resulting from using the given X-band frequencies, we plug in to the well-known equation relating wavelength and frequency, λ= c , ωc where ωc is the center frequency. This gives λ= 3 × 108 m/s = 0.0294m. 10.2 × 109 Hz 27 We expect that the wavelength, λ, will give a limit on the resolution that can reasonably be expected. The total range in each dimension of the images formed here is 3m, with a grid size of 256 × 256 resulting from the FFTs in both dimensions. This implies a resolution of 3 256 = 0.0117m per pixel in each dimension. This resolution, however, is unrealistic, given that it is less than half the wavelength. This resolution limit is what leads to the blurriness of the images, for example, how not all the scatterers are resolved in Figure 2.5. The limit on resolution that results from the FFT size is one of the reasons that target space image summation is used in Section 3.7. 2.11 Matrix Decompositions 2.11.1 The Singular Value Decomposition (SVD) We first present a brief review of the SVD, as it will be used frequently throughout this thesis. The following is adapted from [15]. The singular value decomposition (SVD) of a matrix A ∈ Cm×n is a factorization A = U Σ V ∗, where U ∈ Cm×m is unitary, V ∈ Cn×n is unitary, Σ ∈ Rm×n is diagonal. The diagonal entries σj of Σ are nonnegative and in non-increasing order, so σ1 ≥ σ2 ≥ . . . ≥ σp ≥ 0, where p = min(m, n). 28 2.11.2 The Reduced QR Factorization We now present a brief review of the reduced QR factorization. The following is adapted from [15]. The reduced QR factorization of a matrix A ∈ Cm×n is given by A = Q̂ R̂, where Q̂ is an m×n with orthonormal columns and R̂ is n×n and upper-triangular. The sequence of columns q1 , q2 , . . . has the property span(q1 , q2 , . . . , qj ) = span(a1 , a2 , . . . , aj ), j = 1, . . . , n. The columns a1 , . . . , ak can be expressed as linear combinations of q1 , . . . , qk , a1 = r11 q1 a2 = r12 q1 + r22 q2 a3 = r13 q1 + r23 q2 + r33 q3 .. . an = r1n q1 + r2n q2 + . . . + rnn qn . The full QR factorization of this same matrix A appends m−n more orthonormal columns to Q̂ so that it is an m × m unitary matrix, Q. Rows of zeros are added to R̂ to make the matrix R size m × n. This yields the relation A = Q R. 29 2.12 Basic Problem Formulation We present a brief overview of the basic problem formulation to motivate the subsequent sections. As discussed in Section 1.8, the two methods focused on here are the Phase-Enhanced 3D Snapshot Imaging method and the SVD method. Figures (2.14) and (2.15) give schematics highlighting the key ideas of each method. Figure 2.14: Phase-Enhanced 3D Snapshot Imaging method Figure (2.16) gives a schematic highlighting the goal of the dissertation; to develop a phase-enhanced 3D SVD method that uses range, range-rate, and phase in order to jointly estimate shape and motion while eliminating the need for scatterer correlation. 30 Figure 2.15: SVD Method Figure 2.16: Phase-Enhanced SVD Method CHAPTER 3 Phase-Enhanced 3D Snapshot Imaging Algorithm We present an overview of phase-enhanced interferometric processing. The following development is adapted from [8]. In 3D Inverse Synthetic Aperture Radar (ISAR) imaging, a target-centered coordinate system is employed, from which 2D range-Doppler image planar cuts through the target (as in Figure 2.5) are associated with points in the coordinate system (as in Figure 2.9) using the (assumed) known aspect angle to the target. These 2D images can be combined to form a 3D image of the main scattering centers, or scatterers, on the target. More recently, 3D snapshot imaging was developed. Conventional 2D snapshot imaging involves applying 2D ”snapshot” equations to two extracted radar observables, range and range-rate, to solve for 2D shape. To apply these to 3D, the 2D equations are applied to two different snapshots of data simultaneously and the resulting system of equations is solved in a least squares sense [7]. Another technique currently being explored is Interferometric Synthetic Aperture Radar (IF-SAR). This involves the “overlaying” of nearly identical 2D rangeDoppler images and uses phase differences between these images to estimate the “out-of-plane” height information, from which a 3D image is developed. The phase-enhanced 3D snapshot imaging algorithm combines 3D snapshot imaging and interferometric SAR. Localized microdynamic changes in look angles to the target characterized by Euler’s 6 Degrees-of-Freedom (6DOF) dynamical equations of motion (background given in Section 2.8) result in the nearly identical 2D range-Doppler images needed for interferometric SAR. The conventional 2D snapshot imaging equations are augmented to involve a third equation involving the resulting phase differences. The technique is particularly applicable for targets with smaller numbers (≤ 20) of dominant scatterers. In this section, we present an overview of 2D snapshot imaging, present two derivations of the phase equation, and then show how this equation is augmented 31 32 to the 2D snapshot imaging equations to yield the phase-enhanced 3D snapshot imaging equations. We show how additional data can be incorporated, and present many factors that need to be taken in to account when implementing the resulting equations. These include phase ambiguities, summing composite target space images, and sector processing. 3.1 2D Snapshot Imaging Equations As introduced in Section 2.3, 2D data blocks, or “snapshots”, can be pro- cessed sequentially to develop a sequence of range, range-rate pairs (Rn , Ṙn )m , t = t1 , . . . , tm . Equations (2.15) and (2.16) are applied sequentially to a number of snapshots to develop an image of the scatterers rn . This can be expressed in matrix form: T m · rn = Rn Ṙn , (3.1) m where T is a 2 × 3 transformation matrix given by k̂ T = ˙ . k̂ ˙ Here k̂ and k̂, as defined in (2.17) and (2.18), are expressed as row vectors, rn is a column vector, and (3.1) defines the mapping at time t = tm . Note that this applies to one specific scattering center rn . Applying these equations to each scatterer at each snapshot by direct matrix inversion results in two-dimensional estimates of the scatterers rn = (xn , yn ). This is adapted from [7]. Note that scatterer correlation is not required here since (3.1) is applied to individual scatterers separately at each time tj . Applying the two equations of (3.1) results in a two-dimensional estimate at time tj . 3.2 3D Snapshot Imaging Equations Now consider augmenting equations from another snapshot taken at t = tp to (3.1). Using three or more independent equations will result in a three-dimensional 33 shape estimate. Equation (3.2) gives the augmented equations, Tm Tp Rnm Ṙ nm · rn = Rnp Ṙnp . (3.2) Applying these equations to one scatterer at two snapshots will result in a threedimensional estimate of the scatterer rn = (xn , yn , zn ), as shown in [7]. Note that scatterer correlation is required here since the nth scatterer at t = tm must correspond to the nth scatterer at aspect t = tp . Also note that no restrictions are made on how “close” together the two snapshots must be. 3.3 Phase Equation Here we present two different derivations of the phase equation. The first derivation comes from comparing the equation of the received signal at two snapshots that are “close together” (what this means will be described in Section 3.4. This is adapted from [8]. The second derivation is a new derivation presented here to show that the same result can be obtained directly from the equations of two individual RangeDoppler data blocks. These Range-Doppler data blocks are also assumed to be “close” together. 3.3.1 Derivation from Snapshot Equations As shown in Section 2.3, the received signal from the target can be written in the form (2.7). Let k̂m = k̂1 at time t = t1 so that E1 = X 4π Dn ei λ k̂1 ·rn . (3.3) n Typically, the complex amplitude Dn is a relatively slowly varying function of frequency ω and angles θ(t) and φ(t) and characterizes the geometric theory of diffraction (GTD) diffraction coefficient of the nth scattering center. It is essentially 34 constant from one data block to the next when using sequential block processing, as long as the data blocks are “close” enough together so that no noticeable change occurs in range, range-rate, and amplitude. Consider viewing a target at two consecutive aspects. Let k̂m = k̂2 at t = t2 where t2 > t1 . If we assume k̂2 is “close” to k̂1 , then we say k̂2 = k̂1 + ∆k, (3.4) where ∆k is small. Using (3.4) in the general expression (3.3), the scattered field E2 at time t = t2 can be expressed in the form E2 = X 4π 4π [Dn1 ei λ ∆k·rn ]ei λ k̂1 ·rn (3.5) n Now assume that, since k̂2 is close to k̂1 , the 2D range-Doppler images extracted from each 2D data block centered at times t = t1 and t = t2 , respectively, are nearly identical, so that the individual scatterers in each image can readily be identified and correlated. It is assumed that range, range-rate, and amplitude remain constant; phase is the only quantity that changes. Hence, by extracting the phase of each scattering center from the range-Doppler image, (3.3) and (3.5) can be compared to obtain Phase(I2n ) − Phase(I1n ) = Phase(Dn1 ) − Phase(Dn2 ) + 4π (∆k) · rn λ (3.6) where I1n and I2n represent the complex value of the image pixel of the nth scattering center defined over each respective range-Doppler image plane. If k̂2 and k̂1 are close, then Dn1 ≈ Dn2 , so that (3.6) can be written in the form 4π (∆k) · rn = ∆n (Phase), λ (3.7) ∆n (Phase) = Phase(I2n ) − Phase(I1n ). (3.8) where we define: Equation (3.7) is the third equation that will be used to augment the standard set 35 of 3D snapshot imaging equations. It should be noted that (3.6) - (3.8) as derived are equivalent to the standard characterization of IF-SAR processing. They are repeated here only to emphasize their coupling to the 3D snapshot imaging equations. 3.3.2 Derivation From Range-Doppler Data Block Recall from Section 2.3 the equation of each Range-Doppler data block, I(s0 , R, Ṙ) = X n Z Dn 4π 2 δ( (R − k̂(s) · rn ))Ws0 (s)eis λ Ṙ ds. c (3.9) We Taylor expand k̂(s) about s = s0 , ˙ k̂(s) = k̂(s0 ) + k̂(s0 )(s − s0 ) + · · · , so that I(s0 , R, Ṙ) = X Z 4π 2 ˙ δ( (R − (k̂(s0 ) + k̂(s0 )(s − s0 ) + · · · ) · rn ))Ws0 (s)eis λ Ṙ ds. c (3.10) Dn n We let 2 ˙ s0 = (R − (k̂(s0 ) + k̂(s0 )(s − s0 ) · rn )). c The delta function in (3.10) makes a nonzero contribution approximately when s0 = 0. Using (2.15) and (2.16) at time s0 , setting s0 = 0 gives s = s0 + R − Rn (s0 ) . Ṙn (s0 ) Then denoting the Jacobian resulting from the change of variables from s to s0 by J, we find I(s0 , R, Ṙ) = J X n n (s0 ) R − Rn (s0 ) i(s0 + R−R )( 4π Ṙ) λ Ṙn (s0 ) Dn Ws0 s0 + e . Ṙn (s0 ) (3.11) 36 Isolating the phase of the nth scatterer gives 4π R − Rn (s0 ) Phase(In (s0 , R, Ṙ)) = s0 + Ṙ λ Ṙn (s0 ) 4π Ṙ(R − Rn (s0 ) + s0 Ṙn (s0 )) = . Ṙn (s0 )λ (3.12) The above equation gives the phase for fixed roll angle θ1 . Fixing φ and varying only θ allows us to consider changes in θ in particular instead of k. Now for θ = θ1 + ∆θ = θ2 , we assume that ∆θ is small enough that R(s0 ), Ṙ(s0 ), and Ṙ do not change noticeably. Then ∆n (Phase) = Phase(I2n ) − Phase(I1n ) = 4π Ṙ(Rθ2 − Rθ1 ) . Ṙ(s0 )λ (3.13) Assuming that at extracted peaks we have Ṙ = Ṙ(s0 ) and substituting Rθ2 − Rθ1 = k̂2 · rn − k̂1 · rn = ∆k · rn gives ∆n (Phase) = which is exactly (3.7). 4π (∆k) · rn , λ (3.14) 37 3.4 Phase-Enhanced 3D Snapshot Imaging Equations Note that increasing the number of equations used in (3.1), resulting in (3.2), caused the shape estimate to be three-dimensional instead of two-dimensional. In doing this, correlation between two snapshots is required. However, augmenting (3.1) with a different equation that doesn’t need to be correlated from the previous snapshot would be advantageous. The following methodology is adapted from [8]. We augment (3.1) with the phase equation given in (3.7). We consider the case where the target is rotating (spinning) with the radar stationary, so φ is monotonically increasing and θ is constant. Since Ṙ information is extracted from data along the direction of motion, then for spin-only motion, at constant roll θ0 and using (2.18) with θ̇ = 0, φ̇ sin θ0 φ̂ · rn = Ṙn , which shows that Ṙ is characterized solely by φ̇ for this case. Then since θ and φ change independently, augmenting (3.1) with data blocks associated with changes in θ as in (3.7) will provide additional independent information. In principle, augmenting (3.1) with (3.7) provides three independent equations that can be directly inverted to solve for the three unknowns (xn , yn , zn ). As previously shown, evaluating ∆k directly assuming θ = θ1 + ∆θ shows that (3.7) provides independent information. It can easily be shown that 4π 4π (∆k) · rn = ∆θθ̂ · rn . λ λ Recall that k̂ is a unit vector in the direction of the radar line of sight. Since k̂ is perpendicular to θ̂, φ̂, the three unit vectors (k̂, φ̂, θ̂) form a local (i.e., a function of θ, φ) orthogonal set of unit vectors, defining a local 3D axis set. The three independent equations are given by k̂ · rn = Rn Ṙn φ̂ · rn = φ̇ sin θ λ ∆n (Phase) θ̂ · rn = · , 4π ∆θ (3.15) 38 which clearly delineate the projections of the scattering center location vector rn onto the axes of this localized orthogonal frame. This is shown in Figure 3.1 [8]. Figure 3.1: Localized Orthogonal Coordinate System for 3D PhaseEnhanced Processing We can now rewrite (3.1) in the form T m · rn = Qmn , (3.16) where T is now a 3 × 3 motion matrix given by k̂ T = φ̇ sin θ φ̂ 4π ∆θθ̂ λ (3.17) and Qmn is given by Rn Qmn = Ṙ n , ∆n m where ∆n is shorthand for ∆n (Phase). (3.18) 39 3.5 Augmenting Additional Roll Cuts Although the phase-enhanced 3D snapshot imaging algorithm specifically ex- ploits phase differences between data sets with a very small change in roll between them, it is possible to also include data at other rolls. We can still extract R and Ṙ information for this data, however, we won’t acquire any phase information from it. Instead, we set up this data as in (3.1) and append it to (3.16). If we have u data sets corresponding to small changes in roll and v − u single data sets, the augmented equations become k̂1 φ̇ sin θ φ̂ 1 1 4π ∆θ1 θ̂ 1 λ .. . k̂u φ̇ sin θu φ̂u T · rn = 4π ∆θu θ̂ u λ k̂u+1 φ̇ sin θ u+1 φ̂u+1 .. . k̂v φ̇ sin θv φ̂v Rn1 Ṙ n1 ∆n 1 . . . Rnu Ṙnu · rn = ∆nu R nu+1 Ṙ nu+1 . .. Rnv Ṙnv = Qn . (3.19) This system can then be solved in a least squares sense for the scattering positions rn , rn = (T T T )−1 T T Qn . (3.20) 40 3.6 Phase-Enhanced 3D Snapshot Imaging Considerations We now focus on the implementation of the Phase-Enhanced 3D Snapshot Imaging equations, given in (3.16). We discuss how the change in roll is done in practice, then we discuss the considerations that have to be taken when employing the target centered coordinate system. 3.6.1 Change in Roll Recall that in turntable geometry, “snapshots” are taken at multiple consec- utive aspects at a fixed roll angle. Data can then be taken the same way but for a different fixed roll angle. If the two fixed roll angles are “close” enough together, this similar data can be exploited to gain information about the target in another dimension. Figure 3.2 shows how this data is taken. Figure 3.2: Measurement Collection Method The theory behind the phase-enhanced 3D snapshot imaging method is that the change in phase between data taken at two successive roll angles gives information that can be used as the third independent variable in the snapshot equations. Recall from Section 3.3.1 that k̂1 and k̂2 (corresponding to two consecutive roll angles) must be “close”, i.e., ∆k must be “small”, in order for the phase equation to be useful. This is because if ∆k changes so much that the resulting change in phase is greater than 2π, the result will be ambiguous. This is easily seen by the relation Phase(Aeiα ) = Phase(Aei(α+2π) ). Consequently, we aim to choose a sufficiently small value of ∆k. Now recall from Figure 2.13 that k̂ depends on both θ and φ. When looking at a change in roll, θ, 41 the aspect angle, φ, is kept constant. Therefore, a change in k̂ is equivalent to a change in only θ in this case. Since even 1◦ could be too large of a value for ∆θ, we will use ∆θ = 0.5◦ for simulations. Even smaller would be better, but we keep in mind that it is very difficult to physically rotate a target by even a half a degree, let alone less than half a degree. 3.6.2 Coordinate System It is important to pay attention to the coordinate system used here. The roll angle θ is always measured in the y − z plane from the z-axis down to the point on the target such that 0 ≤ θ ≤ π. To illustrate this, notice that θ = 120◦ for both lower fins in the diagram of the back end of the target in Figure 3.3. Figure 3.3: Apparent Motion of Target Between Two Roll Cuts As can be seen, the three fins are at 0◦ , 120◦ , and 120◦ at roll 0◦ . Note that roll 0◦ corresponds to θ = 90◦ , since θ is measured to the radar and the radar is always in the z = 0 plane for turntable geometry. When roll increases to ∆θ◦ , one would initially think that all points on the target move ∆θ◦ in θ. This diagram illustrates that for points on the right side of the target here, roll does indeed increase by ∆θ◦ , however, roll decreases by ∆θ◦ for points on the left. This is an important condition that must be included when implementing the equations. 42 3.6.3 Implementation As mentioned above, roll 0◦ corresponds to θ = 90◦ . Then with the parameters ∆θ = 0.5◦ , −π ≤φ ≤ π, 0 ≤t ≤ π, φ̇ = 1, (3.17) yields sin φ cos φ 0 , T = cos φ − sin φ 0 0 0 −1 so that Rn ˙n . rn = T −1 · R ∆n Thus we have solved for rn . Keep in mind, however, that the calculated rn will actually have the form rn + ∆r, where the added term represents the estimate error and variance that is inherently added to the truth rn positions. This is due to the limit on resolution imposed by the size of the FFT used to make the Range-Doppler images. Implementing these equations directly results in Figure 3.4a. The color of each point corresponds to its amplitude. Notice the extraneous points reconstructed much higher and lower than the target’s actual location. Even with ∆θ = 0.5◦ , some phase ambiguities are still induced. This will always be the case, however, since for two phases Θ1 ∈ [−π, π] and Θ2 ∈ [−π, π], we find that Θ2 − Θ1 ∈ [−2π, 2π]. In the following section, we find a way to visualize and correct for these induced phase ambiguities. 43 3.6.4 Resolving Phase Ambiguities A way to visualize the induced phase ambiguities is to add 2π to the phase of every reconstructed point and plot it in a different color, then do the same for −2π. This results in Figure 3.4b. It is clear that the extraneous points in the figure on the left indeed do come from these phase jumps. A simple fix, then, is to impose a geometric constraint on the reconstructed data points. The reason that this is possible is because the phase-enhanced 3D snapshot imaging method reconstructs shape in actual target coordinates, so it is known that the center of the target will be at the origin of the coordinate system. As will be seen in Section 4, not all imaging methods reconstruct targets in actual target coordinates, so this is not always possible. The target space filtering algorithm used here is that if a reconstructed height is at least 0.5m greater than the highest target point, then 2π is added to its phase. Similarly, if a reconstructed height is at least 0.5m lower than the lowest target point, 2π is subtracted from its phase. Obviously, 0.5m isn’t the correct height to use for all targets, so this would have to be determined independently for other targets. As a result of this target space filtering algorithm, all reconstructed points are mapped back to their correct locations in target space, as in Figure 3.4c. This is a novel method first presented here. Figure 3.4: a) Reconstructed Positions, b) ±2π Phase Ambiguities c) Target Space Filtered Image Figure 3.5 includes Figure 3.4c at different 2D views so that the main features 44 can be seen. Notice that many scatterers are reconstructed, but the number of scatterers, lack of shadowing, and limited resolution of the Fourier transform prohibit precise estimation. The fins on the target are apparent in the back view. However, the noise in the data due to the limited resolution of the Fourier transform results in some extraneous points being reconstructed above where the top fin should actually be. It also results in many points incorrectly being reconstructed in between scatterers. Figure 3.5: Phase Resolved Target Points, Three 2D Views 45 3.7 Summing Target Space Images The plots in Figure 3.5 are scatter plots of the points calculated directly by inverting the 3 × 3 system of phase-enhanced 3D snapshot imaging equations. As mentioned in Section 2.10, the resolution implied by the grid size resulting from a FFT of length 256 is unrealistic given the wavelength λ = 0.0294m. A more realistic resolution would be larger. For example, assuming resolution of 0.05m implies a grid size of 3 = 60. 0.05 We call each 3D grid point a bin, and sum up all points that fall within each bin. There are two ways of doing this. Amplitudes The first is to sum the amplitudes of all points that fall in to each bin. This allows higher amplitude scatterers to dominate, and has a risk of missing low amplitude scatterers. We call this method “amplitudes”. Counts The second option is to sum only the number of points that fall in to each bin. In this case, lower amplitude scatterers are more likely to be resolved. We call this method “counts”. In either of these cases, we can plot the total sum of each bin at its 3D location. The color of the point at that location will correspond to the total sum there. The images in Figure 3.6 are from the target space reconstruction using the amplitudes method. Using the counts method gives very similar results. Figure 3.6: Target Space Reconstruction, 3D & Three 2D Views As can be seen, not every point is reconstructed, but many main features are. 46 It should be noted that more features could be resolved if a bigger FFT size is used. However, this takes much longer computationally and often causes MATLAB to run out of memory. These images of the target are sometimes referred to as “target space” images, because they are reconstructed in actual target coordinates. This means that the origin corresponds to the center of the target, so the exact location of the target is known. As will be seen when the SVD method is introduced in Section 4, this is not always the case. 3.8 Sector Processing In the peak-finding algorithm, all peaks in the Range-Doppler images that are above a certain threshold are recorded as peaks, then the top N peaks are extracted, where N is the predicted number of scatterers on the target. The peak classification cutoff threshold is measured in decibels (dB) below the maximum intensity in the image. All the processing of the simulated data was performed with a threshold of 30 dB below max. This works well for simulated data of constant amplitude. However, when real radar data is used, amplitudes can vary greatly. In this case, some features can be reconstructed very well, while others are completely missed since their amplitudes are small in comparison to other features. The top fin of this target is a good example here. This inspired the idea of using a variable threshold on the data. This threshold should be low for the look angles that most directly view the top fin and higher in other areas. It is apparent from the reconstructed positions that many extraneous points are picked up. So, instead of using data at all look angles, one option is to perform sector processing. This involves using data from only a few choice look angle sectors, and optimally using a different threshold at each sector. This minimizes some of the unwanted error that has propagated through the system, some still due to the errors induced in the peak-finding algorithm. This is illustrated in Figure 3.7. Included in Figure 3.8 are side views of reconstructions performed using a) just one look angle, b) four look angles, c) the four sectors mentioned previously, and d) all look angles. It is important to note that a good deal of the target can be 47 Figure 3.7: Sector Processing, Pink Areas Correspond to Data Used picked up from just one look angle, since it is indeed a snapshot of the target. These reconstructions show that no features are lost by only using four sectors instead of all look angles, which is less than 10% of the original data. Many points are picked up just from a small number of look angles. Recall that in Section 3.6.3, it was pointed out that each calculated rn position is associated with estimate error and variance. Using more and more data when a relatively good image has already been reconstructed adds even more error. Therefore, if a reasonable image is made with a smaller amount of data, it is redundant to add in more points. Figure 3.9 includes the target space images formed from the simulated data using the sector processing shown in Figure 3.7. They look very similar to the images formed using all look angles, and they actually look a little cleaner (compare to Figure 3.6). This is especially apparent in the back end view. 48 Figure 3.8: Comparison: Sector Processing for Various Amounts of Data Figure 3.9: Target Space Reconstruction Using Sector Processing CHAPTER 4 SVD Method This chapter summarizes the work [11]. 4.1 Derivation In the SVD method, the rigid-body assumption is employed, that is, it is assumed that nothing on the target itself is moving. We start with an equation similar to (2.15), X d̂ = ρ. (4.1) Whereas (2.15) included data for one isolated scatterer, (4.1) includes all scatterers at a given aspect so that the shape matrix X is N × 3, the motion unit vector d̂ is 3 × 1, and the range vector ρ is N × 1. In the phase-enhanced 3D snapshot imaging algorithm, it was required that motion be known a priori. Here, neither shape nor motion is known a priori. The goal is to estimate both shape and motion from the initial data, ρ. Since many variables will be used, a table is included at the end of this section displaying all variables used and their dimensions. At each consecutive aspect, the target moves in a way that can be characterized by rotations and translations, i.e., at the mth aspect, T X Od̂m + d̂m · τ m 1N = ρm . Here, O is a rotation matrix, τ is a translation vector, and 1 is a N × 1 vector of ones. Recall that in the phase-enhanced 3D snapshot imaging algorithm, Rn in (2.15) included no translational information (R0 in Figure 2.2). This isn’t possible in the SVD method, since motion is not known a priori. The method is used on range-only data taken at unknown aspects to the target, and consequently, both X and ρm include translational information. The translational information is removed through centroid removal. To do this, the mean range is subtracted from each range and each target coordinate, 49 50 which allows the translational term to be removed. This yields the equation rk̂ = R, (4.2) where R is a vector consisting of centroid-removed ranges from the radar to each scatterer, r is a matrix whose rows contain the (x, y, z) coordinates of the centroidremoved scatterers, and k̂ = Od̂ is a unit vector view direction. We aim to arrive at an estimate for the shape, r, that doesn’t include the unknown motion, k̂. Solving (4.2) for k̂ yields k̂ = (rT r)−1 rT R. (4.3) Since k̂ is a unit vector, we can apply kk̂k = 1 to (4.3). This gives RT r(rT r)−2 rT R = 1. (4.4) This gives an expression involving just shape, r, and the known range data, R. However, we can’t directly solve this equation since it is a nonlinear equation for r. Notice that any rotated shape matrix, r̃ = r O, could be substituted into (4.4) and it would still hold. Consequently, the unchanging quantity Ω = r(rT r)−2 rT in (4.4) is a matrix characterizing the shape of the target whose relative coordinates remained unchanged through rotation and translation. Such quantities are known as invariants. Here, Ω represents the shape of both X and r. Another set of invariants also representing the shape of X and r is P = r(rT r)−1 rT , since r̃ = r O also does not affect P . 51 Notice that these invariants are overdetermined. There are more equations in the system than there are unknowns to solve for. We introduce a lemma that will be useful in the derivation. Lemma 1. [11] Define R1 (t1 ) R2 (t1 ) . . . RN (t1 ) R (t ) R (t ) . . . R (t ) 2 2 N 2 1 2 φ= .. .. .. . . . R1 (tL ) R2 (tL ) . . . RN (tL ) (4.5) and write the SVD of φ as φ = QAV T, (4.6) where φ has size L × N . If the column rank of the shape matrix r is three and the span of the columns of the motion matrix k̂ is also three, where φ = rk is the desired factorization, then φ has three nonzero singular values so that A is 3 × 3 and P = S(S T S)−1 S T = V V T . We now further examine the structure of Ω, starting by finding its pseudoinverse. The pseudoinverse of matrix M must satisfy the relation M M † = I, where M † denotes the pseudoinverse of M and I is the appropriately sized identity 52 matrix. Computing the pseudoinverse gives Ω Ω† = I r(rT r)−2 rT Ω† = I rT r(rT r)−2 rT Ω† = rT rT Ω† = rT r rT Ω† = r rT . So the pseudoinverse is given by Ω† = r rT . Now since Ω is symmetric, its SVD will also be symmetric. Taking the SVD gives Ω = U Σ UT , where U is an orthogonal matrix. This allows us to find another expression for Ω† , Ω Ω† = I U Σ U T Ω† = I ΣU T Ω† = U T U T Ω† = Σ−1 U T Ω† = U Σ−1 U T . So the pseudoinverse is also given by Ω† = U Σ−1 U T . Here we have used the property that for any orthogonal matrix M , both M T M = I and M M T = I. Now that we have two different expressions for Ω† , we can set them equal to each other, Ω† = r rT = U Σ−1 U T = (U Σ−1/2 )(Σ−1/2 U T ). So we have arrived at an expression for the shape estimate; r = U Σ−1/2 . (4.7) This shape estimate is valid only up to arbitrary rotations, as is shown by the fact 53 that Ω† = r r T = r O O−1 rT = (r O)(r O)T . Now that we have an expression for r, we can rewrite (4.2) as U Σ−1/2 k̂ = R. (4.8) We now further examine the structure of P = r (rT r)−1 rT using the new expression for r. P = r (rT r)−1 rT = U Σ−1/2 ((U Σ−1/2 )T (U Σ−1/2 ))−1 (U Σ−1/2 )T = U Σ−1/2 (Σ−1/2 U T U Σ−1/2 )−1 Σ−1/2 U T = U Σ−1/2 ΣΣ−1/2 U T = U UT . Since by Lemma 1 we have another expression for P in terms of the known quantity V taken from the initial SVD, P = V V T , we see that P = U UT = V V T . This leads to the conclusion that U = V O, so U is an arbitrarily rotated version of V . Then we can rewrite (4.8) as V O Σ−1/2 k̂ = R. (4.9) 54 Now V is already known from the initial SVD (4.6), R is the original data, O is an unknown rotation matrix, Σ−1/2 is a yet unknown matrix of singular values, and k̂ is the yet unknown motion unit vector. Multiplying both sides of (4.9) by V T gives O Σ−1/2 k̂ = V T R ≡ b, (4.10) where b is known. Similarly to what was done to arrive at (4.4), we can now impose the constraint kk̂k = 1. This gives bT O Σ O−1 b = 1. Let W = O Σ O−1 . Then W is a 3 × 3 symmetric matrix with only six nonzero variables. Let w be a column vector of the unknown elements of W , w = [w11 , w12 , w13 , w22 , w23 , w33 ]T . After some manipulation, expanding the system above gives ξw = 1, (4.11) where ξ = [b21 2b1 b2 2b1 b3 b22 2b2 b3 b23 ] is known. Now for each individual observation, recall from (4.10) that b = V T R. Since each column of φ is a transposed R vector from one observation, this gives the relation B = V T φT , where B is a matrix whose columns are the known b vectors from each observation. 55 Then from manipulating the initial SVD (4.6), we see that φ = QAV T φV = QA B = V T φT = A QT , where we have used the fact that AT = A since A is diagonal. Since B is 3 × L, we can denote the three rows of B as B1 , B2 , and B3 . Individual elements of these row vectors correspond to the bi , i = 1 . . . 3 variables in ξ. These can be used to expand the single observation equation (4.11) to a system including data from all observations. Let Ξ= B1 B1 T 2B1 B2 2B1 B3 . B2 B2 2B2 B3 B3 B3 The symbol denotes the Hadamard product, which involves element-wise multiplication of the vectors it acts on. Here Ξ is L × 6 and each of its rows is a ξ observation vector. Using this, (4.11) becomes Ξw = 1L , (4.12) where 1 is an L × 1 vector of ones. This can be directly solved for w, and hence W is found. Taking its SVD gives W = O Σ O−1 , so both O and Σ have been found, and hence so has the shape estimate, r = U Σ−1/2 = V O Σ−1/2 . 56 Table 4.1: SVD Method Variables and Their Dimensions We include Variable X d̂ ρ O τ 1N r k̂ R Ω P φ Q A V U b W w ξ B Bi Ξ 1L a table of each of the matrices involved along with their dimensions. Description Dimension shape matrix N ×3 motion unit vector at one aspect 3×1 range vector at one aspect N ×1 rotation matrix at one aspect 3×3 translation vector at one aspect 3×1 vector of ones N ×1 centroid-removed shape matrix N ×3 unit vector view direction 3×1 centroid-removed ranges N ×1 invariants characterizing shape N ×N invariants characterizing shape N ×N range data matrix at all aspects L×N left unitary matrix resulting from SVD of φ L×3 diagonal matrix resulting from SVD of φ 3×3 right unitary matrix resulting from SVD of φ N ×3 orthogonal matrix resulting from SVD of Ω N ×3 T V R, known 3×1 symmetric matrix, O Σ O−1 3×3 vector of the six unknown elements of W 6×1 2 2 2 [b1 2b1 b2 2b1 b3 b2 2b2 b3 b3 ] 1×6 matrix of b vectors from all aspects 3×L th i row of B 1×L matrix whose columns are Hadamard products of Bi L × 6 vector of ones L×1 This estimate can then be plugged back in to (4.2) to solve for the motion estimate, k̂. 4.2 Methodology Summary The SVD method implementation can be summarized in six steps. 1. Take the SVD of the L × N measurement matrix φ to obtain φ = Q A V T 2. Compute the 3 × L matrix B = A QT 3. Compute the L × 6 matrix Ξ from the rows of B 57 4. Solve the L × 6 system Ξw = 1 to find the 6 elements of the 3 × 3 symmetric matrix W 5. Take the SVD of W , W = O Σ O−1 , to find the shape estimate r = U Σ−1/2 = V O Σ−1/2 6. Plug back in to rk̂ = R to solve for the motion estimate, k̂ 4.3 Geometric Interpretation A key geometric interpretation can be realized here. Computing an SVD of r itself in (4.2), r = Y Φ Z T , gives Y Φ Z T k̂ = R. Multiplying each side by Y T yields Φ Z T k̂ = Y T R = b. Imposing kZ T k̂k = 1 gives bT Φ−2 b = b22 b23 b21 + + = 1, σ12 σ22 σ32 (4.13) where each σi is the ith diagonal entry of Σ, the singular values. So the ”images” lie on an ellipse embedded in RN . This ellipse gives all possible images resulting from the data. CHAPTER 5 Phase-Enhanced SVD Method There are many differences between the Phase-Enhanced 3D Snapshot Imaging algorithm and the 3DMAGI SVD method. A few of the key differences are highlighted in the table below. The attributes in bold are the attributes that the author would like an ideal 3D imaging algorithm to have. The goal of the Phase-Enhanced SVD (PESVD) method is to combine the SVD method and the Phase-Enhanced 3D Snapshot method in order to obtain an optimal hybrid method that has as many of these attributes as possible. In the 3DMAGI SVD method, range is used to jointly solve for shape and motion. An SVD is applied to range data with the goal of recovering its shape and motion components as in (2.15). When range-rate and phase data are available as well, an SVD can instead be applied to the full data matrix (3.18) with the goal of recovering its shape and motion components as in (3.16). Essentially the goal is to Table 5.1: 3D Imaging Method Differences Radar Observables Used Shape Estimate Motion Estimate Rigid Body Assumption Scatterer Correlation Code Run Time Estimate Characteristics Motion Dimension SVD Method Range Estimated up to arbitrary rotations and translations Estimated up to arbitrary rotations and translations Required Required Essentially instant One point for each scatterer 3D Euler Motion 58 Phase-Enhanced 3D Snapshot Range, rangerate, phase Estimated in target coordinates Not estimated; required a priori Not required Not required Iterative (aspect by aspect) Cumulative target space image 2D turntable geometry 59 again solve (3.16), but now without a priori motion knowledge. The system to be jointly solved is r 5.1 k̂ ˙ k̂ R = Ṙ . 4π ∆ ∆k λ (5.1) Scaling We now further examine (5.1). In (3.16), it didn’t matter whether the coefficients in the three equations were contained in the rows of (3.17) or the rows of (3.18), since there was only one unknown matrix to solve for, shape. However, in (5.1), both matrices on the left hand side are unknown. Therefore, if the scaling isn’t predetermined, a scaling factor from the shape matrix could end up in the motion estimate or a scaling factor from the motion matrix could end up in the shape estimate. If the scaling factors in the three equations of (5.1) are different, this will have the effect of scaling each of the three dimensions of r differently. For example, the target’s height, width, and length would each be multiplied by different constants, effectively skewing the shape estimate. Each entry of the motion matrix involves sines and cosines. It is essentially a rotation matrix. By definition, rotation matrices are orthogonal matrices with determinant 1. As of yet, the second and third rows of the motion matrix are arbitrarily scaled. By definition, each k̂ is a unit vector, so the first row is fine. However, the ˙ corresponding derivative, k̂, is not necessarily a unit vector, and neither is ∆k. We have ˙ ˙ ˙ k̂ = ||k̂||k̂0 , (5.2) ∆k = ||∆k||∆k0 , (5.3) and ˙ ˙ where ||k̂|| and ||∆k|| are scaling factors and k̂0 and ∆k0 are unit vectors. This 60 modifies (5.1) to r k̂ ˙ ˙0 ||k̂||k̂ R = Ṙ . 0 4π ∆ ||∆k||∆k λ (5.4) To ensure the scaling is consistent, we move the scaling factors to the right hand side of the equation. This modifies (5.1) to k̂ ˙ 0 r k̂ = ∆k0 R 1 ˙ Ṙ ||k̂|| λ 1 ∆ 4π ||∆k|| . (5.5) The issue is now that when solving (5.5) using R, Ṙ, and ∆ as inputs, the scaling factors are not known a priori. A simple fix for the first scaling factor is to first use just the given range-rate data, Ṙ, and perform the original SVD method on ˙ ˙ it beforehand to find Ṙ = rk̂. Then (5.2) can be solved for ||k̂||. Similarly, the scaling factor λ 1 4π ||∆k|| can be calculated. These scaling factors are then applied to the radar observables on the right hand side of (5.5). Jointly estimating shape and motion using the scaled data matrix will then yield correctly scaled shape and motion estimates. The Phase-Enhanced SVD Method estimates shape and motion only up to arbitrary translations and rotations. Figure 5.1 shows these solutions without the translations. The red corresponds to the truth solution while the blue corresponds to the calculated solution. As can be seen, the calculated solution is a rotated version of the truth solution. They align exactly when multiplied by a rotation matrix. 61 Figure 5.1: Shape and Motion Solutions 5.2 Sample Motion The same target illustrated in Figure 2.8 and Figure 2.9 is used for further simulations. In order for the Phase-Enhanced SVD method to work, the target must have three-dimensional motion. Therefore, turntable geometry can no longer be considered since it is two-dimensional. The target’s motion is given by (2.17) and (2.18), with θ(t) and φ(t) as in (2.13) and (2.14). For the examples throughout this chapter and the next, the specific threedimensional motion is characterized by the following parameters, in radians. κ=1 π θp = 18 π φp = 18 π ψ= 18 This yields the motion shown by the red solution on the right hand side of Figure 5.1. 62 5.3 5.3.1 Imperfect Correlation Simulating Imperfect Correlation Range, range-rate, and phase information taken from radar data will in general not be correlated. Without scatterer correlation, an ideal reconstruction such as the one in Figure 5.1 cannot be obtained. It is useful to examine the effects of imperfect scatterer correlation. We first simulate imperfect correlation as showin in Figure 5.2. Data that is 90% correlated has 10% of its rows randomly mixed up, data that is 80% correlated has 20% of its rows randomly mixed up, and so on. Figure 5.2: Varying Levels of Scatterer Correlation This is done as follows. For data that is p% correlated, (100 − p)% of its rows L) rows need to be rearranged. Since there are L rows, this means that round( 100−p 100 need to be rearranged. So for each n, n = 1 . . . round( 100−p L), row 100 round 100 n 100 − p is randomly rearranged. This is shown for p = 80 in Figure 5.3. For simplicity, only four scatterers are used here. This is a way to easily simulate varying levels of uncorrelation and will be useful in examining its effects. However, it is not a good representation of how data would realistically be uncorrelated. The radar observables range, range-rate, and delta phase are extracted from range vs. range-rate images such as Figure 2.5. They are extracted using peakfinding algorithms such as the one described in Section 2.6. The peak values found at each aspect are inserted into rows of the three data matrices. As the viewing 63 Figure 5.3: Uncorrelated Extracted Radar Observables, p = 80 aspect changes, the scatterers are viewed in different orders, which causes changes in the order of the peak scatterer values. This is shown in Figure 5.4. These are plots of the three radar observables extracted using the peak-finding algorithm. The random rearrangement of the order of scatterers is easily seen. For clarity, we will refer to the aspects at which these rearrangements occur as jump locations. Since all three radar observables are extracted simultaneously from each peak, all their jump locations are the same. Figure 5.4: Uncorrelated Extracted Radar Observables Notice that some outlying points now occur. When using the peak-finding algorithm, points are sometimes recorded in erroneous locations. Also notice that phase ambiguities still occur here. There is no way around this. The two extracted phase values at every aspect each lie within the interval [−π, π]. Because of this, taking their difference to arrive at the delta phase variable will inherently result in a value that lies within the interval [−2π, 2π]. This results 64 in phase jumps of size ±2π in the delta phase variable, in addition to the other jump locations already there from the peak-finding algorithm. Using this information, a much more realistic method of simulating uncorrelated data can be employed. I developed and implemented an algorithm to automatically uncorrelate data in this manner. The algorithm uses a set number of jumps to uncorrelate the data, which can be changed if desired. It picks two tracks at random to be switched. It picks a third track to be switched with probability 0.5 and a fourth track to be switched with probability 0.25. Once all tracks to be switched have been determined, it switches them in a random order. The probabilities and the number of tracks to be switched can easily be changed in the algorithm. Uncorrelated radar observables resulting from this algorithm are illustrated in Figure 5.5. Figure 5.5: Uncorrelated Truth Radar Observables with Random Jumps 65 5.3.2 Effect of Imperfect Correlation Figure 5.6 shows the effect of different levels of correlation where data was uncorrelated by varying percentages. The shape estimate for the 100% correlated case is the correct shape, while the other two are skewed. The motion estimate for the 100% correlated case is the correct motion, while the other two include random clouds of extraneous points from the points that aren’t correlated. All points here are direct outputs of the PESVD algorithm. The seemingly continuous lines in the motion estimate are the correctly calculated motion trajectory, while the points in the cloud around it are incorrect points resulting from extraneous points in the initial data. Figure 5.6: Shape and Motion Estimates for Varying Correlation Percentage Levels 66 Recall from the SVD method that the first three singular values of the matrix A obtained from the initial SVD of the data are ideally the only nonzero singular values of A. This is true when the data is correlated. In order to examine the ideally zero singular values, a target with more scatterers must be used to simulate data so that there are more singular values to look at. A 13-scatterer target was used here. Shown in Figure 5.7 are the distributions of the singular values for varying correlation levels of the initial data. As the data becomes less correlated, the ideally zero singular values steadily increase. The correct correlation is the only set of data that results in the fourth through last singular values being zero, as is desired. Figure 5.7: Singular Value Distributions for Varying Correlation Percentage Levels Figure 5.8 shows the effect of different levels of correlation where data is given with different numbers of jumps. As before, the data with no jumps gives the correct shape and motion estimates. As more jumps occur in the data, the shape estimate becomes worse and the motion estimate ends up with corresponding jumps. Figure 5.9 gives the singular value distributions when increasing numbers of jumps are added to the data. As can be seen, the ideal singular value structure occurs when there are no jumps (100% correlation), then the structure becomes worse as more jumps (more uncorrelation) occur. 67 Figure 5.8: Shape and Motion Estimates for Varying Correlation With Jumps Figure 5.9: Singular Value Distributions for Correlation With Varying Numbers of Jumps 68 5.4 Added Noise 5.4.1 Simulating Added Noise The author devised two different ways to add noise to the radar observables. It should be noted that the only type of noise considered here is noise induced during processing, not noise present pre-processing. This noise results from errors in extracting the radar observables from range vs. range-rate images such as 2.5. 5.4.1.1 The Percentage Method. We first simulate data with p% noise, where p% noise is interpreted as additive white noise where p ||Noise|| ≈ . ||Data|| 100 Each of the three radar observables data matrices are of size L × S, where L is the number of aspects and S is the number of scatterers. All three radar observables are real-valued. We seek to add noise by way of percentages. For example, to add 10% noise to a data matrix, we will multiply each value in the matrix by a random number within the interval [0.90, 1.10]. In this way, the noisy data value will be at least 10% smaller and at most 10% larger than its truth value. We calculate the formula that should be multiplied by each data matrix in in order to add p% noise. The desired formula must yield an interval, such as [0.90, 1.10], that is symmetric and centered about 1. We first find the formula that yields a symmetric interval centered about 0, scale it so that the desired percentage is achieved, then shift it to be centered about 1. In MATLAB, the quantity rand(L, N ) yields an L × N matrix whose elements are numbers randomly chosen from the standard uniform distribution on the open interval (0, 1). Then the quantity 2 rand(L, N ) − 1 (5.6) 69 gives numbers randomly chosen on the interval (−1, 1), which is symmetric about 0. Scaling this interval to achieve the desired percentage yields p (2 rand(L, N ) − 1), 100 which gives numbers randomly chosen on the interval − p p , . 100 100 To ensure the interval is centered about 1, we add 1, p (2 rand(L, N ) − 1) 100 100 − p p = + rand(L, N ). 100 50 1+ (5.7) This shifts the interval to p p 1− ,1 + . 100 100 Multiplying the truth data by (5.7) will add noise of size p. This yields the formula Noisy Data = p 100 − p + rand(L, N ) × Data, 100 50 (5.8) where the matrix multiplication is performed element-by-element. An example of truth radar observables data with 15% noise added is shown in Figure 5.10. Figure 5.10: Truth Radar Observables with p = 15 70 The author initially chose this method of adding noise because it works well for data that is on different scales. If noise was instead added by adding random values to the data, it would be difficult to ensure that noise of comparable magnitude was added to all three radar observables. For example, for the shown data, the delta phase variable varies on [−2π, 2π], while the range-rate variable varies on [−1, 1]. Adding random values on [−1, 1] to the range-rate data would add noise that is approximately equal in magnitude to the data itself, so around 100% noise. Adding random values on [−1, 1] to the delta phase data, however, would add noise that is about 32% of the magnitude of the data. One drawback to this method, however, is that for each individual radar observable, more noise is added to data values of greater magnitude than is added to data values of smaller magnitude. This is apparent in 5.10. The tracks become noisier the greater in magnitude they are. This is not a realistic effect. 5.4.1.2 The Extraneous Points Method. We now look at another method of adding noise to data. As is somewhat noticeable in Figure 5.4, sometimes noise appears in the form of data points randomly being recorded at the wrong location. This results in most of the data being at the correct locations, but random points being randomly offset. These points will be referred to as “extraneous” points. This implementation is very simple. Tracks are chosen at random and extraneous point locations are chosen at random. For each track at each chosen location, a number chosen randomly using (5.6) is added to its value. An input to the algorithm is the number of extraneous points to be added to the data. Figure 5.11 shows radar observables after noise is added using the extraneous points method. 71 Figure 5.11: Truth Radar Observables with Extraneous Points 72 5.4.2 Effect of Added Noise 5.4.2.1 The Percentage Method. Figure 5.12 shows the shape and motion estimates for data with varying percentage levels of noise. Reasonable quantities of added noise don’t seem to have much effect on the shape estimate, however, this blurs the motion estimate. Figure 5.12: Shape and Motion Estimates for Varying Percentage Levels of White Noise Figure 5.13 is a graph of the singular value distributions with increasing percentages of added noise. As predicted, the singular values become increasingly erroneous as increasing percentages of noise are added. 73 Figure 5.13: Singular Value Distributions for Varying Noise Percentages 5.4.2.2 The Extraneous Points Method. We now undergo the same analysis for our other method of adding noise, adding extraneous points to the data. Figure 5.14 shows the shape and motion estimates for data with increasing percentages of extraneous points. The motion estimates are simply rotated versions of each other; the change in scale is needed to accommodate the extraneous motion points. It is important to notice the distinction between the percentages here and the percentages used in Figures 5.12 and 5.13. Previously, data with n% noise was data with noise added of size n% of size of the data, so each noisy point is still a good rough estimate of its corresponding truth point. Here, n% noise means that n% of the points have been moved to be at completely erroneous locations, so each noisy point is in no way guaranteed to be a good estimate of its corresponding truth point. This is why the errors in Figure 5.15 appear to be larger than the errors in Figure 5.13. In reality, both methods of adding noise give similar results for how added noise affect estimates of shape and motion. 74 Figure 5.14: Shape and Motion Estimates for Varying Levels of Extraneous Point Noise 75 Figure 5.15: Singular Value Distributions for Varying Percentages of Extraneous Points CHAPTER 6 Correlation Algorithms We have demonstrated that the PESVD method works very well when given data that is correctly correlated. The goal here was to devise a method that would work on correlated data, but to also minimize the need for scatterer correlation. We do not consider the case of shadowing, also known as disappearing or occluded scatterers, in this work. Occluded scatterers are scatterers that can’t be seen at a given aspect because they are being blocked by other scatterers. This case is currently being pursued by other researchers. We only consider the case where all data is known, but it is uncorrelated in a realistic way and has noise. We devise two methods of correlating data so that uncorrelated data can be used in the PESVD method with much better results. We consider each method first without noise, then with the addition of reasonable amounts of noise. It should be noted that these algorithms can be used to correlate data that can then be used in any imaging algorithm, not just the PESVD method. 6.1 6.1.1 Iterative Algorithm Motivation Recall the effect of imperfect correlation as presented in Section 5.3.2. Figures 5.7 and figure 5.9 help to illustrate the effect of imperfect correlation. These figures both show that the correct correlation for a given set of radar data results in a simple, predictable singular value structure; all singular values after the first three are equal zero. This agrees with Lemma 1. The additional information obtained from these figures is that imperfect correlation affects this singular value structure in a predictable and easily recognizable way. In general, the less correlated a data set is, the greater its ideally nonzero singular values will be. Only one correlation, the correct one, will result in the ideal singular value structure. This observation was the author’s motivation to exploit this known singular 76 77 value structure in order to develop an iterative correlation algorithm that chooses the “best” correlation for a set of data. This section explains the development of the iterative algorithm based on the above observation. A theoretical justification is later provided in 6.3. 6.1.2 Large Number of Scatterers Without Noise We consider truth data uncorrelated as in Figure 5.5, but with a larger number of scatterers. Tracks can quickly become indistinguishable in this case, be it from overlapping, noise, or both. Since all three radar observables need to be correlated but correlating one correlates them all (besides phase jumps in delta phase), we will only include figures of the ranges to illustrate the algorithm. Figure 6.1 shows uncorrelated range data simulated from a target with 13 scatterers. Each black dot indicates where a track jumps. Figure 6.1: 13 Scatterer Data Uncorrelated with Jumps In the absence of a direct way of correlating the data, one method could be to examine every possible correlation that the data could have and choose the best one based on some criteria. Since this data has 13 tracks (corresponding to 13 78 scatterers) taken over 1441 aspects, the total number of possible ways the data could be arranged would be 131441 . It is clearly impossible to test every one of this enormous number of correlations. Since a brute force method is not feasible, we look for a way to cut down the number of possible correlations based on what we know about the data. What we know is • The peak-finding algorithm records scatterers in the same order across aspects until jumps occur, so the “chunks” of data in between jumps are already correlated. • At each aspect, it is fairly simple to identify which tracks have jumped based on range differences. If two tracks happen to switch while intersecting so that there is no noticeable difference in range, the jump can still be identified by range-rate differences at that same location. This means that even if the ranges intersect, they’ll have different slopes so a jump will easily be detected in range-rate. This fact was used by Mark Stuff in [5]. • At each jump location, the correct correlation is already known for the tracks that don’t have jumps; only the tracks with jumps will yield different correlation possibilities. • At each aspect where jumps occur (called a jump location), if n jumps occur, there are n! possible ways the data can be correlated after the jump. Using these known facts to our advantage will significantly reduce the number of possible correlations of the data. The algorithm will first explicitly list all possible correlations. For each possible correlation, it will reorder each jump in the data so that every jump is in the order specified by the given correlation. The PESVD method will then be applied to the reordered data matrix and the singular values recorded. After this has been done for each possible correlation, the correlation that yielded the closest to ideal singular value structure will be chosen as the best possible correlation. The algorithm is as follows. 79 1. The data is first stepped through in order of increasing aspect. If a jump occurs of a size over a given threshold in either range or range-rate, the location is marked as a jump location, k. The aspect at which the jump occurs is recorded, along with the indices of the tracks which had jumps. The total number of jumps at the k th jump location is referred to as mk . 2. The total number of possible correlations is calculated by multiplying the total possible number of permutations at each aspect. This is done through the formula N= K Y mk !. k=1 For example, for the data shown in Figure 6.1, we have N = 2! × 2! × 2! × 2! × 3! = 96. Clearly, this is a much more manageable number of correlations to examine than 131441 . The reason we consider permutations of every jump in the data simultaneously instead of examining one at a time is that the criterion used in the method is the singular values. The singular values come from the SVD of the data matrix resulting from permuting every jump in the data, so considering one jump location at a time would not be useful. 3. List out every possible correlation to be tested in to a matrix, C. The matrix will have N rows, so that each row gives one of the N possible correlations. The matrix will have K columns, so that each row gives one way in which the tracks can be permuted at each of the K jumps. Each element in the k th column will be chosen from the interval [1, mk !], since mk ! is the total possible number of correlations resulting from each jump. Note that we are only listing permutations resulting from the K jump loca- 80 tions, instead of K + 1 permutations resulting from the K + 1 “chunks” of data that K jump locations divide the data into. This is because we’ll take the order in which the data initially appears as being the correct correlation, then find the permutations of the subsequent data that returns it to the correct correlation. This is shown in (6.1), c11 . . . c1K c21 . . . c2K C= . .. .. . cN 1 . . . cN K , (6.1) where the nth column gives the nth possible correlation and where the entry representing the k th jump location of the nth correlation, cnk is contained in the interval cnk ∈ [1, mk !]. (6.2) For example, if only two tracks switch at jump k, then each element in the k th column can only be chosen from the interval [1, 2!], which makes sense because two tracks can only be arranged in two ways; they can stay in the same order or they can switch. If three tracks switch at jump k, then each element in the k th column will be chosen from the interval [1, 3!], so the three tracks can be rearranged in six ways. These cases are shown in Figure 6.2. The values Pk will be explained in the next step. As another example, for the data shown in Figure 6.1, the matrix C will be 96 × 5 and look like Figure 6.3. Here we have included the tracks below the matrix to clarify where the entries come from. Again, the matrices below the tracks will be explained in the next step. The matrix C is formed in MATLAB as follows. The algorithm randomly selects a number from each of the K intervals to form a sequence of K numbers. If this sequence is not already in C, it is appended to C as a new row. If this sequence is already in C, it is not appended to C and the algorithm continues to the next sequence. This continues until C has been completely filled to size N × K. 81 Figure 6.2: 2 and 3 Jump Cases Note that this method of populating C is not an explicit method. The problem of explicitly enumerating each individual correlation is not a simple one. The method of doing it randomly is fast, easy, and avoids this issue. For problems of a much larger magnitude, it could be possible that a method of explicitly enumerating them would be faster. That case is not explored here. 4. The algorithm tests each of the N possible correlations. For a given correlation, it loops through each jump location. At the k th jump location, it forms the matrix P of all the possible ways the jump can be permuted. If only two tracks switch at a jump location, the matrix is P2 = 2 1 1 2 . This means that for any two tracks (T1 , T2 ), their two possible new orders are (T2 , T1 ) and (T1 , T2 ), which is very straightforward. If three tracks switch at a jump location, the matrix becomes 82 Figure 6.3: Matrix C of All Possible Correlations P3 = 3 2 1 3 1 2 2 3 1 . 2 1 3 1 2 3 1 3 2 The rightmost column in Figure 6.2 shows these Pk matrices with their corresponding [1, mk !] intervals. The mk ! rows of each Pk give the mk ! possible 83 permutations of each jump location. Also, the Pk matrices under the tracks in Figure 6.3 are the matrices resulting from the data shown in Figure 6.1. At the nth correlation and k th jump location, the number cnk will give the row of P that should be chosen as the new way to order the tracks after the jump, i.e., the new order is given by P (cnk , :). The number cnk will be in the interval [1, mk !] and will determine which one of the mk ! rows of Pmk is to be used, i.e., we use the cnkth row of P . The tracks are then reordered. The matrix that stores the list of tracks that switch at each jump location then must be updated, as reordering tracks will switch how they are referenced at the next jump location. This is done for each of the K jump locations. Afterward, the Phase-Enhanced SVD algorithm is used on the reordered data. The singular values from the SVD are saved. The algorithm moves on to the next correlation and repeats the process. 5. The singular values were saved from each of the N possible correlations. Figure 6.4 shows what the singular value distributions look like for the data shown in Figure 6.1. The black dots represent the truth singular values. Figure 6.4: Singular Value Distributions from 13 Scatterer Data 84 We now choose the best correlation, where the best correlation is defined as the correlation that results in the best singular value distribution. The singular value distribution that gives the smallest sum of the ideally nonzero singular values is chosen and used to identify the correct correlation. The data is then reordered one last time, in the order given by the chosen correlation. The new correlation is shown for the data shown in Figure 6.1 in Figure 6.5. As can be seen, this method correctly correlates the data. Figure 6.5: 13 Scatterer Data Uncorrelated with Jumps then Correlated with Iterative Algorithm 85 6.1.3 Large Number of Scatterers With Noise We now demonstrate that the algorithm can be used on uncorrelated data with noise in the form of extraneous points as in Figure 6.6. This form of noise is more realistic than noise added by the percentage method, so we only explore this case here. Figure 6.6: 13 Scatterer Data Uncorrelated with Jumps and Extraneous Points A fairly simple addition to the algorithm is used to accommodate noise of this form. In step 1 of the algorithm, jump locations are only actually recorded if more than one jump occurs at them. This ensures that a reasonable number of jump locations are recorded. If only one jump occurs at a jump location, the location is easily identified as having an extraneous point instead of a track switch, so there is no need to record it. In the unlikely case that two extraneous points occur at one jump location, that jump location will then be saved and included in the correlation algorithm. This won’t adversely affect the algorithm, it will just make it take slightly longer. The algorithm then proceeds exactly as before. Figure 6.7 shows what the singular value distributions look like for the data shown in Figure 6.6. Due to the extraneous points, none of the singular value distributions will have the ideal structure. However, as before, selecting the singular value distribution 86 Figure 6.7: Singular Value Distributions, Noisy Data that gives the smallest sum of the ideally nonzero singular values still gives the best possible correlation. Selecting that correlation here and rearranging the data in the specified order results in Figure 6.8. As can be seen, the data has been given the correct correlation. The few extraneous points will affect the shape and motion estimate in the manner described in Section 5.4.2. This is a very reasonable number of extraneous points, though, so a very good shape and motion estimate will result from this. 87 Figure 6.8: Data Uncorrelated with Jumps and Noise then Correlated with Iterative Algorithm 88 6.1.4 Large Number of Scatterers With Phase Ambiguities Without Noise Suppose now that the uncorrelated data we are trying to correlate also has phase ambiguities, as is unavoidable in real data. An example of such data is shown in Figure 6.9. Figure 6.9: Uncorrelated Data With Phase Ambiguities Removing phase ambiguities here is not trivial. Proceeding as in the previous section and initially ignoring the phase jumps so that the data can be correctly correlated before resolving phase jumps is an option, but means that no singular value distribution will give the ideally zero structure, since all singular value distributions will have added errors due to the phase jumps. We instead proceed in a way that will seek to minimize the errors added to all the singular value distributions. After step 1 in the iterative algorithm, we will impose a simple constraint to get rid of the majority of the phase jumps. If a jump of greater than 1.75π occurs, the jump is assumed to be a 2π phase jump and ±2π is added to the value at the jump to map it back to the point closest to the previous point. This will resolve all phase ambiguities besides the ones arising after track switches. The delta phase variable after this constraint is shown in Figure 6.10. The rest of the algorithm is then executed as before. The singular value distributions resulting from it are shown in Figure 6.11. As can only barely be seen here, the small amount of error incurred by the phase ambiguities remaining in the delta phase variable barely affect the singular value distributions at all. The algorithm still picks the correct correlation. At the 89 Figure 6.10: Delta Phase after Phase Constraint end of the algorithm, the same phase constraint as before is applied. Now, the remaining ambiguities will not be affected by track switches. This correctly resolves the rest of the ambiguities and results in the correct delta phase variable, as seen in Figure 6.12. 90 Figure 6.11: Singular Value Distributions from 13 Scatterer Data With Phase Ambiguities Figure 6.12: Delta Phase after Iterative Algorithm Using Phase Constraint 91 6.1.5 Large Number of Scatterers With Phase Ambiguities, Noise, and Filtering We now demonstrate that the same algorithm can be used on data with both phase ambiguities and noise. The few extraneous points in the data after correlation can be smoothed out using median filtering. We use the data shown in Figure 6.13. Figure 6.13: Uncorrelated Data With Phase Ambiguities and Noise After imposing the phase constraint, delta phase improves as shown in Figure 6.14. Figure 6.14: Delta Phase With Noise after Phase Constraint The singular value distribution structure, shown in Figure 6.15, includes more 92 errors than before, as is shown by the ideally nonzero singular values being significantly farther from zero. However, the correct singular value distribution is still chosen as the one that minimizes these values. Figure 6.15: Singular Value Distributions from 13 Scatterer Data With Phase Ambiguities and Noise This results in the correct correlation being applied to the data. The phase ambiguities have been correctly resolved, as shown in Figure 6.16. The data was filtered using simple 1D median filtering in MATLAB that smooths out outlying points in all three radar observables. As can be seen, the iterative algorithm correctly correlates data that has phase ambiguities and reasonable levels of noise. The three correlated radar observables after applying the algorithm are shown in Figure 6.17. 93 Figure 6.16: Delta Phase with Noise after Iterative Algorithm Using Phase Constraint and Filtering Figure 6.17: Radar Observables Correlated with Iterative Algorithm 94 6.2 6.2.1 Nearest Neighbor and Slope Algorithm Small Number of Scatterers Without Noise When fewer scatterers are present, the iterative algorithm cannot be used. This is because there aren’t enough singular values after the first three to analyze, so examining the SVD structure is not useful. We develop a non-iterative algorithm that will perform faster than the iterative algorithm. With a smaller number of scatterers, it will be possible to go through the data sequentially and determine which tracks need to be switched where without testing every possible correlation. This method seems to be a natural way one would go about correlating data. We consider the case of four scatterers that are uncorrelated with random jumps via the algorithm discussed previously. We notice that, again, the data already has all three radar observables coupled so that finding the correct correlation of one radar observable will give the correct correlations for the others, besides phase ambiguities. Due to this we will again just show images of the ranges, until phase ambiguities are discussed. The algorithm is as follows. We begin with just correlating the ranges, then handle the delta phase variable separately to deal with the phase ambiguities. We start with the uncorrelated range data shown in Figure 6.20. This data has three jump locations. 1. We begin by identifying the jump locations. These are recorded whenever there is a jump over a given threshold size in either range or range-rate. For example, if two range tracks intersect and the track switch occurs right at the intersection, a jump location would not be detected in range, since the ranges would be approximately equal. However, their slopes would be different, so checking for a jump in range-rate at that same location would ensure that the jump is identified. Figure 6.19 shows correctly identified jump locations. Along with the jump locations, the track numbers that have jumps are recorded at each location and stored in a matrix for later reference. 2. At each jump location, a simple nearest neighbor algorithm is used to deter- 95 Figure 6.18: Uncorrelated Range Data mine which tracks go together. Each track before the intersection is matched with the track after the intersection that gives the smallest track jump between them. If the jump was detected in range-rate instead of range, the nearest neighbor algorithm is used on range-rate instead of range to correctly switch the tracks. After tracks are switched, the matrix identifying the tracks that jump at each jump location is updated to account for the newly ordered tracks. Applying this simple algorithm results in the data being correctly correlated, as shown in Figure 6.20. 96 Figure 6.19: Range Data After Step 1 Figure 6.20: Range Data After Step 2 97 6.2.2 Small Number of Scatterers With Noise and Phase Ambiguities We consider the case of uncorrelated data from four scatterers with extraneous points. This is a more realistic form of how data would look when extracted using the peak-finding algorithm, for example, as in the data shown in Figure 5.4. We begin with the uncorrelated data shown in Figure 6.21. This data has three jump locations and six extraneous points. Figure 6.21: Uncorrelated Range Data with Extraneous Points We notice that in this case, the previous algorithm is not practical. Each extraneous point would be wrongly recorded as a jump location and the nearest neighbor method would not work. In the case of noise in the form of extraneous points, the algorithm must be modified. The algorithm is as follows. We again begin with just correlating the ranges, as the data is coupled so all three radar observables can be correlated through this. 1. We begin by stepping through all aspects of the ranges. At each aspect, each point is matched up with the point from the previous aspect that is closest to it using a nearest neighbor algorithm. This is depicted in Figure 6.22. We do this because we know that in the presence of noise, it will be much harder to step through the data and handle each jump location individually 98 Figure 6.22: Range Data After Step 1 up front. Notice that if the extraneous points were not present, this step would correctly correlate the data besides track switches at intersections. The extraneous points result in the tracks not being correctly matched up. This step guarantees that all jumps are now either from intersections being incorrectly switched or from extraneous points. 2. The next step is to identify intersections. This is easy for the human eye, but tricker to automate. We start by again stepping through each aspect of the updated ranges. At each aspect, if any two tracks are less than a given threshold apart, a counter is started to measure at how many aspects the tracks remain close together. This window is used to prevent false intersections from being found due to noise. The intersection is found to have ended when the tracks are sufficiently apart. Let P denote the length of the intersection. This is illustrated in Figure 6.23. The center of the intersection is approximated to be at P k − round( ), 2 where k is the aspect at which the intersection ends. 99 Figure 6.23: Identifying an Intersection 3. We now must determine whether or not the tracks at each intersection should be switched. Range information alone is not enough to determine this, since these intersections occur where the ranges are approximately equal. Based on this, an easy way to determine an intersection needs to be switched would be just to search for a jump in range-rate. In the case of noise-free data, this would work. However, a condition depending on data from just one aspect is risky in the case of noisy data. Just one extraneous point could cause the algorithm to not resolve an intersection. Due to this, we examine the range-rates throughout a window about the intersection. We let S denote the length of data to be used on either side of the intersection, so that a total window of size 2S is used about the intersection. An intersection that needs its tracks switched and the corresponding rangerates are showed in Figure 6.24. No jump is seen in range, but a jump is seen in range-rate. The length 2S window about the intersection is seen in the figure. Let C and D be the length 2S vectors of range-rates corresponding to the 100 Figure 6.24: Analyzing an Intersection tracks intersecting in range centered at k − P2 as in Figure 6.24. Let C1 denote the first S elements of C and C2 denote the second S elements of C, and similarly for D. The basic criteria we want to use is that the intersection will be switched if C1 is closer to D2 than it is to C2 . Mathematically, if the condition ||C1 − D2 || < ||C1 − C2 || (6.3) is met, then the tracks at the intersection are switched. Note that using just this one condition is not foolproof. Just a few extraneous points in the range-rates could throw off the condition. Due to this, we seek to impose another condition in hopes that noise will not affect both. Theoretically, the range-rates are the exact derivative of the ranges. Another way to analyze the range-rates would be to actually compute the slope of the ranges. For each aspect j ∈ [2, 2S], we let E(j) = A(j) − A(j − 1) 101 and F (j) = B(j) − B(j − 1). So E and F are scaled versions of the derivative of range as computed by its definition. We seek to smooth any great discontinuities by assuming that the derivatives are linear about the intersection and thus fitting a first order polynomial to both E and F . This is done here using MATLAB’s “polyfit” function. Let Ê and F̂ denote the smoothed vectors formed from fitting first order polynomials to E and F , respectively. This gives the second slope condition. Let Ê1 denote the first S elements of Ê and Ê2 denote the second S elements of Ê, and similarly for F̂ . If the condition ||Ê1 − F̂2 || < ||Ê1 − Ê2 || (6.4) is met, then the tracks at the intersection are switched. Here this condition is exactly the same as (6.3), but it is a condition on the computed derivative instead of the recorded range-rates. Applying these two slope conditions is advantageous since either condition could be compromised by noise, but it is less likely that both will be compromised by noise at the same time. Applying these conditions correctly switches intersections, as shown in Figure 6.25. However, sometimes some are or are not correctly switched depending on how large S is chosen to be. 4. The track switches resulting from extraneous points have still not been corrected. We then impose a simple condition that if two tracks jump a significant amount at a given aspect, it most likely came from an extraneous point and the tracks are switched. This results in these switches being correctly resolved, as is shown in Figure 6.26. This process also simultaneously correctly correlates range-rate and delta phase. Figure 6.27 shows delta phase after this step. The jumps in correlation have been corrected, but the 2π phase ambiguities still remain. 102 Figure 6.25: Range Data After Step 3 5. We now focus on the 2π phase ambiguities in the delta phase variable. Now that correlation has been taken care of, the phase jumps should occur only for each individual track, instead of the tracks jumping in phase and also switching. This makes resolving the phase ambiguities much simpler. The delta phase variable at this point is given by Figure 6.27. Recall that in Section 3.6.4, phase ambiguities were easily resolved using a target space filtering approach. This was for the phase-enhanced 3D snapshot imaging algorithm, where targets were reconstructed in actual target coordinates. Here targets are reconstructed in arbitrary coordinates, so we don’t have any prior knowledge of where the target will be and how to filter accordingly. Because of this, another method must be employed. First, each track is stepped through individually. If a jump of greater than 1.75π occurs, the jump is assumed to be a 2π phase jump and ±2π is added to the value at the jump to map it back to the point closest to the previous point. Along with this, we apply simple 1D median filtering to the data to smooth it. This gets rid of sporadic extraneous points in all three radar observables. 103 Figure 6.26: Range Data After Step 4 The delta phase variable has been corrected, as shown in Figure 6.28. 6. One thing that sometimes occurs is that an entire track is mapped too high or too low due to phase ambiguities, as in Figure 6.29, taken from actual data extracted using the peak-finding algorithm. Because of this, we impose a check at this point. If any individual track has mean greater than 1.5π or less than −1.5π, then −2π or 2π is added to that track so that it gets mapped back to its correct location. After all steps of the algorithm, the data is correctly correlated. This is seen in Figure 6.30. 104 Figure 6.27: Delta Phase Data Before Step 5 Figure 6.28: Delta Phase Data After Step 5 105 Figure 6.29: Track Displaced Due to Phase Ambiguity Figure 6.30: Radar Observables Correlated Through Algorithm 106 6.3 Proof of Theorem Justifying Iterative Algorithm As is suggested in the previous section, it seems that uncorrelated data will always have more than three nonzero singular values. We now prove that this is true. We first state three important lemmas that will be useful in the proof. Lemma 1 [11]. Let A be a matrix of correlated radar range data and write the SVD of A as A = QΣV T. (6.5) If the column rank of the shape matrix r is three and the span of the columns of the motion matrix k̂ is also three (where A = r k is the desired factorization), then A has three nonzero singular values so that Σ is 3 × 3. These conditions on the shape and motion matrix mean that the target must be three-dimensional and is motion must also be three-dimensional. Lemma 2 [14]. Let A ∈ Rm×n be a matrix with rank(A) = r ≤ min(m, n) and Π a permutation matrix such that the first r columns in A Π are linearly independent. Then the QR factorization of A Π will have the form A Π = (Q1 Q2 ) R11 R22 0 0 (6.6) where R11 ∈ Rr×r is upper triangular with positive diagonal elements. Furthermore, the matrices Q and Q form orthogonal bases for the range of A and the null space 1 2 of AT , respectively. The rank of A is r and the null space of AT has rank m − r, so (Q1 Q2 ) has rank m. Lemma 3 [15]. The rank of a matrix A is r, the number of nonzero singular values. Lemma 4 [16]. Let A and B be p × n and p × q matrices, respectively, and [A, B] denote a row block matrix obtained by putting A and B side by side. Then, rank(A) + rank(B) − rank([A, B]) ≤ min(rank(A), rank(B)) (6.7) 107 We prove the following theorem. Theorem. A matrix of radar observables data is uncorrelated if and only if it has more than three nonzero singular values. Proof. Backward: We must show that more than three nonzero singular values implies uncorrelated data. This statement is the contrapositive of Lemma 1. Forward: Let A be an L × N matrix of correlated radar observables data, where L is the number of aspects and N is the number of scatterers. Then each entry aij is the range of the j th scatterer at slow time index i. The data is correlated as shown in Figure 2.11. Assume that the radar observables data results from a target with at least four scatterers, and that the scatterers do not all lie in the same plane so that the target is three-dimensional. Assume that there are more aspects than there are scatterers, so that N < L. By Lemma 1, we know that A has only three nonzero singular values. Then by Lemma 3, A has rank three. We first multiply A by a permutation matrix Π that reorders the columns of A so that the first three are linearly independent, say A Π = Ã. Since A has rank three, à also has rank three. Consequently, all columns of à can be written as a linear combination of the first three columns. Note that the permutation Π does not affect the correlation, since entire columns, not parts of columns, are switched. We take the reduced QR factorization of Ã, à = Q R, as shown below. ã11 . . . ã1N .. .. . ... . .. .. . ... . ãL1 . . . ãLN = q11 . . . q1N .. . . . . . .. .. . . . . . .. r11 r12 r13 . . . . . . r1N r22 r23 . . . . . . r2N r33 . . . . . . r3N (6.8) qL1 . . . qLN Here we have used Lemma 2 which shows that rij = 0 ∀ i > 3. The blank entries of R are equal to zero. Also by Lemma 2, the rank of Q resulting from the full QR factorization would be L. However, since this is a reduced QR factorization, Q has rank N . 108 The first three columns of Q form an orthonormal basis for the column space of Ã, and span(q1 , q2 , q3 ) = span(ã1 , ã2 , ã3 ), where ãj and qj denote the j th columns of à and Q, respectively. We know that it is necessarily the first three columns that this applies to because of the multiplication by the permutation matrix Π above. In other words, any column of à can be written as a linear combination of the first three columns of Q. The coefficients of these linear combinations are given by the corresponding entries of the matrix R. Consistently with Lemma 2, no nonzero entries are needed in R below the first three rows. If à had full rank, its QR factorization would be unique as long as we require that rjj > 0, 1 ≤ j ≤ n. Since à is rank-deficient, more than one QR factorization will be possible. By Lemma 2, the rank of R resulting from these QR factorizations will reveal the rank of Ã. For full rank matrices, “revealing rank” means that the R resulting from the QR factorization has all zero rows below some index, and that index is the rank. For rank-deficient matrices such as Ã, “revealing rank” means that the number of linearly independent rows of R will equal the rank of Ã. Recall how the elements of Q and R as arise from Gram-Schmidt orthogonalization [15]. The column vectors qj , 1 ≤ j ≤ N are given by ã1 r11 ã2 − r12 q1 q2 = r22 ã3 − r13 q1 − r23 q2 q3 = r33 .. . q1 = qn = ãn − Σn−1 i=1 rin qi . rnn (6.9) The coefficients rij are given by rij = q∗i ãj (i 6= j), (6.10) 109 and the coefficients rjj are given by |rjj | = ||ãj − Σj−1 i=1 rij qi ||2 . (6.11) Since the radar observables data results from a target with at least four scatterers, A has at least four distinct columns. It is not possible that any two columns of A will be equal. This is because if two columns were equal, they would have to have resulted from two scatterers that were at exactly the same range during the entire data collection period, in which case the radar would pick them up as one scatterer anyway. Individual values in separate columns will be equal when range tracks intersect, but no two columns will be identical. Consequently, we are guaranteed that any two columns, ãk and ãm , k < m, are not equal. Ensure that n has been chosen such that the nth entries of ãk and ãm are not equal. This means that the nth look aspect corresponds to a view at which the scatterers recorded in the k th and mth columns are not at equal range, so their range tracks are not intersecting. 6.3.1 Case Two Scatterers Switch We consider the case where two scatterers switch at a given aspect here. The case where more than two scatterers switch at a given aspect is discussed in the next section. Switch the nth element of ãk , ãnk , with the nth element of ãm , ãnm . We define the correct correlate to be the ordering in the first row, so we choose n such that 0 n > 1. The uncorrelated matrix resulting from this switch, à , is given by ã11 . . . ã1k . . . ã1m . . . ã1N .. .. ã21 . . ã2N .. .. . ãnm ãnk . .. .. .. .. . . . . ãL1 . . . ... ... 0 ... . (6.12) . . . ãLN Here we have kept the elements of à in terms of the original elements of Ã. We 110 examine how the switch of ãnk and ãnm affects the structure of the matrices Q0 and 0 R0 resulting from the QR decomposition of à . Case k > 3, m > 3 : 0 The first k − 1 columns of à and à are identical. By (6.9), we see that the definitions of qj , 1 ≤ j < k depend only on the first k − 1 columns of Ã. Therefore, the first k − 1 columns of Q0 are equal to the first k − 1 columns of Q. Explicitly, we have q0i = qi , 1 ≤ i < k. (6.13) 111 0 Using (6.13), we can write the QR factorization of à as ã11 . . . ã1k . . . ã1m . . . ã1N .. .. ã21 . . ã2N .. .. . ãnm ãnk . .. .. .. .. . . . . ãL1 . . . q11 q12 = q21 q22 q31 .. . q32 .. . × qn1 qn2 .. .. . . qL1 qL2 r11 r12 r22 ... ... ... . . . ãLN 0 0 0 . . . q1N . . . q1m q13 . . . q1k 0 0 0 . . . q2N . . . q2m q23 . . . q2k 0 0 0 q33 . . . q3k . . . q3m . . . q3N .. .. .. .. .. .. .. . . . . . . . 0 0 0 qn3 . . . qnk . . . qnm . . . qnN .. .. .. .. .. .. .. . . . . . . . 0 0 0 qL3 . . . qLk . . . qLm . . . qLN 0 0 r13 . . . r1k . . . r1m . . . r1N 0 0 r23 . . . r2k . . . r2m . . . r2N 0 0 r33 . . . r3k . . . r3m . . . r3N .. .. . . . . 0 .. rkk .. . 0 rmm (6.14) The unprimed elements in Q0 and R0 are the elements that are equal to the corresponding elements of Q and R. The k th column of Q0 , q0k , as seen in (6.9), depends on a0k . Since a0k 6= ak , q0k is not necessarily equal to qk . Since each successive column of Q0 depends on all previous columns of Q0 , none of the k th through N th columns of Q0 are necessarily equal to their corresponding columns in Q. Therefore, the k th through N th columns of Q0 are all primed. By (6.10), the first k − 1 rows of R0 depend on the first k − 1 columns of Q0 112 0 and all the columns of à . By (6.13), this means that the only elements within the first k − 1 rows of R0 that will be primed will be the ones depending on the columns 0 of à that are different than the corresponding columns of Ã, so the k th and mth columns. Therefore, within the first k − 1 rows of R0 , only the k th and mth columns are primed. Of the elements of R0 below the first k − 1 rows, all that are also below the diagonal are zero since by definition of the QR factorization, R0 must be upper triangular. Of the other elements of R0 below the first k − 1 rows, those that are not in the k th and mth columns must also be zero. This is because for any column a0p , 4 ≤ p ≤ N, p 6= k, p 6= m, we have that ã0p = ap and by matrix multiplication, ãp = r1p q1 + r2p q2 + r3p q3 , so it must be that ã0p = r1p q1 + r2p q2 + r3p q3 . Therefore, since no further columns of Q0 are needed to make ã0p , the coefficients of these columns can be set equal to zero, 0 rip = 0, 4 ≤ p ≤ N, p 6= k, p 6= m, k ≤ i ≤ p. Since we also know that rip = 0, i > 3, we have that 0 rip = rip , 4 ≤ p ≤ N, p 6= k, p 6= m, k ≤ i ≤ p. So we have shown that the first k − 1 columns of Q0 are unprimed, the k th 113 through N th columns of Q0 are primed, and the only primed columns of R0 are the k th and mth columns. We now look for explicit expressions for the primed elements of R0 in order to determine how the switch of ãnm and ãnk effects the rank of R0 , and thus the rank 0 of à . By (6.10), the original elements rik , 1 ≤ i < k are given by rik = q∗i ãk = [q1i q2i . . . qni . . . qLi ]T [ã1k ã2k . . . ãnk . . . ãLk ]. (6.15) 0 Similarly, the elements rik , 1 ≤ i < k are given by ∗ 0 rik = q0 i ã0k = q∗i ã0k = [q1i . . . qni . . . qLi ]T [ã1k . . . ãnm . . . ãLk ] = [q1i . . . qni . . . qLi ]T [ã1k . . . ãnk . . . ãLk ] − qni ãnk + qni ãnm = rik − qni ãnk + qni ãnm = rik + (ãnm − ãnk )qni , 1 ≤ i < k, (6.16) where in the first step we have used (6.13) and in the third step we have used (6.15). The same argument shows that 0 rim = rim + (ãnk − ãnm )qni , 1 ≤ i < k. (6.17) Since both rik = 0 and rim = 0 for 4 ≤ i < k, (6.16) and (6.17) tell us that 0 0 rim = −rik , 4 ≤ i < k. (6.18) Because of (6.13), simple relations between elements of R0 and R as in (6.16) and 114 (6.17) only hold for the first i rows of R0 , where i < k. Then (6.14) becomes ã11 . . . ã1k . . . ã1m . . . ã1N .. .. ã21 . . ã2N .. .. . ãnm ãnk . .. .. .. .. . . . . ãL1 . . . q11 q12 = q21 q22 q31 .. . q32 .. . qn1 qn2 .. .. . . qL1 qL2 ... ... ... . . . ãLN 0 0 0 q13 . . . q1k . . . q1m . . . q1N 0 0 0 q23 . . . q2k . . . q2m . . . q2N 0 0 0 q33 . . . q3k . . . q3m . . . q3N .. .. .. .. .. .. .. . . . . . . . 0 0 0 qn3 . . . qnk . . . qnm . . . qnN .. .. .. .. .. .. .. . . . . . . . 0 0 0 qL3 . . . qLk . . . qLm . . . qLN (6.19) r11 r12 r13 . . . 0 r1k ... 0 r1m . . . r1N × r22 r23 . . . 0 r2k ... 0 r2m r33 . . . 0 r3k ... 0 r3m . . . r2N . . . r3N . 0 r4k .. . 0 −r4k .. . 0 r(k−1)k 0 −r(k−1)k 0 rkk 0 rkm .. . 0 rmm Clearly, by (6.18), the 5th through (k − 1)th rows in R0 are all just scaled versions of the 4th row. Therefore, there is only one linearly independent row in the 4th through (k − 1)th rows of R0 . Consequently, R0 has more than three linearly independent rows instead of just three. Therefore R0 has rank greater than three. Since Q0 has full rank, this shows 0 that à has rank greater than three, so it has more than three nonzero singular 115 values. Any uncorrelated data set can be produced from a correlated one by a sequence of interchanges. As the data becomes more uncorrelated, more values of rij , i > 3 will be required to be nonzero, so the rank will increase each time, increasing the number of nonzero singular values accordingly. So we have shown that uncorrelated data implies that there are more than three nonzero singular values. This completes the proof for the case k > 3. Comparison to Numerical Simulations One may move on to the next case of the proof. However, it is interesting to examine the rest of the elements of R0 to see if any other knowledge can be gained about the structure. This is motivated by the fact that the structure (6.19) is not the structure seen in numerical simulations in MATLAB of the QR structure resulting from a switch of two elements as described in this proof. We know that the original element rkk is zero since k > 3. By (6.11), the equation for rkk is given by 0 = rkk = ||ãk − Σk−1 i=1 rik qi ||2 k−1 2 2 = (ã1k − Σk−1 i=1 rik q1i ) + . . . + (ãnk − Σi=1 rik qni ) + . . . + (ãLk − 2 Σk−1 i=1 rik qLi ) 1/2 . (6.20) Since each squared element in (6.20) cannot be negative, it follows that ãjk − Σk−1 1 ≤ k ≤ L. i=1 rik qji = 0, (6.21) Let χn denote an L × 1 vector that is all zeros besides a 1 in the nth location. The 116 0 element rkk is given by k−1 0 0 rkk = ||ã0k − Σi=1 rik qi ||2 = ||ã0k − Σk−1 i=1 (rik + (ãnm − ãnk )qni )qi ||2 k−1 rik qi − Σk−1 = ||ã0k − Σi=1 i=1 (ãnm − ãnk )qni qi ||2 k−1 = ||ã0k − ãk + ãk − Σk−1 i=1 rik qi − (ãnm − ãnk )Σi=1 qni qi ||2 k−1 = ||ã0k − ãk − (ãnm − ãnk )Σi=1 qni qi ||2 k−1 = ||(ãnm − ãnk )(χn − Σi=1 qni qi )||2 = (ãnm − ãnk ) ||(χn − Σk−1 i=1 qni qi )||2 k−1 k−1 = (ãnm − ãnk ) (Σi=1 qni q1i )2 + (Σi=1 qni q2i )2 + . . . + (1 − 2 2 Σk−1 i=1 qni ) + ... + 2 (Σk−1 i=1 qni qLi ) 1/2 , (6.22) where in the first step we have used (6.16), in the fourth step we have used (6.21), and in the fourth step we have used the fact that ã1k ã 2k . .. 0 ãk − ãk = ãnm . .. ãLk − ã1k 0 0 .. . = ãnk ãnm − ãnk .. .. . . 0 ãLk ã2k .. . = (ãnm − ãnk )χn , where, again, χn denotes an L × 1 vector that is all zeros besides a 1 in the nth location. In (6.22), we notice that the summation is over the first k − 1 columns of Q0 , which are equal to the first k − 1 columns of Q. These columns form an orthogonal basis for the range space of Ã. Therefore, no linear combination of these columns can yield a column entirely of zeros, so it is not possible for all the summation terms in (6.22) to be zero. Therefore, the normed term yields a nonzero contribution. 0 Since n was chosen such that ãnm 6= ãnk , (6.22) then shows that rkk 6= 0. 117 We now review which elements of (6.19) are known. The first k − 1 columns of Q0 are identical to the first k − 1 columns of Q. The k th through N th columns of Q0 are known by (6.9) as long as the corresponding elements of R0 are known. Every column of R0 besides the k th and mth columns is exactly equal to the corresponding column of R. Of the remaining columns, (6.16), (6.17), and (6.22) define every 0 0 . We seek expressions for these remaining elements. , . . . , rmm element besides rkm 0 0 The original two elements switched in à , defined in terms of elements of à , are given by 0 0 0 0 0 0 0 ãnm = qn1 r1k + qn2 r2k + qn3 r3k + qn4 r4k + . . . + qn(k−1) r(k−1)k + qnk rkk (6.23) and 0 0 0 0 0 0 0 ãnk = qn1 r1m + qn2 r2m + qn3 r3m + qn4 r4m + . . . + qn(k−1) r(k−1)m + qnk rkm 0 0 0 0 +qn(k+1) r(k+1)m + . . . + qnm rmm . (6.24) 0 We substitute the expression for rik in (6.16) in the first three elements of the right hand side of (6.23), 0 0 0 qn1 r1k + qn2 r2k + qn3 r3k = qn1 (r1k + (ãnm − ãnk )qn1 ) + qn2 (r2k + (ãnm − ãnk )qn2 ) + qn3 (r3k + (ãnm − ãnk )qn3 ) 2 2 2 = qn1 r1k + qn2 r2k + qn3 r3k + (ãnm − ãnk )(qn1 + qn2 + qn3 ) 2 2 2 + qn2 + qn3 ). = ãnk + (ãnm − ãnk )(qn1 (6.25) 0 Similarly, we substitute the expression for rim in (6.17) in the first three elements 118 of the right hand side of (6.24), 0 0 0 + qn3 r3m qn1 r1m + qn2 r2m = qn1 (r1m + (ãnk − ãnm )qn1 ) + qn2 (r2m + (ãnk − ãnm )qn2 ) + qn3 (r3m + (ãnk − ãnm )qn3 ) 2 2 2 ) + qn3 + qn2 = qn1 r1m + qn2 r2m + qn3 r3m + (ãnk − ãnm )(qn1 2 2 2 = ãnm + (ãnk − ãnm )(qn1 + qn2 + qn3 ). (6.26) 0 0 in (6.23) as + . . . + qn(k−1) r(k−1)m We use (6.18) to write the terms qn4 r4m 0 0 0 0 qn4 r4m + . . . + qn(k−1) r(k−1)m = −qn4 r4k − . . . − qn(k−1) r(k−1)k . (6.27) Adding together (6.23) and (6.24) and using the substitutions given in (6.25), (6.26), and (6.27) yields 0 0 0 0 0 0 0 0 ãnm + ãnk = ãnk + ãnm + qnk rkk + qnk rkm + qn(k+1) r(k+1)m + . . . + qnm rmm 0 0 0 0 0 0 0 0 0 = qnk rkk + qnk rkm + qn(k+1) r(k+1)m + . . . + qnm rmm . (6.28) The QR factorization is unique for full rank matrices, but not for rank-deficient matrices. Even though the QR factorization showed in (6.19) is correct, it is not the only possible QR factorization. Since R0 is rank-deficient and (6.28) is the only 0 0 0 0 that satisfy (6.28) will , . . . , rmm , any set of rkm , . . . , rmm constraint given for rkm provide a sufficient R0 , and therefore another valid QR factorization. A simple 0 0 choice is to set r(k+1)m = . . . = rmm = 0. Then (6.28) yields 0 0 rkm = −rkk . (6.29) The nonzero elements in the k th row of R0 are related to each other in the same way as (6.18). Then the 4th through k th rows of R0 are all linearly dependent, but linearly independent from the first three rows. All subsequent rows are all zeros. 119 With these substitutions, we can write (6.19) as ã11 . . . ã1k . . . ã1m . . . ã1N .. .. ã21 . . ã2N .. .. . ãnm ãnk . .. .. .. .. . . . . ãL1 . . . q11 q12 = q21 q22 q31 .. . q32 .. . qn1 qn2 .. .. . . qL1 qL2 r11 r12 × r22 ... q13 . . . ... 0 q1k ... ... . . . ãLN 0 q1m ... 0 q1N 0 0 0 q23 . . . q2k . . . q2m . . . q2N 0 0 0 q33 . . . q3k . . . q3m . . . q3N .. .. .. .. .. .. .. . . . . . . . 0 0 0 qn3 . . . qnk . . . qnm . . . qnN .. .. .. .. .. .. .. . . . . . . . 0 0 0 qL3 . . . qLk . . . qLm . . . qLN 0 0 r13 . . . r1k . . . r1m . . . r1N 0 0 . . . r2N r23 . . . r2k . . . r2m 0 0 . . . r3N r33 . . . r3k . . . r3m 0 0 . r4k −r4k .. .. . . 0 0 −rkk rkk (6.30) It is now apparent that the rank of the matrix R0 increases by exactly one in this case. It has been verified by the author using MATLAB that when QR factorizations of matrices such as (6.12) are done numerically, the resulting structure is exactly that shown in (6.30). Case k ≤ 3, m ≤ 3 : We now consider the case where both of the scatterers switched are within the first three columns of Ã. Recall that à was multiplied by a permutation matrix Π 0 so that the resulting matrix à ’s first three columns are then linearly independent. 120 We consider three arbitrary linearly independent vectors and show that switching corresponding elements in two of them results in another vector being required to make a full basis. Let the three arbitrary linearly independent vectors be w, x, and y. Let z be a vector in span(w, x, y), so z can be written as a linear combination of w, x, and y, i.e., z = aw + bx + cy, (6.31) for some constants a, b, and c. In the nth row of vectors w and x, switch the entries. Ensure that n is chosen such that w(n) 6= x(n), so that an actual switch in correlation occurs. Also assume that a 6= b, since a = b is the trivial case in which the change in correlation is not detectable. The case a = b corresponds to rij = rkj for some i, j, k. By (6.10), this can only happen if the column qi is identical to the column qk . Since by Lemma 2 Q0 is full rank, no two columns can be identical. Therefore, this trivial case is not possible here. In order for (6.31) to hold for the rows that weren’t switched, a, b, and c must remain the same. However, this results in the nth row yielding the equation ax(n) + bw(n) + cy(n) = z(n). (6.32) However, originally we had aw(n) + bx(n) + cy(n) = z(n). (6.33) In order for both (6.32) and (6.33) to hold, the system of equations a b b a w(n) x(n) = z(n) − cy(n) z(n) − cy(n) (6.34) 121 must hold. Since a 6= b, we find that w(n) x(n) 1 a b z(n) − cy(n) a2 − b 2 b a z(n) − cy(n) 1 (a + b)(z(n) − cy(n)) . = 2 a − b2 (a + b)(z(n) − cy(n)) = (6.35) Therefore, the only unique solution to (6.35) is w(n) = x(n), which makes (6.32) and (6.33) identical. However, n was chosen such that w(n) 6= x(n). This shows that there actually is no unique solution to (6.35), and (6.32) is false. Therefore, a correction term must be added in order for it to be true, ax(n) + bw(n) + cy(n) = z(n) − aw(n) − bx(n) + ax(n) + bw(n) = z(n) + (a − b)(x(n) − w(n)). (6.36) Notice that in the trivial case that a and b happen to be equal, the correction term is equal to zero, as is expected. So the vector z can no longer be expressed as a linear combination of just the three vectors w, x, and y. The switch of elements results in four linearly independent vectors required to span the space; w, x, y, and a vector that is all zeros besides (a − b)(x(n) − w(n)) in the nth entry. Returning to the problem at hand, we see that if at least one element is switched within the first three columns of Ã, one more column will then be required 0 to span the space of à . This means instead of the QR factorization of à having the form (6.8), at least one more entry below the first three rows of R0 will be required to be nonzero to accommodate the four columns of Q0 required to span the space. Consequently, R0 has four linearly independent rows instead of three. There0 fore R0 has rank four. Since Q0 has full rank, this means that à has rank four, so it has four nonzero singular values. As the data becomes more uncorrelated, more values of rij , i > 3 will be required to be nonzero, so the rank will increase each time, increasing the number of nonzero singular values accordingly. 122 So we have shown that uncorrelated data implies that there are more than three nonzero singular values. This completes the proof for the case k ≤ 3, m ≤ 3. Case k ≤ 3, m > 3 : We now consider the case where one of the scatterers switched is within the first three columns of à and one isn’t. We proceed similarly to the case where k ≤ 3 and m ≤ 3. We consider three arbitrary linearly independent vectors and one arbitrary linearly dependent vector. We show that switching an element in one of the linearly independent vectors with its corresponding element in the linearly dependent vector results in another vector being required to make a full basis. Let w, x, y, z again be defined as in (6.31). We ignore the trivial case where a 6= −1, since if a = −1, no noticeable change in correlation will occur since (6.31) will remain unchanged. The constants a, b, and c are chosen arbitrarily so that z is an arbitrary vector that is linearly dependent on the vectors w, x, and y. In the nth row of vectors w and z, switch the entries. Ensure that n is chosen such that w(n) 6= z(n), so that an actual switch in correlation occurs. In order for (6.31) to hold for the rows that weren’t switched, a, b, and c must remain the same after the switch. This results in the nth row yielding the equation az(n) + bx(n) + cy(n) = w(n). (6.37) However, originally we had (6.33). In order for both (6.37) and (6.33) to hold, the system of equations −a 1 1 −a w(n) z(n) = bx(n) + cy(n) bx(n) + cy(n) (6.38) 123 must hold. For a 6= −1, we find that w(n) z(n) 1 a2 − 1 1 = 2 a −1 = a 1 bx(n) + cy(n) bx(n) + cy(n) (a + 1)(bx(n) + cy(n)) . (a + 1)(bx(n) + cy(n)) 1 a (6.39) (6.40) Hence there is no unique solution to (6.40). Then a correction term must be added to (6.37) in order for it to hold, az(n) + bx(n) + cy(n) = z(n) − aw(n) + az(n) + w(n) − w(n) = w(n) + (a + 1)(z(n) − w(n)) (6.41) Notice that in the trivial case that a happens to be equal to −1, the correction term is equal to zero, as is expected. So the vector z can no longer be expressed as a linear combination of just the three vectors w, x, and y. The switch of elements results in four linearly independent vectors required to span the space; w, x, y, and a vector that is all zeros besides (a + 1)(z(n) − w(n)) in the nth entry. Returning to the problem at hand, we see that if at least one element is switched within the first three columns of Ã, one more column will then be required 0 to span the space of à . This means instead of the QR factorization of à having the form (6.8), at least one more entry below the first three rows of R0 will be required to be nonzero to accommodate the four columns of Q0 required to span the space. Consequently, R0 has four linearly independent rows instead of three. There0 fore R0 has rank four. Since Q0 has full rank, this means that à has rank four, so it has four nonzero singular values. As the data becomes more uncorrelated, more values of rij , i > 3 will be required to be nonzero, so the rank will increase each time, increasing the number of nonzero singular values accordingly. So we have shown that uncorrelated data implies that there are more than three nonzero singular values. 124 This completes the proof for the case k ≤ 3, m > 3. Notice that in all three cases, it was shown that switching two elements at the same aspect increases the rank by one. Therefore, we have shown that switching two elements at the same aspect increases the rank regardless of what the indices k and m are. This increase in rank yields an increase in the number of nonzero singular values. This completes the proof for when two scatterers switch. 6.3.2 Case More than Two Scatterers Switch We now examine the case where any number of scatterers are allowed to switch at a given aspect. The proof is very similar to the case where only two scatterers switch. Therefore, not every step is listed explicitly; rather, the proof follows along with the case where only two scatterers switch and the key differences are highlighted. We again begin with the matrix à shown in (6.8). In the nth row of Ã, we arbitrarily switch s elements. The s columns containing the elements that were switched will be referred to as the columns ξ1 , ξ2 , . . . , ξs , where ξ1 < ξ2 < . . . < ξs . We define the correct correlate to be the ordering in the first row, so we choose n such that n > 1. The uncorrelated matrix resulting from 0 this switch, à , is given by ã11 . . . ã1ξ1 . . . ã1ξt . . . ã1ξs . . . ã1N .. .. .. ã21 . . . ã2N .. .. . ãnξ2 . . . ãnξ( t+1) . . . ãnξ1 . .. .. .. .. .. . . . . . ãL1 . . . ... ... ... ... ... . (6.42) . . . ãLN For simplicity, we have replaced the nth element in column ξ1 with the nth element in column ξ2 , the nth element in column ξ2 with nth element in column ξ3 , and so on. The nth element in column ξs is replaced with the nth element in column ξ1 . However, any method of reordering s elements within a given row could be used. 0 Here we have kept the elements of à in terms of the original elements of Ã. 125 We examine how the switch of these elements affects the structure of the matrices 0 Q0 and R0 resulting from the QR factorization of à . Case ξ1 > 3 : The proof is very similar to the case where only two scatterers switch. Following along with the s = 2 case, the same argument that led to (6.13) shows that q0i = qi , 1 ≤ i < ξ1 . ã11 . . . ã1ξ1 . . . ã1ξt . . . ã1ξs . . . ã1N .. .. .. ã21 . . . ã2N .. .. . ãnξ2 . . . ãnξ( t+1) . . . ãnξ1 . .. .. .. .. .. . . . . . ãL1 . . . q11 q12 = q21 q22 q31 .. . q32 .. . × (6.43) qn1 qn2 .. .. . . qL1 qL2 r11 r12 r22 ... ... ... ... ... (6.44) . . . ãLN 0 0 0 0 . . . q1N . . . q1ξ q13 . . . q1ξ . . . q1ξ s t 1 0 0 0 0 . . . q . . . q q23 . . . q2ξ . . . q 2N 2ξs 2ξt 1 0 0 0 0 . . . q . . . q q33 . . . q3ξ . . . q 3N 3ξs 3ξt 1 .. .. .. .. .. .. .. .. .. . . . . . . . . . 0 0 0 0 qn3 . . . qnξ1 . . . qnξt . . . qnξs . . . qnN .. .. .. .. .. .. .. .. .. . . . . . . . . . 0 0 0 0 qL3 . . . qLξ1 . . . qlξt . . . qLξs . . . qLN 0 0 0 r13 . . . r1ξ1 . . . r1ξt . . . r1ξs . . . r1N 0 0 0 r23 . . . r2ξ . . . r . . . r . . . r 2N 2ξt 2ξs 1 0 0 0 r33 . . . r3ξ . . . r . . . r . . . r 3N 3ξt 3ξs 1 .. .. .. . . . . . . .. .. rξ0 1 ξ1 . .. rξ0 t ξt 0 rξs ξs 126 As in the case where only two scatterers switch, the unprimed elements in Q0 and R0 are the elements that are equal to the corresponding elements in Q and R. The same arguments used in that case hold, substituting the new column index ξ1 for k in all arguments. The first difference occurs in the equation analogous to (6.16). 0 , 1 ≤ i ≤ ξ1 are given by The elements riξ 1 0 0 riξ = qi∗ ã0ξ1 1 = q∗i ã0ξ1 = [q1i . . . qni . . . qLi ]T [ã1ξ1 . . . ãnξ2 . . . ãLξ1 ] = [q1i . . . qni . . . qLi ]T [ã1ξ1 . . . ãnξ1 . . . ãLξ1 ] − qni ãnξ1 + qni ãnξ2 = riξ1 − qni ãnξ1 + qni ãnξ2 = riξ1 + (ãnξ2 − ãnξ1 )qni , 1 ≤ i < ξ1 . (6.45) 0 , 1 ≤ i < ξ1 is given by Similarly, we find that any element riξ t 0 riξ = riξt + (ãnξt+1 − ãnξt )qni , 1 ≤ i < ξ1 . t (6.46) Since riξt = 0 for 4 ≤ i < ξ1 and 1 ≤ t ≤ s, (6.45) and (6.46) tell us that 0 0 0 = 0, 4 ≤ i < ξ1 . riξ + riξ + . . . + riξ s 1 2 (6.47) Equations (6.45), (6.46), and (6.47) show that the nonzero elements of the ith row of R0 , 4 ≤ i ≤ ξ1 , are given by qni [(ãnξ2 − ãnξ1 ), (ãnξ3 − ãnξ2 ), . . . , (ãnξ1 − ãnξs )]. (6.48) It is clear that the 4th through (ξ1 − 1)th rows of R0 are all scaled versions of each other. Then they are all linearly dependent, and they increase the rank of R0 . As in the case where only two scatterers switch, this proves the case for ξ1 > 3. Comparison to Numerical Simulations Again, one may move on to the next case of the proof. However, it is interesting to examine the rest of the elements of R0 127 to see if any other knowledge can be gained about the structure. This is motivated by the fact that the structure (6.45) is not the structure seen in numerical simulations in MATLAB of the QR structure resulting from a switch of more than two elements as described in this proof. The equations (6.20) and (6.21) still hold in this case, with k replaced by ξ1 . The same argument used to arrive at (6.22) shows that rξ0 1 ξ1 = (ãnξ2 k−1 2 2 − ãnξ1 ) (Σk−1 i=1 qni q1i ) + (Σi=1 qni q2i ) + . . . + (1 − 2 2 Σk−1 i=1 qni ) + ... + k−1 (Σi=1 qni qLi )2 1/2 . (6.49) The same argument as before shows that rξ0 1 ξ1 6= 0. We now review which elements of (6.30) are known. The first ξ1 − 1 columns of Q0 are identical to the first ξ1 − 1 columns of Q. The ξ1th through N th columns of Q0 are known by (6.9) as long as the corresponding elements of R0 are known. Every column of R0 besides the columns ξt , 1 ≤ t ≤ s, is exactly equal to the corresponding column of R. Of the remaining columns, (6.45), (6.46), and (6.49) define every element besides rξ0 1 ξt , . . . , rξ0 t ξt , 1 < t ≤ s. We seek expressions for these remaining elements. 0 0 The original s elements switched in à , defined in terms of elements of à , are 128 given by 0 0 0 0 0 0 r0 + qnξ + . . . + qn(ξ1 −1) r(ξ + qn4 r4ξ + qn3 r3ξ + qn2 r2ξ ãnξ1 = qn1 r1ξ s s s s 1 ξ1 ξs 1 −1)ξs 0 0 + qn(ξ r0 + . . . + qnξ r0 s ξs ξs 1 +1) (ξ1 +1)ξs (6.50) 0 0 0 0 0 0 ãnξ2 = qn1 r1ξ + qn2 r2ξ + qn3 r3ξ + qn4 r4ξ + . . . + qn(ξ1 −1) r(ξ + qnξ r0 1 1 1 1 1 −1)ξ1 1 ξ1 ξ1 (6.51) .. . 0 0 0 0 0 0 r0 + qnξ + . . . + qn(ξ1 −1) r(ξ + qn4 r4ξ + qn3 r3ξ ãnξ(t+1) = qn1 r1ξ + qn2 r2ξ t t t t 1 ξ1 ξt 1 −1)ξt 0 0 + qn(ξ r0 + . . . + qnξ r0 t ξt ξt 1 +1) (ξ1 +1)ξt .. . (6.52) 0 0 0 0 0 ãnξs = qn1 r1ξ + qn2 r2ξ + qn3 r3ξ + qn4 r4ξ + . . . + qn(ξ1 −1) r(ξ 1 −1)ξ(s−1) (s−1) (s−1) (s−1) (s−1) 0 0 0 + qnξ r0 + qn(ξ r0 + . . . + qnξ r0 . 1 ξ1 ξ(s−1) 1 +1) (ξ1 +1)ξ(s−1) (s−1) ξ(s−1) ξ(s−1) (6.53) 0 We substitute the expression for riξ in (6.45) in the first three elements on the right 1 hand side of (6.51), 0 0 0 qn1 r1ξ + qn2 r2ξ + qn3 r3ξ 1 1 1 = qn1 (r1ξ1 + (ãnξ2 − ãnξ1 )qn1 ) + qn2 (r2ξ1 + (ãnξ2 − ãnξ1 )qn2 ) + qn3 (r3ξ1 + (ãnξ2 − ãnξ1 )qn3 ) 2 2 2 = qn1 r1ξ1 + qn2 r2ξ1 + qn3 r3ξ1 + (ãnξ2 − ãnξ1 )(qn1 + qn2 + qn3 ) 2 2 2 = ãnξ1 + (ãnξ2 − ãnξ1 )(qn1 + qn2 + qn3 ). (6.54) 0 Similarly, we substitute the expression for riξ in (6.46) in the first three elements t 129 on the right hand side of (6.52), 0 0 0 qn1 r1ξ + qn2 r2ξ + qn3 r3ξ t t t = qn1 (r1ξt + (ãnξt+1 − ãnξt )qn1 ) + qn2 (r2ξt + (ãnξt+1 − ãnξt )qn2 ) + qn3 (r3ξt + (ãnξt+1 − ãnξt )qn3 ) 2 2 2 ) + qn3 + qn2 = qn1 r1ξt + qn2 r2ξt + qn3 r3ξt + (ãnξt+1 − ãnξt )(qn1 2 2 2 = ãnξt + (ãnξt+1 − ãnξt )(qn1 + qn2 + qn3 ). (6.55) Then adding together all equations (6.50) through (6.53) using the substitutions shown in (6.54) and (6.55) yields ãnξ1 + . . . + ãnξs = 2 2 2 ãnξ1 + . . . + ãnξs + (ãnξ2 − ãnξ1 )(qn1 + qn2 + qn3 ) 2 2 2 2 2 2 + (ãnξ3 − ãnξ2 )(qn1 + qn2 + qn3 ) + . . . + (ãnξs − ãnξs−1 )(qn1 + qn2 + qn3 ) 2 2 2 + . . . + (ãnξs − ãnξs−1 )(qn1 + qn2 + qn3 ) 0 0 0 0 0 0 + qn4 (r4ξ + r4ξ + . . . + r4ξ ) + qn5 (r5ξ + r5ξ + . . . + r5ξ )... s s 1 2 1 2 0 0 0 + qn(ξ1 −1) (r(ξ ) + r(ξ + . . . + r(ξ 1 −1)ξ1 1 −1)ξ2 1 −1)ξs 0 0 0 0 + qnξ r0 + . . . + qnξ r0 + qn(ξ r0 + . . . + qnξ r0 . s ξs ξs 1 ξ1 ξ1 1 ξ1 ξs 1 +1) (ξ1 +1)ξs (6.56) After cancellation of terms and substituting (6.47), (6.56) reduces to 0 0 0 0 0 = qnξ r0 + . . . + qnξ r0 + qn(ξ r0 + . . . + qnξ r0 . s ξs ξs 1 ξ1 ξ1 1 ξ1 ξs 1 +1) (ξ1 +1)ξs (6.57) This equation is analogous to (6.28). Again, the QR factorization is not unique for rank-deficient matrices and (6.57) is the only constraint given for rξ0 1 ξ1 , . . . , rξ0 1 ξs , . . . , rξ0 s ξs . Then a simple choice is to 0 set all riξ = 0 for i > ξ1 and let rξ0 1 ξj satisfy (6.47). This can be done while ensuring j that the ξith row is a scaled version of the 4th through (ξ1 − 1)th rows so that the 130 following structure is achieved, ã11 . . . ã1ξ1 . . . ã1ξt . . . ã1ξs . . . ã1N .. .. .. ã21 . . . ã2N .. .. . ãnξ2 . . . ãnξ( t+1) . . . ãnξ1 . .. .. .. .. .. . . . . . ãL1 . . . q11 q12 = q21 q22 q31 .. . q32 .. . qn1 qn2 .. .. . . qL1 qL2 r11 r12 × r22 ... ... q13 . . . 0 q1ξ 1 ... ... ... 0 q1ξ t ... ... . . . ãLN 0 q1ξ s ... 0 q1N 0 0 0 0 q23 . . . q2ξ . . . q . . . q . . . q 2N 2ξt 2ξs 1 0 0 0 0 q33 . . . q3ξ . . . q . . . q . . . q 3N 3ξt 3ξs 1 .. .. .. .. .. .. .. .. .. . . . . . . . . . 0 0 0 0 qn3 . . . qnξ1 . . . qnξt . . . qnξs . . . qnN .. .. .. .. .. .. .. .. .. . . . . . . . . . 0 0 0 0 qL3 . . . qLξ1 . . . qlξt . . . qLξs . . . qLN 0 0 0 r13 . . . r1ξ1 . . . r1ξt . . . r1ξs . . . r1N 0 0 0 r23 . . . r2ξ1 . . . r2ξt . . . r2ξs . . . r2N 0 0 0 r33 . . . r3ξ . . . r . . . r . . . r 3N 3ξ 3ξ t s 1 . .. .. .. . . . 0 0 0 rξ1 ξ1 rξ1 ξt rξ1 ξs (6.58) All of the 4th through ξ1th rows of R0 are linearly dependent, but linearly independent from the first three rows. This results in R0 having rank four. It has been verified by the author using MATLAB that when QR factorizations of matrices such as (6.42) are done numerically, the resulting structure is exactly that shown in (6.58). This completes the proof for the case ξ1 > 3. Case ξs ≤ 3 : We now consider the case where all the scatterers switched are within the first three columns of Ã. Clearly the only way this can happen is if only three scatterers switch. Again, recall that A was multiplied by a permutation matrix Π so that the 131 resulting matrix Ã’s first three columns are then linearly dependent. We consider three arbitrary linearly independent vectors and show that switching corresponding elements in all three of them results in another vector being required to make a full basis. Let the three arbitrary linearly independent vectors w, x, y, and z be defined as in (6.31). Instead of switching the nth entries of the three vectors all at once, we consider first the switch of just two elements as in (6.32), then make another switch of two elements so that afterward, all three have been switched. Ensure that n is chosen such that w(n) 6= x(n) 6= y(n), so that an actual switch in correlation occurs. Also assume that a 6= b 6= c, since a = b = c is the trivial case in which the change in correlation is not detectable. The case a = b = c corresponds to rij = rkj = rlj for some i, j, k, l. By (6.10), this can only happen if qi = qj = qk . Since by Lemma 2 Q0 is full rank, no three columns can be identical. Therefore, this trivial case is not possible here. Recall that after a switch of w(n) and x(n), (6.33) was modified to (6.36). To consider the case where three scatterers switch, we now switch w(n) and y(n) in (6.36). Proceeding as before, this results in ax(n) + by(n) + cw(n) = z(n) + (a − b)(w(n) − x(n)) − bw(n) − cy(n) + by(n) + cw(n) = z(n) + (a − b)(w(n) − x(n)) + (b − c)(y(n) − w(n)). (6.59) As before, this results in an additional vector being required to span the space. This vector is all zeros besides (b − c)(y(n) − w(n)) in the nth entry. Therefore, this additional vector is linearly dependent with the first additional vector that was needed to span the space, the vector that was all zeros besides (a − b)(w(n) − x(n)) in the nth entry. Returning to the problem at hand, we see that if all three elements are switched within the first three columns of Ã, one more column will then be required to span 0 the space of à . Again, as described in the proof of the case where only two scatterers 0 switch in the first three columns of Ã, this results in à having four nonzero singular 132 values. So switching three scatterers in the first three columns has the same effect on rank as switching just two scatterers in the first three columns. This completes the proof for the case ξs ≤ 3. Case ξ1 ≤ 3, ξs > 3 : We lastly consider the case where more than three scatterers switch, and the scatterers can be both within the first three columns and in the 4th through N th columns. A switch of more than three scatterers can be thought of as a sequence of two-scatterer switches. We break down the switches in to the order 1. A switch of any number of scatterers from the first three columns with scatterers from the 4th through N th columns, 2. A switch of scatterers within the first three columns, 3. A switch of scatterers within the 4th through N th columns. In the case where two scatterers switch and k ≤ 3, m > 3, we saw that this increased rank by one. A quick review of this proof easily shows that if another switch occurred in this same manner in the same row after the first switch, the result would be that another vector that is all zeros besides an entry in the nth location would be required to span the space. However, this vector is linearly dependent with the fourth vector that was already found to be required to span the space after the first switch. Clearly, no matter how many switches of this nature occur, each will require another vector to span the space, and each subsequent vector will be linearly dependent on the ones before it. So no matter how many switches of this nature occur, rank will increase only by one. In the case where only two scatterers switch for k > 3, m > 3, we saw that this increased rank by one. In the case where more than two scatterers switch for ξ1 > 3, we saw that despite this increase in the number of switches, rank still increased by only one. Similarly, in the case where only two scatterers switch for k ≤ 3, m ≤ 3, we saw that this increased rank by one. In the case where three scatterers switch for 133 ξ3 < 3, we saw that this additional switch in the first three columns did not increase the overall rank by more than initial switch; rank still increased by only one. Based on these previous conclusions, it follows that when more than three scatterers switch, and the scatterers can be both within the first three columns and in the 4th through N th columns, rank still increases by only one. This completes the proof for when ξ1 ≤ 3, ξs > 3. Notice that in all three cases, it was shown that switching any number of elements at the same aspect increases the rank by one. Therefore, we have shown that switching any number of elements at the same aspect increases the rank regardless of what the indices ξi are. This increase in rank yields an increase in the number of nonzero singular values. This completes the proof for the case where more than two scatterers switch at a single aspect. 6.3.3 Case Scatterers Switch at Multiple Aspects All the previous cases have shown that no matter how many scatterers are switched at a given aspect, the rank of the matrix will increase by one, increasing the number of nonzero singular values by one. We now show that no further changes in correlation at subsequent aspects will 0 ever decrease the rank. Consider partitioning the matrix à into two matrices, the first of which being the first n rows and the second of which being the (n + 1)th through Lth rows, 0 à = 0 0 . Ã1 Ã2 (6.60) 0 0 Since the only switch of scatterers in à occurs in its nth row, the matrix Ã2 is 0 correlated, and therefore by Lemma 1, it has rank 3. As shown in this proof, Ã1 has rank 4. Notice that Lemma 4 applies to matrices partitioned side by side, instead of 0 one above the other, as the matrix à is. 134 0 0 Rearranging (6.7) and substituting in Ã1 and Ã2 gives 0 0 0 0 0 0 rank([Ã1 , Ã2 ]) ≥ rank(Ã1 ) + rank(Ã2 ) − min(rank(Ã1 ), rank(Ã2 )). (6.61) By basic properties of matrices, 0T 0 0T 0 rank(Ã1 ) = rank(Ã1 ) (6.62) and rank(Ã2 ) = rank(Ã2 ). 0T (6.63) 0T We can plug Ã1 and Ã1 in to (6.61) to find that 0T 0T 0T 0T 0T 0T rank([Ã1 , Ã2 ]) ≥ rank(Ã1 ) + rank(Ã2 ) − min(rank(Ã1 ), rank(Ã2 )). (6.64) Plugging (6.62) and (6.63) in to (6.64) yields 0T 0T 0T 0T 0 0 0 0 rank([Ã1 , Ã2 ]) ≥ rank(Ã1 ) + rank(Ã2 ) − min(rank(Ã1 ), rank(Ã2 )). Since 0T 0T rank([Ã1 , Ã2 ]) = rank([Ã1 , Ã2 ]T ) = rank 0 0 = rank(Ã0 ), Ã1 Ã2 we find that 0 0 0 0 0 rank(à ) ≥ rank(Ã1 ) + rank(Ã2 ) − min(rank(Ã1 ), rank(Ã2 )). 0 0 (6.65) 0 Since rank(Ã1 ) = 4 and rank(Ã2 ) = 3, (6.65) shows that rank(à ) ≥ 4. This is the 0 for the case where we know that Ã2 is correlated. 0 If, however, Ã2 becomes uncorrelated at multiple aspects, its rank may change. 135 0 We let rank(Ã2 ) = r, where 1 ≤ r ≤ N . Then (6.65) becomes 4, 1 ≤ r ≤ 3 0 rank(à ) ≥ 4 + r − min(4, r) = 4, r = 4 r, 4 < r ≤ N. (6.66) 0 No matter what changes in correlation occur in Ã2 and regardless of if its rank 0 increases or decreases, (6.66) shows that rank(à ) cannot decrease. Therefore, no changes in correlation at any subsequent aspects can decrease the rank and we will always have that 0 rank(à ) ≥ 4. (6.67) The above cases and these observations together are sufficient to justify the use of the iterative algorithm in practice. It has been shown that uncorrelated data at just one aspect increases the rank from 3 to 4, and that this rank cannot then decrease at any subsequent aspects. As a result, uncorrelated data will always have at least four nonzero singular values (and more are possible resulting from changes in correlation at subsequent aspects). Therefore, this proves that a matrix of radar observables data is uncorrelated if and only if it has more than three nonzero singular values. 6.4 Comparison of Algorithms The two proposed methods are useful in different circumstances. If there are a small number of scatterers and not too much noise, then the nearest neighbor and slope algorithm can easily be employed. This runs very quickly and gives good results. When dealing with a large number of scatterers, finding the optimal correlation is not as straightforward as this. This is when the iterative algorithm can be used. It takes longer to run, but it is the most effective option in this case. When the number of scatterers is small, the iterative algorithm cannot be used anyway. This is because the algorithm depends on there being enough singular values after the first three that the algorithm can operate based on minimizing them. 136 If there are only four scatterers, this means the algorithm would be minimizing just the one last singular value, which would not be very effective. Figure 6.31 shows a singular value distribution from a target with four scatterers. Figure 6.31: Singular Value Distribution for 4 Scatterer Target As can be seen, all of the last singular values are zero. The iterative algorithm would be useless in this case, since it would be trying to find the minimum of a set of values that are all zero. Figure 6.32 shows a singular value distribution from a target with five scatterers. Here the algorithm correctly correlates uncorrelated data in the case of only five scatterers. Notice that the algorithm had to minimize only one value, since the last singular value seems to always be zero. From this, we can conclude that the iterative algorithm can be used whenever there are more than four scatterers. The nearest neighbor and slope algorithm quickly breaks down when more than a few scatterers are present, since many tracks can have similar slopes at the same aspect. The method seems to work well for up to five scatterers, but doesn’t do as well as more scatterers are added in. One thing to notice is that many ballistic targets will have a very small number of scatterers, since many of them are simple shapes such as cones. It is a very 137 Figure 6.32: Singular Value Distribution for 5 Scatterer Target reasonable assumption that a ballistic target could have four or five scatterers. Both of the correlation algorithms can be very useful for targets of this nature. A key attribute to notice is that the iterative algorithm uses all data at once, while the nearest neighbor and slope algorithm doesn’t necessarily have to. The number of scatterers is manageable enough in the nearest neighbor and slope algorithm that the data can be looked at sequentially. If needed, the PESVD method could be run periodically throughout the algorithm and the shape estimate modified as more data is acquired. This is a feature that the iterative algorithm does not have. The iterative algorithm is useful when there is so much data that a sequential method is impractical. The reason the iterative algorithm can’t correlate data sequentially is that the SVD needs to be applied to all the data as a whole in order to analyze its singular value structure. Due to this, it makes no sense to correlate the data aspect by aspect using an SVD. CHAPTER 7 Discussion and Conclusions 7.1 Summary An introduction is provided that gives a brief background of radar and explains the ideas of SAR, ISAR, and IFSAR, and scatterer correlation. An overview of previous work is given in order to explain where the work of this thesis fits in with other previous work. Background is given on key ideas required to understand the work of this thesis, including electromagnetic wave propagation and scatterers. Background is given on how radar data is collected and processed in order to arrive at the radar observables used for radar imaging: range, range-rate, and delta phase. This is done using Fourier processing in order to arrive at Range-Doppler data blocks from which the radar observables are extracted. Scatterer correlation is introduced, which is a key concept in this thesis. Background is given on the sample target used and how radar data was simulated from it for this thesis, and resolution constraints are discussed. A brief review of the Singular Value Decomposition (SVD) and QR factorization is given. An overview of the phase-enhanced 3D snapshot imaging method is given, along with an overview of how it fits in with the general 2D and 3D snapshot imaging equations. A derivation of the phase equation, as presented in [7], is provided, along with a new alternate derivation of the phase equation directly from the equation of the Range-Doppler data block. This is shown to verify that it agrees with the phase equation derived directly from the snapshot equations. The phase-enhanced 3D snapshot imaging method estimates shape given uncorrelated range, range-rate, and delta phase data and a priori motion. Considerations that must be taken in to account when implementing the phaseenhanced 3D snapshot imaging algorithm are presented. This includes specifics related to the coordinate system that is used. A new method of handling the phase ambiguities that inherently arise when using this method is presented. A summary 138 139 of the 3DMAGI SVD method is also presented. This method estimates both shape and motion given correlated range data. The hybrid phase-enhanced SVD (PESVD) method is then presented. This method combines the phase-enhanced 3D snapshot imaging algorithm and the 3DMAGI SVD method in to a hybrid method that uses range, range-rate, and delta phase data in order to estimate both shape and motion. In order to use this method, data must still be correlated (this motivates the next section). Scaling considerations are presented in order to ensure that the shape and motion estimates are reconstructed with accurate scaling. The effects of imperfect correlation and added noise are investigated. Two correlation algorithms are presented. The iterative correlation algorithm is applicable for targets with more than four scatterers. It can be used in the presence of noise and can resolve phase ambiguities. The method iterates on a set of possible correlations for radar data and chooses the optimal correlation based on a condition on the singular values resulting from the PESVD method. The second algorithm is applicable to targets with fewer scatterers. It uses the ideal continuity of the radar observables and the fact that range-rate data should ideally be the derivative of range data in order to methodically correlate radar observables data. Both methods can handle reasonable quantities of noise, but of the two, the iterative algorithm is better at handling noise. Lastly, a mathematical proof is provided of a theorem that justifies the use of the iterative algorithm. The proof is presented for the case when two scatters switch at a given aspect in order to make data uncorrelated, and for the case when more than two scatterers switch at a given aspect. These separate cases are included because the two-scatterer case is very straightforward and easy to follow. It is important to note the limitations of this work: • Can’t handle occluded scatterers • Effects of preprocessing noise haven’t been investigated • Can’t handle slipping scatterers • Can’t handle speculars 140 • Resolution is limited by FFT size 7.2 Main Contributions Previous work and the author’s new work are intermixed throughout the thesis. Therefore, a some of the author’s main contributions are listed below. • Developed a method to combine the Phase-Enhanced 3D Snapshot Imaging Algorithm and the 3DMAGI SVD Method in to a hybrid method that uses range, range-rate, and delta phase to estimate both shape and motion. • Developed an alternate derivation of the phase equation directly from the equation of the Range-Doppler data block and verified that it agrees with the phase equation derived from the snapshot equations. • Developed a method to resolve phase ambiguities in delta phase data taken when motion is known so that targets are reconstructed in target coordinates. • Developed a method to simulate realistically uncorrelated radar observables data with realistic noise and phase ambiguities. • Developed a methodical correlation algorithm that eliminates the need for scatterer correlation in the case of a small number of scatterers in the presence of noise and can resolve phase ambiguities. • Developed an iterative correlation algorithm that eliminates the need for scatterer correlation in the case of a large number of scatterers in the presence of noise and can resolve phase ambiguities. • Proved a theorem relating the number of nonzero singular values to whether or not a matrix of radar observables data is correlated. This theorem justifies the use of the iterative algorithm. 7.3 Future Work There are many possible future research topics that result from the work of this thesis. A few are listed below. 141 • Extend PESVD method to incorporate data from multiple sensors. The ability to extend this method to multistatics would be very useful. • Investigate the effect of preprocessing noise on the PESVD method. For example, varying Signal-to-Noise Ratios (SNRs) could be tested. It would be especially interesting to investigate what SNR results in noise so high that one of the three singular values used in the PESVD method results completely from noise, rendering the method is unusable. LITERATURE CITED [1] C. Groetsch, Inverse Problems. Washington, DC: The Mathematical Association of America, 1999. [2] W. Carrara et al., Spotlight Synthetic Aperture Radar: Signal Processing Algorithms. Norwood, MA: Artech Print on Demand, 1995. [3] M. Cheney and B. Borden, Fundamentals of Radar Imaging. Philadelphia, PA: SIAM, 2009. [4] C. Ozdemir, Inverse Synthetic Aperture Radar Imaging With MATLAB Algorithms. Hoboken, NJ: Wiley Series in Microwave and Optical Engineering, 2012. [5] M. Stuff, “Three-dimensional Analysis of Moving Target Radar Signals: Methods and Implications for ATR and Feature Aided Tracking”, in Proc. SPIE3721 Algorithms for Synthetic Aperture Radar Imagery VI, 1999, pp. 485-496. [6] A. Grigis and J. Sjostrand, Microlocal Analysis for Differential Operators, an Introduction (London Mathematical Society Lecture Note Series 196). Cambridge, UK: Cambridge University Press, 1994. [7] J. Mayhan et al., “High Resolution 3D ’Snapshot’ ISAR Imaging and Feature Extraction”. IEEE Trans. Aerosp. and Electron. Syst., vol. 37, no. 2, pp. 630-641, 2001. [8] J. Mayhan, “Phase-Enhanced 3D Snapshot ISAR Imaging and Interferometric SAR,” MIT Lincoln Laboratory, Lexington, MA, Tech. Rep. 1135, 2009. [9] C. Tomasi and T. Kanade, “Shape and Motion from Image Streams: A Factorization Method”. Carnegie Mellon University, Pittsburgh, PA, Tech. Rep. CMU-CS-91-105, 1991. 142 143 [10] M. Stuff, “Derivation and Estimation of Euclidean Invariants of Far Field Range Data,” Ph.D. Dissertation, Dept. Stats., Univ. of Michigan, Ann Arbor, 2002. [11] M. Ferrara et al., “Shape and Motion Reconstruction from 3D-to-1D Orthographically Projected Data via Object-Image Relations”. IEEE Trans. Pattern Anal. Mach. Intell., vol. 31, no. 10, pp. 1906-1912, 2009. [12] R. Sullivan, Radar Foundations for Imaging and Advanced Concepts. Herndon, VA: SciTech Publishing, 2004. [13] R. Becker, Introduction to Theoretical Mechanics. New York, NY: McGraw-Hill Book Company, Inc, 1954. [14] A. Bjorck and G. Dahlquist, Numerical Methods in Scientific Computing, Volume II. Philadelphia, PA: SIAM, 2008. [15] L. Trefethen and D. Bau, Numerical Linear Algebra. Philadelphia, PA: SIAM, 1997. [16] H. Yanai et al., Projection Matrices, Generalized Inverse Matrices, and Singular Value Decomposition. New York, NY: Springer, 2011.