Deconvolution Photometry of QSO 0957+561 A,B Candidata Scientiarum thesis by Anna Kathinka Dalland Evans Institute of Theoretical Astrophysics University of Oslo Norway November 2003 The years teach much which the days never knew. Ralph Waldo Emerson In memory Til Mormor, den tøffeste av alle & To Granny & Grandad, as promised Cover picture: North is up, East is left on all four images. Top left: Observed image of the quasar components. Top right and bottom left: Residual images. Bottom right: Deconvolved image. Thanks! One of the symptoms of an approaching nervous breakdown is the belief that one’s work is terribly important. Bertrand Russell The work with this thesis turned out to be different than expected. It was a genuine relief when I discovered that the frustrating ‘obstacles’ in my way were actually part of my thesis work rather than merely annoying hindrances. As the road to finishing this work has been rather long, many people have helped me on the way. The first load of ‘thank yous’ goes to my supervisor Rolf Stabell, for taking me on as a student in the first place, for sending me off to Belgium and to the Canary Islands, for letting me go observing at the NOT, and for pestering me throughout with his exquisite exactness in writing (and speaking). Ingunn Burud, thanks for all the help and support you have provided since the first days in Liège, and for becoming my good friend as well as helper. My most sincere and heartfelt thanks also goes to Frédéric Courbin, for swift and steady guidance on using the programs during the last crucial stages. I thank Wesley N. Colley and Rudy Schild for providing me with photometric results for Q0957 for a comparative analysis, as well as Luis Goicoechea and Aurora Ullán Nieto for sending me their GLITP data for the same purpose. From the social side of life, I would like to thank ‘the gang’ in room 303 at ITA, for the good old days: Benedicte Selmer, Jan-Erik Ovaldsen, Kjetil Kjernsmo and Torstein Sæbø. As it turned out, we were the last group of students to enjoy the luxury of that room before it was confiscated and turned into a (mostly empty) conference room. I would also like to thank my new gang in the prison cell on the 1st floor, where we sat ‘temporarily’ while the institute was besieged by renovation workers. We managed to work practically on top of each other for close to a year without any murders being committed. Thanks to Eirik Mysen and Martin Ytre-Eide for providing comic relief, and to Morten Wergeland Hansen for not throwing me out of the window. The largest thanks of all goes to Mari Anne Killie: You have ii followed this work and all my ups and downs on a daily basis, and I could not have done without you! You’re the kind of girl that fits in with my world. I also have to mention a few other astro-people. Thanks to Håkon Dahle and Vidar Aarum-Ulvås, for your enduring patience with my IRAF questions and for helping me probe my light curves. To Jan-Erik Ovaldsen (again!) and Jan Teuber of AstroConsult A/S I am extremely grateful for help and counselling, for useful comments on my written material and for pointing out many of the seemingly small but oh so crucial details involved in image processing. Thanks to Øystein Elgarøy, for your always clear and thorough explanations to my various questions. To Egil Leer, thank you for asking me ‘how I am doing today’, and for being genuinely interested. I would also like to thank Anlaug Amanda Kaas at the NOT who has helped me with various technical details via e-mail from La Palma. For fun&games, thanks to the rest of the institute’s X-mas party organising committee 2001 not mentioned elsewhere: Astrid Fossum, Eirik Endeve and Torben Leifsen. I cannot resist thanking Google, the friend in need when you are all alone, and Cambridge Dictionaries Online. To my friends outside the astrophysical community, I would like to thank you all for being there for me and for helping me keep my sanity which would otherwise have been gravely endangered after spending all this time with computer nerds. Special thanks to Aina Kristin Seland Nilsen, for remaining my oldest friend in Oslo and for being a rock whenever I freak out. To Øyvind Kristiansen, for late evenings and early mornings. To Benedicte Karen Fasmer Waaler, for encouragement and for everything else, especially for the year 1995–96. To Henriette Nilsson, for coffee and opera in the morning. And to Rikke Kristin Gåsholt, for throwing great parties! I would also like to thank my fantastic mathematics teacher at Bergen Katedralskole, Anne Berit Holme, and Geir Ole Øverby for helping me trust my own sense of logic when I returned to the natural sciences. To my closest family: My mother Tordis Dalland, my father Bernard John Evans, my brother Peter John Evans, my uncle Arild Jensen and Pus the Cat, thank you all for your interest and support. Thanks are also due to my newly acquired extended family: Kjell Myhren, Bozena Elizabeth Jakubowska, Rosallina and Toby, for taking such good care of my parents. Finally: Terje Fredvik - my IDL guru, my sparring partner in discussions on observational astronomy, my playmate and friend. It has been such a lot of fun sharing this with you, both at work and off, and you know how invaluable you have been in pushing me on and getting me through. Thanks for your endless support, for reading the thesis, for holding my hand and especially for being the one thing during this time that was always right. I’m on another planet with you! Contents Preface i 1 Introduction 1.1 General outline of the thesis . . . . . . . . . . . . . . . . . . . 2 Studying QSO 0957+561 A,B 2.1 Data sets . . . . . . . . . . . . 2.1.1 The field . . . . . . . . . 2.2 Gravitational lensing of quasars 2.2.1 Time delay . . . . . . . 2.2.2 Microlensing . . . . . . . 2.3 A brief time delay history . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 Pre-processing the ALFOSC data set 3.1 A look at the equipment: The CCD . . . . . . . . . . 3.1.1 The ALFOSC CCD . . . . . . . . . . . . . . . 3.2 Bias and overscan . . . . . . . . . . . . . . . . . . . . 3.3 Flat-fielding . . . . . . . . . . . . . . . . . . . . . . . 3.3.1 Inspecting the flats . . . . . . . . . . . . . . . 3.3.2 The moving blob . . . . . . . . . . . . . . . . 3.4 Other preparations . . . . . . . . . . . . . . . . . . . 3.4.1 Combining images . . . . . . . . . . . . . . . 3.4.2 Scaling, aligning and background subtraction . 3.4.3 Bad Columns . . . . . . . . . . . . . . . . . . 3.4.4 The importance of careful pre-processing . . . 4 The MCS procedure 4.1 Convolution . . . . . . . . . . 4.1.1 The sampling theorem 4.2 Deconvolution . . . . . . . . . 4.2.1 Traditional methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 2 . . . . . . 5 5 7 7 8 10 12 . . . . . . . . . . . 13 13 14 16 18 20 21 24 24 25 27 29 . . . . 31 31 34 35 35 iv CONTENTS 4.2.2 Deconvolution with correct sampling: 4.3 Preparations for running MCS . . . . . . . . 4.4 The aim of MCS . . . . . . . . . . . . . . . 4.5 General outline of MCS . . . . . . . . . . . 4.5.1 Extracting data frames . . . . . . . . 4.5.2 Constructing the PSF . . . . . . . . 4.5.3 Deconvolution with the profile s . . . 4.5.4 Residual frames . . . . . . . . . . . . 4.6 Simultaneous deconvolution . . . . . . . . . 4.6.1 Final deconvolution results . . . . . . 4.6.2 Obstacles . . . . . . . . . . . . . . . 4.6.3 Testing MCS . . . . . . . . . . . . . a . . . . . . . . . . . preview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 36 38 39 41 41 47 49 50 52 53 55 5 Photometry from the ALFOSC data 5.1 Basic principles of photometric studies . . . . . . . 5.1.1 Standard systems and reference magnitudes 5.1.2 Colours . . . . . . . . . . . . . . . . . . . . 5.1.3 Photometry from MCS . . . . . . . . . . . . 5.2 Photometric results . . . . . . . . . . . . . . . . . . 5.2.1 Error estimations . . . . . . . . . . . . . . . 5.2.2 Light curves from March 2001 . . . . . . . . 5.2.3 Light curves from January 2000 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 59 61 62 63 64 64 66 76 6 Discussion 6.1 Testing the results . . . . . . . . . . 6.1.1 A simple χ2 test . . . . . . . . 6.1.2 Seeing . . . . . . . . . . . . . 6.1.3 Possible influence of the Moon 6.1.4 Colour diagrams . . . . . . . 6.2 Comparing the time shifted data . . 6.3 Zero lag correlation . . . . . . . . . . 6.4 Comparisons with other results . . . 6.4.1 Results from Ovaldsen et. al. . 6.4.2 Results from Colley et. al. . . 6.4.3 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81 81 81 84 88 93 95 98 99 100 102 105 7 Reduction and photometry of 7.1 Pre-processing . . . . . . . . 7.2 Running MCS . . . . . . . . 7.3 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . the StanCam . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . data 107 . . . . . . . . . . 107 . . . . . . . . . . 109 . . . . . . . . . . 109 CONTENTS v 8 Conclusions 115 8.1 Future studies . . . . . . . . . . . . . . . . . . . . . . . . . . . 117 Bibliography A The Image Reduction and Analysis Facility (IRAF) A.1 Using IRAF . . . . . . . . . . . . . . . . . . . . . . . A.2 Pre-processing with IRAF . . . . . . . . . . . . . . . A.2.1 Zerocombine . . . . . . . . . . . . . . . . . . . A.2.2 Ccdproc and Flatcombine . . . . . . . . . . . . A.3 Other tasks using IRAF . . . . . . . . . . . . . . . . A.3.1 Image scaling and background subtraction . . A.3.2 Stacking the images . . . . . . . . . . . . . . . 119 . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123 . 123 . 124 . 124 . 124 . 126 . 126 . 126 B Text files associated with the MCS programs 131 C Tables of output files from the MCS programs 137 D Miscellaneous 139 D.1 Integrating the Gaussian function . . . . . . . . . . . . . . . . 139 vi CONTENTS Chapter 1 Introduction “The time has come,” The Walrus said, “To talk of many things” Lewis Carroll, Through the Looking-Glass and what Alice found there Observational astronomy is one of the earliest human activities that can be described as scientific, at least in the sense of making predictions based on careful studies of natural phenomena. The most famous early examples are probably the ancient Oriental astronomers, the Greeks and the Aztec civilisation. The impressive knowledge these societies had of the movements of the stars and planets never stops to fascinate and intrigue us. They not only gathered information from observations, they used this data to calculate the distance around the Earth and make predictions about the future movements of celestial objects. It was, however, the business of the philosophers of ancient Greece to discover ‘how the world really works’. The main function of astronomers, through antiquity and the Middle Ages, was to provide the mathematical basis for astrology. Not until the period of what is known as The Scientific Revolution1 did the natural sciences emerge as an autonomous field of research. During this epoch, the foundation for the scientific method was laid. Speculation and theory about the nature of the universe on the one hand, and the empirical gathering of information on the other, were joined in a mutually dependent system. Another aspect of this period is the technological development that occurred. With the invention of the telescope in the early 1600s came the first apparatus that enhanced the sensitivity of the organic light detector – the human eye. Intriguing observations stimulated the development of bigger and better telescopes. The ensuing astronomical discoveries in turn inspired 1 Classically dated as 1543 – 1687, that is, the years between the death of Copernicus (and the publication of his work De revolutionibus orbium cælestium) and Newton’s Philosophiae Naturalis Principia Mathematica. 2 Introduction the building of even better telescopes. This interplay between technology and science is of course a continuing process, but it is notably marked by sporadic progress following particular inventions opening up hitherto untouched territories. When Joseph Fraunhofer (1787–1826) invented the prism spectroscope, he initiated the investigation of the chemical composition of the Sun. When gelatine-based photographic emulsions and the introduction of silver bromide, which enhanced the light sensitivity of photographic film, became common in the late 19th century, astronomers quickly started applying this to make better observational records. This trend has continued from the Renaissance to the present day in an ever-increasing spiral. Each improvement of scientific apparatus helps to provide answers to old questions as well as raising a whole host of new ones, which in turn drives the quest for better equipment. Better equipment is one way of obtaining better observations. Another approach, which has accelerated in importance in later years with advancing computer technology, is the development of new and better image processing routines. With this approach, images are improved in quality after the observation itself has been conducted. Here we see again the effect technological progress has on scientific observations. Developing computer programs that reduce noise and enhance the resolution of images, effectively increases and hopefully improves the science we can get out of our observations. Many devices to this effect are created today. 1.1 General outline of the thesis Accurate photometry is crucial to obtaining reliable measurements of, e.g., the time delay between images of a multiply lensed quasar. Precision photometry of QSO 0957+561A,B is the main objective of this thesis work. We will be using a relatively new method of improving image quality. The MCS routine, named after its developers Pierre Magain, Frédéric Courbin and Sandrine Sohy, uses deconvolution with correct sampling to improve the resolution of astronomical images, without the drawbacks traditional deconvolution routines have suffered from, see Magain et al. (1998). The method is described in Chapter 4. We will utilise MCS to obtain photometric results from our data sets of the gravitationally lensed quasar QSO 0957+561 and some adjoining field stars. In Chapter 2 we describe the field of our observations and the quasar we are studying, and we briefly discuss the principles of gravitational lensing in general as well as some of its applications. The data sets we use were obtained by us and other groups during several observing sessions, from the Nordic Optical Telescope situated on the Roque de los 1.1 General outline of the thesis 3 Muchachos on La Palma. The main part of this thesis consists of analysing two data sets from observations with the ALFOSC instrument. Before starting to extract photometry from the images, we must correct, as well as can be achieved, for effects introduced by the passage of light through the detector. This pre-processing is described in Chapter 3. Results from the photometry itself follow immediately after the description of the MCS programs, in Chapter 5. A detailed discussion and testing of the results is presented in Chapter 6. In addition, we have analysed some images provided by the GLITP cooperation from the instrument StanCam at the NOT. The results from this data set are presented in Chapter 7. Finally, a short summary of our main conclusions is provided in Chapter 8. Hoping that it can be of assistance to others who want to start penetrating the world of IRAF reductions, we include in Appendix A a brief description of the tasks and parameters we have used during the course of this work. This is not in any way intended as a general introduction to the many reduction options available in IRAF, but more as a guideline for students and others seeking to do similar work as what we have done here. Since few such overviews exist (see, however, Appendix A in Dahle, 1994), we found this relevant to include. 4 Introduction Chapter 2 Studying QSO 0957+561 A,B The truth is rarely pure and never simple. Oscar Wilde, The Importance of Being Earnest, Act I Since its discovery and identification as a gravitationally lensed double object by Walsh et al. (1979), the quasi-stellar object scientifically named QSO 0957+561 and familiarly dubbed the ‘Twin Quasar’ has been the target of much study and debate. Shortly after its discovery, changes in brightness of the quasar images were discovered and the system became the first for which an attempt to measure the time delay was reported, see the introduction in Kundic et al. (1995) for a list of references. Accuracy of the photometric results is crucial when using light curves to measure the time delay and eventually e.g. calculate the Hubble parameter H0 (see Section 2.2). This thesis work is a detailed study of optical monitoring of the quasar, from three sets of observations. As our data were unsuited for time delay measurements, we focus on precision photometry. In this chapter, we first describe our data sets and discuss some basic elements from the field of gravitational lensing and its uses as a cosmological probe. In the last section, we recapture briefly the history of the Twin Quasar and recent efforts to measure the time delay. 2.1 Data sets In this section, we briefly describe the data sets we have reduced and which form the observational basis for this thesis work. All observations were obtained at the Nordic Optical Telescope (NOT) situated at the Observatorio del Roque de los Muchachos on the island of La Palma in the Canary Islands. The data and the CCDs are described in further detail in Section 3.1.1 and Chapter 7. 6 Studying QSO 0957+561 A,B We have analysed two sets of observations from the ALFOSC instrument (The Andalucia Faint Object Spectrograph and Camera).1 Our main set of observations consisted of 997 images, 806 in the R filter and 191 in V . We obtained these images during 4 nights of photometric observing conditions from March 14th–18th 2001.2 This constitutes the main data set for this thesis work. Our second set of observations from ALFOSC were from an observing run in January 2000.3 This set consisted of data from a single night, January 25th. We had a total of 171 R and 39 V images from this night. The data from these two observing sessions was obtained as part of the ‘QuOC Around the Clock’4 campaign 2000/2001, a collaborating network involving 10 participating observatories (12 in 2001). The campaign was initiated by Wesley Colley and Rudy Schild, with the aim of closely monitoring QSO 0957+561 in order to determine more accurately the time delay of the system. The quasar was observed continuously for a total of 10 nights in January 2000 and again in March 2001. Colley et al. (2002) and Colley et al. (2003) describe the official QuOC results. The observations were, as requested from the QuOC team, run in cycles of 60 minutes of imaging in the R band, 15 minutes in V, 60 minutes R, 15 minutes V, ... 60 minutes R. The finding chart in Figure 2.1 identifies five standard stars that the QuOC team asked to make sure were unsaturated; the brightest is star G. The ideal exposure would bring this star to about 90% of the saturation limit. We have also reduced a third data set, which was obtained from the GLITP5 collaboration. Luis Goicoechea and Aurora Ullán Nieto kindly provided us with a set of data from the StanCam instrument at the NOT, a total of 75 R and 53 V images. The dates on these observations ranged from February 4th to March 30th 2000. With some modifications, the data from each night consisted of two R frames and one V frames, see Chapter 7 for details. 1 ALFOSC is owned by the Instituto de Astrofisica de Andalucia (IAA) and operated at the Nordic Optical Telescope under agreement between IAA and the NBIfAFG of the Astronomical Observatory of Copenhagen. 2 Observers were Kjetil Kjernsmo, Øyvind Saanum and Anna Kathinka Dalland Evans. 3 Observers were Kjetil Kjernsmo and Øyvind Saanum. 4 Quasar Observing Consortium for Around the Clock Monitoring. 5 The Gravitational Lenses International Time Project. 2.2 Gravitational lensing of quasars 2.1.1 7 The field Figure 2.1 shows the central field of all our image frames from the ALFOSC data. We see that there are several stars in the field, the brightest are for historical reasons marked with letters. There are two images of the quasar, labeled A and B, with a separation of 6.2 00 . The main lensing galaxy, close to the centre of a galaxy cluster, is an elliptical (cD) galaxy, at redshift zL = 0.36. The redshift of the quasar is zQSO = 1.41. The centre of the galaxy is located only ∼1 00 from the centre of the B image, slightly to the North–East of B. Figure 2.1: The central part of our image frames, showing the quasar components A and B of QSO 0957+561 and the ‘classic’ field stars F, G, H, E, D as well as the extra stars labeled X, R, P and Q. This image is 3.8 0 × 3.0 0 . North is up, East is left. 2.2 Gravitational lensing of quasars According to Einstein’s General Theory of Relativity, matter bends light. This prediction was first verified during the total solar eclipse of 1919, when the apparent position of stars just outside the solar limb were found to have temporarily changed due to the gravitational field of the sun. 8 Studying QSO 0957+561 A,B Light rays travelling towards us from distant celestial sources, can pass through large concentrations of matter such as galaxies or galaxy clusters. The trajectories of photons passing through gravity fields, are curved as the photons follow the geodesics of curved space-time. From our point of view, the intervening mass between us and the source may be seen to act as a lens, splitting and distorting the image of the distant object. The manner of this distortion will, in addition to the amount of intervening mass and the length of the distances involved, depend on the exact geometrical configuration of the triumvirate observer – lensing mass – lensed object. The nature of Quasi Stellar Objects (‘quasars’ or QSOs) is one of the great enigmas in modern cosmology. Quasars are very remote and extremely luminous, apparently starlike objects with strong emission lines. They are thought to be the central parts of active galaxies. The intense radiation emitted from quasars is often variable in time. Their high redshifts, along with their intense luminosity, makes them excellent candidates for gravitational lensing. That they are far away increases the probability that their light will pass through large mass concentrations on the way to our detectors. That they are so luminous is a prerequisite to us observing them at all at redshifts up to (so far) 6.4.6 2.2.1 Time delay The geometrical difference between the lensed and unlensed photon paths introduces a time-lag between the arrival times of the lensed and unlensed photons at the position of the observer. This time-lag is called the geometrical time delay. While passing in the immediate vicinity of the gravity field of the lens, the light is also affected by a gravitational time delay. A ‘lensed’ photon will be seen by an observer with a total time delay which is the sum of these two effects, with respect to the observation of the same photon if it were not lensed. When two or more images of the source are observed, it is possible to compare the arrival times of the lensed images and to determine a ‘relative time delay’. As the unlensed source is never visible, this is in fact the only truly measurable quantity, rather than the actual time delay between the lensed and unlensed paths. Refsdal (1964) proposed an original method to use gravitational lensing as a tool to measure the Hubble parameter H0 . This application of gravitational lensing is one of the main motivating factors for the detailed photometric monitoring of lensed quasars like the Twin Quasar. The ‘Refsdal method’, or ‘wavefront method’ of determining H0 through 6 Sloan Digital Sky Survey: http://www.sdss.org/news/releases/20030109.quasar.html 2.2 Gravitational lensing of quasars 9 the time delay, is based on the deformation of wavefronts as light waves move from the source, past the deflector to the observer. This deformation is caused by curvature effects and time retardation described above. Each passage of the wavefront at the observer corresponds to an image perpendicular to the wavefront, forming multiple images. All points located on a wavefront have identical propagation times. The distance between the sections of the wavefront at the observer gives the time delay between the corresponding images. We use a double imaged system as an example, and assume redshifts small enough so that Hubble’s law is valid and an axially symmetric lens. In Figure 2.2 we show a wavefront diagram for this case. Note that the angle between wavefront and line-of-sight should be 90 degrees. The wavefronts cross each other at the symmetry point E, where they represent the same light propagation time. For an observer O located a distance x from the symmetry axis, the time delay must be equal to the distance between the wavefronts at the observer, divided by the speed of light (c). An expression for H0 or, equivalently, for the Hubble age τ0 of our universe, can be derived from Figure 2.2 and the assumption of a deflecting law of the type α̂ ∝ |ξ|(−1) , where α̂ is the deflection angle of a light ray passing within a distance ξ of the lens. The expression for H0 derived in Refsdal and Surdej (1994), is given by τ0 = H0−1 = 2(zs − zd )∆t . zd zs θAB (θA + θB )(2 − ) (2.1) Note that this equation is solely in terms of observable quantities. zs and zd are, respectively, the redshifts of the source and the lens. The relative time delay is ∆t, θA and θB are the angular distances from image to lens, and θAB is the angular separation between the images A and B. For large redshifts, the modification of Hubble’s law makes it necessary to introduce a correction factor in Equation 2.1. Kayser and Refsdal (1983) showed that the correction factor is less than 10–20%. The main difficulty in calculating H0 from QSO 0957+561, is the determination of . This is very difficult for the Twin Quasar, since the surrounding galaxy cluster, as well as the central galaxy, acts as a lens. Conversely, Equation 2.1 can be used to constrain the mass distribution in a given lens, once H0 has been determined form other methods, or from other QSO systems. 10 B :9:9 :9:9 !" Studying QSO 0957+561 A,B ;<; ;<; 87< 87 < S #$ %& (' *)*) 0 0// 34 65 12 34 65 , +,+ .- .- ξ ξ M B A A θS θB θA ~ ~θ A B θA B WAVE FRONT E ~ ~θ A B c ∆t O x Figure 2.2: The wavefront method applied to two selected light rays for the determination of the time delay. Reproduced from a figure in Refsdal and Surdej (1994). See text for details. 2.2.2 Microlensing In Figure 2.2 we showed how light from a distant source was lensed by an axially symmetric lens, represented as a spherical point (M ). As a result of the deflection, the observer sees two images of the source. In reality, the lensing 2.2 Gravitational lensing of quasars 11 galaxy is an extended object containing billions of stars. These stars, and other compact objects (MACHOs), may act as microlenses, perturbing the light bundles so that additional images appear. However, as the typical image separation is proportional to the mass of the lens, these micro-images have a separation of the order of micro-arcseconds, hence the term microlensing. These additional images are well below the resolution of present optical telescopes, so they cannot be seen directly. However, the micro-images change in strength and number with the relative transverse motion between the observer, lens and source. This may cause brightness variations between the macro-images. What we observe, therefore, is a ‘twinkling’ of the quasar components. The detection of microlensing can be useful in several respects. It can be used to extract information about the mass of the microlenses in the lensing galaxy, as well as the size of the continuum and the line emitting regions in the quasar. We see why lensed objects are so valuable a tool. For a single quasar, we do not know the intrinsic variability, as we do not have a complete theory of the nature of quasars. At present, our only chance to prove microlensing is to make use of macrolensed multiple quasars. Intrinsic variability in the quasar will appear (time shifted) in all lensed images. Microlensing, on the other hand, appears independently in the different images of the quasar. In order to use a macrolensed object to prove the existence of microlensing, we have to know the time delay so that we can separate the recurring variations from the independent ones. The usual procedure is to time shift the light curves of the quasar components and subtract one from the other.7 This produces a difference light curve which contains the microlensing signal. A difficult issue is that the time delay itself may be extra difficult to measure if the light curves are a mixture of intrinsic variations and microlensing. Gravitational lensing itself is achromatic: the strength of the deflection of the light rays is not dependent on wavelength. As the macro-images we observe are images of the same quasar, the magnitude difference in different bands should therefore be equal. If they are observed to be different, the cause could be differential absorption along the different light paths, or it could be caused by microlensing. Microlensing can produce an apparent chromatic effect. This is because the average amplification factor may depend on source size, which in turn depends on wavelength. This can cause different parts of the quasar to be magnified differently, which we may observe as a difference in flux ratio at different wavelengths between the components. We return to this issue in Section 6.2. 7 One must also shift the curves in magnitude to correct for the different magnifications. 12 2.3 Studying QSO 0957+561 A,B A brief time delay history Measuring the time delay between the images of lensed quasars is not a trivial task. In spite of extensive monitoring and investigations by several groups, the time delay between the two components of QSO 0957+561 has proved difficult to determine. Regular monitoring over time periods longer than the time delay is made difficult by the long time delay of the system, which is over a year. As long as microlensing signals are not clearly distinguished from intrinsic variations, microlensing complicates time delay measurements. More than 15 years after its discovery, there were two favoured time delay values: ∼540 days and ∼415 days. Kundic et al. (1997) convincingly rejected the higher value, finding a best-fit delay of 417 ± 3 days. Burud et al. (2001), using the public data set published by Serra-Ricart et al. (1999), found a value of 423 ± 9 days. In recent years, reports of time delays between 416– 425 days have been reported. Again, two leading candidates seem to emerge: 417 days (e.g. Kundic et al., 1997, Colley and Schild, 2000) and 425 days (e.g. Serra-Ricart et al., 1999, Ovaldsen et al., 2003a). Chapter 3 Pre-processing the ALFOSC data set Facts do not cease to exist just because they are ignored. Aldous Huxley, Proper Studies A key feature of observational astronomy has been the keeping of records: the preservation of data archives for future reference. In the earliest of times, buildings and monuments were designed that had special functions on solstices and equinoxes, thus passing on information about the movements of the heavenly bodies. Observers started drawing what they saw, either through their own eyes or, later, by looking through the first telescopes. The inventions of the camera and photographic techniques spurred the development of record-keeping further. In this chapter, we will take a look at the modern way: Images are obtained at the telescope through electronic devices and converted to digital numbers which can then be processed, or reduced, by computer programs. The first step of data reduction is to minimise noise and enhance image quality, before attempting to extract any scientific information. We will describe how we performed this pre-processing on our set of data, as well as discuss some other necessary preparations. Our main focus is the pre-processing of the main ALFOSC data set, from March 2001. Pre-processing the set from January 2000 followed essentially the same lines. 3.1 A look at the equipment: The CCD The great advantage of the photographic plate is that it can build up a picture of a faint object by accumulating light on its emulsion for a long period of time. Charge-coupled devices, or CCDs, have further revolutionised modern 14 Pre-processing the ALFOSC data set astronomy. The first astronomical image from a CCD was produced in 1976,1 and today the CCD is in use at most professional observatories as well as in many amateur telescopes. The CCD has a supremely high quantum efficiency (QE), that is, ability to detect and store light signals. The QE is formally defined as the ratio of incoming photons to those actually detected. The CCD has a broad spectral range where it is sensitive to light (the bandpass). Typical CCD QE curves today peak near 90%, while maintaining a 60% or more efficiency over two thirds of their total spectral range. A CCD is an array of a finite number of individual, regularly spaced, ‘picture elements’, or pixels, made of silicon. These pixels detect the light reaching them by converting photons to electrons or ADUs via the photoelectric effect.2 Present day CCDs in use in astrophysical contexts normally come in sizes of equal length and width, ranging from 512 × 512 up to 4096 × 4096 pixels. The basic physical principle behind the CCD is the photoelectric effect: A photon with the target energy3 is absorbed by the silicon, which is forced to give up a valence electron. This photo-generated electron then migrates into the conduction band, is collected by a gate structure within the pixel and held in a potential well until the end of the exposure, a process known as charge storage. After the exposure, a change of voltage potentials transfers the charge collected within each pixel row by row to an output register where the charge is measured as a voltage and converted into an output digital number. This number is referred to as counts or ADUs. The number of collected electrons (or received photons) needed to produce 1 ADU is termed the gain of the CCD. For a CCD, there exists a linear relation between input (detected photons) and output (digital number), where the gain of the CCD is the proportionality constant. Because of the properties of silicon, early CCDs were less sensitive in blue wavelengths. Since then, various CCD coatings have been developed to remedy this, usually consisting of organic phosphors that convert incident light into longer wavelengths. 3.1.1 The ALFOSC CCD Two of the data sets discussed in this thesis were obtained using the instrument ALFOSC.4 We record here a few technical details about the CCD and 1 Scientists from the Jet Propulsion Laboratory imaged the planet Uranus at a wavelength of 890 nm (Janesick and Blouke, 1987). 2 ADU = analog-to-digital unit. 3 Silicon has a useful photoelectric effect range of 1.1 to about 10 eV, covering the near-IR to the soft X-ray region (Howell, 2000 and references therein). 4 The Andalucia Faint Object Spectrograph and Camera 3.1 A look at the equipment: The CCD 15 the images that were obtained. Our third set of data, obtained from the StanCam instrument, is described in Chapter 7. The CCD detector consists of 2048 × 2048 pixels. Such a relatively large device gives a rather long read-out time, which is the time it takes for the CCD to register the image. As we did not need such a large field-of-view, the object frames were cut in the y direction in order to save read-out time. The dimension of our object frames was thus reduced to 2048 × 1200 pixels and our read-out time to ∼50 seconds. 2048 pixels 1200 pixels 2048 pixels Figure 3.1: Schematic illustration of how a total of 848 rows on the CCD was ‘cut off’ in order to save read-out time. The entire field shown in Fig. 2.1 is contained within the 2048 × 1200 pixels shown here. The CCD has a gain of 1 electron per ADU, a read-out noise of 6 electrons per pixel and a pixel scale of 0.188 00 per pixel. As mentioned, the ALFOSC data was obtained as part of the ‘QuOC Around the Clock’ campaign. To accommodate the requested observation cycles in R and V , see Section 2.1, our frames therefore had varying exposure times, ranging from 32 to 240 seconds in the R band and from 36 to 175 seconds in the V band, but with an approximate mean for each night of 60 seconds in R and 80 seconds in V. Indeed, even after cutting part of the CCD as shown in Figure 3.1, our exposure time was quite often shorter than the read-out time. The telescope was stepped slightly (a few pixels) in the x and y direction between successive exposures to allow us to average out any small-scale nonuniformities. This also allowed us to remove unwanted stars from the sky flats, see Section 3.3. 16 Pre-processing the ALFOSC data set 3.2 Bias and overscan To avoid negative values in the output image, CCDs are set up to add a positive pedestal level of a few hundred ADUs to each image. The output signal is thus ‘biased’ by this level. The first step in pre-processing data is to subtract this level from the image frames, both flats (see Section 3.3) and object frames. We obtained sets of 10 bias frames each night.5 A bias frame is simply a read-out of unexposed CCD pixels: It has an exposure time of zero seconds, and is taken with the shutter closed. Individual bias frames are averaged so as to eliminate cosmic rays,6 read-out noise variations and random fluctuations. An averaged bias frame with no overall structure in the image would show a Gaussian distribution in a histogram, see for example Howell (2000). The top panel of Figure 3.2 shows a histogram of how our pixel values were distributed for a typical, averaged bias. We have also fitted a Gaussian distribution with three coefficients; height, width and centre of the function, to the data. The fit is not very good in the top panel of Figure 3.2. There is a surplus of values below the average value, compared to the Gaussian fit. This is because there is a column-wise structure in our (averaged) bias. The left side of the bias frame has a higher value, by approximately 5 ADU, than the right side. This structure can be seen in Figure 3.3, where we have plotted the average of each column of a bias frame. To illustrate that the bad Gaussian fit is a result of the structure in the bias, we removed the structure on a typical averaged bias frame by subtracting a polynomial of degree 2, and then we fitted a Gaussian again. The result is shown in the bottom panel of Figure 3.2. We have searched in the literature for theories about what causes the structure in the bias frames, but have been unsuccessful in coming up with an answer. Overscan strips are columns of ‘pseudo-pixels’ added to and stored with each image frame. This region, lying outside the physical area of the CCD, is generated by sending additional clock cycles to the CCD output electronics after reading out the real CCD. Like a bias frame, the overscan region gives a measure of the zero offset level and the uncertainty of this level. Figure 3.4 shows that the variation across the overscan region of a given frame is typically only a few ADUs, but that the level can vary considerably through one night. In Figure 3.4, we see that the level has increased through the night 5 The exception was the night of March 14th–15th when we obtained only 5 bias frames. ‘Cosmic’ rays can sometimes be caused by weakly radioactive materials used in the construction of CCD dewars (Florentin-Nielsen, Anderson & Nielsen, 1995). The elimination of these ‘hot’ pixels in the combination of individual bias frames are due to the rejection option minmax in IRAF, see Appendix A 6 3.2 Bias and overscan 17 Figure 3.2: A typical pixel value distribution of an averaged bias frame, from the night of March 15th–16th. We have plotted the number of pixels vs. pixel ADU value. The mean bias level offset of the data is 363.55 with standard deviation 14.26. Overplotted is a Gaussian fit to the data. Top: The Gaussian has its centre at a pixel value of 364.57, and the standard deviation of the Gaussian is 2.77 ADU (the larger standard deviation of the actual data is caused by ∼500 pixels having values below 350, which do not appear in the figure). Bottom: The structure in the bias frame has been removed by subtracting a polynomial of degree 2. We notice a good Gaussian fit. The mean of the Gaussian is 364.34, the standard deviation is 1.79. by ∼15 ADU. From the image headers, we checked the CCD temperature throughout the nights, but found no systematic changes, so the increase in overscan level remains a puzzle. When performing bias subtraction and flat field division in IRAF (see appendix A), we can monitor any changes in pedestal level from graphs of the overscan region such as Figure 3.4. A bias frame includes more information than the overscan region, as it represents any two-dimensional structure that may exist in the CCD bias level. If this variation across the frame is very low, it may be sufficient to subtract only the mean bias level provided by the overscan region. However, as we had some structure in our bias frames, see Figure 3.3, we used subtrac- 18 Pre-processing the ALFOSC data set Figure 3.3: Variation across an (averaged) bias frame. The average of each column is plotted. The high values in the rightmost part is the overscan region, see also Figure 3.4 tion of the bias frame on our data. To perform bias frame subtraction for an object frame, IRAF will first subtract an average of the overscan region from the given frame before subtracting the variation, corrected for each overscan level, across the frame. The variation across the frame is provided by the (averaged) bias frame. Dark frames are images taken with the shutter closed, but for some time period, usually equal to that of the object frames. These frames are used to measure the thermal noise in a CCD, which is the production of electrons caused by the random motion of atoms. However, the efficient cooling of CCDs using liquid nitrogen makes the dark current practically negligible, especially for relatively short exposure times like ours. 3.3 Flat-fielding Across a CCD, there are variations in the individual pixels’ response to light. The quantum efficiency and gain of each pixel is slightly different from that of its neighbours. In order to correct for these pixel-to-pixel variations, we obtain flat field images. Flat fields are exposures with uniform illumination and a high signal-to-noise ratio,7 providing a read-out of the internal variation of the pixels within the CCD. Dome flats are exposures of the illuminated inside 7 See Section 3.4. 3.3 Flat-fielding 19 Figure 3.4: The overscan region from two different frames. The lower line is from a frame taken early at night, while the upper is from a frame taken early next morning, at the end of the observing run. The average value of each column is plotted. of the telescope dome, while sky flats are exposures of the bright twilight sky in the morning or (preferably) evening. A flat-fielded object image is obtained by first subtracting the bias level and then dividing by the bias-subtracted flat field image. In equational form, this is: Final Reduced Object Frame = Raw Object Frame − Bias Frame Flat Field Frame − Bias Frame The flat field frame is normalised to a mean value of 1 before division. A pixel in the flat frame with less than average value will thus have a normalised value < 1. Division by this value will cause the corresponding pixel value in the object frame to increase, which evens out pixel variations, as this pixel had, from the outset, a lesser value than it would have had if all pixels reacted alike. A dust grain on the filter or on the window of the CCD dewar will cause a shadow on the image frames. Flat field images usually show a collection of dark ‘doughnut’ shapes on a brighter background. The doughnut shape is caused by an out-of-focus negative image of the telescope main mirror, with the central obstruction caused by the secondary mirror. 20 Pre-processing the ALFOSC data set While the main role of flat field images is to remove pixel-to-pixel variations, they will also compensate for any vignetting, i.e. unintended darkening of corners, and for time-varying dust accumulation that may occur on the dewar window and/or filters within the optical path. Flat field calibration frames are needed for each wavelength region (R and V filter in our case) and for each observing night. This is because the variation in the pixels’ response itself varies with wavelength, and because conditions may change from night to night (which in fact proved to be a bit of a challenge, see section 3.3.2). 3.3.1 Inspecting the flats The image pre-processing was done using the IRAF (Image Reduction and Analysis Facility) environment. See appendix A for a short discussion on IRAF, including details on what parameters we used for the different tasks. We obtained both dome and sky flats at the observatory. However, the dome flats turned out to be very uneven: Parts of each doughnut were much darker than other parts. We are not sure what caused this, but we could see from inspection of some of the brighter ‘raw’ object frames that the doughnuts were even in appearance, and we suspected that something was probably wrong with our dome flats. We could only speculate as to what caused this, perhaps the lamp used to illuminate the telescope dome was badly placed. Dome flats from the StanCam observations also showed this effect, see Section 7.1. A ‘master flat’ is obtained by combining several individual flats. To further test the dome flats, we constructed one master flat consisting entirely of sky flats, one entirely of dome flats and one of a combination of the two flat types. After examining the results we decided that the sky flats were superior and discarded our dome flats. This of course reduced our number of flats considerably. However, in all cases except one we managed to retain five or more good sky flats for each filter and each night, which suffices for a decent flat-fielding. As a rule of thumb, 5 flats or more is acceptable although it is far better to have a few good flats than many bad ones. On the one occasion (night of the 14th–15th, filter V ) that we had only three good flat frames left, we adjusted the combination parameters accordingly, see appendix A for details. When combining images, one can form a type of average value of each pixel in all frames. Not all pixel values need to be taken into account when combining images, indeed not all of them should be. Cosmic rays, stars appearing on the sky flats etc. result in some pixel values deviating noticeably and erroneously from the rest. Reckoning with this value could wrongly offset the statistics of the combined frame. Therefore, a rejection criterion 3.3 Flat-fielding 21 of some sort produces a more realistic combination. In combining n images the routine must decide at each pixel which of the n data values, if any, to reject in forming the average. In IRAF, there are a number of different rejection operations to choose from. Depending on what frames that are to be combined, different criteria for rejecting ‘deviating’ pixels can be chosen. The up side of using sky flats is that the colour of the night sky is usually closer to the colour of the actual object frames. The down side is that stars may be present on sky flats, which can render them unusable. Since we used only sky flats, we wanted to be sure to use the most appropriate combination parameters for eliminating stars from the flat frames. We examined all sky flats individually, looking for prominent stars located in sensitive parts of the images (i.e. around the areas where the quasar components and the field stars shown in Figure 2.1 would be positioned) that could not be eliminated. We did have to discard a couple of flats with too many stars. On those nights when the telescope had been stepped between exposures, faint stars could be removed by using the median instead of the average when combining individual flats. For the nights when the telescope had not been stepped between flat field exposures, we had to discard sky flats containing stars. The flat field and the bias images utilised the full extent of the CCD, i.e. they were not cut in the y direction during the observing run. We therefore had to cut these frames accordingly, to extract the correct section of the flat field that would correspond exactly to the (pre-cut) object frame. We made a short IDL program that cut the flat fields appropriately. Figure 3.1 shows how the full CCD detector was cut when obtaining the data frames and subsequently how we had to cut the flat fields afterwards, to make them correspond to the object frames. Figure 3.5 shows the variation across a typical flat frame. Here, we have plotted the average of each column in the frame. We see some edge effects, but this is a trimmed flat frame so the severe edge effects have been trimmed off. The dips and troughs are discussed in section 3.4.3. 3.3.2 The moving blob There were a lot of out-of-focus dust grains causing doughnut shapes on our flat frames. After close inspection of the flats for our third night, the night of the 16th–17th, we found that while the 5 sky flats from the evening were internally consistent in structure, the flats from the early morning were not: A strangely shaped, quadruple doughnut had shifted quite considerably in position. This apparently sudden shift in position of the blob was apparently not affected by filter changes (R or V ), and it occurred twice: once for the night 22 Pre-processing the ALFOSC data set Figure 3.5: An example of variation across a typical flat frame. The graph shows the average ADU value of all columns. The two sharp declines at x = 1232 and x ∼ 1370, as well as the apparent vertical line at x ∼ 1786, will be discussed in Section 3.4.3. This figure shows a combined flat, so the extreme low values at the far edges have been trimmed off. The ordinate on the left show the counts on the images (ADU), while the ordinate on the right and the abscissa show the y and x axes of the image frame. 16th–17th and then again for the night 17th–18th. We discussed several theories as to what this ‘object’ might be. We now think the blob may be a conglomeration of 3–4 grains of dust, giving rise to several doughnuts situated very close together. This would make it heavier, and it is thus a possibility that this ‘giant dust grain’ has been blown by a chance wind and, the individual dust grains sticking together, has flipped and rolled over, while the other, lighter ones stayed in place. Although we have not investigated this in more detail, this is the most likely hypothesis concerning the nature of the moving blob that we came up with. We decided to leave the matter of the nature of the blob here, and concentrate on how to avoid any problems it might cause. On the night of the 16th–17th, the shift in position of the blob occured while we were taking our object frames. The logical solution was to regroup 3.3 Flat-fielding 23 Figure 3.6: The moving blob. Top: Master flat from the evening of the 16th. The ‘blob’ is in the bottom right-hand quadrant. Bottom: Master flat from the morning of the 18th. Notice how the blob has moved in relation to the other, stable doughnuts. Also visible on these frames are some of the bad columns and the circular area of reduced sensitivity at x ∼ 1370, y ∼ 170 discussed in Section 3.4.3. the flats, producing two master flats with the blob in the two different positions. We proceeded to flat the object frames with their respective flat field images. The night of the 17th–18th provided more of a challenge. On this night, the blob moved very early in the evening, while we were taking our first flat frames. This meant that our only flats with the blob in the same place as on (all of) the object frames were from the morning of the 18th. Also, many 24 Pre-processing the ALFOSC data set of the flats from the morning seemed to contain quite a few stars. To get enough good flats (i.e. not containing stars), we first considered using flats from both the evening and the morning, which would mean combining flats with the blob in different positions. As the object frames were stepped in position between each exposure, the coordinates of the stars were different on all 301 frames. To observe the blob on all these frames, we made a film that showed all object frames in succession, with a rectangular box indicating the two positions of the blob. We concluded that the stars F, G, H, X and R, see Figure 2.1, were never positioned in the areas affected by the blob. Stars D and E, however, were affected on several frames. As we wanted to use these stars in the photometry, we ended up using only flats from the morning. This meant we only had 4 flat frames in R and 5 in V , but we considered this the best option as we did not want to jeopardise the D and E stars. 3.4 Other preparations The signal of an image is produced by the object being imaged. The signalto-noise ratio (S/N), measures the amount of signal in relation to the amount of noise, which is important for the degree of detail that can be seen in the image. For performing deconvolution of images, a high S/N ratio is needed. We therefore decided to make a summation of 10 images at a time, which corresponds to increasing the exposure time in our final image, which again increases the S/N ratio. With an exposure time of 60 seconds for each individual image, this would mean a ‘new’ exposure time of 600 seconds, or 10 minutes. 3.4.1 Combining images Combining images, either by taking the sum, the mean or the median of several individual exposures, can be useful for several reasons, especially for increasing the S/N ratio. The standard deviation is the simplest measure of the noise in the image. The standard deviation (σ) of a stochastic variable x, or rather its probability distribution, is defined by: σ= p hx2 i − hxi2 , where ‘hxi’ denotes the mean of the variable x. The standard deviation per 3.4 Other preparations 25 pixel,8 when summing N images, is given by: q 2 σper pixel = σ12 + σ22 + . . . + σN (3.1) In terms of describing electromagnetic radiation as photons, the rate of arrival of individual photons is a random process. Independently of the nature of the detector there is an irreducible level of fluctuations, known as Poisson noise. There are several ways of computing the signal (S) divided by the noise (N ). For sources that behave according to √ Poisson statistics, a signal level of N∗ photons has an associated 1σ error of N∗ . That is, if the noise is dominated by the Poisson noise from the signal itself, we have that: p S N∗ '√ = N∗ . N N∗ Additional noise comes from sources such as background sky level, light pollution, heat in the camera, the process of reading the data out of the camera etc. For images dominated by Poisson noise, summing several images will increase the noise (see the expression for σper pixel above), but will increase the signal more, causing the signal-to-noise ratio (S/N) to increase. An additional motivation for combining images was the relative shortness of our exposure times. Also, there are several steps in the MCS procedure that has to be performed manually for each image that is to be deconvolved. Reducing our number of images from close to 1000 to less than 100 (93), was therefore very helpful. We should also mention that the number of images for each filter and each night did not always add up to 10. The rejection operation we used in the combination procedure, see appendix A, favoured a large number of images, so instead of having, e.g., 3 images in the last batch and 10 in the second last, we evened them out, giving 7+6 instead of 10+3. If we had 10 images in the second last batch and only one or two in the last, we combined the 11 or 12 individual images to one image. 3.4.2 Scaling, aligning and background subtraction Turbulence in the atmosphere gives rise to a broadening of the PSF of an image, known as seeing. As conditions change through the night, so does the 8 By standard deviation per pixel, we mean an estimate of what the standard deviation (σ) would be in this one pixel, if we were able to repeat the observation an infinite number of times. 26 Pre-processing the ALFOSC data set PSF. Seeing is easily measured by the FWHM of stars. In our images, the seeing varied from 0.7 00 to 1.5 00 . We can think of an astronomical image as being a composite of a background level with a signal (i.e. a star) on top. The background, or ‘sky’ level, depends on the exposure time and also on the observing conditions: the time of night, whether the moon is present, clouds and dust in the atmosphere etc. The signal, on the other hand, increases linearly with the exposure time and does not (ideally) depend on when we made the observation.9 Variations in background level and in ellipticity, varying seeing conditions and different exposure times can produce unwanted effects when combining images, especially in the wings of stellar profiles. We found that combining images without background subtraction, could lead to ‘holes’ in the outer regions of stars. The σ rejection method we used when combining images (see Appendix A), would also be affected if we combined images with different background levels. Higher signal means higher noise, so an image with higher level would contribute with an ‘artificially’ high σ in Equation 3.1, affecting the number of pixels that are rejected. We therefore subtracted the background from all images before combining. As a measure of the background level in an image, we used the median of the entire image, as this will be completely dominated by the background. In Section 3.1, we mentioned that the average exposure time for the R and V frames were, respectively, approximately 60 and 80 seconds. We decided to scale the images to these exposure times. This also proved useful later, when we plotted magnitudes versus sky level, see Section 6.1.3. Aligning As mentioned, the object frames were stepped a few pixels in the x and y direction. Before combining the images we therefore had to align them so that the stars had the same coordinates before combining. We aligned the frames so that 10 and 10 frames had identical coordinates. To perform the alignment, we needed coordinates for as many stars as possible from all frames, in standardised text files to use as input to the IRAF aligning routine. This was supplied by the first part of Jan-Erik Ovaldsen and Jan Teuber’s photometry program (Ovaldsen, 2002), giving high-precision coordinates for those of the stars F, G, H, E, D, X, R that were unsaturated and not too close to bad columns (see Section 3.4.3) in each image. The procedure performs an initial centring by fitting a paraboloid to the nine 9 In reality of course, the signal will be affected by factors such as an increase of air mass as the star sinks towards the horizon. 3.4 Other preparations 27 central pixels, i.e. the pixel with maximum intensity and the eight surrounding pixels. Then, an iterative and more robust centring algorithm improves the first centroid by correlating the PSF with several synthetic images of the star in question – each offset by sub-pixel distances. The scaling, aligning and background subtraction were performed in IRAF. We note that the alignment procedure, using bilinear interpolation, will tend to smooth the images somewhat. The automatic alignment procedure would often result in large residual values, that is, differences between the reference and the transformed coordinates of up to ∼0.5 pixels. We therefore used an interactive feature in order to remove from the transformation the stars on these images that caused higher residual values, possibly due to bad pixels or cosmic rays, see Appendix A for details. This procedure effectively reduced the residual values from the alignment to around ∼0.03 pixels. 3.4.3 Bad Columns According to the online specification of the ALFOSC detector,10 there are a total of 16 charge traps and bad columns on the CCD, as well as many dark ‘specks’ scattered across the imaging area. The specks are areas a few pixels wide with reduced sensitivity, of diverse origin. Some of them are probably results of defects in the coating of the CCD. All the specks are removed by flat-fielding. Bad columns are columns of pixels that for some reason do not work properly. There were several bad columns on the CCD we used. Two of the bad columns11 were a particular nuisance, as they were close to our target objects. Although we tried to avoid these particular bad columns when observing, we did not always succeed in doing so, as the bad columns depended on illumination and are not always clearly visible on the frames. As we discovered after the observations had been completed, the images of the quasar components and some of the field stars came close to and even directly hit the bad columns on ∼10% of the images. We wanted to discard images where a bad column came too close to the point sources. Tests performed by Jan-Erik Ovaldsen, replacing the bad column by an average of the two immediately surrounding columns, placed a limit on how close to a field star or quasar component a bad column could come before the interpolation yield an unsatisfactory result, i.e. the flux of the star was severely affected by the interpolation. Near a star, the gradients 10 Note that the ALFOSC CCD has recently been replaced (September 24th 2003). The specifications for the CCD we used for our observations are still located at http://www.not.iac.es/technical/astronomical/detectors/ CCD7/PreCommissioningReport/ 11 Coordinates x = 1232 and x = 1409 on our (pre-processed) frames. 28 Pre-processing the ALFOSC data set across columns are so great that simple interpolation will not give a satisfactory result. A total of 86 frames, 78 R and 8 V , were rejected from our data set because a bad column came too close to an object of interest. It might have been possible to try and figure out a complex interpolation scheme to deal with bad columns this close to stars, but we chose not to as we had an abundance of images in the first place. Rejection seemed the safest choice. When running the MCS deconvolution, we extract smaller images of 64 × 64 pixels around the objects of interest. These ‘sub-images’ should be free of bad columns. When combining 10 successive images, we made use of procedures that reject ‘out of bounds’ pixels. An average of the values of each pixel is computed and values deviating from the average by some factor are rejected, see Appendix A for details. As we aligned the images before combining them, the bad columns had different coordinates on every frame. We therefore initially trusted the rejection method to eliminate the bad columns. However, the bad columns did not all disappear on the combined frames. With hindsight, we realized that we should have interpolated the bad columns before combination, while they still had the same coordinates on every frame. The explanation for why the rejection method did not work as expected, could be that we used the ‘σ clipping’ rejection algorithm mentioned in Section 3.4.2, which rejects pixels with values higher or lower than a specified σ value (usually 2 or 3σ). When a bad pixel is taken into the computations, the σ computed will be artificially large because of this deviant value. This leads to a threshold value which is too high and therefore too few or none of the pixels will be rejected. As it involved quite a large amount of work to do the interactive aligning over again, we decided not to. However, when deconvolving, we had to inspect the (93) combined images, and could then check for any remaining bad columns in the extracted sub-images. We could then replace these pixels by a relatively high value in the ‘noise map’ produced in MCS, which would cause the deconvolution routine to disregard these particular pixels, see Section 4.5.1. Referring back to Figure 3.5, we notice some bad columns and also an area several pixels wide of reduced sensitivity. What appears to be a vertical line at x = 1785 to x = 1787, is actually a steep incline followed by an equally steep decline. See Fig. 3.7 for a view of a magnified portion of Fig. 3.5. There are two bright bad columns followed by one dark. These columns did not affect our observations, as they were never in the vicinity of any of the objects we studied. In Figure 3.5 we also see a trough in the graph around x ∼ 1370. This corresponds to a circular spot of low QE on the CCD, approximately 9 pixels in radius. The spot can also be seen in Figure 3.6. The lowest value of this 3.4 Other preparations 29 Figure 3.7: Bad columns. Closeup of Figure 3.5. spot, in the centre, has a value of ∼20% of the surroundings. By comparison, the doughnuts have only 2-4% lower value. The spot does not flat-field away properly on all images, and we suspect it might be a defect in the chip. However, no objects of importance to our investigations come near this spot on any of the frames. The sharp decline at x = 1232, one of the bad columns that did give us some problems, can also be seen in Figure 3.5. 3.4.4 The importance of careful pre-processing In concluding this chapter, we want to stress that careful pre-processing of data is a crucial step towards obtaining accurate photometry. Simple flaws in the CCD, or a diversity of other conditions, can seriously affect the results if not noted and, where possible, corrected for. Careful inspection of the bias, flats (darks where applicable) and object frames is important, as well as choosing the particular pre-processing methods that fit our particular observations. Understanding the principles of the noise affecting the images is paramount to obtaining reliable results. 30 Pre-processing the ALFOSC data set Chapter 4 The MCS procedure Everything you say to me Takes me one step closer to the edge... Linkin Park, One step closer In this chapter, we describe the basic ideas behind the MCS deconvolution procedure, named after its developers Pierre Magain, Frédéric Courbin and Sandrine Sohy. A few elements from Fourier theory and digital image processing theory are mentioned in Section 4.1. To illustrate the general principles of the MCS procedure, we first discuss deconvolving one individual image, and in Section 4.6 we expand our discussion to the deconvolution of several images simultaneously. Simultaneous deconvolution has the advantage of utilising the combined S/N ratio of all the images. We also describe in some detail the process of running the programs. As a certain amount of experience is needed in order to run MCS in the most efficient manner, we spent quite some time familiarising ourselves with the different routines. This part of the chapter is therefore partly intended as a guide for future users of MCS. 4.1 Convolution As light passes through our atmosphere and our detectors, it is distorted and blurred. This is a result of atmospheric turbulence, the finite resolving power of our detectors and instrumental noise. The light distribution of point sources such as stars, is therefore broader and the resolution of the images we obtain is poorer than it would have been if we were observing outside the Earth’s atmosphere with an ‘ideal’ telescope (i.e. a telescope with infinite resolving power). The mathematical representation of this broadening involves perceiving the original light distribution as convolved with a function representing the 32 The MCS procedure total instrumental profile. This function is termed the point spread function (PSF). The imaging equation describing the broadening is: d(~x) = t(~x) ∗ f (~x) + n(~x) , (4.1) where t(~x) is the total PSF for the image, f (~x) and d(~x) are, respectively, the original and observed light distributions and n(~x) is the noise affecting the data. The symbol ‘ * ’ signifies convolution. Convolution is a process which ‘folds’ functions together. In the onedimensional, continuous case it is expressed mathematically as: Z ∞ y(t) = f (t) ∗ g(t) = f (α)g(t − α) d(α) , −∞ where α is an integration variable. To find y(t) at a specified t, the integrand f (α)g(t − α) is first computed as a function of α. Then, integration with respect to α is performed, resulting in y(t). These mathematical operations have simple graphical interpretations, see Figure 4.1 for illustration. First, plot f (α) and the flipped and shifted g(t−α) on the α axis, where t is fixed. Second, multiply the two signals and integrate the resulting function of α from −∞ to ∞. These operations can be repeated for every value of t of interest. In essence, the Fourier transform decomposes a function into sinusoids of different frequency which sum to the original function. The Fourier transform F (ω) of a continuous function f (t), where t ∈ R, is given by: Z ∞ 1 f (t) e−iωt dt . F (ω) = √ 2π −∞ The transform is here expressed using the angular frequency ω(= 2πν) and the time t. We have restricted ourselves to the one-dimensional case since it is sufficient for our illustrative purposes. Inversely, we can obtain f (t) from F (ω): Z ∞ 1 F (ω) eiωt dω . f (t) = √ 2π −∞ The expressions for F (ω) and f (t) are called a Fourier transform pair, and they translate between the Fourier (or frequency) domain and the ‘time’ domain. One of the theorems that constitute a basic link between the Fourier and the spatial domain is called the convolution theorem. It simply states that the convolution of two functions in the ‘time’ domain corresponds to 4.1 Convolution 33 g( α ) f( α ) 1 1 α 1 α 1 -1 t α 1 1 t α y(t) t Figure 4.1: Convolution of f and g. Top: The two functions f (α) and g(α). Middle: The flipped g for t = 12 and t = 52 . Imagine g sliding towards the right, across f . The marked area shows the multiplication of f (α) with g(t − α) for these specific values of t. Bottom: Integrating the products for all values of t of interest produces the function y(t) which equals the convolution of f (t) and g(t). The two points originating from integration of the marked areas in the middle panel, are shown as filled dots on y(t). ordinary multiplication in the Fourier domain. In other words, f (t) ∗ g(t) and F (ω)G(ω) constitute a Fourier transform pair. We have that: f (t) ∗ g(t) ⇔ F (ω)G(ω) , 34 The MCS procedure and vice versa: f (t)g(t) ⇔ F (ω) ∗ G(ω) . The task of deconvolving Equation (4.1) thus seems in principle quite an easy task when using Fourier transforms, as the equation in the Fourier domain is simply: D(ω) = T (ω)F (ω) + N (ω) , where the functions D, T , F and N are the Fourier transforms of d, t, f and n from Eq. 4.1 respectively. To recover F we then simply put: F (ω) = D(ω) − N (ω) . T (ω) (4.2) When one has obtained F (ω), one can simply transform back to the spatial domain to find f (x). However, on practically all modern light detectors used in astrophysics the observed light distribution is not continuous but sampled and this introduces an important factor to be considered when solving Equation (4.2). Also, the noise N (ω) is unknown and has to be modelled. 4.1.1 The sampling theorem Most astronomical detectors in use today are electronic charge-coupled devices (CCDs), see Section 3.1. This means that the light we observe with our instruments is sampled, i.e. known only at regularly spaced intervals, while the original light distribution from the star is continuous. The sampling theorem was first described by Shannon (1949). It is an important theorem, determining the maximum sampling interval allowed if a continuous function is to be regained completely from discrete samples. The theorem states that if there exists a cutoff frequency ω0 above which the Fourier transform of the function is zero, then the function can be fully specified by sampling values spaced at equal intervals not greater than 1/2ω0 . Imagine a continuous function f (x), whose Fourier transform is bandlimited as described above. That is, the Fourier transform F (ω) vanishes for values of ω outside the interval [−ω0 , ω0 ]. To obtain a sampled version of this function, multiply f (x) by a ‘sampling function’ g(x) consisting of a train of impulses (δ-functions) an arbitrary ∆x units apart. By the convolution theorem described in the previous section, multiplication in the spatial domain corresponds to convolution in the Fourier domain. The Fourier transform of the product f (x)g(x) is periodic, with a period (∆x)−1 . If the individual repetitions of F (ω) overlap, this will contaminate the recovery of f (x). The centre of the overlapped region in the first period occurs at ω = (2∆x)−1 4.2 Deconvolution 35 if this quantity is less than ω0 . To be able to isolate F (ω), the following requirement has to be satisfied: ∆x ≤ 1 . 2ω0 (4.3) The minimum sampling frequency ω0 is called the Nyquist frequency. This is the sparsest sampling allowed if one is to avoid overlapping in the Fourier domain. 4.2 Deconvolution The problem of improving the resolution of astronomical images by regaining (some of the) precision that has been lost, is the inverse problem of deconvolving the images. This corresponds to recovering the function f (~x) in Equation 4.1. Deconvolution tries to recover high frequencies, which will tend to increase the noise. We now turn to a description of the MCS programs, and how we used them on our data of QSO 0957+561. 4.2.1 Traditional methods Being an inverse problem, deconvolution is ill-posed, meaning that there are a large number of possible solutions compatible with the original light distribution. Some additional constraint must be applied in order to chose between the plausible solutions. Several methods have been developed to apply constraints to this problem. A typical method consists of minimising the χ2 of the differences between model and data, in addition to applying a constraint imposing smoothness of the solution. The original light distribution fi is then the solution that minimises the function N N 2 X 1 X S1 = + λH(f1 , ..., fN ) . t f − d ij i i σ 2 j=1 i=1 i (4.4) The first part of S1 is the sum of the χ2 with σi the standard deviation at the ith sampling point. N is the number of sampling points, dj , fj and nj are vector components giving the sampled values of d(~x), f (~x) and n(~x) in Equation (4.1) respectively, and tij is the value at point j of the PSF centred on point i. The χ2 represents sampled deconvolved model, reconvolved with the PSF, minus the original data points. In the second part of S1 , H is a smoothing function and λ is a Lagrange parameter that is determined so that the reconstructed model is statistically compatible with the data (χ2 ≈ N ). 36 The MCS procedure It is also very useful to consider any prior knowledge in order to choose the correct solution. For example, the maximum entropy method automatically rejects all solutions with negative values as no negative flux can be detected. This is also the case for another method, the Richardson-Lucy iterative algorithm (Richardson, 1972; Lucy, 1974). Traditional deconvolution methods tend to introduce artifacts in the images, as a result of violating the sampling theorem. For example, ‘ringing’ artifacts can be produced around point sources in the image. The relative intensities of different parts of the image are not conserved either, which precludes quantitative photometry from the image. The main problem with classical deconvolution algorithms is that they try to deconvolve to infinite resolution. However, if the observed signal is sampled so that the sampling theorem is obeyed, the deconvolved data will generally violate that same theorem. Increasing the resolution means recovering higher Fourier frequencies, thus increasing the cutoff frequency, which by Equation 4.3 makes the correct sampling denser. 4.2.2 Deconvolution with correct sampling: a preview The MCS algorithm is based on the principle that sampled data cannot be fully deconvolved without violating the sampling theorem. It differs from other deconvolution methods in that it does not deconvolve with the total PSF, but with a narrower PSF chosen to accommodate the sampling theorem. In other words, since the data are sampled we should not try to deconvolve the image to achieve a resolution from an ideal telescope in perfect surroundings (i.e. a space telescope located above the atmosphere with infinite resolving power), but only a resolution as would be obtained with a better telescope. Along with the technique for smoothing the background, see Section 4.4 and 4.5.4, this removes the ringing artifacts. The method also allows for accurate photometric and astrometric measurements on the deconvolved frame. Figure 4.2 shows an example of a deconvolved image of the components of QSO 0957+561. The left panel shows the original image, and the right panel the deconvolved image. 4.3 Preparations for running MCS Running MCS on such a relatively large number of images requires some preparation. The background must be the same in all images, so they must be background subtracted. We also need the approximate positions of the objects of interest, and the standard deviation in the background of each 4.3 Preparations for running MCS 37 Figure 4.2: Left: The extracted image from the original, combined (R) frame. North is up, East is left. Exposure time for the combined image is 600 seconds, seeing 0.75 00 , plate scale 0.188 00 /pixel. Right: The final deconvolved frame. FWHM is 2 pixels. The deconvolved image is subsampled by a factor 2. See text for details. image, which we measured with IRAF. When simultaneously deconvolving many images, see Section 4.6, we aligned all the images to the same reference image, as the simultaneous deconvolution requires all objects to have equal coordinates within one pixel. Starlight reaching us is scattered and absorbed by molecules in the Earth’s atmosphere. As the star moves across the sky during each night, it is affected by varying amounts of air mass that the light has to pass through in order to reach us, and the apparent brightness will decrease as the air mass increases. The air mass depends upon the zenith distance of the object, as shown in Figure 4.3. As a first approximation, the air mass varies with sec(z). At angles more than about 60 degrees from the zenith, this approximation is less accurate due to the curvature of the atmosphere and increased refraction. Changes in air mass or cirrus affect the quasar and reference stars in the same way. These effects cannot be easily modelled, but they can be accounted for by normalising by a star in the field. By normalising, we correct for the effects of any changes in cirrus or air mass that may occur with time. We also normalised the other field stars we were interested in deconvolving, by the same reference star. We chose the star D in Figure 2.1 as our normalising star. In addition to choosing a normalising star, we must choose a star to create the PSF from. In case of PSF variations across the field, the star chosen for constructing the PSF should be close to the objects we are interested in. We considered using the two relatively weak stars P and Q as PSF stars, but in the end decided to use the star E, as P and Q are considerably dimmer than the quasar components. This star is bright, so it has a high S/N ratio, but it is not too close to the saturation limit. It is also relatively close to the quasar components and the other field stars. We performed some tests with different PSF stars to see if the residuals varied much from star to star, but 38 The MCS procedure Zenith * Star path length ~ sec(z) z Horizon Figure 4.3: Changes in air mass are proportional to sec(z). This figure shows a simplified, plane-parallel atmosphere. found that they were quite similar. So the PSF does not appear to change much across the CCD. In the MCS procedure, one may also choose to construct the PSF from several stars. If several stars are used, the different residuals after Moffat subtraction, see Section 4.5.2, should be approximately equal. We deconvolve to a FWHM of 2 pixels, which makes the construction of the PSF, s, easier compared to deconvolving to a broader FWHM, see Section 4.5.2. Our deconvolutions will thus try to recover the sharpest possible light distribution allowed by the sampling theorem. 4.4 The aim of MCS The MCS deconvolution method does not use the total PSF t(~x) of an image, but a narrower function s(~x) chosen so as not to violate the sampling theorem. The two PSFs are related by the convolution t(~x) = r(~x) ∗ s(~x) . (4.5) Here, t(~x) is the total PSF of the image, and s(~x) is the narrower function that we want to use for deconvolving our images. To create the function s(~x) we use the relation above and first deconvolve t(~x) with a pre-chosen function r(~x). This analytical function r(~x) must have a certain width, i.e. not be point-like. In the MCS algorithm r(~x) is set to be a Gaussian distribution, 4.5 General outline of MCS 39 whose FWHM must be at least 2 pixels in order not to violate the sampling theorem. In the final deconvolution of the image, see Eq. 4.7, the narrower function s(~x) is used instead of the total PSF t(~x). Instead of using t(~x) to deconvolve the observed image, thereby aiming to achieve a ‘perfect’ pointlike resolution, the narrower function s(~x) is used to deconvolve the observed image to the resolution of r(~x). See Figure 4.4 for a comparison of the total (t) and the narrower (s) point spread functions. The MCS method does not aim at obtaining the point-like distribution f , but instead a distribution, r, with a certain width. For this purpose we must deconvolve the data with the PSF, s, which we have constructed, instead of t, which is the original PSF. The deconvolved image will thus have an improved but not an infinite resolution. A method like the maximum entropy deconvolution looks for the solution where the entropy is maximum, which corresponds to the flattest solution compatible with the image. Understandably, this would tend to underestimate the intensity of point sources. In the MCS method, the knowledge of the shape of the point sources in the deconvolved image is a very strong prior knowledge that can be used to constrain the solution f (~x). Since the profile r is known and chosen by the user, the deconvolved image f can be decomposed into a sum of point sources plus a diffuse background, see Figures 4.4 and 4.7. The program returns positions ~ck and intensities ak for the M point sources, while the ‘background’ consists of a deconvolved numerical image for all extended sources h(~x). The solution f (~x) can thus be written f (~x) = h(~x) + M X k=1 4.5 ak r(~x − ~ck ) . (4.6) General outline of MCS The MCS procedure consists of 4 FORTRAN 77 programs that are run individually and consecutively. From these programs we obtain as output 5 image files per PSF star, as well as 14 other files, both images and text files. The PSF stars are the stars chosen for the construction of the PSF, s. Thus, if we run the program on one image using 3 PSF stars, we end up with a total of 14 + 5 ∗ 3 = 29 output files. This is the ‘grand total’ number of output files, where the output from one program is generally used as input in the next. Some of these output files are relics from older versions of the programs and are not used but still given as output, see the overview in Appendix C. The MCS programs are, in running order: 40 The MCS procedure d = t * f d = s * r Figure 4.4: Above: d is the observed distribution, t is the total PSF and f is the original (point-like) distribution. Below: s is the narrower PSF that we want to use in the deconvolution in order to obtain our chosen solution r from our observations, r being the Gaussian function with a FWHM that we choose (≥ 2 pixels). d and t have the same width and can be interchanged for practical purposes. 1. extract.f 2. psfm prive.f 3. psf prive.f 4. dec prive.f These programs use a total of 4 input text files, to be filled in by the user, at various stages (see examples of input files in Appendix B). In addition, 2 sub-routines used for setting the image sizes and the number of point sources in the image to be deconvolved must be present and filled in correctly. We will now go through each stage of the MCS deconvolution process, and for illustrative purposes first describe how to deconvolve a single image. The simultaneous deconvolution of several frames is one of the main advantages of the MCS method, and will be described in later sections. 4.5 General outline of MCS 4.5.1 41 Extracting data frames The program extract.f extracts frames of optional size from the original image. We have set the frame size to 64 × 64 pixels, which encompasses both the individual field stars and the quasar images. The program takes as input the text file extract.txt, where we enter the coordinates of the centres of the objects we want to extract. This means that we have to specify approximate coordinates of the stars we want to use for constructing the PSF, s, and of the object image itself (the quasar components or a star). The program extract.f gives as output the extracted images psfnn.fits, that is, in numerical order psf01.fits, psf02.fits . . . , of the PSF stars, normalised to a peak intensity of 1, and the extracted image g.fits of the object. The routine extract.f also calculates an estimate of the photon noise in each pixel of the normalised PSF frames, and returns this as a ‘noise map’ image, termed psfsignn.fits and sig.fits for the PSF and object frames respectively. For a sky subtracted image, the noise in one pixel expressed as a standard deviation σimage is given by: q 2 σimage = I/g + σsky , where σsky is the standard deviation in the background, which we have to measure on all image frames. I is the flux of the pixel and g is the gain of the detector. In extract.txt, the user can also choose to use ‘small pixels’ which means that the pixel size in the deconvolved image will be half that of the original frame. This ‘subsampling’ allows for a finer resolution in the deconvolved image. The plate scale of the detector is 0.188 00 per pixel. The pixel size of the deconvolved image is therefore 0.188 00 /2 = 0.094 00 . 4.5.2 Constructing the PSF The profiles of astronomical point sources, typically imaged on two-dimensional arrays, are referred to as point spread functions or PSFs. A star image is inherently complex and, in practise, the telescope and detectors themselves impose their own signatures on the stellar images recorded. There are therefore many ways to encode the detailed morphology of star images, and the subject has been much debated, e.g. in Moffat (1969), King (1971) and Stetson et al. (1990). To obtain a good stellar profile, it is customary to construct the PSF as a combination of an analytical function and an empirical lookup table of residuals from the analytic fit. The assumption is that this will provide good representations of the data itself. 42 The MCS procedure In this section, we describe the two programs in the MCS procedure that construct the PSF, points 2 and 3 in the list in Section 4.5.1. The analytical part is a Moffat function. The numerical part consists of the residuals from the analytic fit. Both parts are obtained by a deconvolution combined with a χ2 -minimisation. After a short discussion of the type of analytical function used, we will describe how the two parts of the PSF are constructed. The Moffat Function The analytical part of stellar profiles can be approximated by several mathematical functions. Three functions in particular have been found to be useful descriptions of at least parts of a stellar brightness profile: the Gaussian G(r), the modified Lorentzian L(r) and the Moffat M (r) function. In their simplest form, these functions can be written as 1 r2 G(r) ∝ e( 2a2 ) 1 L(r) ∝ 2 1 + ( ar 2 )β 1 M (r) ∝ 2 (1 + ar 2 )β In these equations, r is the radius of the distribution and a and β are fitting parameters. All of these functions have been tried in various combinations. A typical stellar image has a sharp peak and rather broad wings. The Gaussian is a good representation of a stellar profile in the central parts, but not so good in the wings as these are low for a Gaussian. When an analytic function is used to represent a star image, either in its entirety or in combination with a look-up table of residuals, it is useful to generalise that function with additional parameters which allow for the inclusion of ellipticity in the PSF, i.e. non-circular images produced by e.g. telescope aberrations. The MCS programs utilise such a Moffat function as the analytical part of the PSF. In generalised form, this can be written M∝ 1 (1 + x2 α2x + y2 α2y + xy β ) αxy , for a profile centred in origo. In this equation, the parameters αx , αy and αxy can be adjusted to account for ellipticity and rotation of the profile. By 1 For a more detailed treatment of the Gaussian distribution, see appendix D. 4.5 General outline of MCS 43 rotation, we mean that the elliptical Moffat is allowed to be inclined with respect to the rows and columns of the detector. The Moffat function used in the MCS programming is explicitly described by a m φβ , where φ = 1 + b1 (x − cmx )2 + b2 (y − cmy )2 + b3 (x − cmx )(y − cmy ) . The centre of the function is (cmx , cmy ). The intensity is am , while b1 , b2 , b3 and β are shape parameters. Comparing the two expressions for a generalised Moffat function, we see that b1 and b2 represent the relation between the semi-major and -minor axes in an ellipse, with b1 and b2 corresponding to 1/αx2 and 1/αy2 respectively, while b3 = 1/αxy represents any rotation of the ellipse. The exponent β adjusts the width of the distribution in relation to the wings. In contrast to the Gaussian function, where the σ alone determines this relationship, so that the relation between centre and wings are ‘locked’, the Moffat can have a very sharp peak and still significant wings according to β. Figure 4.5 shows a Moffat function with typical values for b1 , b2 , b3 and β returned from the program. 1 0.8 0.6 0.4 0.2 0 –15 15 –10 10 5 –5 0 x 0 5 –5 10 15 –10 y Figure 4.5: A normalised Moffat function with typical values for the shape parameters b1 , b2 , b3 and β returned from the program. 44 The MCS procedure Figure 4.6: A typical image of the Moffat, mofc.fits, returned from the program psfm prive.f. Some ellipticity and rotation can be seen. The analytical part of the PSF This program takes as input the text file psfmof.txt. The program fits an analytical Moffat distribution to the PSF stars extracted in the preceding program. By χ2 -minimisation of the difference between the extracted star and the Moffat (convolved with the function r), an image mofc.fits of the Moffat with a resolution of s is obtained and returned as output from the program. This is the analytical part of the point spread function. Figure 4.6 shows what a typical Moffat image looks like. The Moffat profile, reconvolved2 with the profile r to a resolution of t, is subtracted from the PSF stars psfnn.fits and the residuals are returned as images difcnn.fits. That is, difc01.fits = psf01.fits − mofc.fits, etc. After running this program, we examine the data residuals difcnn.fits. If more than one star is used, we check that the residuals are approximately equal in structure and values. If one residual image differs noticeably from the rest, this star is discarded as PSF star and only stars with approximately equal residuals are retained. In psfmof.txt we enter the starting value for the rotation parameter b 3 and the starting β value. The β is closely correlated to the b-parameters,3 and the program will perform iterations until it finds a satisfactory fit for all shape 2 To subtract the Moffat from the data residuals, these must have the same resolution, namely that of t. 3 See Stetson et al. (1990). 4.5 General outline of MCS 45 parameters. We also enter the number of stars that are used in constructing the PSF, the FWHM value in the observed image and the FWHM value that we wish to obtain after deconvolution, i.e. the width of r. Both FWHM values must be entered in the coordinate system of the deconvolved image. That is, if we are using small pixels we must multiply the observed FWHM value by 2. Finally, we enter the maximum intensity (=1) and coordinates of the PSF stars in the extracted images. The numerical part of the PSF The next program also takes the text file psfmof.txt as input. The final aim of this program is to add the residuals from the Moffat fit to the analytical function, in order to construct the final PSF, s, see Figure 4.7. In order to do this, the residuals and the Moffat must have the same resolution. The Moffat already has the resolution of s. The data residuals however, have the resolution of t. Referring to Figure 4.4, we see that to obtain a resolution of s we have to deconvolve the residuals difcnn.fits, which have the resolution of t, with the Gaussian r. This has to be done numerically, since we are deconvolving data residuals and not an analytical function. This deconvolution is easier when r is sharper, which is why we chose to deconvolve to 2 pixels FWHM, as we found some correlation in the noise in the modelled residuals when using 3 pixels FWHM as the width of r. The program psf prive.f performs this deconvolution and, in this process, has to interpolate and smooth the data residuals. Another χ2 -minimisation is performed, minimising the difference between the data residuals and the modelled residuals (convolved with r). The program will ask the user for values of ‘inner’ (λ1) and ‘outer’ (λ2) smoothing parameters that must be specified, as well as the radius of the circle that separates the areas where these smoothings apply. This radius should be 3–4 times the FWHM before deconvolution. Typical smoothing values are ∼ 10 for low smoothing and up to 100 000 for heavy smoothing, but can in principle vary from close to 0 to 106 . The modelled residuals, that is, the numerical part of the PSF with a resolution of s, is returned from the program as an image psff.fits. For comparison with the actual data residuals, an image of the modelled residuals with resolution t is also made, by reconvolving psff.fits with r. This image is returned as output psfr.fits, which can be compared directly to the actual data residuals difcnn.fits as these now have the same resolution (t). We want to smooth the model of the residuals to avoid noise. However, 46 The MCS procedure the residuals must also be comparable to the data residuals that we are trying to model. That is, psfr.fits must have approximately the same values as difcnn.fits. That they are comparable in value must be checked manually after the program has been run. The program is re-run with different values of λ1 and λ2 until the values match. One has to find the highest smoothing possible that will still give a psfr.fits sharp enough compared to difcnn.fits (and a psff.fits that is not too noisy). Next, the program masks the outer parts of the modelled residuals as these consist mostly of noise. The signal outside the region defined by a mask radius is deleted and replaced by zero. The masking radius is supplied by the user who investigates the images and decides where to set the mask. PSF = Moffat + residuals s.fits = mofc.fits + psffond.fits Figure 4.7: psf prive.f: Constructing the point spread function (freehand drawing for illustrative purposes only). The masked model of the residuals is returned as the image psffond.fits. The final part of the program psf prive.f performs this addition: s.fits (final PSF) = mofc.fits (analytical part) + psffond.fits (numerical part) – see Figure 4.7. We now have the PSF, s, that we will use in the final deconvolution. It is normalised to a total flux of one, and with the pixels re-ordered so that the peak intensity of the PSF falls on the lower left corner of the image, with coordinates (1,1). This re-ordering is simply done for technical reasons, to facilitate the programming which uses FFT (Fast Fourier Transform). 4.5 General outline of MCS 4.5.3 47 Deconvolution with the profile s The last MCS program, dec prive.f, performs the final deconvolution of the data image g.fits. This program takes as input the image g.fits and its associated noise map sig.fits, as well as the PSF s.fits created during the two previous stages. The program tries to minimise a complicated function of many variables using two χ2 -minimising algorithms: minimi and conjugate gradient. By minimising this function, see Eq. 4.7, the program tries to minimise the differences between the deconvolved image, reconvolved with the PSF, and the original data, to make as close a fit as possible. The minimisation method assumes that all variables are independent and minimises an N dimensional function along N independent directions. Minimisation is done along the N directions until each individual derivative changes sign. This is done iteratively. When the derivatives of the function relative to one variable changes sign, the step is divided by two and the minimisations continue in the opposite direction. The step is increased by 10% if the derivative does not change sign between the next iterations. The iterations stop when the minimum of the function does not decrease significantly or even increases instead of decreasing. The algorithm minimi is the main algorithm, while conjugate gradient fine-tunes the result. There are two input text files to fill in before running this final program: deconv.txt for the deconvolution parameters and in.txt for the photometry. In these text files the user has to enter the initial parameters, the step values and the maximum step size for the minimising algorithms, the smoothing strength to be applied to the extended component of the image (the background), the peak intensity of the point sources after deconvolution as well as the coordinates of these point sources. The function that is minimised is, in one dimension: N h N N N M i2 i2 X X X X 1 hX hi − rij hj (4.7) S2 = sij (hj + ak r(xj −ck ) )−di +λ σ2 i=1 j=1 i=1 i j=1 k=1 The function S2 is minimised with respect to the unknowns hj (j = 1 . . . N ), ak and ck (k = 1 . . . M ). The first term is the χ2 fit, and the second term is the smoothing function for the background. We can see from S2 that the image is decomposed into a sum of point sources and a diffuse background, see also Equation 4.6. We also have that: N is the number of pixels (sampling points), M is the number of point sources, 48 The MCS procedure σi is the standard deviation of the image intensity measured at the ith sampling point, k indicates a point source with intensity ak and position ck , h is the background, λ is a Lagrange parameter for the background which is determined so that the reconstructed model is statistically compatible with the data (λ can also be a varying function across the image frame). Note that in the first term of Equation 4.7, the deconvolution is performed using the calculated PSF, s, instead of the total PSF t. Comparing Equations 4.4 and 4.7, we also see that the solution f has been replaced by Equation 4.6. We assume that what is not point sources is a smooth background. In the second term of Equation 4.7, λ gives the intensity of the smoothing of the background. The value of λ is entered into deconv.txt. In contrast to our previous smoothing parameter in the program psf prive.f, higher values of λ will here decrease the smoothing. Since the width of the Gaussian profile of the point sources in the deconvolved image is chosen by the user, and thus known a priori, the peak intensity is uniquely determined. An approximate value for the peak intensity after deconvolution is entered into in.txt. This can easily be calculated, as the deconvolution preserves fluxes. Approximating the light distribution with a Gaussian, the flux F of a point source is given by F = 2πσ 2 Ipeak . (4.8) This expression is derived from integrating a Gaussian function, see Appendix D for the calculation. Figure 4.8 is an illustration of the relationship between flux and peak intensity before and after deconvolution. Since F is conserved, we get σ1 σ12 Ipeak,1 = σ22 Ipeak,2 ⇒ Ipeak,2 = Ipeak,1 ( )2 , (4.9) σ2 where the indices 1 and 2 refer to values before and after deconvolution, respectively. We also have that σ relates linearly to the FWHM by the relation 4 √ (4.10) FWHM = 2σ 2ln2 . Thus, we can calculate the intensity after deconvolution, Ipeak,2 , by using Equations 4.9 and 4.10. To obtain σ1 , we measure the FWHM of the point sources in the original image (before deconvolution), whereas σ2 is 4 See Appendix D for details. 4.5 General outline of MCS 49 the FWHM after deconvolution, the FWHM we have chosen: the profile r whose width has to be ≥ 2 pixels at FWHM. We measure the peak intensities for the point sources, multiply by the ratio of the σs and enter this in the input text file in.txt. I (peak) I (peak) deconvolution FWHM FWHM Figure 4.8: The FWHM decreases and the peak intensity increases after deconvolution, but the flux (the area under the graph) is preserved. The output from the program dec prive.f is given as two text files, out.txt and in2.txt, as well as the following images: dec01.fits, back.fits, resi01.fits and resi sm01.fits. The image dec01.fits shows the two point sources (+ the background) after deconvolution. The image back.fits contains only the background. In our case of studying the QSO 0957+561 system, the galaxy should appear here, especially if we deconvolve several frames simultaneously to obtain more information on the background. The text file out.txt contains the output information from the program (source intensity and coordinates), as well as some information on the χ2 reached and the number of iterations used to reach it. The text file in2.txt contains the output photometry but in the same FORTRAN format as the input text file in.txt. This is to facilitate re-running of the program. The deconvolution may be run several times, once with relatively large steps to improve too rough estimates of the initial parameters, and then with smaller steps, using the results obtained in the first run as new input. There are some additional parameters to set in in.txt and deconv.txt if several images are to be deconvolved simultaneously, see Section 4.6. We obtain a numerical model of the galaxy, which we could improve by using an option in the program to introduce a variable λ map. 4.5.4 Residual frames The quality of the results from the MCS deconvolution can be checked from ‘residual maps’: images showing the residual in each pixel, i.e. the match 50 The MCS procedure between the deconvolved model image (reconvolved with the PSF, s) and the data. Examples of two residual maps, from our final deconvolution results, can be seen in the top right and bottom left panels of Figure 4.9. The image resi01.fits is simply the difference between the data g.fits and model dec01.fits, in units of the photon noise (that is, divided by sig.fits). The image resi sm01.fits is resi01.fits squared: it is the χ 2 value in each pixel, smoothed and reconvolved with the PSF, s. The smoothing is done with an appropriate function so that any value is replaced by a weighted mean on a neighbourhood containing a few dozen pixels. This smoothing and reconvolution makes it easier for the eye to see the structure in the fit and to recognise objects that may be fitted less well, e.g. the galaxy. We see that there are some residuals where the point sources are located. The parameter λ in deconv.txt determines the smoothing strength. This value is adjusted by the user and the deconvolution is re–run until resi sm01.fits is as flat as possible with a mean value of 1. If we have smoothed too much (low λ value in deconv.txt), this image will have a mean value significantly above 1, and if we smooth too little (high λ value), the mean value will be lower than 1. We try to smooth so as to obtain a value in this image of approximately 1. In the point sources the value will be slightly higher, but we try to obtain as good a fit as possible by checking this image. As deconvolution tries to recover high frequencies, it will increase the noise in the images. The last term in Equation 4.7 tries to minimise this increased noise. The background h is smoothed with the profile r, revealing the frequencies in the background that are too high for the chosen sampling. These higher frequencies are then subtracted from the background, thus ridding the image of artifacts resulting from too high frequencies. 4.6 Simultaneous deconvolution In addition to deconvolving a single frame, the MCS deconvolution can be used to simultaneously deconvolve several individual frames. Indeed this is one of the main points of the MCS method. When deconvolving simultaneously, one takes advantage of the S/N of all the images to construct a background, hence obtaining a numerical model of the galaxy. Each data frame has its own PSF s1 , s2 , . . . sn . The extended sources and the positions of the point sources are forced to be equal in all images. We find the deconvolved image that is the best fit between each data image and the deconvolved model image (reconvolved with the PSF), using 4.6 Simultaneous deconvolution 51 the respective point spread functions and sigma images (one for each image). In this way, the solution is compatible with all the images. The intensities of the point sources are allowed to vary, while the positions are constrained by all the images used. We obtain n deconvolved images that are essentially one image, the only difference between them being the peak intensities. When running the simultaneous deconvolution, initial intensities for the point sources are entered for each frame. In order to detect intrinsic variations in the quasar components, the intensities of the point sources are allowed to converge to different values from one frame to another. One set of initial positions are entered for all the frames together, which is the reason we had to align all g.fits images to within one pixel of each other. We also enter one guess peak intensity for all images in the set, calculated from an image with representative seeing. The minimisation performed during simultaneous deconvolution can be expressed by the following: N N M i2 X X λ1 h X ~ a1,k r(~xj − ~ck + ~δ1 ) − d1,i S3 = s1,ij α1 h(~x + δ1 ) + β1 + i σ i=1 1,i j=1 k=1 + .. . N N M i2 X X λ2 h X s2,ij α2 h(~x + ~δ2 ) + β2 + a2,k r(~xj − ~ck + ~δ2 ) − d2,i σ i i=1 2,i j=1 k=1 N M N i2 X X λn h X ~ ~ s2,ij αn h(~x + δn ) + βn + an,k r(~xj − ~ck + δn ) − dn,i + σ i i=1 n,i j=1 k=1 + N h X i=1 hi − N X j=1 rij hj i2 (4.11) The function S3 that is minimised, consists of a deconvolution part and a smoothing term (the last term in Equation 4.11). Three new parameters are included when deconvolving several frames: α, β and δ. The background h can be adjusted for each image by a multiplicative factor α and an additive term β. The factor α can correct for effects due to different exposure times and varying transparency of the atmosphere, while β corrects for errors in the sky–subtraction. To account for small shifts in position from frame to frame, a δx and δy parameter are entered in in.txt. These affect the positions of the centres of the point sources during the deconvolution in the following way. xn = x 1 + δ x 52 The MCS procedure yn = y 1 + δ y where n indicates the frame being deconvolved, and the subscript 1 refers to the first image in the list which is chosen as reference image. The reason we need these shift parameters, is that the centres of the point sources, after aligning the frames to within one pixel of each other, will be located in different places within this one pixel. The dithering in x and y improves the sampling by combining several offset samples of the same light distribution, similar to a technique known as variable-pixel linear reconstruction, or ‘drizzling’, which is used on WFPC2 images from the Hubble Space Telescope to optimise the information obtained from undersampled images, see Fruchter and Hook (2002). This technique is implemented in the MCS deconvolution method. 4.6.1 Final deconvolution results For our two final, simultaneous deconvolutions of the March 2001 data, we used our 73 R and 15 V combined image frames. We experienced some problems when deconvolving both the quasar components and the stars. Our results show relatively large residuals, although this varied considerably from frame to frame. In general, the amount of residuals did not seem to affect the photometry, although there were exceptions to this in extreme cases, see Section 5.2.2. The deconvolved image also shows some artifacts, which we comment on in this section. Figure 4.9 shows our final deconvolution results of the quasar components, in the R band. The top left panel shows the original, extracted image of the two quasar components A and B. The top right and bottom left panels show the two residual images discussed in Section 4.5.4, and the bottom right panel shows the final, deconvolved image. The galaxy can be seen to the North– East of the B component. In the deconvolved image, the galaxy is separated from the B component. There is some noise in the deconvolved image, as well as some ‘holes’ or rings around the point sources. This may be a result of deconvolving to a FWHM value of 2 pixels, the maximum deconvolution allowed. However, we chose this FWHM in order to facilitate the construction of the PSF, see Section 4.5.2. One of the main problems we encountered in the deconvolution process, apart from the residuals, was to get enough flux into the galaxy. When examining preliminary deconvolution results, we would often find that the galaxy was flat in the central parts. This indicated that the program, for some reason, had stopped iterating before completely building up the galaxy. Our 4.6 Simultaneous deconvolution 53 solution to this was to construct the sharp centre of the galaxy in the first run, using low smoothing and relatively large steps for the background, see Section 4.5.3. In the second deconvolution run, using the output solution from the first run as input, we used smaller steps to fine-tune the result and applied more smoothing. The point sources in the deconvolved image, lower right panel in Figure 4.9, does not seem perfectly symmetric. Examining the image back.fits, we see that the background is not completely zero where the points sources are located. There are values of maximum ∼3% of the maximum peak intensity after deconvolution, at the positions of the point sources in the background image. This causes the point sources in the deconvolved image to appear non-circular. We attribute the residuals mainly to errors in the PSF fit, although we did not find evidence of a large PSF variation across the CCD, see Section 4.3. It may be that the PSF star was too bright compared to the quasar components, and that a PSF of a more comparable brightness to these objects would have produced less residuals. When analysing different objects, there have been some indications by users of MCS that the PSF star chosen in the MCS programs should not be too bright compared to the object that is being deconvolved (Ingunn Burud, private communication). In our case, there were no ‘ideal’ PSF star candidates. Our PSF star, E, is approximately 4 times as bright in flux as the QSO A component. The stars P and Q, which we considered for PSF stars, are about a quarter of the brightness of A. On the other hand, we had even larger residuals when deconvolving the star H, up to 10 times larger values in the χ2 image, but this star is about twice as bright as E. When deconvolving the bright field stars we needed to apply more smoothing. Residuals from the deconvolution of star R were comparable to those from the deconvolution of the quasar. 4.6.2 Obstacles When deconvolving the March 2001 data in the V filter, the δx and δy increased inexplicably a little above 1 on two frames. We tested several combinations of step sizes without any improvement. There did not seem to be any problems with the rms values of the aligning from these images, so in the end we simply discarded them, regarding this as the safest choice. After discarding the images, MCS ran smoothly on the rest of the images. The same thing happened in R, but only affecting one frame, which was also discarded from the final deconvolution. Sometimes the program psfm prive.f would stop iterating. Changing the start value for β helped. 54 The MCS procedure Figure 4.9: North is up, East is left on all four images. Top left: Observed, extracted image of the quasar components. Top right: Difference between data and model, divided by the photon noise. Bottom left: The squared, reconvolved and smoothed residual image. Bottom right: Final, deconvolved image. Maximum value in the background, i.e. in the galaxy, is ∼5 ADU. For the observed and deconvolved images, the colour table used produces the colour black for values above 5, and white for values below zero. For the lower left residual image the levels were set to 0 and 10, and for the upper right residual image the levels are –10 and 10. There is a certain amount of ‘art’ involved in running MCS efficiently, in the sense that the user has to familiarise him- or herself with how things should ‘look’, which takes some experience and time. We found the interplay between smoothing strength and step size to be very important when deconvolving. The effect of a given step on the function to be minimised is stronger with high smoothing than with low smoothing. This implies that, in general, heavier smoothing requires smaller step sizes. In our initial runs of MCS, we found strong correlation between quasar 4.6 Simultaneous deconvolution 55 components and the stars, at the level of several tenths of magnitudes. We found that we had cut too much of the PSF, so we re-made the PSFs using almost no mask. As a consequence, we produced noisier PSFs. However, the residuals from the Moffat fit were so extensive that we had to make the mask very large. Essentially no mask was used unless to cut off a cosmic. Changing the PSFs in this manner, as well as running more images simultaneously and running the deconvolution program for a long time with many iterations, effectively reduced the correlation to the millimagnitude level. In about 20% of the images from the main data set, we experienced some problems when running psf prive.f. The modelled residuals that were produced were far too weak compared to the actual data residuals, with values at least 10 times too weak. On these images, we used P and Q as PSF stars instead. It is possible that large gradients in the bright PSF star E caused the problem. We noted the images, and examined all light curves without finding any deviant behaviour for these particular points. 4.6.3 Testing MCS As we start the deconvolution with equal peak intensities for all point sources, it may be argued that this is the reason for the small standard deviation in the results presented in Chapter 5. To test whether the starting values we enter in the photometry parameter file for the deconvolution had any significance for our results, i.e. for the magnitude levels and the dispersion of the data points, we performed two tests. In the following, what we refer to as our ‘reference run’ is not the final deconvolution results discussed in Section 4.6.1 but a very similar result that we obtained earlier. First, we switched the guess peak intensities after deconvolution for the two quasar components A and B. For this test, the values for A and B were still the same for all images, but their internal ratio was changed, i.e. B brighter than A as input instead of vice versa. For the second test, we selected a wide range of guess peak intensities for A and B. The output values from the deconvolution programs, two numbers representing the peak intensities of the quasar components, for technical reasons relative to some normalisation factors, had a standard deviation of approximately 0.7% of their mean value. The range of guess peak intensities we entered for this test, varied from 4 to 44% of the mean output value from the reference run of MCS. In other words, the variations we introduced artificially were much larger than the variations we had obtained from the reference run. We now wanted to see if MCS could return the relatively constant magnitudes for A and B, and whether the ratio between A and B would be affected by introducing these variations. 56 The MCS procedure Figure 4.10: Results of testing MCS for stability with different entries for the guess peak intensity after deconvolution. Left: The change in magnitude level of the A component is very small, both for test 1 (dashed line) and for test 2 (dotted line). Right: The change in magnitude level for the B component is very small for test 1 but larger for test 2. The results, displaying the change in magnitudes for the two tests, are shown in Figure 4.10. For the reference run, the mean magnitude difference between the two components was 0.10 mag, with A the brightest component. For the first test, MCS quickly switched back to A being brighter than B. The components ended up being just slightly closer (A brighter than B by 0.09 mag) than for the reference run. The magnitude levels were approximately the same as before, for both components. For the second test, A ended up being brighter than B by a mean value of 0.06, mostly due to the somewhat larger change in magnitude of the B component, see Figure 4.10. Table 4.1: Standard deviation of the A and B magnitudes for the reference run and for each test. Reference Test 1 Test 2 A component 0.0077 0.0078 0.0077 B component 0.0095 0.0093 0.0090 We see from Table 4.1 that the low dispersion in magnitudes from the reference run was regained very well in both tests. We note with some surprise that the biggest change in standard deviation is that component B shows a 4.6 Simultaneous deconvolution 57 lower standard deviation in test 2 than in the reference run. All in all, we note that the program is very robust against ‘artificially introduced’ variations, consistently producing magnitudes with low standard deviation. Note that our tests were performed with the rest of the input text files being equal, i.e. with the same step sizes for the different parameters. 58 The MCS procedure Chapter 5 Photometry from the ALFOSC data May I borrow some of the light that you possess? The 3rd and the Mortal, Oceana In this chapter, we first discuss a few basic principles of photometry studies. We explain how we obtained magnitudes from the results of MCS, and how we have estimated the errors in these magnitudes. In the main part of the chapter we present our photometric results from the ALFOSC data, from our main data set from March 2001, as well as from the one night in January 2000. We display light curves showing our results. A discussion of the results, as well as a comparative analysis with the findings of Ovaldsen et al. (2003b) and Colley et al. (2003), follows in Chapter 6. 5.1 Basic principles of photometric studies Photometry, or the study of brightness measurements, is one of the fundamental branches of astronomy. The flux F of an astronomical object relates directly to what we measure with a telescope. The monochromatic flux of energy through a surface, or into a detector, is the amount of energy per unit time passing through a unit area of the surface per unit frequency interval. When performing photometric investigations, the amount of flux emitted from an object at different wavelengths is studied, as well as how this flux changes in time. Photometry is important for objects as various as planets, stars and galaxies, and the time scales of interest range from single observations to careful monitoring over months and years. For the purposes of this work, we are solely interested in studying stars and other objects that appear to us as point sources, i.e. the twin quasar. 60 Photometry from the ALFOSC data Apparent magnitude, that is, the magnitude we observe here on Earth, is defined by F m = −2.5 log10 , (5.1) F0 where F0 is an arbitrary ‘standard’ flux. The magnitudes of any two stars are then related by F 1 m1 − m2 = −2.5 log10 , (5.2) F2 where m1 and m2 are the magnitudes of the two stars and F1 and F2 are their respective fluxes. Note that m1 > m2 when F2 > F1 , that is, brighter objects have numerically smaller magnitudes. Because the magnitude scale is a logarithmic scale, the mean value of a series of magnitudes is not equal to the mean value of the corresponding fluxes. To calculate the mean value of magnitudes corresponding to the fluxes, we should first convert the magnitudes into fluxes, using F = 10(−0.4m) , calculate the mean of these fluxes and then convert this number back into magnitudes. In this thesis, we specify when we calculate directly the mean value of the magnitudes. The results we display in the tables in this chapter, would not be affected by how we calculate the mean, as the difference in our case comes only at the 5th decimal point. When calculating the magnification ratio in Section 6.2, we also checked that the two methods provided equal results to the given accuracy. There are two main types of photometric techniques, aperture and pointspread function (PSF) photometry. The aperture technique makes no assumption about the actual shape of the source PSF. The observed counts within a chosen aperture, centred on the source, are collected and summed together. The contribution from the background level is estimated by nearby pixels and subtracted, and the net result constitutes the signal from the source. Although sounding deceptively simple, there are many intricate steps involved, including determining the image centre and the size of the aperture, correcting for partial pixels that are intersected by the (usually circular) aperture boundary and deciding how and where to measure the background level. Aperture photometry works best with a group of stars that are not too crowded nor too faint. If this prerequisite is not present, PSF photometry is a more appropriate technique. As the name implies, this technique assumes that all point sources on the CCD frame have the same shape, barring distortions introduced by camera optics. This shape differs from a point source δ–function because the light from the point source is blurred when passing through the Earth’s atmosphere and the telescope. Performing PSF photometry consists of constructing the point-spread function and scaling 5.1 Basic principles of photometric studies 61 it to fit the different point sources on the image. There are many ways of constructing a PSF, some of which were mentioned in Section 4.5.2. There are of course a myriad of photometric techniques derived in some way from these two main types. Described in Ovaldsen (2002), is a hybrid method consisting of ‘PSF cleaning’ of a nearby source followed by aperture flux measurement of the other. The deconvolution method we have adopted in this thesis work, utilises a PSF constructed from one or several stars on the frame and increases the resolution of the images by deconvolving them as well as exploiting the S/N ratio of several images simultaneously. 5.1.1 Standard systems and reference magnitudes In modern usage, a photometric system 1 consists of a set of discreet wavelength bands and a set of primary standard stars. Each wavelength band has a known sensitivity to incident radiation, defined by the detectors and filters used. The standard stars define the magnitude scale of the system. A catalogue of primary standard stars for a given photometric system is usually published when the system is defined. Sometimes, for widely used systems, further catalogues of ‘secondary’ standards will be compiled by making observations calibrated with the original primary standards. Many catalogues of photometric standard stars exist, e.g. the widely used Landolt catalogue. Reference magnitudes encountered in the literature are derived, step by step, from the first standards. Since this process can result in the reference magnitudes differing from each other, the system one has chosen must be specified. The magnitudes we produce are instrumental, i.e. they are dependent on the combination of the bandpass of the filters and the response of the CCD. This means that the magnitudes are specific to our data set. In order to compare results to that of others, it is usually necessary to transform the instrumental magnitudes into a standard photometric system. This is done by re-observing the standard stars in the system and comparing the instrumental and standard magnitudes. In this thesis, we are mostly interested in measuring variations and obtaining relative photometric measurements. Ovaldsen et al. (2003b) have used the same ALFOSC data set that we have. In chapter 6, we wish to compare our magnitudes with those obtained by these authors and by Colley et al. (2003), as well as comparing our relative photometry, e.g. the internal ratio between A and B, and any variations in the light curves. We therefore 1 For example the Johnson-Morgan UBV system, the Strömgren system etc. 62 Photometry from the ALFOSC data calibrate our magnitudes using the same values for the reference magnitudes as these authors. This does not produce absolute photometry in the strict sense, but it places our data on a common scale so we can compare results directly. Our reference magnitudes are listed in Tables 5.1 and 5.2. The R references were obtained by Ovaldsen et al. (2003b), by a bootstrapping procedure of magnitudes originally supplied by Rudy Schild (private communication). The reference V magnitudes were obtained by the same authors using V − R colour terms from Serra-Ricart et al. (1999). In our combined images, the brightest stars (F, G and X, see Figure 2.1) were often saturated on at least one of the 10 individual images we stacked. As a consequence of this, we did not use these stars in our photometric study, but their magnitudes are still listed here as reference. Table 5.1: Reference magnitudes in the R filter for 7 field stars, see Figure 2.1 for location of the stars. Star Ref. mag F 13.758 G 13.731 H 13.972 E 14.778 D 14.513 X 13.425 R 16.329 Table 5.2: Reference magnitudes in the V filter for 7 field stars, see Figure 2.1 for location of the stars. Star Ref. mag 5.1.2 F 14.093 G 14.054 H 14.497 E 15.247 D 14.948 X 13.770 R 17.086 Colours The intensity of the light emitted by stars and other astronomical objects varies strongly with wavelength. The apparent magnitude observed for a given star by a detector therefore depends on the shape of the filter response function, i.e. the width of the observed wavelength band and the transmittance as a function of wavelength. A quantitative measure of the colour of a star is given by its colour index (CI). This is defined as the difference between magnitudes at different wavelengths: CI = m(λ1 ) − m(λ2 ) . 5.1 Basic principles of photometric studies 63 Because the colour index is a magnitude difference, it corresponds by Equation 5.2 to a flux ratio. The star Vega, or α Lyrae, is defined to have zero magnitude at all wavelengths. We note that the star we used for normalisation (star D in Figure 2.1) has a different colour than the quasar. In fact, all the field stars are redder than the quasar components. In Section 6.1.4 we perform some simple tests based on colour diagrams. 5.1.3 Photometry from MCS The MCS procedure was described in Chapter 4. We have used MCS as a means of obtaining photometry from the QSO 0957+561 system. In this section we describe how we calculate object magnitudes from the MCS output. The photometric output information from MCS consists of a list of the maximum intensities of the point sources. In Section 4.5.3 we mentioned the relationship between the flux and peak intensity of a star, where the star is approximated by a Gaussian distribution. We show in Appendix D that this is given by F = 2πσ 2 Ipeak . When using subsampled pixels, we effectively divide each pixel into four smaller pixels. We therefore divide the expression above by four, giving 1 F = πσ 2 Ipeak . 2 Inserting this expression for the flux into Equation 5.2, we obtain, by comparison with a reference star with a known standard magnitude: mobject − mstar = −2.5log10 F object Fstar = −2.5log10 I object Istar , (5.3) where the subscript ‘object’ denotes the object we are investigating, whether a quasar component or another star. The σs cancel each other in the fraction, as they are the same for the object and the reference star. This σ is related by Eq. 4.10 to the FWHM we choose in the MCS programs. In Figure 4.4, this is the width r of the Gaussian distribution. Since we use the same PSF to deconvolve the quasar components and the reference star, this σ is the same for Fobject and Fstar . The MCS procedure requires the background to be the same in all images. To deconvolve the quasar, which has a galaxy in the ‘background’, we therefore normalise the images by a ‘normalising star’, which is assumed to be constant. This also has the effect of removing air mass and cirrus effects, 64 Photometry from the ALFOSC data as described in Section 4.3. The reference star we chose for normalisation was star D. The direct output from MCS thus consisted of a list of peak intensities representing the ratio Iobject /ID . We therefore merely had to use Equation 5.3 directly, giving: mobject = −2.5log10 I object ID + mD . (5.4) We considered using the mean of several reference stars for calibration purposes, which is a good idea if we have several reference stars available which are known to be constant. If one of them should vary, however, we would introduce systematic effects in the calibration. A detailed study of the behaviour of the field stars in Figure 2.1 with special focus on variability is still to be done. As mentioned earlier, we also have a reduced number of available stars because of saturation of the brightest field stars. We finally chose one reference star (star D), to use for both normalisation and calibration as described. We deconvolved other stars in the field to check for correlated temporal intensity variations of the stars and the quasar components, which would indicate that the chosen reference star was variable. We first suspected this was the case, see Section 4.6.2 for a description of the strong correlation from our preliminary runs. In our final runs, we did not find any indications of variability in the reference star. 5.2 Photometric results In this section, we present our photometric results from the ALFOSC data sets, from March 2001 and January 2000. The results are discussed and analysed in more detail in Chapter 6. All results are from the simultaneous deconvolution. 5.2.1 Error estimations We calculated our errors for the quasar component magnitudes as the sum of a systematic part (attributed mainly to PSF errors) and a random part (the photon noise). 2 2 2 σtot = σsyst + σphoton (5.5) The star R is of similar brightness as the quasar components. Assuming that this star is not variable, we used it to compute the systematic part of the errors. These systematics are an a posteriori measure of how well the PSF fits the object we study (in this case the star R). The underlying 5.2 Photometric results 65 assumption is that the systematic errors are the same for the star and the quasar components. We used the standard deviation around the mean value of the flux of R from each night as a measure of σtot for the star. See Table 5.7 for a list of the standard deviations of the magnitudes. The standard deviation of the star was calculated after variations due to air mass etc. had been removed through normalisation. For the night of March 14th–15th, the last two data points in the top panel (R band) of Figure 5.5, showing the star’s light curve, differed significantly from the rest. We thought it unlikely that these points were representative of this star’s assumed constant behaviour. When calculating the errors for the quasar components for the remaining images this night, we therefore calculated the systematic errors from the star without these two points. When calculating the error bars for the two ‘deviating’ points themselves, we used the standard deviation of all points that night. This produces, appropriately, larger error bars for the last two points. We calculated σphoton for the star R (in ADUs) by using the following formula: s Flux(star R) 2 σphoton (star R) = + npix σsky . (5.6) g In this formula, for a sky subtracted image, g is the gain of the CCD and σsky is the standard deviation in the background. When squared, σsky approximates the original background level. The first term underneath the square root represents the contribution to the total photon noise from the source itself, in this case the star R. The second term is the contribution from the background. In order to estimate the background contribution from inside an aperture of approximately the size of the object, we multiply σsky by the number of pixels inside this aperture, npix . The size of the aperture was set in the following way: We measured the distance from the centre to the point where the intensity of the object was ∼10% of the peak intensity, and used this as radius in the circular aperture. For the systematic errors in star R, we therefore have, from Equation 5.5: 2 2 2 σsyst (star R) = σtot (star R) − σphoton (star R) , (5.7) 2 2 where σtot is measured from our results and σphoton is calculated from Equation 5.6. Since we did not use star R in constructing the PSF, the systematic errors calculated in this way will include the errors introduced by the difference between PSF and the object we are deconvolving. We add these systematics, in quadrature, to the errors from the photon noise for the quasar components: 66 Photometry from the ALFOSC data 2 2 2 σtot (QSO) = σsyst (star R) + σphoton (QSO) (5.8) Using an equation like Eq. 5.6 to calculate the photon noise for the quasar 2 components, we can now estimate the total errors σtot (QSO). When calculating these errors, we found that they were dominated by systematics, i.e. the systematic part of Equation 5.8 was much larger than the random part. This we attribute to the rather large residuals from the deconvolution. Error bars for the other field stars were calculated in the same manner. To convert the total error bars from flux to magnitude, we ‘differentiate’ Equation 5.1: d(mag) = −2.5 d(log10 (Flux)) . Using the following relation to differentiate the logarithm: log10 (x) = ln(x) , ln(10) where ln(x) is the natural logarithm, we obtain d(mag) = −2.5 5.2.2 dFlux 1 . Flux ln(10) (5.9) Light curves from March 2001 In this section, we present our results from our main data set; the four nights in March 2001. Tables 5.3 and 5.4 display the mean of the R magnitudes each night for the quasar components A and B, respectively. Also given are the standard deviations of the magnitudes each night. The individual error estimates were calculated in the manner discussed in Section 5.2.1. The mean of these individual errors for each night, is also given in the tables. Table 5.3: Mean values of the R magnitudes for the A component of QSO 0957+561 for the four nights, the standard deviation of the magnitudes σ mag and the mean error (average of individual estimated errors) for each night. 2001 2001 2001 2001 Night March 14–15 March 15–16 March 16–17 March 17–18 Mean R mag 16.422 16.428 16.432 16.436 σmag 0.0026 0.0023 0.0031 0.0032 Mean error 0.0034 0.0033 0.0030 0.0042 5.2 Photometric results 67 Table 5.4: Mean value of the R magnitudes for the B component of QSO 0957+561 for the four nights, the standard deviation of the magnitudes σ mag and the mean error (average of individual estimated errors) for each night. 2001 2001 2001 2001 Night March 14–15 March 15–16 March 16–17 March 17–18 Mean R mag 16.523 16.529 16.534 16.540 σmag 0.0041 0.0033 0.0046 0.0033 Mean error 0.0037 0.0036 0.0033 0.0046 Figure 5.1 displays the light curves for the A and B quasar components for the four nights in March 2001. The horizontal dashed line is the mean of all data points for all nights, and is meant solely as a guide to the eye. Error bars are omitted for clarity, and are shown in Figure 5.2. The most striking feature of Figure 5.1 is the near constant magnitudes, for both A and B. As the B component and the galaxy are partially overlapping, it is harder to separate this component from the background. This is probably reflected in the slightly larger standard deviation each night for component B than for the A component. A slight, systematic decrease in brightness from night to night, in both components, can be seen. Subtracting the mean R magnitude of the night 14th–15th, from the mean magnitude of the night 17th–18th, yields 0.014 mag for the A component, and 0.017 mag for the B component. Comparing this decrease between the nights, to the standard deviation within each night, we see that this change in magnitudes is more than 3σmag , which we consider a significant decrease. Whether the decrease is ‘real’ in the sense of being intrinsic to the quasar or could be caused by observational effects, will be discussed further in Chapter 6. We will here mention the main points of interest. That the decrease is so similar in both A and B does not particularly favour the theory that the variation is intrinsic. The two components are shown from observations of the same night, making the curves theoretically unrelated. The light rays from the two components have not travelled the same distance and represent different times in the quasar’s ‘emission history’. It is of course a possibility that both components happen to exhibit the same variation at the same time. However, the decreasing magnitude from night to night is not reflected in the behaviour within each night. Although it is on a very small scale, there seems to be a variation within each night, similar in 68 Photometry from the ALFOSC data Figure 5.1: Light curves in the R band for the quasar components A and B during the four nights. The horizontal dashed line represents the mean of all data points and is meant as a guide to the eye. Notice the decrease in brightness from night to night, in both A and B. A and B. In Figure 5.1 we see a variation which resembles a ‘U’ shape each night. In Figure 5.2 we have plotted light curves for each night individually, including the individual error bars for each point. The variations within each night, though on a small scale, could still be significant compared to the (1σtot ) error bars. Since the ‘U’ shape each night is discernible every night, the probability that it is a significant variation increases, although the variations themselves are of a scale comparable to the error bars. We see from Figure 5.1 that the maximum magnitude value each night occurs around midnight. We therefore performed a quick calculation intended to probe the scale of the nightly variations. Subtracting the mean of the three data points in the middle of each night, from the mean of the three points at the start of that night, we obtain values ranging from 0.0005 to 0.0068. Performing the same calculation, but replacing the first data points each night with the last three points of the night, we find changes in magni- 5.2 Photometric results 69 Figure 5.2: Light curves in the R band for the quasar components A and B for each night, and with individual error bars. As in Figure 5.1, data points for the A component are represented by a cross (×) and points for B by a diamond shape (). tude ranging from 0.0008 to 0.0103. Compared to the average error bars in Tables 5.3 and 5.4, we see that both computations show some variations on a larger scale than the error estimations: exceeding 2 and even 3σtot for the B component. This exercise was not intended to be an exact measure of the possible variations within each night, but is included as a rough estimate. We find that the variations within each night, although possibly consistent with a straight line fit, are on a significant enough scale, especially as they occur in all nights, to warrant further investigations. These we will pursue in Chapter 6. In the V filter, we had fewer images, approximately 50 each night, leading to 4–5 data points in the corresponding light curves, see Figures 5.3 and 5.4. 70 Photometry from the ALFOSC data We also had to discard two images because of aligning problems, as mentioned in Section 4.6.2. Results for the quasar V images are listed in Tables 5.5 and 5.6. The standard deviation of the values for each night is given in the tables, although we had very few points. We note that the quasar components are weaker in V than in R. Table 5.5: Mean values of the V magnitudes for the A component of QSO 0957+561 for the four nights, the standard deviation of the magnitudes σ mag and the mean error (average of individual estimated errors) for each night. 2001 2001 2001 2001 Night March 14–15 March 15–16 March 16–17 March 17–18 Mean V mag 16.833 16.833 16.839 16.843 σmag 0.0098 0.0102 0.0033 0.0033 Mean error 0.0068 0.0058 0.0034 0.0060 Table 5.6: Mean values of the V magnitudes for the B component of QSO 0957+561 for the four nights, the standard deviation of the magnitudes σ mag and the mean error (average of individual estimated errors) for each night. 2001 2001 2001 2001 Night March 14–15 March 15–16 March 16–17 March 17–18 Mean V mag 16.921 16.918 16.927 16.932 σmag 0.0098 0.0099 0.0029 0.0038 Mean error 0.0073 0.0062 0.0037 0.0066 The temporal sampling in the V filter is not good enough to conclude anything about magnitude variations internal to each night. However, we do notice a decrease in brightness from the first night to the last, the same trend as in the R band. The overall decrease in V is 0.010 and 0.011 mag for the A and B component, respectively. The tendency of a decline in the V brightness is thus slightly less pronounced than for the R filter. In both R and V we have a slightly larger decline in the B component than in A. The average colour index V − R of all four nights is 0.407 ± 0.010 for component A and 0.393 ± 0.012 for B. Changes in colour from night to night will be discussed in Section 6.1.4. Future studies should include more images in the V filter in order to discuss colour changes within each night. 5.2 Photometric results 71 We note that both quasar components seem to exhibit similar behaviour, in both R and V . This ‘zero lag correlation’ will be discussed in Section 6.3. Figure 5.3: Light curves in the V band for the quasar components A and B during the four nights. The horizontal dashed line represents the mean of all data points and is meant as a guide to the eye. Notice the decrease in brightness from the first night to the last, for both A and B. Results from the field stars We have also deconvolved and produced light curves from four stars in the field: the stars R, H, P and Q, see Figures 5.5, 5.6 and 5.7. This is done in order to have – hopefully stable – reference objects. An important issue when studying the reference stars, is to see whether they exhibit similar variations to the ones we observe in the quasar components. An additional reason for deconvolving the star R was to estimate the systematic errors, as explained in Section 5.2.1. In this section, we first discuss our results for the star R, and then the results from the other stars. 72 Photometry from the ALFOSC data Figure 5.4: Light curves in the V band, for the quasar components A and B for each night individually. As in previous figures, data points for the A component are represented by a cross (×) and points for B by a diamond shape (). In Table 5.7 we present the results from star R. As we estimated the errors for the first night using the standard deviation of R excluding the last two points, see Section 5.2.1, this is the value we have entered in the table. Including these points produces a larger standard deviation, of 0.0049 mag, while the mean level of the night is unchanged. In the R filter, for both A and B components, the error bars are largest on the night of March 17th–18th, which can be seen in Figure 5.2. As our errors are dominated by systematics, this seems to be the result of the larger standard deviation, in R, of star R on this night. Similarly, for V , the smaller standard deviation on the night of March 16th–17th, has resulted in smaller error bars this night for the quasar components in V , see Figure 5.4. The magnitude of star R seems fairly stable from one night to the next. 5.2 Photometric results 73 The distribution of points within each night seem more random than for the quasar. We can discern no ‘U’ shape, but in some cases the brightness seems to be increasing during the night, see e.g. the last three nights in the R band. That we do not find the same variations for the star R as for the quasar components, indicates that the quasar variations are not caused by a varying normalisation star, as this would probably have caused R to vary in a similar fashion. Again, there are too few points in the V band to say much about variations internal to each night. Table 5.7: Mean values of the R and V magnitudes for the star R for the four nights, and the standard deviation of the magnitudes. This star, assumed constant in time, was used to estimate the errors for the quasar components and other stars, see Section 5.2.1 for details. 2001 2001 2001 2001 Night March 14–15 March 15–16 March 16–17 March 17–18 Mean R mag 16.389 16.393 16.392 16.388 σmag (R) 0.0029 0.0032 0.0030 0.0039 Mean V mag 17.092 17.092 17.098 17.097 σmag (V ) 0.0087 0.0072 0.0043 0.0074 We compare the magnitude levels for the star R, in R and V , with the reference magnitudes listed in Tables 5.1 and 5.2. Taking the mean of all four nights as our value obtained for star R, we get a difference in magnitude of m(star R) − mref (star R) = 0.0615 mag in the R band and, similarly, 0.0088 mag in V . So, our magnitudes for the reference star are slightly higher (i.e. less bright) than the reference values. For our further studies of the variations in the light curves of the quasar components, both the nightly ‘U’ shape and the decrease from night to night, the lack of similar variations in star R tells us something about what the variations are probably not be caused by. For example, if the variations in the quasar light curves are to be interpreted as the product of changes in observing conditions, these changes must affect the star and the quasar components differently, as we do not observe any of the same variations in Figure 5.5. This motivates our discussion of the differences in colour between the stars and the quasar, in Chapter 6. Light curves in the R band for the field stars H, P and Q are shown in Figure 5.6, and Table 5.8 provides the mean magnitudes of these stars for each of the four nights. For the star H, there is no systematic decrease in magnitudes from night to night comparable to that observed for the quasar components. The star appears very constant until the last night, when we 74 Photometry from the ALFOSC data Figure 5.5: Light curve in the R and V bands for the star R during the four nights. The horizontal dashed line represents the mean of all data points and is meant as a guide to the eye. Notice there is no systematic decrease in the level for each night, as opposed to the systematic decrease seen in the quasar components, see Figures 5.1 and 5.3. see an inexplicable drop in brightness of 0.01 mag. The stars P and Q show a total decrease in brightness of 0.012 mag and 0.014 mag, respectively. This decrease is systematic, similar to that of the quasar components. The error bars for P and Q are larger than for the quasar components, which is appropriate as these stars are much weaker. The bright star H on the other hand, has error bars of the order of ∼0.4 mmag, which is too small to be seen in the figure. Figure 5.7 displays light curves for the field stars H, P and Q in the V band, and Table 5.9 shows the mean V magnitudes of the stars for each of the four nights. As for the star R, there are too few points in V to conclude 5.2 Photometric results 75 Table 5.8: Mean values of the R magnitudes for the stars H, P and Q for the four nights. Numbers in parenthesis are the standard deviation of the magnitudes in mmag. 2001 2001 2001 2001 Night March 14–15 March 15–16 March 16–17 March 17–18 H (R mag) 13.966 (0.0031) 13.968 (0.0018) 13.969 (0.0023) 13.976 (0.0020) P (R mag) 17.960 (0.0060) 17.965 (0.0042) 17.969 (0.0068) 17.972 (0.0061) Q (R mag) 18.238 (0.0072) 18.243 (0.0057) 18.248 (0.0064) 18.252 (0.0071) anything about variations internal to each night. Taking the mean value of each night shows a decrease in brightness from the first night to the last of 0.008 mag for P and 0.015 mag for Q. The light curve for star H contains even fewer points, so we do not comment further on the variations here, but include the graph for completeness. In both R and V , the star H was saturated on some images, and the corresponding points have been discarded from the light curve. In addition, the deconvolution of three of the extracted images of H in the V band, resulted in extreme χ2 residuals, of values ∼30 times that of the other images. The corresponding points on the light curve deviated from the rest by ∼0.5 mag, so the points were discarded from the final light curve of the star. Again, the error bars for star H are too small to be seen, they are of the order ∼0.8 mmag. Table 5.9: Mean values of the V magnitudes for the stars H, P and Q for the four nights. Numbers in parenthesis are the standard deviation of the magnitudes in mmag. 2001 2001 2001 2001 Night March 14–15 March 15–16 March 16–17 March 17–18 H (V mag) 14.642 (0.0032) 14.663 (—) 14.669 (0.0199) 14.658 (0.0125) P (V mag) 18.836 (0.0088) 18.827 (0.0107) 18.841 (0.0070) 18.844 (0.0080) Q (V mag) 19.261 (0.0078) 19.265 (0.0146) 19.277 (0.0062) 19.276 (0.0117) The stars are generally redder than the quasar components, which could be a cause of dissimilar variations in magnitude. A discussion of colour indices of the quasar components and the reference stars is presented in Section 6.1.4. 76 5.2.3 Photometry from the ALFOSC data Light curves from January 2000 We now turn to our second set (chronologically the first) of observations from the NOT using ALFOSC, originally 171 R and 39 V images from January 25th 2000. We combined the individual images in stacks of 10, in the same manner as for the March 2001 data, to 15 R and 4 V images. The time separation between the two sets is 416 days, from January 25th 2000 to March 16th 2001, which is within the range of the currently held time delay. Unfortunately, weather conditions proved a severe hindrance for the observers. Although ten days of NOT observing time was scheduled for this session, nine nights contributed either no data or very few images of poor quality. We have thus restrained our analysis to the one night with good observing conditions, January 25th–26th. Unfortunately, a determination of time delay from this data set is excluded, due to several facts. One is the weak magnitude variations we observe, the other is the bad observing conditions in January 2000. If more data from the January 2000 session had been obtained, we could have investigated the light curves for short time scale variations that could possibly lead to a determination of the time delay. We would have looked for similarities in the behaviour of the (first arriving) A component in January 2000, compared to the B component in March 2001. The small variations within each night seem to depend heavily on observing conditions and/or reduction errors, so to determine the time delay we would probably require variations in magnitude on the time scale of several days. If we could have discerned a correlation in the behaviour of the two components, after time shifting one of them, we could have entered our findings into the still current debate about which value of the time delay is most appropriate. As it is, we could not hope to find any events in the one day available to us, that would be reflected in our data from March 2001. We have, however, analysed the existing images from the one night in January 2000. This leads to our discussion of the colour indices and difference in magnitude of the quasar components, in Section 6.2. For both R and V bands, we find that the components’ relative brightness has shifted between the two observing sessions, see Table 5.10 and Figure 5.8. In January 2000, we find that B is brighter than A, which is the opposite situation of that in March 2001. In the R band the B component is dimmer by ∼0.02 mag in January 2000 than in March 2001, which is less than the change in R for the star R, which is assumed to be constant. Component A, on the other hand, is dimmer by all of ∼0.2 mag in January 2000. So, chronologically, component A has brightened considerably more than B from January 2000 to March 5.2 Photometric results 77 Table 5.10: Mean values of the R and V magnitudes for the A and B components of QSO 0957+561 for the one night in January 2000, the standard deviation of the magnitudes σmag and the mean errors (average of individual estimated errors) for each night. Comp A B Mean R mag 16.632 16.555 σmag 0.0038 0.0051 Mean error 0.0043 0.0040 Comp A B Mean V mag 16.989 16.864 σmag 0.0045 0.0041 Mean error 0.0037 0.0033 Table 5.11: Mean values of the R and V magnitudes for the star R for the one night in January. Night 2000 January 25 Mean R mag 16.354 σmag (R) 0.0034 Mean V mag 17.078 σmag (V ) 0.0040 2001. B is more or less stable. In the V filter, the components seem to have interchanged magnitude values, although the A component has changed most in this band also. A is ∼0.15 mag dimmer in January 2000, while B is ∼0.06 mag brighter. In Table 5.11 and Figure 5.9 we show our results for the star R on January 25th 2000. 78 Photometry from the ALFOSC data Figure 5.6: Light curve in the R band for the stars H (top), P (middle) and Q (bottom) during the four nights in March 2001. The horizontal dashed line represents the mean of all data points and is meant as a guide to the eye. Some points for star H have been discarded, see text for details. Notice the systematic decrease in the mean level of each night for the stars P and Q, similar to the systematic decrease in the quasar components, see Figure 5.1. The star H shows no such systematic decrease, but a sudden, unexplained ‘drop’ on the last night. 5.2 Photometric results 79 Figure 5.7: Light curve in the V band for the stars H (top), P (middle) and Q (bottom) during the four nights in March 2001. The horizontal dashed line represents the mean of all data points and is meant as a guide to the eye. Some points for star H have been discarded, see text for details. For stars P and Q, a decrease in the mean level from the first to the last night can be seen, comparable to that of the quasar components. 80 Photometry from the ALFOSC data Figure 5.8: Light curves for the quasar components A and B during January 25th–26th 2000, for the R (left panel) and V band (right panel). Notice that the relative strengths of the A and B components are shifted compared to the results from March 2001. Figure 5.9: Light curves for the star R in R and V during January 25th 2000. Chapter 6 Discussion Man: An argument is an intellectual process. Contradiction is just the automatic gainsaying of any statement the other person makes. Arguer: No it isn’t Monty Python, The Argument Clinic In this chapter we discuss in more detail the results presented in Chapter 5, with primary focus on our main data set, from March 2001. We also compare our results to those of Ovaldsen et al. (2003b) and Colley et al. (2003). 6.1 Testing the results In this section, we discuss the significance and possible causes of the variations in our light curves, concentrating on the data in the R band from March 2001. Two different variations seen in the light curves are discussed: The possible ‘U’ shaped variation within each night, and the decrease in brightness from night to night. We performed several tests to see if we could find any connection between these variations and changes in observing conditions, such as seeing and moonlight. 6.1.1 A simple χ2 test The variations seen within each night in the light curve from March 2001 (Figure 5.1) seem to be at least comparable to the estimated errors, see Section 5.2.1. To investigate these variations more quantitatively, we performed a reduced χ2 test for goodness of fit to each night separately. In this way, we wanted to examine whether the nightly ‘U’ shapes are consistent with a horizontal, straight line fit. 82 Discussion The χ2 is a statistical quantity that characterises the dispersion of the observed frequencies from the expected frequencies. The χ2 is defined by χ2 ≡ N X (yi − y(xi ))2 n=0 σi2 , (6.1) where N is the number of data points and σi are the associated errors for each point i. The basis for this test is the hypothesis that the optimum description of a set of data is the one that minimises the weighted sum of the squares of the deviation of the data yi from the fitting function y(xi ). The numerator in Equation 6.1 is a measure of the spread of the observations, while the denominator is a measure of the expected spread. The expectation value for χ2 is hχ2 i = ν = n − nc , where ν is the number of degrees of freedom which is equal to the number of data points n minus the number of parameters describing y(xi ). For our simple model of a horizontal straight line, we have nc = 1. The reduced χ2 is conveniently defined as χ2ν ≡ χ2 /ν, and has expectation value hχ2ν i = 1. Values of χ2ν much larger than 1 may indicate an underestimation of the errors, or it may be the result of real deviations from the assumed distribution. Values significantly below 1 indicates overestimated errors. We calculated the reduced χ2 : χ2ν ≡ N X n=0 (yi − y(xi ))2 1 , N −1 σi2 (6.2) where the normalising factor is equal to the number of data points each night minus one, ν = N − 1. The data points yi are the magnitudes for each night, and y(xi ) = C, where C is constant, is the constant magnitude level that we tried fitting to the data. We calculated χ2ν for many values of the free parameter describing the straight line model. The resulting graphs are shown in Figs. 6.1 and 6.2 for the A and B component, respectively. Looking in statistical tables, e.g. in Bevington and Robinson (2003), we find the probability of exceeding the minimum values of χ2ν given in the figures, for the corresponding values of ν. For example, the first night a minimum value of χ2ν = 0.6696 for ν = 18 degrees of freedom gives a probability between 0.80 and 0.90 of obtaining χ2ν ≥ 0.6696. In other words, the chance of obtaining such a low value of χ2ν by chance, is only between 10 and 20%. We note that the values of χ2ν are almost all well below or very close to 1, with the exception of the bottom left panel in Figure 6.2. Unfortunately, this means that our test is inconclusive with regard to the question of variability within each night. In the example mentioned, the probability that 6.1 Testing the results 83 Figure 6.1: Reduced χ2 for the A component for the four nights in March 2001. Top left: March 14th–15th. Top right: March 15th–16th. Bottom left: March 16th–17th. Bottom right: March 17th–18th. the straight line would give such a good fit to the data, even if it were not the correct model, is between 10 and 20%, which means the model might be considered an unlikely good fit and we can conclude that our errors seem to be overestimated. From Eqs. 5.7 and 5.8, we have that 2 2 2 2 σtot (QSO) = [σtot (star R) − σphoton (star R)] + σphoton (QSO) . (6.3) The photon noise of star R and the quasar components were approximately equal, as was expected since they have very similar flux. From Equation 6.3 we see that in this case the photon noise from star R and from the quasar components will approximately cancel each other. As our errors were dominated by systematics, the most likely cause for the possible overestimation 2 is that σtot (star R) is too big. However, this quantity is measured directly from our results for the star, and we corrected for the two points we had any reason to consider atypical for the assumed constant star, see Section 5.2.1. In the third panel of Figure 6.2 the value of χ2ν corresponds to a probability of 0.02 of exceeding χ2ν = 1.8263. By itself this would seem to indicate a 84 Discussion Figure 6.2: Reduced χ2 for the B component for the four nights in March 2001. Top left: March 14th–15th. Top right: March 15th–16th. Bottom left: March 16th–17th. Bottom right: March 17th–18th. model very inadequate at representing the physical situation, and a very good chance that we have significant variability this night. However, as the rest of the values for χ2ν are so low, the high value in this particular case seems slightly suspicious as the estimated errors for all nights are approximately equal. We note that the highest value of χ2ν in Figure 6.1 is from the same night. 6.1.2 Seeing The seeing of the March 2001 observations varied from roughly 0.700 to 1.500 , with a median value of 0.800 , see Figure 6.3. There are few points with FWHM over 1.200 . We calculated the seeing using the FWHM value of a selection of stars in the field, measured with IRAF. These were rather rough estimates of the FWHM that we entered into psfmof.txt when constructing the PSFs for each image. We measured the seeing on our combined images, which were sums of 10 individual images, each with exposure time scaled to 60 seconds. 6.1 Testing the results 85 Figure 6.3: Changing seeing conditions within each night. To check if any systematic effects were introduced when seeing conditions varied, we have in Figures 6.4 and 6.5 plotted magnitudes versus seeing for all four nights, for both quasar components and the star R. The figures show no marked effects, neither in R nor in V . Fitting a straight line, mag = a ∗ seeing + b to the points, yields practically straight lines, as shown in the figure. Values for a are −2.5 mmag/00 and −6.7 mmag/00 for the A and B components in the R filter, see the top panel in Figure 6.4. Corresponding values in the V filter are a = 7.4 mmag/00 and a = −0.5 mmag/00 . For the star R, the values of the straight line fit to the data are a = 2.2 mmag/00 and a = 3.8 mmag/00 for the R and V filter, respectively. To further investigate any relationships between magnitude and seeing, we have in Figure 6.6 plotted the R magnitude of the quasar components versus seeing, for each of the four nights individually. We fitted straight lines to the data points, and obtained values for the gradient a ranging from −11.0 mmag/00 to 3.0 mmag/00 for the A component, and −13.1 mmag/00 to 1.9 mmag/00 for the B component. We find no evidence of any general 86 Discussion Figure 6.4: Top: R magnitude for the A (×) and B () quasar components versus seeing (FWHM) for all four nights in March 2001. Bottom: V magnitude versus seeing. A straight line is fitted to the data. Figure 6.5: Top: R magnitude for the R star versus seeing (FWHM). Bottom: V magnitude versus seeing. A straight line is fitted to the data. 6.1 Testing the results 87 Figure 6.6: R magnitude for component A (top four panels) and component B (bottom four panels), versus seeing, for each night individually. A straight line is fitted to the data. We see no general trend of changing magnitude with seeing. tendency for the magnitude of either quasar component to vary with seeing within each night. Although seeing generally increases towards morning, see Figure 6.3, it does not seem that this could have caused the observed increase in brightness at the end of each night. 88 Discussion 6.1.3 Possible influence of the Moon Systematic effects introduced by insufficient removal of the background can depend upon the colour of the object. The moon was present during our observations in March 2001. We wanted to investigate whether the presence of the moon had any effect on our results. Table 6.1 shows the time of Moon rise, how much of the Moon that was illuminated each night seen from Earth and the altitude of the Moon at 6 a.m., the approximate maximum altitude the Moon reached before morning twilight. Table 6.1: Presence of the Moon during the March 2001 observations. Night March 14–15 March 15–16 March 16–17 March 17–18 Moon rise (UT) 00:22 01:18 02:13 03:04 % illumination 67.5 57.8 48.0 38.6 Alt. at 6 a. m. (degrees) 42 38 33 27 When the Moon is present, the background level is higher, especially at blue wavelengths, than when the Moon is absent. In Figure 6.7 we have plotted the background level for each night, against time. As mentioned in Section 3.4.2, the background level of each image was subtracted before adding the images in stacks of 10, each individual image being scaled to the same exposure time. By adding the levels that we originally subtracted, for each stack, we obtained the background levels of the combined images. The images that were not stacks of 10, see Section 3.4.2, were corrected for, by taking the mean of the background levels in the stack and multiplying by 10. The vertical dotted line in Figure 6.7 represents the time (UT) of Moon rise each night. The rise in sky level immediately after Moon rise is easily seen. Note that the range on the y axis is reduced for the bottom panels. On the first night, the Moon rises to a maximum altitude of 42 degrees and the sky level is over 5000 ADU. On the last night, the moon only rises to 27 degrees above the horizon, producing a sky level of approximately 3000 ADU. Comparing this to around 3 a.m. on the first night, when the Moon was also 27 degrees above the horizon, we see that the background level again has a value of ∼3000 ADU. To determine whether the moonlight had any effect on our magnitudes, we constructed a scatter plot showing R and V magnitude versus background (sky) level for the March 2001 data, see Figure 6.8. 6.1 Testing the results 89 Figure 6.7: Sky level variation for the four nights in March 2001. The dotted line represents Moon rise. The Earth’s atmosphere emits photons by fluorescence, a process sometimes referred to as airglow, see Léna (1988). Daily photochemical dissociation causes electrons to recombine with ions which produces fluorescence photons. The main sources of this line emission are OH molecules, oxygen (O), nitrogen (N) and sodium (Na). After sunset, there is still some emission in the atmosphere, especially at red wavelengths. This emission probably affects our data in a different way than the moonlight. We therefore removed points from images in R taken less than 2 hours after astronomical evening twilight, from Figure 6.8. The last data point from the night March 14th– 15th was also removed, as the background level in this image was noticeably higher than for all other points (level at 22400 ADU). This extremely high background probably originates from morning twilight rather than the Moon. In Figure 6.8, we have plotted magnitude (both R and V ) against background level for the two quasar components and for the reference star R. A straight line was fitted to the points, which is also shown in the figure. As the V plots contain quite few points, we will mainly discuss the results in the 90 Discussion Figure 6.8: Scatter plots showing the relationship between magnitude and background level. Top: R magnitude for the A (left) and B (right) component. A marked decrease in magnitude, i.e. increasing brightness, with increasing sky level is illustrated by the fitted line. Middle: V magnitude for the A (left) and B (right) component. The fitted line shows increasing magnitude with increasing sky level, but there are very few points. Bottom: R (left) and V (right) magnitude for the R star. The straight line fitted to the R plot shows a decreasing magnitude trend with increasing sky level, but the gradient is clearly smaller than for the quasar images. Again, the V plot contains too few points to conclude much. R band. We merely remark that, in V , the fitted line shows a decrease in brightness with increasing sky level, for the quasar components, and the opposite situation for the star R. However, we see from the figure that only two points contribute to the ‘lowering’ of the fitted line, so we cannot conclude whether this is coincidental or a real trend. In the R band, we see a clear difference in the gradients of the fitted lines for the quasar components and the star, although the fits have considerable spread. For the A component, the gradient of the fitted line is 2.2 times larger than the gradient of the line fitted to star R. For component B, the gradient is 3.1 times larger than that of the star. 6.1 Testing the results 91 The increase in background level apparently influences the quasar components in a different way than it does the star. This may come from the fact that the star is redder than the quasar. Assuming the Moon is causing the trends, we can say that the moonlight affects the quasar and the star very differently, possibly because of their different colours. It might also be that insufficient galaxy modelling is causing this effect for the quasar components. However, as the B component is much closer to the galaxy than A, one would in that case perhaps expect to find a more marked difference in the two top panels in Figure 6.8. It would be very interesting to see what tendencies we could find in V , from both quasar and star, with more images in this band. A future study of this effect should include better temporal sampling in V so as to obtain more data points for this filter. Figure 6.9: A model light curve showing the distribution of points with the relation between magnitude and background level shown in Figure 6.8 removed. Symbols used: × and for the A and B component, respectively, for the points not included in the scatter plots in Figure 6.8. The symbols + and are used for the modelled points for the A and B component, respectively. In order to better see how the light curves in Figure 5.1 would have looked 92 Discussion without the connection between background level and magnitude we have just shown, we constructed a model light curve for the quasar components. This is shown in Figure 6.9. The data points originating from the images that were removed from the moonlight analysis, as they were obtained less than two hours after astronomical twilight, are plotted with the same symbol conventions for the A and B components used in Chapter 5. The new points that have been modelled, have corresponding symbols rotated 45 degrees (i.e. a plus sign for the A component and a square for the B component). To construct the model curve, we subtracted the effect of the straight line fit from the original magnitudes, using the following equation: mi = moriginal − [f (xi ) − f (xfix )] . Here, mi is the magnitude of the model curve and moriginal is the actually observed magnitude. The straight line we fitted is f (x). We assume this is the correct relation between magnitude and background level, and calculate f (xi ), the magnitude corresponding to background xi from the equation of the straight line, and f (xfix ), which is the magnitude corresponding to the lowest background level in Figure 6.8. We thus obtain the magnitude mi that we would have had, if the Moon had not been present, assuming that the Moon is causing the increase in background level and that the straight line fit is the correct relation between increasing background level and increasing magnitude. Figure 6.9 also shows that the straight line fit is not very exact. We see from Figure 6.8 that there is considerable spread in the points. The drop we see in brightness at the end of the first night, is an example of the mismatch of the model. It corresponds to the fact that there are a few points in Figure 5.1 on this night that actually decrease in brightness towards the end of the night. The residuals from the deconvolution for these images were not particularly large compared to other images. However, disregarding these last points on the first night, we see that we have with some success ‘straightened’ the curve, at least for the first two nights. The last two nights are a bit messier, but we recall from Figure 6.7 that the background levels here are lower as the Moon rises later in the night. The conclusion seems to be that the ‘U’ shape each night can be a result of background effects caused by the Moon. However, we can see in Figure 6.9 that our model is not in perfect agreement with the data. There might be other aspects, e.g. atmospheric changes, which introduces other effects that mix with that of the moonlight to form a more complex situation than we have modelled. 6.1 Testing the results 6.1.4 93 Colour diagrams We wanted to study whether the variations we see in quasar brightness from March 2001 correlate at all with changes in colour. This could indicate that the observing conditions changed. We were also interested in studying the colours of the reference stars, and compare this with the light curves presented in Section 5.2.2. As our temporal sampling in V was not good enough, we did not investigate variations with colour within each night. Only the possible colour dependency of the decline in brightness from night to night was studied. We calculated V − R for the four nights in March 2001, using the average magnitude of each night for each filter. We checked that the results of calculating the mean directly from the magnitudes and via the flux, gave identical results to the given accuracy. The colour indices of the quasar components and the star R are plotted in Figure 6.10. Our symbol conventions still holds: The A component is represented by a cross (×), the B component by a diamond shape () and the R star by a triangle (M). Figure 6.10: V − R plotted against time for the quasar components (top panel) and the star R (bottom panel). The A component is represented by a cross (×) and the B component by a diamond shape (). The star R is represented by a triangle (M). 94 Discussion When subtracting two magnitudes, the errors increase. The error bars in Figure 6.10 were calculated in the following manner: q σV −R = σV2 + σR2 , where σ is the standard deviation of the magnitudes within each night. For all four nights, the mean value of V − R is 0.407 ± 0.010 for the A component and 0.393 ± 0.012 for B. Differences in colour between the A and B component could indicate differential extinction along the light paths, or possibly microlensing. Extinction should influence the light from B most, because light from this component travels through the lensing galaxy. Scattering by material in the galaxy affects blue wavelengths most,1 so we would expect this to cause the light from B to be redder than the light from A. We find the opposite: A is slightly redder than B. However, the difference in V and R magnitudes is very small. The V −R values for the A component range from 0.405±0.010 to 0.410± 0.010 for the four nights. For the B component, the range is 0.392 ± 0.005 to 0.400 ± 0.010. In other words, the changes we see of colour versus time are within the error bars. The ‘dip’ from the first to the second night in Figure 6.10, constitutes 0.005 magnitudes for (V − R)A and 0.008 for (V − R)B . The corresponding dip for the R star is (V − R)R = 0.004. This suggests a possible, observational influence that produces the same colour change, towards the blue, in all three objects between the two nights. Changes in colour might be due to atmospheric conditions, perhaps in combination with moonlight. However, the variations in colour are not significant compared to the error bars. We also note that we do not find any correlating change in colour index with the stepwise decrease in brightness from night to night. As the light curves for the stars P and Q show a similar systematic decrease from night to night as that of the quasar components, we wanted to examine the colours of these stars. The field stars R and H were known to be redder than the quasar components, but we did not have any reference magnitudes for P and Q. If the colour of the objects influence the variations from night to night, we might expect the stars P and Q to be bluer than R and H, more similar in colour, perhaps, to the quasar components. In Table 6.2, we show the mean V − R value, for all four nights, of the quasar components and all field stars we have studied. Our results show that the stars P and Q were in fact significantly redder than the stars R and H. We therefore did not find any evidence that the 1 The amount of scattering that occurs to a beam of light is dependent upon the size of the particles and the wavelength of the light. Blue light has a shorter wavelength than red light, and is therefore scattered more. 6.2 Comparing the time shifted data 95 Table 6.2: Mean values of the quasar components A and B, and the field stars R, H, P and Q, for the four nights in March 2001. The standard deviation of the V − R values are also given, computed in the same manner as the error bars in Figure 6.10. Object V −R σV −R A 0.407 0.010 B 0.393 0.012 R 0.705 0.008 H 0.748 0.016 P 0.870 0.013 Q 1.024 0.015 decrease in brightness from night to night is correlated with the colour of the object. Pondering as to what else could cause P and Q to behave so similar to the quasar components, we wondered whether the decreasing tendency has something to do with the brightness of the object. The A component is approximately four times brighter in flux than P and five times brighter than Q. The star H, on the other hand, is almost ten times brighter than A. Conceivably, the nightly decrease is seen only in weak objects. However, the star R is only slightly brighter than A, so in that case we would expect to see the decreasing tendency in this star also, which we do not. In Figure 6.11, we show a colour magnitude diagram of V −R versus R for the quasar components and the reference star R. The groups of four points are results from the four nights in March 2001, while the isolated points are from the one night in January 2000. 6.2 Comparing the time shifted data Gravitational lenses are achromatic. The angle that light rays are bent by the intervening mass is not dependent on wavelength. From this it follows that the magnitude difference between the images of a gravitationally lensed quasar should be the same in different bands, at the same point in the quasar’s emission history. As our data sets (March 2001 and January 2000) are separated in time by the approximate time delay, the time shifted difference in magnitude between the two components of QSO 0SO 957+561 provides the optical continuum magnification ratio. A is the first arriving component, followed by B a number of days equal to the time delay later. We therefore calculate the magnitude difference between the components, using the magnitude of A from the chronologically first observing session, January 2000, and the magnitude of B from March 2001. We calculate the 96 Discussion Figure 6.11: V − R plotted against R magnitude for the quasar components (top panel) and the star R (bottom panel). The symbol conventions for the QSO components and the star are the same as in Figure 6.10. The groups of four points close together are data from March 2001, while the isolated points are from January 2000. magnitude difference in both R and V : mB, 2001 (R) − mA, 2000 (R) mB, 2001 (V ) − mA, 2000 (V ) In these calculations, we have used the mean level of all four nights for the 2001 data, although we observed a decrease in brightness from night to night of 0.014 and 0.017 magnitudes for A and B, respectively. For the 2000 data we used the mean value of the one available night. Converting these magnitude 6.2 Comparing the time shifted data 97 differences into B/A flux magnification ratios, we get 1.096 ± 0.022 in R and 1.061 ± 0.024 in V . To calculate the errors for the magnification ratio, we inverted and differentiated Equation 5.1. As mentioned in Section 2.2.2, a discrepancy between the magnifications in R and V can be caused by either differential absorption along the different light paths, or by microlensing. However, the difference between the magnification ratios is within the error estimations, so it is not very significant. We also calculate the colour indices of A and B and find: mA, 2000 (V ) − mA, mB, 2001 (V ) − mB, 2000 (R) = 0.357 ± 0.006 2001 (R) = 0.393 ± 0.012 where the errors were calculated in the same manner as in Section 6.1.4. We see that, for the time shifted data, B is slightly redder than A. The lensing galaxy is significantly redder than the QSO. Therefore, one would expect to find that B is redder than A if no galaxy correction has been done. If the galaxy was perfectly corrected for, one might still expect B to be redder than A, as the blue light has probably been scattered more as described in Section 6.1.4. Compared to our V − R calculations in Section 6.1.4, the situation has changed. Without time shifting, we found A slightly redder than B, although within the errors. We do not know of any reason why the quasar itself should change colour. It is probable that something else has caused the change. Microlensing by compact objects in the lensing galaxy or galaxy cluster, can cause the quasar components to become brighter, or weaker. We see from Figure 6.11 that component A becomes brighter in R by ∼0.2 mag from January 2000 to March 2001. Component B has approximately the same magnitude in March 2001 as in January 2000, it brightens by only ∼0.02 mag. If the changes in brightness were the result of microlensing, we would expect that as the quasar brightens, it would also change in colour towards the blue. This is because microlensing would mainly amplify the smaller, central parts of the quasar, which are also the bluest. From Figure 6.11 and Table 6.2, we find that V − R for the A component in January 2000 was 0.357 ± 0.006, while in March 2001 the value was 0.407 ± 0.010. In other words, as the A component brightens, it also becomes more red. This makes it unlikely that the change in brightness of component A is the result of microlensing, and more probable that this is an intrinsic brightness variation in the quasar. As the brightness change for the A and B component are so dissimilar, it is probably not a result of different observing conditions in January and March. We do not have any explanation for the change in colour, aside from speculations involving quasar models with radiating ‘spots’. 98 6.3 Discussion Zero lag correlation Because of the time delay, the light curves for the two quasar images are in theory uncorrelated, and should exhibit independent behaviour. The appearance of correlation without correcting for the time delay, is therefore noted with some apprehension. There have been several reports on a ‘zero lag correlation’ between the A and B images, for time-scales of weeks and months (Kundic et al., 1995; Ovaldsen, 2002; Ovaldsen et al., 2003a) as well as for shorter time-scales (Ovaldsen et al., 2003b; Colley et al., 2003). In Ovaldsen et al. (2003b), the authors find a small brightness decrease in the mean magnitude level from night to night, using the same ALFOSC data set that we use in this thesis. Their decrease is similar to what we have found, although our decrease is about twice as large in terms of magnitudes. From March 14th to March 18th, Ovaldsen et al. found a decrease of 0.005 R mag for A and 0.008 mag for B. For the V filter, their decreases are 0.006 mag for both A and B. This small decrease is suggested by the authors as a possible manifestation of the correlation at zero lag. Colley et al. (2003), in addition to the same ALFOSC data, use data from 11 other observatories. From their results, ranging from March 12th to March 21st, they report a surprising zero lag agreement of the curves of A and B. They interpret this as evidence of an as yet uncorrected photometric problem at the few millimagnitude level. Referring to Figure 5.1 and Figure 5.2, we see that the nightly variations in both components show a tendency to correlate at zero lag. For example, the magnitudes increases from around midnight towards the morning. As it has been argued for by e.g. the moonlight test, this is probably caused by some or more observational and/or photometric effects. Our discussion in Section 6.1.3 has shown that the presence of moonlight is a possible explanation for the small increase in brightness at the end of each night. Only five nights of optical monitoring does not suffice to draw any firm conclusions about this effect. However, the correlation between the components at zero lag within each night, as well as its possible connection with moonlight, constitutes an indication that our variations within each night may be caused by observational effects rather than intrinsic variations in the quasar. The zero lag correlation within each night can also be seen in the V band, see Figures 5.3 and 5.4. The decrease in brightness from night to night could also be a variety of this unexplained correlation at zero lag, as suggested by Ovaldsen et al. (2003b). We also tentatively note, that if our model of quasar magnitude versus sky level is correct, it could be possible that the variation from night 6.4 Comparisons with other results 99 to night really is reflected in the nightly observations. If so, that could mean either i) evidence of zero lag correlation on a time scale of the four observation nights, or ii) both quasar components happen to exhibit very similar decreases in flux during these four nights. A more sophisticated model than our straight line fit should probably be constructed before making any definite claims. 6.4 Comparisons with other results In this section, we directly compare some of our results to those of Ovaldsen et al. (2003b), and to results kindly provided by Wesley Colley (private communication). The latter constitute the NOT part of the official QuOC results (Colley et al., 2002, 2003). Tables 6.3 and 6.4 show results from the three groups in the R band for the quasar components. Our results, presented in the second column in the table, are the same as in Tables 5.3 and 5.4. We only compare our results in R, as the QuOC team has not published any results using the V images. Table 6.3: Our mean values of the R magnitudes for the A component of QSO 0957+561 for the four nights, the results of Ovaldsen et al. (2003b) and results from the data provided by Wesley Colley for the same nights. Numbers in parenthesis are the standard deviation of the magnitudes in mmag. 2001 2001 2001 2001 Night March 14–15 March 15–16 March 16–17 March 17–18 Mean R mag 16.422 (2.6) 16.428 (2.3) 16.432 (3.1) 16.436 (3.2) Ovaldsen 16.394 (5.7) 16.397 (4.4) 16.399 (4.0) 16.399 (5.1) Colley 16.374 (7.6) 16.377 (4.8) 16.374 (5.0) 16.378 (5.9) A note on the calibrations: As mentioned in Section 5.1.1, we have used the exact same reference magnitudes as those of Ovaldsen et al., see Tables 5.1 and 5.2. The latter authors use a mean of several reference stars for the calibration, whereas we use one star (D). The reference magnitudes were supplied, to two decimal points, by Rudy Schild, but have been slightly modified by the bootstrapping applied by Ovaldsen et al. Of the stars supplied by Schild, only H differ at all from the reference magnitudes in Table 5.1, by ∼0.01 mag. We therefore note that there is a possibility that a small systematic difference between our own results and those of Colley et al. has been 100 Discussion Table 6.4: Our mean values of the R magnitudes for the B component of QSO 0957+561 for the four nights, the results of Ovaldsen et al. (2003b) and results from the data provided by Wesley Colley for the same nights. Numbers in parenthesis are the standard deviation of the magnitudes in mmag. 2001 2001 2001 2001 Night March 14–15 March 15–16 March 16–17 March 17–18 Mean R mag 16.523 (4.1) 16.529 (3.3) 16.534 (4.6) 16.540 (3.3) Ovaldsen 16.486 (6.2) 16.491 (4.0) 16.495 (4.4) 16.494 (5.7) Colley 16.296 (5.9) 16.299 (4.2) 16.297 (4.5) 16.300 (5.5) introduced by the difference in one of the reference magnitudes we believe was used by these authors. 6.4.1 Results from Ovaldsen et. al. Ovaldsen et al. (2003b) discuss the detailed variations in brightness for QSO 0957+561, employing the newly developed aperture photometry scheme from Ovaldsen (2002). The method involves subtracting a galaxy model from each image frame. Then, one of the quasar components is ‘cleaned’ by subtracting a PSF, followed by aperture photometry of the other component, and vice versa. The PSF subtraction is done in order to correct for crosscontamination between the quasar components. In Figure 6.12 we show their results for the A and B component, for R in 2001. In order to facilitate comparisons, this figure is presented in the same format as our light curves in Section 5.2. The figure shows all 673 individual data points, without binning. Error bars are omitted for clarity. The mean value of individual formal error bars for A is 5.6 mmag, and for B 6.2 mmag. These were typical errors. Mean values of magnitudes from each night as well as the standard deviation, are shown in Tables 6.3 and 6.4. We have several findings that agree with those of Ovaldsen et al. This is interesting and reassuring, as these authors have used the exact same ALFOSC data sets as we have done, but a very different method. Both groups find that, for March 2001, A is brighter than B in both R and V by approximately 0.1 mag. We also find the opposite situation, B brighter than A, in January 2000, both groups finding that B is brighter than A by ∼0.07 mag in R and ∼0.12 mag in V . Table 6.5 shows a comparison between the magnitude differences of the A and B components for the two observing 6.4 Comparisons with other results 101 sessions. Another similarity between our results is the systematic night–to–night decrease in brightness, seen in March 2001 in both quasar components, see Figure 6.12 and Tables 6.3 and 6.4. The overall decrease in brightness, from the first to the last night, is 0.005 mag for A and 0.008 mag for B. Compared to our own total decreases of 0.014 mag for A and 0.017 mag for B, we see that the trend is the same but our decrease is considerably more than that of Ovaldsen et al. It is interesting that both methods yield a systematic decrease. Table 6.5: Differences in magnitudes between the two QSO components in the R and V filters, for our results and the results of Ovaldsen et al. The mean magnitudes of each night of component A have been subtracted from the mean of each night of component B. Values from our results, subscripted EV, are shown in the second and third columns for R and V respectively, while the results from Ovaldsen et al. (2003b), with subscripts OV, are displayed in the fourth and fifth columns, also for R and V . Night 2000 January 25–26 2001 March 14–15 2001 March 15–16 2001 March 16–17 2001 March 17–18 (R)EV –0.077 0.101 0.101 0.102 0.104 (V )EV –0.125 0.088 0.085 0.088 0.089 (R)OV –0.069 0.092 0.094 0.096 0.095 (V )OV –0.115 0.088 0.087 0.087 0.088 From Tables 6.3 and 6.4 we see that the magnitudes in Ovaldsen et al. are between 0.03–0.04 mag brighter than our results. Table 6.5 shows that the magnitude difference between the two quasar components remain stable for the four nights, both for our results and those in Ovaldsen et al. The differences in magnitudes between the components are also very similar for the two methods, compare columns two and four, and columns three and five, in Table 6.5. For the star R, there is a difference in brightness between our results and Ovaldsen et al. of ∼0.06 mag. In this case also, Ovaldsen et al. have the brightest values, so this seems to be a systematic difference introduced by the two photometric reduction methods. We find a discrepancy in the colour indices for A and B between the two groups. For the time shifted results, Ovaldsen et al. find that A is redder than B, while we have the opposite: B is redder than A. V − R magnitudes for A and B from Ovaldsen et al., are 0.357 ± 0.007 and 0.318 ± 0.009, 102 Discussion Figure 6.12: Light curve from the results of Ovaldsen et al. (2003b) for the quasar components A and B in R during the four nights in March 2001. The horizontal dashed line represents the mean of all data points and is meant as a guide to the eye. Notice the systematic decrease in brightness from night to night, similar to our results, although less pronounced. respectively. In Section 6.2 we stated our V − R values for A and B: 0.357 ± 0.006 and 0.393 ± 0.012. In other words, the colour indices for component A are practically identical for us and Ovaldsen et al., while for B they differ by ∼0.08 mag, our results showing a higher V − R value, thus a B component more red in colour. Since the values for A are so similar, it is probably the different galaxy corrections that makes our results differ, as the galaxy affects the B component the most. Ovaldsen et al. subtract a galaxy model, while we produce a numerical model of the galaxy through deconvolution. 6.4.2 Results from Colley et. al. In the paper reporting the results from the QuOC campaign’s second session, Colley et al. (2003) mention the data from the NOT, i.e. our main data set in this thesis, as being part of the ‘back-bone’ data, providing many frames 6.4 Comparisons with other results 103 with excellent statistics. On request, Wesley Colley kindly provided us with results from their reduction of the NOT data in the R band. Colley et al. applied aperture photometry with a single correction for aperture cross-talk, i.e. the contribution of light from component A to the B aperture and vice versa, and galaxy contamination, as a function of seeing. We have plotted the results of Colley et al. in Figure 6.13, choosing to plot the 743 individual data points as this is how we showed the results of Ovaldsen et al. in Section 6.4.1. In Colley et al., hourly means of the magnitudes are used. As in Figure 6.12, error bars are omitted for clarity. Mean of individual errors in magnitude, provided by Wesley Colley, is 4.1 mmag for the A component and 3.9 mmag for B. We note that the number of data points are 70 more than those of Ovaldsen et al., probably due to different rejection criteria. Figure 6.13: Light curves for the quasar components A and B in R during the four nights in March 2001, from the results supplied to us by Wesley Colley. The horizontal dashed line represents the mean of all data points and is meant as a guide to the eye. Notice there is no systematic decrease in magnitude from night to night, and that the B component is stronger than A. Comparing the brightness ratio between A and B in Figure 6.13, we 104 Discussion see that, in this case, we have the opposite situation of that in our own results and the results from Ovaldsen et al. The B component – with a mean magnitude for all four nights of 16.298 mag – is here brighter than A, which has a mean of 16.375 mag. This discrepancy is due to the fact that the method used by Colley et al. involves a correction for the galaxy as a function of seeing, but not a galaxy subtraction. As the galaxy contributes ∼18% to the aperture photometry of component B (Rudy Schild, private communication), the result is that B appears brighter than A. Converting the uncorrected B magnitudes into fluxes, subtracting 18% and then averaging over all points, yields a mean value of mB, corrected = 16.513 mag, a value in good agreement with our own B magnitude. With this value for the corrected B magnitude, the internal ratio between the components A and B also agree with our results. The difference in magnitude between the two components for the QuOC results in March 2001, become mB, corrected − mA = 0.138 mag. This value is a little higher than the value we have obtained, see Table 6.5. However, the internal brightness ratio between A and B are now approximately the same for all three groups, with A the brightest component. For January 2000, the mean magnitudes from the QuOC results, without subtracting the galaxy contribution, are 16.46 mag for the A component and 16.29 mag for B (supplied by Rudy Schild to two decimal points, private communication). That is, B is very similar to the (uncorrected) March 2001 value. The A component is brighter in 2001 by ∼0.1 mag, while B is less bright in 2001 by merely ∼0.01 mag. In fact, all three groups find that the A component has changed much more than B during the ∼416 days between the two observing sessions. As mentioned in Section 6.2, taking the mean of all four nights for the results in March 2001, we find that the A component is ∼0.2 mag brighter in 2001 than in 2000, while B is only ∼0.02 mag brighter in 2001, see Tables 5.3, 5.4 and 5.10. For Ovaldsen et al. (2003b), A brightens by ∼0.1 mag from 2000 to 2001, while B has become ∼0.04 mag less bright. Assuming the contribution of the galaxy to the B aperture in January 2000 is still 18% in the results supplied by Schild, we subtracted the galaxy contribution and note that, for January 2001, mB, corrected − mA = 0.05. In other words, the A and B components are closer in magnitude than in 2001, but A is still brighter than B. Both we and Ovaldsen et al. found B to be brighter than A in January 2000, see Section 6.4.1. We also note from Figure 6.13 that there is no evidence of the systematic decrease from night to night comparable to that observed by us, see Section 5.2, and to a somewhat lesser extent by Ovaldsen et al., see Section 6.4.1. Although there is a difference in mean magnitude of ∼0.04 mag between the first and last nights in both components, in Figure 6.13, the mean of each night shows a decrease in brightness between the first and second nights, 6.4 Comparisons with other results 105 then an increase from the second to the third night, followed by another decrease from the third to the fourth night. This pattern is discernible for both components, see Tables 6.3 and 6.4. 6.4.3 Summary In general, our results support several of the findings of Ovaldsen et al. (2003b), as well as those of Colley et al. (2003) after subtracting the galaxy contamination. As regards the decrease in brightness from night to night, our results agree with the findings of Ovaldsen et al. (2003b). It also supports these authors’ values for the brightness ratio of the quasar images for the March 2001 data: A is stronger than B with approximately 0.1 mag in both R and V . For the R band, this is also the result we found in the QuOC data after subtracting the galaxy influence in the B aperture. Furthermore, our analysis also supports the shift in relative brightness reported by Ovaldsen et al. The night–to–night decrease in magnitudes seen in the observations from March 2001 are puzzling, because the decrease is not reflected in the variations internal to each night and because the decrease is so similar in both components (although we did find a decrease that was 0.003 mag larger in the B component). We have shown that moonlight may be causing the ‘U’ shape in each night. There is thus a possibility that the night to night decrease is evidence of intrinsic behaviour, and the fact that we do not see this decrease during the night could be due to some observational conditions. For example, moonlight and changing atmospheric conditions could prevent us from observing it. However, the fact that the decrease from night to night is so similar in both components, suggests the possibility that it is an observational effect. It remains a puzzle why no such systematic decrease in brightness from night to night is observed in the results from Colley et al. (2003). The trend of increase/decrease in brightness from one night to next are the same in both components for Colley et al., which could indicate that the night–to– night changes are manifestations of the correlation at zero lag. However, we note that our results and those of Ovaldsen et al. both show the systematic decrease from night to night, and that our decreases are larger in magnitude and more significant compared to the error bars. We have also, in Section 5.2.2, found the same trend for the reference stars P and Q, while stars R and H did not show this trend. As regards the time shifted colour indices, our results show that component B is redder than A, while Ovaldsen et al. have the opposite result. This is probably due to differences the galaxy corrections. 106 Discussion Chapter 7 Reduction and photometry of the StanCam data The most exciting phrase to hear in science, the one that heralds new discoveries, is not ‘Eureka!’ (I found it!) but ‘That’s funny ...’ Isaac Asimov The Gravitational Lenses International Time Project (GLITP), is a cooperative venture between lensing communities in several European countries as well as the Harvard–Smithsonian Center for Astrophysics, the Anglo Australian Observatory and the Canada–France–Hawaii Telescope Corporation. Members monitor and analyse gravitationally lensed objects. In the field of optical monitoring, the group has focused particularly on the systems QSO 0957+561 and QSO 2237+0305, the so-called Einstein Cross. Ullán et al. (2003) discuss the daily, weekly and monthly variability of QSO 0957 based on GLITP observations from the instrument StanCam at the NOT from February 3rd to March 31st 2000, in the R and V bands. On request, Luis Goicoechea and Aurora Ullàn Nieto kindly provided us with this data set. In this chapter, we present our reduction of their R images, as an example of using the MCS procedure on a very different data set than our main observations. As we reduced the StanCam data in a similar way as that described previously, we will here mainly mention the differences between the two reduction processes. 7.1 Pre-processing The StanCam CCD is 1024 × 1024 pixels, with a gain of 1.68 e− /ADU and a read-out noise of 6.5 e− /pixel. An overscan section of 50 columns is added to each side of the frames. 108 Reduction and photometry of the StanCam data There were a total of 75 R images in the data set, but we discarded one image due to a bad line of ∼80% lower pixel value crossing one of the quasar components. As a rule, there were two R images from each observing night, although this number varied between one and four. The images had been obtained approximately every second or third night through February and March 2000. Exposure times were 150 seconds for most of the images. A total of 17 images had exposure times of 60, 100, 180 or 300 seconds. Seeing conditions varied from approximately 0.500 to 2.800 , measured by the FWHM of point sources. The field did not show all the stars present in our ALFOSC images, see Figure 2.1. However, most of the StanCam images included stars H, E, D, P and R. As explained in Chapter 3 and Appendix A, we have used IRAF for the flat-fielding of the object frames. Most of the bias frames had a lot of structure. Many of them were filled with a wide variety of ‘stripes’ of different widths, with somewhat different ADU counts from the rest of the frames. As there was an overscan section on all frames, we decided in this case to discard the bias frames and subtract only the overscan level. Using the interactive overscan fitting in IRAF on a selection of the images, we determined an appropriate overscan section that did not exhibit any gradient. There were sky flats on 2 out of 39 nights, so we were forced to use mainly dome flats. The doughnuts on the dome flats were uneven, as we also found for the ALFOSC flat frames in Chapter 3. The rotator angle had been shifted 90 degrees at irregular intervals when obtaining the dome flats, which resulted in a rotation of the irregular doughnuts on the frames. Assuming that this rotation was done in order to correct for the uneven doughnuts, we tried combining dome flats with different rotation angles. This resulted in more even doughnuts, so we decided to combine dome flats with different rotation angles. When executing the task flatcombine in IRAF, we used the average combination choice with rejection method avsigclip, see Appendix A. Whenever possible, we constructed one master flat for each filter each night. However, not all nights contained flats, so we were sometimes forced to use dome flats from the night immediately before or after. Examining all flat frames carefully, we noted spurious artifacts occurring at irregular intervals: bright or dark columns, or even broad, black stripes down the middle of the frames. We discarded all flats containing these artifacts. After flat-fielding the images, we subtracted the sky background in each image, taking the median of each image as a measure of the background level. 7.2 Running MCS 7.2 109 Running MCS Using the same techniques as explained in Chapter 3 and Appendix A, we aligned all images to the same coordinates as one reference image. This was done in order to run the simultaneous deconvolution. The PSF star and the normalisation star were chosen to be E and D, respectively, the same as in our previous analysis. The residuals from the Moffat fit were similar to those from the ALFOSC data, i.e. quite extensive residuals, so we again utilised a wide mask of ∼60 subsampled pixels radius. When running the program that constructs the analytical Moffat adjustment to the PSF star, we again experienced that the program was sensitive to the β value in the parameter input text file psfmof.txt, as mentioned in Section 4.6.2. However, it was not easy to distinguish a direct relationship between the input and the final β value that would make the program run smoothly. For example, the program would crash when given a starting β value of 3, but would run if β was changed to 1.3, producing a final β value of 2.7. We did not experience any problems with modelling the residuals from the Moffat fit, as mentioned for the ALFOSC images in Section 4.6.2. 7.3 Results Light curves for the quasar components A and B, with error bars, are shown in Figure 7.1. As in our previous work, we used the normalisation star D for calibrating the magnitudes. The error bars shown in the figure were computed as described in Section 5.2.1. The star R was present on only 37 out of 74 frames. When computing the error bars for the frames where R was missing, we ‘simulated’ star R using the quasar component A. Taking the average of the fluxes of R on the frames where the star was present, we found that star R was 1.28 times brighter than component A. Star R was then simulated on the frames where it was missing, by multiplying the flux of A by 1.28. Two points were discarded when computing the standard deviation of R for the error estimations. The flux of star R on these images varied by more than 25% from the flux on the other images. As the star is assumed constant, we did not consider these two frames to show a correct value for the star. When calculating error bars for the quasar components on these same frames, the standard deviation of all R fluxes was used, producing very large error bars for the two frames. In fact, the error bars were too large to be included in Figure 7.1. The two points, one from February 4th and the 110 Reduction and photometry of the StanCam data other from March 2nd, are plotted without error bars in the figure. The standard deviation of star R, with the two deviant points removed and the simulated stars included, was 25% larger for the StanCam data than for the ALFOSC data. The stacked ALFOSC data were sums of 10 individual images, each with exposure time ∼60 seconds, so they had longer ‘combined’ exposure times than the StanCam images. This has probably resulted in the rather sizeable error bars for the A and B components for the StanCam results, typically ∼0.03 mag, approximately 10 times larger in mag than the error bars for the quasar components in Chapter 5. Noting that the start of this observing run is only a few days after our one night of ALFOSC observations in January 2000, we compare the magnitudes of A and B obtained here with our previous results. Our mean R values for January 25th were 16.632 for A and 16.555 for B, see Table 5.10. Corresponding mean A and B magnitudes from Figure 7.1, are 16.582±0.035 mag and 16.515 ± 0.025 mag, where the errors are the standard deviation of the magnitudes of all points. Our result from January 25th are in reasonable good agreement with these values, although not within the errors expressed as the standard deviation of the mean. Our A and B magnitudes from January are ∼0.05 mag dimmer than the mean of the StanCam data points in Figure 7.1. They also agree that B is brighter than A by approximately ∼0.07 mag. Ullán et al. (2003) obtain their final light curves using the psfphot technique (e.g. McLeod et al., 1998). The method uses PSF fitting to extract the fluxes, and assumes a de Vaucouleurs profile,1 convolved with a PSF, as the lens galaxy model. Our light curves look qualitatively similar to the results of Ullán et al. Over the entire observation period, component A seems to increase somewhat in brightness, by ∼0.04 mag, while there is a possible decrease in component B of ∼0.02 mag. Ullán et al. concluded they did not find evidence of any daily variability of the quasar. The daily variations were very similar in both quasar components, and they conclude that these correlated variations are due to some ‘observational noise’. From Figure 7.1, we see that our results also show evidence of zero lag correlation between the two components, especially in the first part of the period. However, as Ullán et al. also find, there seems to be uncorrelated gradients in both A and B with respect to time scales of several weeks. As Ullán et al. performed a straight line fit to their data points, we have also attempted this, in order to compare the two results. We first note a few 1 The de Vaucouleurs profile is an empirical luminosity profile of elliptical galaxies, given by I(r) ∝ Ie exp[−(r/re )], where Ie is the intensity at the effective radius re . 7.3 Results 111 differences between the two data sets. Ullán et al. combined images for each night, while we have included all individual points. They also rejected points from 10 nights, based on criteria inherent in their photometry method. The χ2 residuals from the deconvolution did not provide us with reasons to discard any images, so we included them all in Figure 7.1. On request, Aurora Ullàn provided us with a list of the images they had discarded, which we compared to our own results. The points of conspicuously large magnitude in Figure 7.1 from February 7th were on the list of discarded images. These points differ noticeably from all other points, for both components, indicating some observational cause for the deviation. As we perform the straight line fit solely in order to compare our results with those of Ullán et al., we found it reasonable to reject these two points when fitting the line. The straight line fits are shown in Figure 7.2. The gradient for component A is –0.63 mmag/day, and for component B +0.57 mmag/day. Gradients found by Ullán et al. are –0.8 mmag/day for A and +0.3 mmag/day for B. Our results show similar gradients, although a slightly smaller absolute value for the A component and a relatively larger value for B. Discrepancies in our results probably originate from the different photometry methods, as well as the difference in data sets. In conclusion of this chapter, we note that the components seem to be moving closer in terms of brightness values. The gradient in component A is towards brighter values, while B seems to become weaker. If this trend continued, it could point to the shift that, according to our results in Chapter 5, has occurred in March 2001, when component A is brighter than B. It would indeed be interesting to analyse more data from the year 2000 and the beginning of 2001, to see if this shift in the brightness ratio constitutes a continuous trend, or an abrupt change. Preliminary results from aperture and PSF photometry using data obtained at the Spanish Calar Alto 1.5 meter telescope between February and June 2003, show that B is again brighter than A in both R and V . Results for the R band is ∼16.58 mag for the A component and ∼16.47 mag for B. In the V band the results are ∼16.95 mag for A and ∼16.70 mag for B. These results are obtained from Jan-Erik Ovaldsen and Aurora Ullán Nieto (private communication), using the photometry methods of Ovaldsen et al. (2003b). 112 Reduction and photometry of the StanCam data Figure 7.1: Light curves in the R band for the quasar components A and B from the StanCam data. As in Chapter 5, the A component is represented by a cross (×) and B by a diamond (). 7.3 Results 113 Figure 7.2: Straight line fits to the data points in the R band for the quasar components A and B from the StanCam data. Two points seen in Figure 7.1 are rejected in the line fits, see text for details. The gradients for the two lines are –0.63 mmag/day for component A, and +0.57 mmag/day for component B. 114 Reduction and photometry of the StanCam data Chapter 8 Conclusions I could be bounded in a nutshell and count myself a king of infinite space were it not that I have bad dreams. William Shakespeare, Hamlet, Prince of Denmark, Act II, Scene II In this thesis, we have presented photometric results from three data sets, all obtained at the Nordic Optical Telescope. Two of the data sets contained observations from January–March 2000, and one set was from four nights in March 2001. Optical monitoring of QSO 0957+561 has proved a fascinating study in itself, especially when comparing our results to that of others who have used the same data but different methods of analysis. Applications of time delay measurements and studies of microlensing are diverse and important, but they require firm observational ground. This thesis work has been a contribution to that platform from which one can probe the distribution of matter in lensing galaxies, constrain the different regions of quasars and even calculate the Hubble parameter. The main part of this work has been a detailed analysis of the light curves of QSO 0957+561 during a relatively short time period, four nights in March 2001. Comparison with the one night of observations in January 2000 provided an opportunity to calculate the time shifted colour indices and magnification ratio of the quasar components, which we compared to results in Ovaldsen et al. (2003b). Analysis of the StanCam data provided us with an opportunity to study variations of the quasar on a time scale of two months. For our main data set, from March 2001, we studied the light curves in detail. Our investigations of the influence of the Moon in Section 6.1.3 showed that moonlight is a possible explanation for the increase in brightness 116 Conclusions seen in both components towards the end of each night. We have no simple explanation for why moonlight should have this effect. Chromatic effects seem to be important, as the star R is not influenced in the same degree as the quasar components. The effect may be caused by the fact that the quasar is bluer than star R. It may also be an effect arising from our normalisation of the quasar with star D, which is also more red than the quasar. If the moonlight increases the brightness of the quasar components towards the end of each night, it could explain the slight increase in the last part of each night, which starts approximately at the time of moon-rise. Referring to the model light curve in Figure 6.9, we note that the last points on, e.g., the first night, that have low brightness (large magnitude value) might be caused by inadequacies in the model. However, disregarding some of these points, it is tempting to suggest that the sinking tendency throughout the four nights could be reflected in the observations within each night. Other, unknown observational effects might also be influencing the data, which further studies will hopefully shed some light on. This could mean that the trend seen from night to night is evidence of intrinsic variation in the quasar, and that both quasar components happen to decrease during this period. The aber of interpreting the light curves as sinking steadily, is of course the fact that the tendencies are so similar in the theoretically unrelated A and B components. Although we acknowledge that this might be interpreted as too much of a coincidence, we also remark that it is not impossible. After all, both components must exhibit some behaviour at all times, e.g. a nightto-night decrease in brightness, and the observed decreasing tendency is a bit larger (0.003 mag) in the B component than in A. On the other hand, we note that two of the field stars, P and Q, exhibited a very similar decrease in brightness from night to night as that of the quasar components. This could be a further indication that the decrease has origin in something other than the quasar itself. What could cause such an effect in both the quasar components and P and Q is uncertain. The two stars showing the same effect as the (bluer) quasar components, are the reddest stars we have studied. The stars H and R, that are closer in colour to the quasar components, do not show this systematic decrease. This indicates that the cause of the decrease, if based on colour differences, is complex. It could also indicate that the explanation has to be found elsewhere. Apart from the possible decrease in brightness for both components, we have not found any evidence of intrinsic night-to-night variability in the ALFOSC data. For the StanCam data there is a possible gradient on the time scale of several weeks, towards increasing brightness for component A and decreasing for B. The most conspicuous brightness variation seen in our results, however, is 8.1 Future studies 117 the change from 2000 to 2001. Within this period, the two components have shifted in relative brightness, from B brighter than A in 2000 to A brighter than B in 2001. 8.1 Future studies First of all, it would be intriguing to study further the colour variations of the quasar components and field stars. Both the variations we have studied, the ‘U’ shape within each night and the decreasing brightness from night to night, could be further analysed with more information on colour variations. For example, it could be interesting to study the spectra of the quasar components. As this thesis work has shown that stars of different colour seem to behave differently, possibly because of reacting differently to changes in observing conditions such as moonlight, it would be informative to obtain specific information about variations in the continuum and line emitting regions of the quasar, how the intensities in these regions change in time and, possibly, how they respond to changes in observing conditions. The observations we used for the main part of this thesis work, i.e. the ALFOSC data, were obtained from directions from the QuOC team, see Section 2.1, designed for their time delay analysis. We find that the studies we ended up performing could have had a more optimally designed observation schedule. As the observations were done on an hourly basis (60 minutes R, 15 minutes V etc.), there was a small and irregular number of V images. One 15 minute session with V could yield 9 frames while most gave us 3–6, when images with bad columns had been subtracted from the data set, see Section 3.4.3. When summing 10 and 10 images, we thus ended up combining V frames separated by time intervals of ∼1 hour. As we had few data points in V , and they were separated by large intervals, we could not study detailed variations in V − R within each night. If planning the observations for this thesis work from scratch, it would be better to observe, e.g. 10 R frames followed by 10 V frames etc. in cycles. This procedure could in principle have provided us with a better time resolution for studying colour variations, especially if we consistently avoided the bad column so we did not have to reject so many images. Also, on the technical side, we would have stepped all the sky flats to remove stars more easily, and we would have checked the dome flats in the beginning of the observation run and if they seemed bad, we would take only sky flats. As regards the brightness variations seen in the course of one year, more observations from April 2000 and onwards, if available, could reveal how the shift in relative brightness between the two components occured, whether a 118 Conclusions slow trend – perhaps starting with the trend we see in the StanCam data – or a sudden change. The preliminary results we mentioned in Section 7.3 indicate that the relative brightness of the components has shifted again, which would also be interesting to study further. As a last remark, we note that if we had obtained more data from the January 2000 session, we could possibly have used our data sets to make a time delay analysis, as data from the two observing sessions were separated by ∼416 days. Since different groups report time delays ranging from ∼417 to ∼425 days, see Section 2.3, observing sessions lasting for several days, seperated by an approximate time delay, might be necessary to resolve the issue. Bibliography Bevington, P. R. and Robinson, D. K.: 2003, Data Reduction and Error Analysis for the Physical Sciences, Mc Graw-Hill Higher Education, 3rd ed. Burud, I., Magain, P., Sohy, S., and Hjorth, J.: 2001, A&A 380, 805 Colley, W. N. and Schild, R. E.: 2000, ApJ 540, 104 Colley, W. N., Schild, R. E., Abajas, C., Alcalde, D., Aslan, Z., Barrena, R., Dudinov, V., Khamitov, I., Kjernsmo, K., Lee, H. J., Lee, J., Lee, M. G., Licandro, J., Maoz, D., Mediavilla, E., Motta, V., Muñoz, J., Oscoz, A., Serra-Ricart, M., Sinelnikov, I., Stabell, R., Teuber, J., and Zheleznyak, A.: 2002, ApJ 565, 105 Colley, W. N., Schild, R. E., Abajas, C., Alcalde, D., Aslan, Z., Bikmaev, I., Chavushyan, V., Chinarro, L., Cournoyer, J., Crowe, R., Dudinov, V., Evans, A. K. D., Jeon, Y., Goicoechea, L. J., Golbasi, O., Khamitov, I., Kjernsmo, K., Lee, H. J., Lee, J., Lee, K., Lee, M. G., Lopez-Cruz, O., Mediavilla, E., Moffat, A. F. J., Mujica, R., Ullan, A., Muñoz, J., Oscoz, A., Park, M., Purves, N., Saanum, O., Sakhibullin, N., Serra-Ricart, M., Sinelnikov, I., Stabell, R., Stockton, A., Teuber, J., Thompson, R., Woo, H., and Zheleznyak, A.: 2003, ApJ 587, 71 Dahle, H.: 1994, Master’s thesis, Institute of Theoretical Astrophysics, University of Oslo Fruchter, A. S. and Hook, R. N.: 2002, PASP 114, 144 Howell, S. B.: 2000, Handbook of CCD Astronomy, Cambridge University Press Janesick, J. and Blouke, M.: 1987, S&T 74, 238 Kayser, R. and Refsdal, S.: 1983, A&A 128, 156 120 BIBLIOGRAPHY King, I. R.: 1971, PASP 83, 199 Kundic, T., Colley, W. N., Gott, J. R. I., Malhotra, S., Pen, U., Rhoads, J. E., Stanek, K. Z., Turner, E. L., and Wambsganss, J.: 1995, ApJ 455, L5+ Kundic, T., Turner, E. L., Colley, W. N., Gott, J. R. I., Rhoads, J. E., Wang, Y., Bergeron, L. E., Gloria, K. A., Long, D. C., Malhotra, S., and Wambsganss, J.: 1997, ApJ 482, 75 Léna, P.: 1988, Observational Astrophysics, Springer-Verlag Lucy, L. B.: 1974, AJ 79, 745 Magain, P., Courbin, F., and Sohy, S.: 1998, ApJ 494, 472 McLeod, B. A., Bernstein, G. M., Rieke, M. J., and Weedman, D. W.: 1998, AJ 115, 1377 Moffat, A. F. J.: 1969, A&A 3, 455 Ovaldsen, J.-E.: 2002, Master’s thesis, Institute of Theoretical Astrophysics, University of Oslo Ovaldsen, J.-E., Teuber, J., Schild, R. E., and Stabell, R.: 2003a, A&A 402, 891 Ovaldsen, J.-E., Teuber, J., Stabell, R., and Evans, A. K. D.: 2003b, MNRAS 345, 795 Refsdal, S.: 1964, MNRAS 128, 307 Refsdal, S. and Surdej, J.: 1994, Reports of Progress in Physics 57, 117 Richardson, W. H.: 1972, Optical Society of America Journal 62, 55 Rottmann, K.: 1995, Matematisk Formelsamling, Bracan Forlag Serra-Ricart, M., Oscoz, A., Sanchı́s, T., Mediavilla, E., Goicoechea, L. J., Licandro, J., Alcalde, D., and Gil-Merino, R.: 1999, ApJ 526, 40 Shannon, C. J.: 1949, in Proc. I. R. E., Vol. 37, p. 10 Stetson, P. B., Davis, L. E., and Crabtree, D. R.: 1990, in ASP Conf. Ser. 8: CCDs in astronomy, pp 289–304 BIBLIOGRAPHY 121 Ullán, A., Goicoechea, L. J., Muñoz, J. A., Mediavilla, E., Serra-Ricart, M., Puga, E., Alcalde, D., Oscoz, A., and Barrena, R.: 2003, GLITP optical monitoring of QSO 0957+561: V R light curves and variability, MNRAS (in press) Walsh, D., Carswell, R. F., and Weymann, R. J.: 1979, Nature 279, 381 122 BIBLIOGRAPHY Appendix A The Image Reduction and Analysis Facility (IRAF) IRAF is a general purpose software system for the reduction and analysis of astronomical data, developed by the National Optical Astronomy Observatories (NOAO) in Tucson, Arizona. The following is not in any way meant to be an exhaustive discussion of the reduction possibilities in IRAF. It is a description of the tasks we have used in the course of this work, in more detail than included in previous chapters. A.1 Using IRAF IRAF tasks are commands that perform certain operations. The parameters for the tasks determine how these operations will be done. There are two types of IRAF parameters: required (‘query’) and hidden parameters, see figure A.1. Practically speaking, using IRAF consists of deciding upon and filling in different parameters for the tasks you want to execute. The epar1 [taskname] command opens the editing environment. There are certain default values for each task. In the following, we will primarily discuss the parameters we used when they differ from the default values. An IRAF task does not have to be repeated manually for each image. One way of repeating the same task for several images, is to provide a list of images as the input parameter. This enables IRAF to perform the same task on all images in the list. A list is indicated by the prescript @. We made use of lists for the tasks zerocombine, ccdproc and flatcombine. Another method of dealing with many images, is to make a small script, which performs the same task for many images while allowing the user to 1 ‘epar’ = ‘edit parameters’. 124 The Image Reduction and Analysis Facility (IRAF) Figure A.1: Example of environment for editing parameters in IRAF, for the task imarith. The first four parameters are query parameters: They must be specified in order to execute the task. The parameters in parentheses are the hidden parameters: If they are not specified, the current default values will be used. vary parameters if necessary. We made use of scripts when running the tasks geomap, gregister and imcombine. A.2 A.2.1 Pre-processing with IRAF Zerocombine As described in Section 3.2, we need to combine our bias frames to avoid random fluctuations and ‘cosmics’ present in the individual frames. For this purpose we used the IRAF task zerocombine (the image type code for bias frames in IRAF is ‘zero’). For this task, the rejection type minmax was used to reject any pixels with deviating value, before combinaion. This method rejects the n low and n high pixels, where n is specified by the user. We used the default value of nhigh=1, nlow=0. Thus, the bias frames for each night were averaged, but with the highest value being ignored when forming the average for each given pixel. A.2.2 Ccdproc and Flatcombine The task ccdproc processes CCD images to correct and calibrate for detector defects, bias level, dark counts and fringing. We have used the task for two purposes: preparing the flat frames for the task flatcombine and the flat field calibration of the object frames themselves. In the editing environment of ccdproc, we enter the ‘trim section’, which is the part of the image that will be saved, and the ‘bias section’, which is the A.2 Pre-processing with IRAF 125 section of the raw image that will be used to fit the bias level. For the March 2001 data, we used a trim section of [50:2000, 5:1195] and a bias section of [2053:2088, 5:1195]. The task flatcombine performs the combination of a list of individual flat frames. We used the combine option average when combining most of the flats, but median on the night of March 17th–18th. On this night, we combined sky flats that had been shifted in x and y, and used the median option to remove the effects of a few stars appearing on the flat frames. The reject parameter selects a type of rejection operation to be applied to the pixels when combining the flats. A rejection algorithm is useful when forming the average or median of n pixels, as one or several of the pixels may be affected by a cosmic ray, or for some other reason have an unrealistic deviant value. We compared two rejection options for flatcombine: avsigclip and crreject. Both rejection methods assume that the image values are proportional to the number of photons recorded. Pixels above or below the median or average by N σ are rejected, where N has a default value of 3 but can be modified by the user. We chose the median here, to avoid that extreme-valued pixels should dominate the calculation. Values deviating by more than the specified σ threshold factors are rejected. These steps are repeated, until no further pixels are rejected or there are fewer than three pixels. The avsigclip method determines σ empirically, by making an estimate of the expected σ value based upon the data itself. The crreject method on the other hand, determines σ from the CCD noise characteristics (i.e. readout noise and gain) that we provide a priori. For every night, we made one image with each reject parameter, and subtracted these images from each other. The residuals we obtained were random fluctuations above and below zero. Since no distinguishable pattern was apparent, we concluded that the routines were more or less equivalent for our data. We used avsigclip when combining most of our flats. On the one occasion when we had only 3 flats (night of 14th–15th in the V filter), we used crreject as this option uses the noise characteristics we provide instead of calculating them from the image itself. Assuming that this is the best option when few images are available, we used it on the one occasion when we had only 3 individual flats. 126 A.3 The Image Reduction and Analysis Facility (IRAF) Other tasks using IRAF Apart from the flat-fielding, there were several other operations that we performed on the images using different tasks in IRAF. One of them was ‘stacking’ the images to obtain a better S/N ratio. In addition, we used IRAF to perform arithmetical operations on the images like subtracting the background, normalisation, multiplication etc. A.3.1 Image scaling and background subtraction To perform arithmetical operations on the images, we used the task imarith, see Figure A.1. We made a short IDL routine that read exposure times from the image headers to a text file, and we then made lists of the image names and the exposure times, so we could use imarith on all images (of each night and each filter) simultaneously. The background subtraction, described in Section 3.4.2, was also done using imarith, and by making lists of the image names and of the background level in each image. In this way the background subtraction could be done efficiently on all ∼1000 images. A.3.2 Stacking the images As described in Section 3.4.2, we wanted to sum the images in batches of 10. The way we proceeded to do this was to first form an average frame from 10 images, and then multiply this image by 10. This provided us with the S/N ratio of a much longer exposure time than that of our individual frames. Aligning the images using geomap and gregister There are many ways of aligning, or registering, images in IRAF, ranging from simple x and y transformations to more complex coordinate transformations. We have used the tasks geomap and gregister for both our alignments. First, we aligned images using simple shifts in x and y, to construct our ‘combined’ image frame from 10 individual frames. Second, we aligned all combined images, in preparation for the simultaneous deconvolution, using a parameter that computed the rotation as well as shifts in x and y. Aligning the images twice like this is not ideal, but circumstances forced us to do this. The task geomap computes the spatial transformations required to map input coordinates to a reference coordinate system. The format of the input file to geomap includes one object per line, with the x and y coordinates of the reference image followed by the x and y coordinates of the image to be A.3 Other tasks using IRAF 127 registered. These input files were supplied to us by Jan-Erik Ovaldsen, as described in Section 3.4.2. The transformations have the form xin = f (xref , yref ) yin = g(xref , yref ) , and are stored in a text database file. The functions f and g are by default a power series polynomial surface of order 2. For our simple case of a shift in x and y (fitgeom = shift, see below), this becomes xin = a + xref yin = b + yref , where a and b give the shift in the x and y direction respectively. Geomap may be run interactively by switching on the interactive parameter. In interactive mode, we have the option of changing the fit parameters and displaying the data graphically, using keystroke commands, until a satisfactory fit has been achieved. We checked manually each shift and rejected stars that had large residual values, that is, differences between the reference and the transformed coordinates. As remarked in Section 3.4.2, this resulted in satisfactory residuals. We performed a few tests to see how the peak value and the flux of stars changed after the transformation, and found that the peak value could change by ∼5% while the total flux only changed by ∼0.3%. The peak value changes more than the flux, as the values of the pixels are distributed a bit differently after the transformation. For the second alignment, we used rxyscale as the fitgeom parameter (‘fitting geometry’), thus taking any rotation of the images into account when aligning images. This second alignment was done because the images had to be aligned so that the centre of the point sources fell within the same pixel on each frame, for performing the simultaneous deconvolution. We chose 6 unsaturated stars for the transformation, choosing stars that were well ‘spread out’, i.e. located at different positions in x and y coordinates on the frame. On the first night, we noticed a sudden rotation from R image number 195 to 196. The rest of the images this night, about 20, were also rotated in the same way with respect to all previous. We avoided this problem in the first alignment by only combining images not rotated in relation to each other. However, this (unexplained) rotation needed to be taken into account when aligning images for the simultaneous deconvolution, if we were to utilise the 128 The Image Reduction and Analysis Facility (IRAF) last two stacked images. This was achieved by using the rxyscale option for fitgeom. In retrospect, we realise we should have first aligned all images to the same coordinates, using rxyscale, and then combined 10 and 10 images. In this way, we would only have needed to align the images once. However, tests showed that even with the rxyscale option, the automatic alignment produced relatively large residuals, and we did not have time to perform the interactive alignment on all ∼1000 individual images over again. The output file created by geomap, containing the transformation parameters, is used as input to the task gregister, which performs the actual transformation. The transformation is stored as the coefficients of a polynomial surface. The transformed coordinates at intermediate pixel values are determined by bilinear interpolation. The parameter geometry determines the type of geometry to be applied by geomap. We found that with the default option geometric, which computes the ‘full transformation’, the task had a tendency to ‘overfit’, i.e. introduce unnecessary effects, so we changed this to the option linear which computes shifts and rotations, which sufficed for our purposes. We ran geomap and gregister using simple IRAF scripts. A script is a text file containing a sequence of IRAF commands, and are well suited for performing the same task on a large number of images. The script was edited in emacs, and executed by typing ‘cl < (script name)’. Combining the images using imcombine After aligning the images to their reference coordinates, we used the task imcombine to construct an average image frame from the 10 individual images. As regards the reject parameter, the same options as for flatcombine apply. This time, however, we could not use avsigclip or crreject as these methods assume a linear relation between ADU and photons, and we had already subtracted the background from the image frames before forming the average, see Section 3.4.2. When we subtracted the background we violated the linearity criterium. Instead, we used the reject option sigclip. This method computes the median (or average) at each output pixel, excluding the highest and lowest value, and then computes the σ about this estimate. For this method to work well, it is recommended that at least 10 images are used, which is just what we were using. In the last sets for each night and filter, we therefore distributed the images evenly in the two last sets. That is, 10 + 4 number of images in the last two sets was changed to 7 + 7. We again used a simple IRAF script to combine the images efficiently. The input images to be combined were spectified by a list. We made one list A.3 Other tasks using IRAF 129 for each set of 10 images to be combined. After combining, we used imarith to multiply all the (combined) images by 10. 130 The Image Reduction and Analysis Facility (IRAF) Appendix B Text files associated with the MCS programs To illustrate the discussion of the MCS deconvolution programs in Chapter 4, we include examples of the input text files that must be filled in by the user of the programs. To illustrate the format, we also include an example of one of the the output text files. Figure B.1: extract.txt: The input text file for extract.f 132 Text files associated with the MCS programs Figure B.1 displays the input text file for the MCS program that extracts sub-images of optional size (la taille des images a extraire) from the image. See Section 4.5.1 for a description of the program. Figure B.2: psfmof.txt: The input text file for psfm prive.f and psf prive.f Figure B.2 displays the input text file for the two MCS programs that are used for constructing the PSF, see Section 4.5.2. Figures B.3 and B.4 show the two input text files for the program that performs the final deconvolution, see Section 4.5.3. Figure B.3 shows the text file for the photometry parameters and Figure B.4 the input deconvolution parameters. Figure B.5 shows an example of an output text file, with information on the positions and peak intensities, in units specific to the normalisations applied and to this particular image, as well as information on the χ2 reached and specifications on the minimising routines. The output text file in2.txt contains the same information on the intensities and positions reached, but in the format of Figure B.3. 133 Figure B.3: in.txt: Input text file for dec prive.f 134 Text files associated with the MCS programs Figure B.4: deconv.txt: Input text file for dec prive.f 135 Figure B.5: out.txt: An example of an output text file from dec prive.f 136 Text files associated with the MCS programs Appendix C Tables of output files from the MCS programs We include an overview of the output files from the programs. Files to the right of the double line in Tables C.2 and C.3 were used in earlier versions of the program and are still given as output but not used by us. The left column shows the name of the FORTRAN 77 program in bold text. The first program extracts images of the objects of interest: the object to be deconvolved and the stars that are to be used for constructing the PSF. Table C.1: The program extract.f and its output files. The extracted images are located in the second column, and their associated noise maps in the third column, see Section 4.5.1 extract.f g.fits psfnn.fits sig.fits psfsignn.fits The next two programs aim at constructing the PSF, s, with an improved, but not infinite, resolution. Table C.2: The program psfm prive.f and its output files: The residuals difcnn.fits are the residuals remaining after subtracting the moffat mofc.fits from the stars psfnn.fits, see Table C.1 and Section 4.5.2. psfm prive.f difcnn.fits mofc.fits sigcnn.fits 138 Tables of output files from the MCS programs Table C.3: The program psf prive.f and its output files. The constructed PSF is s.fits. The other files in the second and third column are the modeled residuals, see Section 4.5.2 for details. psf prive.f psfr.fits psff.fits psffond.fits s.fits xixinn.fits t.fits The last program performs the actual deconvolution with the PSF that was constructed in the preceeding two programs. Table C.4: The program dec prive.f and its output files. The deconvolved images are decnn.fits. The intensity and position of the point sources are listed in the text file out.txt. The residual frames are resinn.fits and resi smnn.fits. The background of the deconvolved image is returned as back.fits. See Section 4.5.3 for details. dec prive.f decnn.fits resinn.fits.fits back.fits resi smnn.fits out.txt in2.txt Appendix D Miscellaneous D.1 Integrating the Gaussian function Point sources in the deconvolved images from the MCS procedure are represented as Gaussian distributions. In this appendix, we show how we can obtain Eq. D.3, the relationship between flux and peak intensity of a Gaussian that we used in Section 4.5.3. In two dimensions, the Gaussian function is given by I(x, y) = Ipeak e−[(x−µx ) 2 +(y−µ y) 2 ]/2σ 2 . (D.1) The centre of the function is (µx , µy ) and the standard deviation is σ. We have assumed that the function is circular, i.e. σ = σx = σy . The intensity of the distribution is I. When the function is centred in origo, i.e. µx = µy = 0, we easily see that the peak value of I is given by x = y = 0 ⇒ I = Ipeak . The flux is obtained by intergrating Equation D.1. We can therefore write Z ∞Z ∞ 2 2 2 F = Ipeak e−(x +y )/2σ dx dy , (D.2) −∞ −∞ where F is the flux of the star. Considering the double integral, we see that it can be written as the multiplication of two terms, since x and y are uncorrelated. That is, we can write Z ∞ Z ∞ Z ∞Z ∞ 2 y2 − x2 −(x2 +y 2 )/2σ 2 e− 2σ2 dy , dx dy = e 2σ dx e −∞ −∞ −∞ −∞ Equation D.2 can thus be written F = Ipeak Z ∞ e −∞ − x2 2σ 2 dx 2 . (D.3) 140 Miscellaneous From Rottmann (1995), this integral is given by Z ∞ √ x2 e− 2σ2 dx = 2π σ . −∞ This brings us from Eq. D.3 to the following relation between flux and peak intensity: F = 2πσ 2 Ipeak . To obtain Eq. 4.10, the relation between the FWHM and standard deviation, also used in Section 4.5.3, we use the one-dimensional Gaussian. Setting Ipeak = 1, we solve −(x0 −µ)2 1 e 2σ2 = f (xmax ) , 2 where x0 is the x position at FWHM. We have that f (xmax ) = 1 at x0 = µ. From this, we get −(x0 −µ)2 1 1 e 2σ2 = f (µ) = . 2 2 Taking the natural logarithm on both sides of the equation we obtain and thus 1 (x0 − µ)2 = ln = −ln2 , − 2σ 2 2 √ x0 = ±σ 2ln2 + µ . The full width at half maximum is therefore given by √ FWHM = x+ − x− = 2σ 2ln2 ≈ 2.3548σ.