wfc3cte_alpha

advertisement
Instructions for Using the Alpha-Release of the WFC3/UVIS
Pixel-based CTE Correction
Jay Anderson
STScI
February 19, 2013
1. Abstract
The current version of the pixel-based CTE correction for WFC3/UVIS is available only
as a stand-alone FORTRAN program. This routine is currently an “alpha” release,
meaning that we reserve the right to make major changes to it. The routine does work
reasonably well in the domain where it can: where the source is bright or where the
background is moderate. Faint sources on low backgrounds will always be very difficult
to correct for, as they tend to experience losses that are a large fraction of the initial
counts. It will always be impossible to reconstruct something from nothing.
This routine is not yet a part of the pipeline and it will probably be several months before
it is included in the pipeline. Also, at this point, there are also no additional reference
files (such as de-trailed darks) that can be used in conjunction with it.
2. Construction of the model
A comprehensive description of the construction of the model will eventually follow in a
separate ISR, but it is worthwhile to provide a brief history here. The model is generally
based on the pixel-based correction developed by Anderson & Bedin (2010) for ACS,
which itself was an extension of the model constructed by Massey et al. (2010) for their
reduction of the COSMOS data.
There are two aspects to the CTE model: (1) how charge is lost from one pixel and (2)
how it is released into the upstream pixels. The basic model assumes that charge “traps”
are distributed throughout the detector, and each trap is able to grab a particular electron
out of the packets that pass through it (either the first, second, fifth, five-hundredth, etc.).
Since it is impossible to know exactly what traps can be found in which pixels, we make
the assumption that each of the 2070 pixels up the column has an identical spectrum of
fractional traps. The current UVIS detector has, on average, about 500 traps in each
column. Most of these traps will affect only large pixel packets, but unfortunately there
are enough traps that grab the first few electrons to make CTE losses pathological when
sources are faint on a low background. An iterative approach was used to construct the
model.
2.1 The initial model The first iteration was similar to what was described in Anderson
& Bedin (2010) for ACS. Warm pixels (WPs) serve as convenient delta-function
sources for probing how charge is transferred down the detector. We examined the trails
behind the few relatively bright WPs that UVIS has. The area under each trail told us
the number of traps that impacted the WP during its journey (the losses), and the shape of
the trail told us how charge gets released. This provided the initial version of the model.
2.2 Probing losses in the smallest packets. Unfortunately, it is very hard to measure
the trails behind WPs that have less than 50 electrons. This is partly because the losses
are large (and no longer perturbations) and partly because the trail gets lost in the noise.
In an effort to pin the model down for the smaller charge packets — where CTE losses
are most critical — we took a set of short-dark and long-dark exposures. The short-dark
exposures were 100s and the long-dark exposures were the standard 900s. To study the
CTE losses from small packets, we first apply our initial CTE correction to the long-dark
stacks. This gives us a pretty good idea of how much flux started out in each of the
medium-bright WPs (say, those with 75 to 1,000 electrons). We can then multiply these
images by 0.1111 (100/900, the ratio of exposure times) to get the expected number of
counts in each short-dark exposure. We then compare the observed number of counts in
the short dark with the expected number to get a direct picture of CTE losses as a
function of the number of transfers each WP experienced. This allows us to study
electron packets from 10 to 125 electrons in size.
The above procedure allowed us to see how many electrons a charge packet loses on
(essentially) zero background. We found the losses to be even greater than expected. A
pixel packet that started out with 80 electrons at the top of the chip ended up with less
than 30 by the time it reached the readout register; and a packet with 50 ended up with
less than 10. We examined the trends for charge packets of different size and found that
packets with between 1 and 10 electrons continued to lose more and more electrons as the
packet got larger. But once packets reached the size of 12 e−, their losses appeared to be
largely independent of packet size. This was true until they reached the size of 50 e−or
so, at which point they slowly started losing more again.
These trends told us that there aren’t many traps that grab electrons #13 through #50.
Thus, if an image has a background of about 12 electrons, then perhaps that could keep
many of the traps filled and significantly mitigate the CTE losses.
2.3 The efficacy of some background. To test this optimistic hypothesis, we took a
series of observations of the center of globular cluster Omega Centauri, with pairs of
short and deep exposures (10s and 700s) through F336W. This is not unlike what was
done above to study WPs in the dark exposures. The reason for this target/filter choice
was that the field contains a nice, flat distribution of stars. In the short exposures, there
are not many stars brighter then S/N ~ 50, but below this (which is the location of the
turnoff) there is a relatively flat distribution with magnitude. These image pairs allowed
us to assess losses directly by comparing the observed counts in the short exposures
against the predictions from the scaled-down deep exposures. We also varied the
Figure 1: This figure shows the observed trends that were used to constrain the model, and the model itself.
Each of the seven panels corresponds to a set of WPs that the long darks tell us should contain 10, 20, … 80
electrons in the short darks. This WP is then observed after a different number of transfers (250, 750, 1250, and
1750) and the total number of electrons (background + WP) is shown. The black trends show the behavior on
very low background, the green on a background of ~2, and the blue on a background of 12. The data were
taken in September 2012.
background using the post-flash option. This gave us a direct assessment of how various
background levels shield sources from CTE losses. This study confirmed the efficacy of
12 electrons background. (See Anderson et al 2013).
Finally, we took a series of short-long darks with various backgrounds to help us pin
down the model in the context of background mitigation. Again, we constructed a model
of the dark current in each pixel from the de-trailed long dark exposures (using our initial
readout model). Then we examined how the various WPs lost flux as a function of the
number of transfers and the background level. Figure 1 shows the WP trends that were
observed in the short-dark exposures taken with three different levels of background (0
e−, 2e−, and 12e−). We actually fit the model to 10 different levels of background, but for
clarity we show only three here.
The black curves in the panels from left to right show what happens to WPs that start out
with 10, 20, 30, 40, 50, 60 and 80 electrons on a detector with no background. Losses
are 70% or greater for stars that are transferred all the way down the chip, even for WPs
that start with 80 electrons. The green curve shows the results for backgrounds of about
2 electrons. Losses are still large, but are down by perhaps a factor of two from the
pathological zero-background case. The blue curve shows the losses for WPs on a
background of 12. Losses are less than 20% for all WPs.
The model, indicated by the dashed curves, does a nice job describing most of these
trends. At the very lowest end (a small WP on a low background), the model overpredicts the losses. The model has a monotonic function with one degree of freedom
(the number of traps encountered per 2048 transfers as a function of packet size) to
constrain this two-dimensional distribution of losses versus WP intensity and
background. The errors at the very low end indicate some inadequacy in the model
algorithm, but in fairness, it is very hard to reconstruct sources in this regime anyway.
2.4 The current UVIS model. The current UVIS model algorithm is somewhat different
from the algorithm that is operating in the ACS pipeline. Whereas the ACS model
allowed traps to affect fractional pixel levels and work on real-number pixel arrays, the
new model explicitly deals only with integer numbers of electrons. The new model is
also specified in a somewhat simpler manner than the ACS model. Whereas the ACS
model specified the “trap density” at various electron levels, the current model is simply
specified by the cumulative number of traps as a function of packet-size in electrons. We
show the cumulative trap distribution in Figure 2, below.
Figure 2: From the current model: the cumulative number of accessible traps per 2048 pixels-up-the-column,
as a function of packet size. The left plot shows the small-packet region with linear scaling, and the right plot
shows the full range of packet sizes with log-log scaling. The marginal losses are essentially the slope of the
curve.
It turns out that the trail profile for WFC3/UVIS does not appear to be a perceptible
function of packet size. We found that 20% of the trapped electrons are released after the
first transfer, 8.5% in the second transfer, 6.75% in the third about 1.5% in the tenth,
0.1% in the fiftieth, and the trail goes to zero after about 60 pixels. There are not a lot of
bright warm pixels in the UVIS detector, so it is not trivial to follow the trail out to large
distances.
2.5 Dealing with “readout CRs”. It often happens that cosmic rays (CRs) strike the
detector during readout. A full-chip readout takes about 90s, so an exposure that is 810s
will have 10% of its CRs hit during readout. CRs that hit during readout do not undergo
the same number of transfers as the electron packet they arrive at the readout-register
with; they undergo fewer. For instance, if a CR hits the detector in pixel [200,200] at the
time when pixel [200,2000] is passing through it, then the CR-added electrons will
undergo one tenth the implied number of parallel transfers. It is clear that if we treat all
electrons as having undergone the number of transfers implied by their vertical pixel
location, then we will overestimate the amount of trailing suffered for these CRs, and
hence will over-subtract the trails.
Like the ACS model, the current UVIS model also searches for over-subtracted trails.
We define an over-subtracted trail as either: a single pixel value below −10 e−, two
consecutive pixels totaling −12 e−, or three totaling −15 e−. When we detect such an
over-subtracted trail, we iteratively reduce the local CTE scaling by 25% until the trail is
no longer negative. This does not identify all readout-CRs, but it does deal with many of
them. For images that have backgrounds greater than 10 or so, this will still end up oversubtracting CRs a bit, since we allow their trails to be subtracted down to −10, rather than
to 0. It would be possible to have the algorithm use the background sky value rather than
zero, as the baseline, below which it looks for over-subtracted trails.
2.6 CTE reconstruction with very low backgrounds. The fact that many of UVIS’s
observations have been taken with low background makes CTE reconstruction
particularly challenging. This is a regime that ACS has not had to face, since even “bias”
exposures with no integration time collect ~5 electrons dark-current in their top-row
pixels by the time that pixel completes its 2048 parallel transfers. UVIS pixels currently
get about 0.5 electron dark current during readout; they got even less in the past.
We mentioned above that a source with only a few electrons loses almost all of them
during the 2000 parallel transfers down the detector. This makes it very hard to
reconstruct faint sources on low backgrounds, as there is not much evidence in the readout image that there was anything there. This is particularly true when we fold in the
contribution of the 3 e- readnoise. A source that loses so many electrons that it cannot
stand out above the readnoise will be impossible to reconstruct. A corollary of this is that
if we ignore the fact that our observed images have readnoise and try to determine what
original image could get pushed through the read-out model to produce the observed
pixel distribution, we will end up with an image that is perhaps 10× noisier than the
original observation1. It is clear that readnoise-mitigation will be even more important
for UVIS than it was for ACS.
A further complication with low backgrounds is that the exact number of electrons in the
background can make a big difference in the CTE losses. A 25 e- WP on no background
will lose more than 20 electrons (80%), while the same WP on a background of 3 will
lose perhaps 10 electrons (40%). It is therefore critical to accurately estimate the
background so that sources can be reconstructed as accurately as possible. It is clear that
this estimate of the background must be much more accurate than the readnoise (±3
electrons) allows us to know the number of electrons in any given pixel.
For all the above reasons, the new model includes an improved readnoise-mitigation
algorithm. In brief, the goal of the algorithm is to identify the smoothest possible image
that is consistent with being the observed image plus readnoise. The reconstruction
algorithm acts on this smooth image to make a conservative estimate of how charge may
have been transferred from one pixel to another in the real image during readout. The
negative of this transfer is then added to the original image.
3. Download
The routine can be downloaded by visiting the following website:
http://www.stsci.edu/~jayander/X/EXPORT_WFC3UV_CTE
There are two FORTRAN programs available for download from this directory.
Instructions on compiling the programs and on their parameters are given in the
README file and in comments at the top of the files. We will just give a very brief
description here.
The first program is the correction routine itself, named wfc3uv_ctereverse. It
takes a _raw.fits file and generates what is being called a _rac.fits file. This
output file is as similar as possible to the original raw file, except that it is real*4 instead
of an unsigned integer*2, and has had its electrons re-arranged in accordance with a
model of how charge likely got redistributed during readout. These rac files should be
able to be run through the CALWFC3 pipeline as if they were normal raw files. The
routine also has the option (use “FLC+”) of taking a raw file and an flt file. It will
determine from the raw file how the electrons need to be redistributed and will apply this
1
It is easy to see where this comes from. If an observed image is full of empty pixels and
has one pixel far from the readout has one electron, then in order to read-out such a
distribution, we would need to start with (say) 10 electrons, 9 of which would get lost
along the way. This would allow us to read out a value of “1” for the target pixel, but the
model would also have to account for the fact that (say) “2” electrons would have been
added to the upstream pixel, and single electrons to seven other upstream pixels. The
original image would have to look something like: (0 +10 -2 -1 -1 -1 -1 -1 -1 -1) to be
read out as (0 1 0 0 0 0 0 0 0 0). Generalizing this to an image full of readnoise, where
each pixel can vary by ±3, it is clear that the de-blurred result will be very noisy.
to the flt image, producing an flc file, which can be used for drizzle or other
standard exposure-level image analysis. The routine has several flags and the comments
at the beginning of it should help you use it. It is worth noting that the routine cannot
operate directly on the flt file, since that file may have had post-flash electrons
subtracted, and the reconstruction routine needs to know about them, since they lessen
the CTE blurring. Depending on the background, it can take 15 minutes to an hour to
construct the correction for a raw image.
The second program is named wfc3uv_forward, and it applies the forward CTE
modeling. It takes a file in what I call my “z” format (8412×2070, real*4, with each
2103×2070 amplifier arranged with its parallel readout direction down and its serial
readout direction to the left) as an original distribution of pixels. The routine then
simulates the image what would result if this distribution were read out (with or without
readnoise added afterward). Instead of starting with a pre-readout image, this routine is
also able to take a list of sources (x,y and total flux) and a sky background and simulate
the initial image and the image that would be read-out. There are comments at the
beginning of the program to provide more information on how to run it. We should note
that Figure 1 shows us that the current model over-predicts losses when the background
is below 5 electrons and the source has less than 20 electrons per pixel. As such, this
forward model should not be used to predict in detail what happens to very faint sources
on very low backgrounds. We are working on ways to adjust the model to accommodate
this inadequacy. It is not simply a modification of the parameters, but rather will require
some fine-tuning of the algorithm itself.
4. How accurate are the corrections under various circumstances?
We have run the pixel-reconstruction algorithm on the short-dark pairs of Omega Cen
images (from §2.3 above) and done aperture photometry and fit PSFs for positions. The
results are shown in Figure 3 for photometry and Figure 4 for astrometry.
Figure 3: Results of simple 5×5-pixel aperture photometry on the short and long images of Omega Cen, for
stars more than 1500 pixels from the readout register. We used the long exposures to predict the number of
counts in the short exposures. The pairs of exposures were taken at the same pointing, so that no aperture
correction is needed to compare the numbers of counts. Each point represents a single star. The median trends
are shown in the middle line, and the inter quartiles are shown above and below. The top rows of panels show
the losses in the uncorrected exposures. The magnitudes are given in instrumental units, −2.5 log10(number of
electrons), such that −10 is S/N ~ 100 and −5 is S/N ~ 10. When the sky background is zero, losses go from 0.1
magnitude at −10 to more than 0.5 magnitude at −7. When the background is 12 electrons, losses are never
more than 0.15 magnitude. The photometry on the corrected images is shown in the bottom panels.
Figure 3 shows that the pixel-based correction does a very nice job on images with some
amount of sky, but when the sky goes to zero, the correction becomes increasingly
inadequate as the source gets fainter. This is not surprising. In order to avoid readnoise
amplification, we had to be conservative in terms of what could be a source and what
could be noise. Even when the cores of faint sources could be distinguished from readnoise — say, for sources with 100 total counts (instrumental magnitude of −5) should
have 20 counts in their central pixels, which is 6 times the readnoise — the surrounding
pixels would not necessarily be identified as having significant flux, and as such would
not have the full correction applied to them.
Figure 4 shows the same results, but for astrometry. Again, the correction is quite good.
This gives us the hope that if the correction restores total flux and position for most stars,
it might also do a good job preserving shape. This, however, remains to be tested.
Figure 4: Same as the previous figure, but for astrometry. Positions were fit with an empirical library PSF for
F336W. Astrometry is clearly restored for most stars on most backgrounds.
5. Remaining issues
One issue that has come to light during our post-flash investigation is that
there are quite a few pixels that appear to have several traps in them. We
know this because the electrons that get added during the post-flash (or
background electrons in a science exposure) get trapped and do not make it
out of the pixel during the first transfer. For some reason, these pixels often
appear to be paired with WPs below them, which makes it somewhat hard to
interpret the trails behind WPs on images with background. The UVIS team
is investigating this phenomenon.
Figure 5: This is a close up of a post-flashed dark stack with a background of ~12 electrons. The whitest pixels
have values of ~2 electrons. The left panel shows a part of the image near the readout register, and on the right
shows a part far from the readout register. Black corresponds to higher values. Clear the WPs are CTEblurred on the right but not on the left. The white spots are pixels significantly below the background. These
also are sharp on the left and more blurred on the right.
The fact that these “holes” appear capable of holding several electrons may
hint to us that traps may not occur one at a time randomly on the detector,
but they may well be bunched up.
6. Anticipated improvements
The correction described here treats every pixel as having exactly the same
distribution of traps. The ACS algorithm is a bit more sophisticated than
this. Ogaz et al (in prep) studied the parallel overscan pixels for each
ACS/WFC column to determine a rough scaling for how many traps each
column had relative to the average.
We plan to do a similar exercise for WFC3/UVIS, but we can go even
farther. We can use the charge-injection procedure to add ~15,000 electrons
to lines separated by 10, 17, or 25 pixels. Examining the trails behind these
lines will allow us to estimate not only how many traps can be found in each
column, but where along the column the traps can be found. Of course, with
such a high injection level, we are not able to probe exactly which electrons
(first, hundredth, etc) each trap impacts, we can only estimate the total
number of traps. It might be possible to use the scan mode to move stars of
different brightness across the detector at a variety of rows to estimate the
loss in each column. The WFC3/UVIS team is exploring these options for
specifying the CTE model more locally.
References
Anderson et al 2013, “The Efficacy of Post-Flash for Mitigating CTE Losses in
WFC3/UVIS images”
http://www.stsci.edu/hst/wfc3/ins_performance/CTE/ANDERSON_UVIS_POST
FLASH_EFFICACY.pdf
Anderson, J. & Bedin, L. R. 2010 PASP 122 1035 “An Empirical Pixel-Based Corretion
for Imperfect CTE. I. HST’s Advanced Camera for Surveys”
Massey, R. et al. 2010 MNRAS 401 371
MacKenty & Smith 2013. “CTE White Paper”
http://www.stsci.edu/hst/wfc3/ins_performance/CTE/CTE_White_Paper.pdf
Ogaz, Sara et al. ISR in prep.
Download