Development of a light detection system ... over the solar spectrum and sun course simulations ...

Development of a light detection system for bidirectional measurements
over the solar spectrum and sun course simulations with scale models
by
Courtney A. Browne
S. B. Mechanical Engineering
MIT, 2004
SUBMITTED TO THE DEPARTMENT OF MECHANICAL ENGINEERING IN
PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF
SCIENCE MASTER OF MECHANICAL ENGINEERING (S.M.M.E.)
AT THE
MASSACHUSETTS INSTITUTE OF TECHNOLOGY
JUNE 2006
C 2006 Massachusetts Institute of Technology. All rights reserved.
Signature of Author:
Department of Mechanical Engineering
June 7, 2006
Certified by:
Certified by:
,-06
lDC
Marilyne Andersn
Assistant Professor of Building Technology
Thesis Advisor
Leon R. Glicksman
Professor of Building Technology and Mechanical Engineering
Thesis Reader
Ac~ep-c
MA SSACHUSETTS INSTTUTE
OF TECHNOLOGY
JU L14 2006
MeIni
~pKjvv
BARKER
Development of a light detection system for bidirectional
measurements over the solar spectrum and sun course
simulations with scale models
*Courtney Browne
June 5, 2006
Abstract
The use of natural light in building structures can increase energy efficiency and lead
to more sustainable architecture. To encourage such use of natural light, a dual experimental device is being developed at MIT to help evaluate the effectiveness of various
daylighting approaches, to be used as a goniophotometer for materials and coatings analysis and as a heliodon for studying scale models. The goniophotometer will be used to
conduct detailed assessments of the bidirectional transmission or reflecting distribution
function ("BT(R)DF") properties of building materials, using a CCD camera to produce
a luminance map of the emerging light distribution. The heliodon mode will be used to
as an educational tool to perform qualitative evaluations of shadow patterns by simulating sunlight illumination on scale models. This thesis focuses on several aspects of this
larger project. This thesis first describes the design of an illumination system appropriate for both functions of the joint goniophotometer/heliodon. This thesis then describes
the design and manufacture of a light collection system for the goniophotometer mode,
specifically the design and fabrication of an acrylic semi-ellipsoid with a half-mirrored
coating that focuses the collected light at the CCD camera used for collection and analysis. Finally, this thesis describes the calibration of the light detection system (the color
CCD camera) to make its spectral sensitivity match that of the human eye. With this
calibration, the CCD camera will be useful not only as a component of the goniophotometer/heliodon system, but may also be adapted to serve as a freestanding multi-point
luminance meter for the characterization of BT(R)DFs for various materials of interest.
Acknowledgements
This thesis is based upon work jointly supported by the National Science Foundation
under Grant No. 0533269 and by the Massachusetts Institute of Technology. Any opinions, findings and conclusions or recommendations expressed in this thesis are those of the
author and do not necessarily reflect the views of the National Science Foundation (NSF).
I would like to thank my advisor, Dr. Marilyne Andersen, for her guidance throughout
this project. Additionally, I would like to thank American Tooling and Engineering, Inc.,
Spartech PDC, and Tanury Industries for all of their patience and their wonderful service
in fabricating the semi-ellipsoid. Also, Steve Kaye at Kayelites was incredibly helpful
in determining the correct spotlight for our application. Furthermore, I am grateful to
Zofia Gajdos for her thoughtful editing of my horrific grammar and sentence structure
and Victor Lum for his patient help with the programming of the beam shaper. Finally, I
would like to thank my husband, Ira Phillips, for his support- from keeping me company
on long nights to explaining MATLAB programming- I couldn't have done the project
without you!
1
Contents
1
Introduction
1.1 The need for natural light . . . .
1.1.1 Economic concerns . . . .
1.1.2 Psychological effects . . .
1.1.3 Physiological effects . . . .
1.1.4 Rendering concerns . . . .
1.2 Existing light directing strategies
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
9
9
9
10
11
12
15
19
19
20
21
25
26
27
28
2
Existing technology
2.1 Existing goniophotometers . . . . . . . . . . . . . . . . . . .
2.1.1 Scanning goniophotometers . . . . . . . . . . . . . .
2.1.2 Projection goniophotometers . . . . . . . . . . . . . .
2.2 Shortcomings of existing goniophotometers . . . . . . . . . .
2.3 Existing Heliodons for Scale Models Solar Simulation . . . .
2.3.1 Current Heliodons . . . . . . . . . . . . . . . . . . .
2.4 Development of an ellipsoidal goniophotometer and heliodon
3
Design and Development of the architecture of the goniophotometer/heliodon
29
system
29
3.1 Principle of goniophotometer/helidon operation . . . . . . . . . . . . . .
29
3.2 The dual illumination system . . . . . . . . . . . . . . . . . . . . . . . .
30
3.2.1 Placement of light sources . . . . . . . . . . . . . . . . . . . . . .
3.2.2 The heliodon light source . . . . . . . . . . . . . . . . . . . . . . .
31
34
3.2.3 The goniophotometer light source . . . . . . . . . . . . . . . . . .
3.2.4 Beam Shaper . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
39
4
Data collection
4.1 Development of a Semi-ellipsoid
4.2
4.1.1
4.1.2
Principle . . . . . . . . . . .
Design . . . . . . . . . . . .
4.1.3
Development
4.1.4
Validation . . . . . . . . . .
. . . . . . . .
Calibration of a color CCD camera to behave as a multipoint luminance
m eter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.2.1
Spectral Calibration . . . . . . . . . . . . . . . . . . . . . . . . .
2
46
46
46
47
49
53
54
55
4.2.2
4.2.3
Spectral calibration for continuous spectra . . . . . . . . . . . . .
Photometric Calibration . . . . . . . . . . . . . . . . . . . . . . .
68
71
5
Conclusion
73
6
Appendix
6.1 Positioning the Dedolight.
. . . . . ..
. . . . . . . . . . . . . . . . . . .
75
75
6.2
Positioning the beam shaper . . . . ..
. . . . . . . . . . . . . . . . . . .
75
6.3
Matlab Code
. . . . . . . . . . . . ..
. . . . . . . . . . . . . . . . . . .
76
6.3.1
Speclntegrate.m . . . . . . . ..
. . . . . . . . . . . . . . . . . . .
76
6.3.2
6.3.3
calcam.m . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . .
turnOffBadPixels.m . . . . . . . . . . . . . . . . . . . . . . . . . .
80
81
6.3.4
6.3.5
datamean.m . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
compare.m . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . .
81
82
6.3.6
opentif.m
. . . . . . . . . . . . . . . . . . .
83
6.3.7
integrate.m
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
84
6.3.8
6.3.9
6.3.10
6.3.11
6.3.12
photo2radio.m
radio2photo.m
rad2XYZ2.m
rgb2XYZ.m .
XYZ2rgb.m .
. . . . . . . . . . ..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
. . . . . . .
. .. . . . .
. . . . . . .
. . . . . . .
...
. . . .
3
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
84
85
86
87
88
List of Figures
1.1
The Kresge Chapel at MIT uses natual light in novel and beautiful ways.
1.2
The light in Kresge Chapel emerges from the moat to dance on the interior
walls. Unfortunately, electric lights, which are often used, cancel out this
... . ......................
....
effect . . . . . .
The altar in Kresge Chapel has a diffuse skylight over it which makes the
white marble appear almost supernatural. . . . . . . . . . . . . . . . . .
This advanced fenestration system has reflective blinds which can redirect
light in useful ways [Andersen et al., 2005b]. . . . . . . . . . . . . . . . .
This light pipe has a reflector in the dome to collect light from the exterior
of the building and send it down the pipe. In addition, the reflective
coating is located at the back of the dome to maximize collection of light
1.3
1.4
1.5
1.6
2.1
2.2
2.3
2.5
2.6
14
14
16
at low sun angles [oikos@ , 2006] . . . . . . . . . . . . . . . . . . . . . . .
17
The anadolic system is mounted on the exterior of the building, extending
into the interior. Light ducts allow the light to exit, permitting it to penetrate deep within the space. This photograph shows an implemented
anidolic system at Ecole Polytechnique Federale de Lausanne (EPFL)
[Scartezzini and Courret, 2002]. . . . . . . . . . . . . . . . . . . . . . . .
18
The ISE Goniophotometer is a scanning goniophotometer in which the
light detector rides on a rail which is rotated about the sample [ApianBennewitz and von der Hardt, 1998]. . . . . . . . . . . . . . . . . . . . .
pab@opto Goniophotometer is commercially available and allows flexibility
in detecter and light source selection [Apian-Bennewitz, 2006]. . . . . . .
The concept of the goniophotometer at the Universite de Rennes 1 in
France. A reflective box fits over the sample as the projection device
[D eniel, 2002]. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.4
13
The EPFL Goniophotometer uses a six-sided structure as its projection
mechanism. The CCD camera images each of these screens and combines
the results to characterize all incident angles [Andersen et al., 2005a]. . .
The LBNL goniophotometer uses a half-mirrored, hollow hemisphere as its
projection device. The CCD camera images the interior of the hemisphere
and is able to record data for all incident angles in one data collection
[Ward , 1992]. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
The heliodon built by the PEC group of UC Berkeley at the PG&E Energy
21
22
23
24
25
Center in San Francisco and a scale model being tested with the heliodon
[U C Berkeley, 2006].
. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4
27
3.1
3.2
3.3
3.4
3.5
3.6
The structure of the goniophotometer/heliodon [Ljubicic, 2005]. . . . . .
The final room layout . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
If the light source is too big or mounted too low, the heliodon itself might
block its light path. The light rays from the spotlight will hit the back
side of the heliodon instead of the mirror. If this is the case, then the light
will never reach its intended destination on the front side of the heliodon.
Mole Richardson Mole Beam . . . . . . . . . . . . . . . . . . . . . . . . .
An illuminance map of the Molebeam. The dip in intensity in the center
30
31
is due to the HMI bulb in the spotlight . . . . . . . . . . . . . . . . . . .
33
The shadow test- a transparent ruler was held three inches and one foot off
the ground in sunlight, outdoors, at noon and with the Molebeam on the
heliodon. a) outdoors, 1', b) outdoors, 3", c) Molebeam, 1', d) Molebeam,
3"
3.7
3.8
3.9
32
33
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
35
Spectrum of natural sunlight from [Wikipedia, 2006f] and the spectrum of
the Molebeam (the jagged curve) measured with an Ocean Optics spec36
trom eter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
The emission spectrum of a xenon lamp compared to natural light [xen, 2006].
The dotted line is natural light and the solid line is the xenon spectrum.
36
of
the
and
the
specifications
The xenon lamps available from Hamamatsu
L2274 150 watt lamp we purchased [AIM Digital Imaging, 2006b]. ....
37
3.10 Principle of point source and paraboloid . . . . . . . . . . . . . . . . . .
3.11 The xenon lam p . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
37
38
3.12 400 watt Dedolight spotlight . . . . . . . . . . . . . . . . . . . . . . . . .
39
3.13 Spectrum of the 400 watt Dedolight (the jagged curve) compared to the
spectrum of sunlight. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
40
3.14 Illuminance map of the Dedolight . . . . . . . . . . . . . . . . . . . . . .
40
3.15 The shadow test- a transparent ruler was held three inches and one foot
off the ground under sunlight, at noon, outdoors and with the Dedolight
on the goniophotometer. a) outdoors, 1', b) outdoors, 3", c) Dedolight, 1',
. . . . . . . . . . . . . . . .
41
3.16 The apparent beam on the goniophotometer with the beam shaper . . . .
3.17 The beam shaper "shapes" the light from the Dedolight so that the apparent beam on the goniophotometer is always circular. . . . . . . . . . . . .
3.18 The motion control chip circuit . . . . . . . . . . . . . . . . . . . . . . .
42
d) D edolight, 3" . . . . . . . . . . . . . . .
4.1
4.2
4.3
4.4
4.5
Optical principle of ellipses. Light from one focus is emitted, reflected off
the mirrored ellipsoid surface, and focused back to the other focus. . . . .
The sample and CCD camera with fisheye lens are located at the foci of
the semi-ellipsoid, along the major axis of the base. . . . . . . . . . . . .
The holding mechanism for the CCD camera. This setup allows the camera's fisheye lens to be flush with the table at the focus of the semi-ellipsoid
The final dimensions of the semi-ellipsoid . . . . . . . . . . . . . . . . . .
The data sheet sent to the fabrication company. . . . . . . . . . . . . . .
5
43
43
47
48
48
49
50
4.6
4.7
4.8
4.9
Four holes were placed on the edge of the semi-ellipsoid to accurately
position it on the goniophotometer. Two are located on the flange between
the semi-ellipsoid and the lip and serve to accurately position the semiellipsoid. The other two holes are found on the lip and their purpose is to
attach the semi-ellipsoid to the goniophotometer. . . . . . . . . . . . . .
51
The general reflection of aluminum as a function of wavelength [Griot, 20061. 52
In reflection experiments (a), light travels through the semi-ellipsoid, reflects off the sample, and is focused with the semi-ellipsoid to the camera.
In transmission experiments (b), light travels through the sample and is
focused with the semi-ellipsoid to the camera. . . . . . . . . . . . . . . .
52
The CIE Standard Colorimetric Observer functions: how a human perceives the color of monochromatic light. These curves indicate how sensitive the human eye's receptor's are for red, green, and blue for the visible
spectrum of light [Wikipedia, 2006d]. . . . . . . . . . . . . . . . . . . . .
4.10 The composite V(A) curve gives the overall sensitivity of the human eye
to radiation at different wavelengths [Wikipedia, 2006e]. It is not divided
into three colors as with Figure 4.9. . . . . . . . . . . . . . . . . . . . . .
4.11 Spectral calibration experimental setup. Light passes from a Labsphere
tungsten-halogen calibrated light source through a Spectral Products CM 110
im monochromator to a Labsphere SRS-99-010 pure white reflectance
standard. Reflected light is measured with a Minolta LS-110 luminance
meter and a Kappa DX20 color CCD camera. . . . . . . . . . . . . . . .
4.12 The RGB sensitivity of the CCD camera (top) and its associated errors in
averaging pixels (bottom ). . . . . . . . . . . . . . . . . . . . . . . . . . .
4.13 The Labsphere tungsten-halogen source does not emit light in wavelengths
close to the U V . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.14 The RGB sensitivity of the CCD camera (top) and its associated errors
in averaging pixels (bottom) when the xenon source is used instead of the
tungsten-halogen source. Note that the signal starts at 380 nm instead
of 430 nm (as in Figure 4.12) because the xenon source emits these wavelengths and the tungsten-halogen source does not. In future work, the
xenon source should be used for camera calibration in order to obtain an
accurate portrayal of camera behavior for these wavelengths. . . . . . . .
4.15 New spectral calibration experimental setup with the spectrometer instead
of the lum inance m eter . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.16 k is determined by dividing the luminance for the monochromatic light
at a given wavelength by the Y value determined in the above procedure.
This k value can now scale the X, Y and Z values so that they will compare
to what the camera imaged. . . . . . . . . . . . . . . . . . . . . . . . . .
4.17 The smaller curve (based on the RGBs the camera captured) must be
mapped onto the larger curve (the XYZs derived from the amount of light
present) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
6
56
56
57
59
60
61
63
66
67
4.18 The proposal of how the CIE Standard Observer function curves should be
divided by the filter system. Within these ranges, the XYZ values will be
unique, allowing monochromatic light with the same XYZs to be assigned
to a certain wavelength. The XYZs of quasi-monochromatic light should
be the summation of the XYZs of monochromatic light from these ranges.
69
The commands written for the camera calibration and how they relate.
SpecIntegrate is the root command that calls all of the other functions. .
76
6.1
7
List of Tables
2.1
A comparison of current goniophotometers in terms of their features how
they experimentally assess fagade components and fenestration systems
[Andersen and de Boer, 2006.
. . . . . . . . . . . . . . . . . . . . . . . .
8
26
Chapter 1
Introduction
1.1
The need for natural light
Buildings often use electrical energy to light a space when natural light could be used
instead. Recently, there has been an increased interest in understanding sophisticated
daylighting strategies, such as advanced fenestration systems, because of their ability to
exploit natural light (from the sun) for use in buildings. Advanced fenestration systems
can redirect natural light so that it is used more efficiently within a building. They might
block light at eye level which can cause glare and instead redirect it to the ceiling where
it can reflect, diffusing within the room. Using natural light has obvious advantages.
It reduces the need for electric lighting, which lowers energy costs and benefits the environment. Also, humans are more productive and healthy when they are exposed to
natural light rather than electric light. Natural light is the obvious choice; however, new
strategies are needed to increase its effective use.
1.1.1
Economic concerns
In 2001, it was estimated by the US Department of Energy, Energy Information Administration (EIA), that residential and commercial buildings used 39% of the United States'
total energy consumed [US Department of Energy, 2006]. Even in China, an emerging
economy, buildings account for 27.6% of the country's energy usage [Lam et al., 2006.
In the United States, estimates vary from 25-40% of the 39% being due to use of artificial lighting [Krarti et al., 2005] [Li and Tsang, 2005]. In China, two-thirds of the 27.6%
is due to electric lighting and heating and HVAC systems [Lam et al., 2006]. In Israel,
typical office buildings on a summer day use as much as a third of their total energy
consumed on artificial lighting [C.E.Ochoa and Capeluto, 2006].
In Southeast Asia, experts believe effective use of daylight could result in an overall energy reduction of 20% [Zain-Ahmed et al., 2002]. Even simple daylighting strategies such
as altering the size of a window can result in a reduction of 10% [Zain-Ahmed et al., 2002].
Commercially available daylight systems that are properly installed in areas with considerable daylight can result in savings of up to 45% [Granderson and Agogino, 2006].
Another study showed that daylighting strategy in conjunction with a "daylight-linked
automatic lighting control" which adjusts the amount of electric lighting in response
9
to the amount of natural light in a space, energy savings can be as much as 70%
[Li and Lam, 2003].
Additionally, use of natural light can decrease energy consumption in terms of cooling
and heating a building. In the summer, heat gains from electric lighting can cause a large
percentage of the total building cooling load. Because electric lighting produces heat
which can be alleviated by a building's HVAC system, the use of natural lighting provides
the potential to decrease this cooling load [Li and Lam, 20011. Furthermore, in the winter,
natural light can provide solar gains through the windows, decreasing a building's heating
load. However maximizing natural light does have disadvantages; solar gains that were
desirable in the winter can become a nuisance in the summer, increasing a building's
cooling load [Franzetti et al., 2004]. However, even after taking these factors into account,
it is estimated that, in general, any savings in electric lighting due to natural light usage
results in a subsequent reduction of one-third of the cost of operating the HVAC system
due to the reduced heating and cooling loads [Lam et al., 2006]. For optimal reduction
in HVAC and electric lighting costs, all of these variables must be taken into account
when designing a lighting strategy for a specific building [Franzetti et al., 2004]. But
most importantly, one must have the tools to analyze all variables in the equation.
1.1.2
Psychological effects
People are drawn to light; they will choose the more illuminated path when traveling
around a barrier [Taylor and Socov, 1974]. Another study indicated that people tend to
sit where there is more light, finding that people in a cafeteria preferred to sit facing
bright areas. When the pattern of light was changed to highlight a different area, more
people sat in the newly illuminated areas [Rea, 2000]. In a classroom setting, subjects
of a study would cross a room to sit at an illuminated desk. However, when the desk
was not highlighted, the subjects would most often sit at the desk closest to the door
[Yorks and Ginthner, 1987].
Natural light helps humans to maintain mental stability. For example, serotonin, a
neurotransmitter, has long been used to treat depression and aggressive behavior. A study
in Australia found a direct correlation between the luminosity of a given day and serotonin
levels [Bj6rksten et al., 2005]. Furthermore, lack of exposure to natural light may cause
symptoms of seasonal effective disorder (SAD). This condition occurs primarily during
the winter when individuals are not exposed to sufficient sunlight. Symptoms of SAD
include "overeating, oversleeping, carbohydrate cravings and weight gain, withdrawal
from friends and family, decreased sex drive, lack of energy, and in severe cases, feelings
of clinical depression" [Women's Health Weekly, 2005]. Unfortunately, symptoms similar
to those of SAD can occur any time during the year if a person spends too much time
indoors, where there is often not enough natural light. Even though people can use
expensive machines to simulate natural light to prevent this disorder, the obvious and
easiest solution is to maximize ones exposure to natural light.
Finally, people are more productive in areas lit by natural light. Companies have begun to recognize that an efficient use of energy and daylighting promotes their corporate
image and provides a better environment for their workers which can result in increased
productivity [Li and Lam, 2001]. A study of Elementary school students in 2003 found
10
that learning rates of students in classrooms with the most daylight were as much 21%
better than students in classrooms with the least daylight [Group, 2003b]. In an office
setting studies have shown that "Call Center" workers in cubicles with the best possible window view processed phone calls 6-12% faster than workers with no view at all
[Group, 2003c]. Additionally, the same study found that office workers with the best
views performed 10-25% better on tests assessing memory recall and mental function
than those office workers who had no view. Studies of department store sales indicated
that stores utilizing skylights to improve natural lighting within the store experienced
sales of as much as 40% more than those not exposed to natural light [Libby, 2003].
Additionally, during the California power crisis in 2001 when retailers in the state were
required to operate their stores at half lighting power, stores from this same department
store chain with skylights experienced an average of 5.5% increase in sales over their
usual totals [Group, 2003a]. These studies demonstrate the importance of natural light
in mental health and productivity.
1.1.3
Physiological effects
People are biologically drawn to natural light. Not receiving sufficient exposure to natural light can cause adverse physiological effects. For example, lack of exposure to natural
light, and more specifically, UV wavelengths, reduces the production of vitamin D. Vitamin D is important in many bodily functions. For instance, it aids in calcium uptake,
which is essential for strong bones. Studies have shown that in northern latitudes (where
there is less access to sunlight because of the suns solar course), vitamin D deficiency is
endemic. Vitamin D deficiency has also been associated with rickets (a childhood disease that deforms bones), osteoporosis, rheumatoid arthritis, multiple sclerosis, and even
cancer. Studies have also shown that rickets and the other diseases mentioned above are
much more common at higher latitudes, which suggests that a lower exposure to natural
light can, by affecting vitamin D production, cause illness. Although vitamin D is found
in certain foods, like fish, it is most efficiently made when the human body is exposed
to natural light [Raymond and Adler, 2005]. This is another reason that exposure to
sunlight is necessary for our health.
Additionally, exposure to sunlight helps to set our "biological clock." Many complex
organisms are governed by their circadian clock which prepares their bodies for activities that occur throughout the day. Mammals have evolved so that they are biologically
prepared to accomplish specific tasks at particular times of the day. For instance, predatory animals must have sufficient energy for hunting supplied to their muscles during the
time of day that they typically hunt. These animals have evolved to function on an approximate 24 hour cycle corresponding to the sun's course [Albrecht and Eichele, 2003].
Humans also perform better when their circadian clocks are regulated properly. Humans
appear to require exposure to natural light to notify their bodies that it is time to work.
Conversely, melatonin, a neurotransmitter responsible for regulating human sleep cycle, is
produced in response to a given day's photoperiod (the duration of light throughout that
day) [Sumovai et al., 2004]. These examples clearly indicate the importance of exposure
to natural light for human physiological health.
Finally, natural lighting is more comfortable to the human eye because physiologically,
11
we are better able to process white light emitted by the sun [Verilux, 2006]. In incandescent lighting, the spectrum is comprised primarily of red and near-infrared wavelengths.
Therefore, it has a warm glow. In fluorescent lighting, something that has become more
popular with new energy efficient light bulbs, the light often appears to have a blue tinge.
This effect, due to the short wavelengths that comprise many fluorescent lights, can cause
eye strain. The short wavelengths have scattering characteristics that make it difficult for
human eyes to focus [W.S. Strouse Watt, 2006]. Additionally, electric lighting can exhibit
"flickering," an effect resulting from variation in voltage supplied to the light source; this
can cause further eye strain in humans. When magnetic or electronic ballasts are used to
supply electricity to the fluorescent bulbs, the frequency of these fluctuations increases to
a degree that they are much less noticeable to humans [CCOHS, 2006]. A study concluded
that use of these ballasts decreased complaints of eye strain by 50%. Additionally, workers on higher levels of an office building who were exposed to more natural light because
of less sunlight obstruction from neighboring buildings complained of fewer symptoms,
like headaches, associated with eye strain from electric lighting [Wilkins et al., 1989].
Although technology is improving, natural light is still best.
1.1.4
Rendering concerns
Natural light and light sources that mimic it are considered by many to have a better
quality of light. The spectrum of a light source effects the way one perceives colors under
that light source. If two objects have the same pigment, an observer might perceive their
color differently under two different light sources [Rea, 20003. If the light source has a
blue-ish tinge (as with many fluorescent lamps), objects viewed under this light can also
have a blue-ish tinge depending on the objects properties and comparative environment.
Natural lighting allows one to view colors under conditions human are most adept at
processing. Skylights are often used in retail situations to increase sales because the
natural light they provide improves color rendition of the goods for sale [oikos@, 2006]
[Heschong et al., 2002].
In museum, artwork is most often lit with light from windows or light sources that
simulate natural light. However, in one instance, an artist, Dan Flavin used fluorescent
lights to highlight the harsh, minimalist ideas in his artwork [Scott, 2006]. He realized
that this light source would alter the appearance of his artwork.
Additionally, it is important that the level of light be appropriate in a space for the task
intended to be performed in that space. In an office environment, increased use of computers means that different light levels are needed for one work space [Osterhaus, 2005].
A computer needs a lower light level to maintain contrast, while a desk needs higher light
level in order to read printed pages. However, these two tasks are often performed by
the same person in the same workspace. If the visual conditions of the workspace are
inappropriate, it may be difficult to do one's work. Light redirecting strategies can be
employed to direct light to the desk, and redirect it away from the computer monitor, but
presently, there is not a clear strategy on how to implement the light directing systems
[Osterhaus, 2005]. Information about how the light directing strategies, such as advanced
fenestration systems will add to this knowledge, aiding in better working conditions.
Another beneficial effect of natural light is that it can enhance the space's aesthetic
12
Figure 1.1: The Kresge Chapel at MIT uses natual light in novel and beautiful ways.
value. The sunlight will highlight different portions of the space as the sun moves throughout the day. It may filter through parts of the building, leaving beautiful patterns where
it finally hits the ground. At the chapel at MIT (known as Kresge Chapel, Figure 1.1)
designed by Eero Saarinen, this effect is at its height.
The chapel is surrounded by a small moat. Inside the chapel, there is a gap between
the outer wall and a small interior wall that rises only 3 feet. The gap between the
exterior wall and this interior wall is glazed so that there is a horizontal hidden window.
These windows allow the reflection off the water in the moat to penetrate within the
chapel. To further exaggerate the light reflection, the walls are built with uneven bricks
so that the light has something extra on which to bounce off (Figure 1.2).
Even with this beautiful effect, electric lighting is often used in the chapel. Unfortunately, when artificial light is used, the overhead electric lighting cancels out the lighting
from the hidden window and prevents it from reverberating off the walls in its intended,
beautiful display (Figure 1.2).
Furthermore, there is an expansive, diffuse skylight over the altar (Figure 1.3). The
placement and quality of the skylight makes it seem as though the light is supernatural.
In addition, its supernatural quality is exaggerated by the use of white marble in the
altar, which makes the light seem brighter than it actually is. Also, there is a sculpture
which hangs behind the altar and under the skylight that reflects the light from above.
These techniques use natural light in a novel and beautiful way. With better understanding of natural light, more advanced effects might even be possible.
13
Figure 1.2: The light in Kresge Chapel emerges from the moat to dance on the interior walls. Unfortunately, electric lights, which are often used, cancel out this effect.
Figure 1.3: The altar in Kresge Chapel has a diffuse skylight over it which makes the white marble
appear almost supernatural.
14
1.2
Existing light directing strategies
Although it is important for humans to be exposed to natural light during the day, it
is often not possible for them to be outdoors in order to experience this light. Some
may have indoor jobs or find it too cold to go outside. Fortunately, windows can provide
natural light to indoor spaces, although the light they can provide is limited by factors
such as the building and window geometry, and a building's orientation and exterior
environment. Conventional windows alone rarely permit natural light to penetrate deep
within an enclosed space.
Because of these limitations, designers have turned to light directing strategies to
increase the amount of usable natural light within a building. A large variety of strategies
have been developed.
Technologies to utilize diffuse light
While natural light in buildings is desirable, direct sunlight through a window can be
a nuisance. A study of office workers found that performance decreased by 15-21% in
cubicles with greater glare potential [Group, 2003a]. In these situations, occupants will
often block the light with shading devices, decreasing or eliminating natural light, and
thus, its benefits. However, if the direct light can be converted into diffuse light, occupants
will be less likely to use shading devices to block the light.
For example, anisotropic coatings on windows allow for angular selection of light which
can decrease direct solar radiation while permitting transmission of a large portion of
diffuse natural light [Sullivan et al., 1998].
"Smart glazings," a special coating on windows, change their behavior in response
to different environmental stimuli. They can block light in the summer when the solar gains are undesirable and permit light to travel through the window in the winter when its cooler. A type of smart glazing, holographic optical elements (HOE), allow solar control by reflecting or redirecting light incident on a window coated with
HOE [James and Bahaj, 2005b]. Direct sunlight is reflected and diffuse light is allowed
to pass through, improving occupant comfort in terms of light level and temperature
[James and Bahaj, 2005a]. The holograms are printed on transparent films and sandwiched between two panes of glass and can be used for a variety of purposes such as
"gratings, zone plates, lenses, mirrors and any other type of optical element"
[H.F.O. Muller, 2005]. Use of HOE can reduce the temperature in a sunroom by 6.10
when as little as 61% of the windows are coated [James and Bahaj, 2005a].
Technologies that redirect light
Reflective coatings on blinds reduce absorption of light when it strikes the blinds, allowing
the light to bounce off the material and travel deeper into the space (Figure 1.4). These
are often used in advanced fenestration systems [Andersen et al., 2005b], a particular
interest to the area of research this thesis covers.
Furthermore, laser-etched glass and prismatic panels can reorient the direction of
incoming light so that it will hit the ceiling and reflect farther into the room, moving
light deeper into the space [Andersen, 2004] [Sweitzer, 1993].
15
--..--
.
Figure 1.4: This advanced fenestration system has reflective blinds which can redirect light in useful
ways [Andersen et al., 2005b].
Another technology, "microstructured sun-shading devices" are microstructures that
are partially coated to have special reflective, absorptive, and transmissive properties and
can be used in glazings such as a light-shading device. Their use allows redirection of
light in desirable ways [Walze et al., 20053.
Technologies that reflect unwanted light
Low-emissivity glazings are another example of interesting light redirecting strategies;
they permit visible natural light while reflecting infrared [Sweitzer, 1993.
Other devices to select or redirect light
While the following devices are not directly characterizable by this line of research, they
are important in advanced daylighting strategies.
"Light pipes," tubes mounted on the exterior of a building that have a reflective
coatings, let direct light from the sun bounce within the pipe with little absorption,
allowing the light to exit from the pipe farther into the space than traditional windows
permit (Figure 1.5). At the end of the pipe, there is a diffuser which spreads the light
throughout the space. This technology was invented to address the shortcomings of
skylights- light pipes do not have to be mounted directly above the area to be lit; they
can direct light to the area through reflective coatings within the pipe [oikos@, 2006].
Anidolic light ducts are another light directing strategy which transmits diffuse light
further into a building than is possible with windows (Figure 1.6). They are similar
to light pipes in that they collect light from the exterior of a building and, through a
reflective conducting structure, transmit light farther into the building than it would
16
Figure 1.5: This light pipe has a reflector in the dome to collect light from the exterior of the building
and send it down the pipe. In addition, the reflective coating is located at the back of the dome to
maximize collection of light at low sun angles [oikos@, 2006]
travel with windows. However, anidolic light ducts are mounted on the side of a building,
rather than on the roof, as with light pipes. Also, they can utilize diffuse light while light
pipes exploit direct light. Through careful design of the shape of the system, the number
of reflections within the system can be reduced, and thus, the effectiveness increased
(every time a light ray reflects, it loses some of its energy to absorption). Simulations of
this system suggested that the daylight factor could be increased by a factor of as much
as 2.7 in urban settings with the use of these systems[Scartezzini and Courret, 2002].
Although several light directing methods exist to be used in daylighting strategies,
studies have shown that only 10% of commercial buildings employ tactics to increase
natural light [Krarti et al., 2005]. On the other hand, as much as 50% use energy efficient lighting fixtures [Krarti et al., 2005]. These statistics clearly show that commercial
building designers and operators are interested in energy efficient strategies for their
buildings. But, due to a lack of information on daylighting strategies, building designers
and operators are reluctant to incorporate the technology into their buildings. Although
simple raytracing can produce generalizations about how these special coatings redirect
light and some have even been characterized by empirical measurement devices, it is likely
that they could be used more often and more effectively in buildings if more information
were available about all varieties.
17
\
Roller bhnd
0A7
gL-
.....9
..........
Arydolk cekwnts
......
Double
Figure 1.6: The anadolic system is mounted on the exterior of the building, extending into the interior.
Light ducts allow the light to exit, permitting it to penetrate deep within the space. This photograph shows an implemented anidolic system at Ecole Polytechnique F6d6rale de Lausanne (EPFL)
[Scartezzini and Courret, 2002].
18
Chapter 2
Existing technology
2.1
Existing goniophotometers
To plan effectively for daylighting strategies, we must have information about how solar
shading or daylight redirecting systems will behave under different lighting conditions.
In particular, physical properties such as directional reflectance and transmission information on advanced fenestration systems will further daylighting design. Manufacturers of these advanced lighting systems must be able to test their products to validate
and to improve upon them. Additionally, architects and building managers need general information about these systems in order to use them effectively in building design.
Also, programmers of light simulation software can integrate this information into their
programs to produce better lighting simulations, and thus, aid the manufacturers, architects, and building managers by providing them with more accurate lighting information
[Andersen and de Boer, 2006].
Simulation and modeling techniques have been performed multiple times to forecast
energy savings for different daylighting strategies [Sullivan et al., 1998],
[Li and Lam, 2001], [Franzetti et al., 2004]. However, quantitative experimental information is needed to maximize the utility of these models; experimental characterization can
give insight to systems that are not fully understood, improving the simulation and modeling techniques. Measurement devices are generally able to achieve acceptable accuracy
(specific accuracies discussed in the following chapter), but they are not perfect. However, even partially flawed experimental data is important. Reflection and transmission
characteristics of some materials are not known because their properties are too difficult
to accurately model mathematically. Experimental data can help fill this void. These
approaches taken together can improve understanding of these materials more than either
approach could on its own.
For advanced fenestration systems, the quantitative information can be described by
the photometric property described by Bidirectional Transmission (or Reflection) Distribution Functions (BTDFs or BRDFs). A BT(R)DF conveys the emerging light distribution for a given incident direction for a given material. Mathematically, it is expressed
as
L
1
BT(R)DF=-=ou[]
E
steradian
19
where
emerging luminance =[-2
and Es = incident illuminance =[2"]
BT(R)DFs are angle-dependent in both incidence and transmission or reflection
([Andersen and Scartezzini, 2003]). It gives the ratio of light that a given surface reflects
or transmits back into its environment along a particular direction to the amount of light
arriving at that surface. Several methods for determining the BT(R)DFs are currently
in use. In this project, we are developing a new type of goniophotometer to measure the
directional reflectance and transmittance used in calculating BT(R)DFs.
There are several types of goniophotometers, but all operate by collecting information
about emerging luminance for a given direction. There are two main ways to accomplish
this task. One is to scan all points in space for emerging luminance from a sample by
moving a detector. The other is to collect this data with the aid of a stationary multipoint detector in combination with a light collection system, thus reducing the data
acquisition time.
2.1.1
Scanning goniophotometers
Scanning goniophotometers are the most common. They can be easier to develop since
the detector signal can be easily converted into luminance. However, this strength is also
the source of the scanning goniophotometer's weakness. Because these goniophotometers
use a mobile detector, the data produced cannot be continuous. Therefore, data for the
areas in between the sample points must be interpolated. The possibility exists that
highly dynamic sets would not be properly characterized because of this interpolation
[Andersen and de Boer, 2006].
The first goniophotometer was developed at Lawrence Berkeley National Laboratory
(LBNL). It was used to test multi-layer fenestration systems. This data was used in
solar heat gain simulations. Comparison to existing data sets for solar heat gain values showed that the data obtained from the goniophotometer helped the simulations to
achieve accuracy within 10% [Andersen and de Boer, 2006].
At Fraunhofer Institute for Solar Energy Systems (ISE), their original goniophotometer consisted of two parallel ground steel shafts used as rails on which a linear detector rides around the sample of interest (Figure 2.1). They produced a second design
that improved upon the track system for the detector, reducing the measurement error
[Apian-Bennewitz and von der Hardt, 1998]. They were able to improve upon LBNL's
design by allowing flexibility in the dimensions of the sample to be tested. Also, they
worked on refining angular resolution in the data acquisition. Although the work on
improving angular resolution was successful, it did increase the data acquisition time.
The results obtained from this goniophotometer were tested against Ulbricht integrating
sphere measurements, and it was found that there was only a 20% discrepancy. Also,
when the goniophotometer data was judged against data from ray-tracing simulations,
another estimation tool in the field, the differences were as low as 5% for small incident
angles. However, the discrepancies increased up to 61% when data was compared for angles greater than 60' [Apian-Bennewitz and von der Hardt, 1998]. Currently, researchers
at ISE are working to improve their goniophotometer's performance with a better light
source and a moveable CCD camera as the data collection device, rather than a discrete
20
I 111
.
- -
__
; !pi_
__
-_
__7 1_
-
_
--
-
motor for moving detecour Iider
g detector kr
xeni...
lamipX
motor for vertical axis
Figure 2.1: The ISE Goniophotometer is a scanning goniophotometer in which the light detector rides
on a rail which is rotated about the sample [Apian-Bennewitz and von der Hardt, 1998].
detector [Andersen and de Boer, 2006].
pab@opto is an optical consulting company who have designed and produced a commercially available scanning goniophotometer(Figure 2.2). It is unique in that it allows
flexibility in detection devices and light sources used with the system
[Apian-Bennewitz, 2006].
At the University of Technology Sydney (UTS), they developed a goniophotometer
that operates in a similar way to the ISE goniophotometer. The TNO Building and
Construction Research in Delft, The Netherlands, also built a similar design. Their
machine was meant to characterize transparent insulating (TI) materials. Its data was
tested against data from an integrating sphere and the difference was only 10% for most
samples (although it was as high as 20% for others) [Andersen and de Boer, 2006].
Researchers at Cardiff University, UK proposed a novel twist to existing goniophotometers. They would like to add spectral assessment capabilities to future machines,
allowing one to perform an analysis of wavelength-selective glazings. In their existing
machine, the transmitted light in this goniophotometer is focused, with the aid of an offaxis parabolic reflector, to a optical fiber bundle [Breitenbach and Rosenfeld, 1998]. This
advancement permits sample sizes much larger than any existing goniophotometer. The
measurements obtained from this goniophotometer were compared to integrating sphere
results and the difference was about 10% [Andersen and de Boer, 2006].
Finally, researchers at the Technical University in Berlin (TUB) have built a goniophotometer with a spiral, scanning design [Aydinli, 1996]. However, it is now being replaced
by a new system of multiple sensors on a rotating arc [Andersen and de Boer, 2006].
2.1.2
Projection goniophotometers
Data collection of all incident or emerging angles utilizing scanning goniophotometers such
as those described above can be quite time consuming, taking anywhere from four to thirty
days to obtain [Andersen and de Boer, 2006]. Most scanning devices take approximately
21
__ __
I
Figure 2.2: pab@opto Goniophotometer is commercially available and allows flexibility in detecter and
light source selection [Apian-Bennewitz, 2006].
four days, but a device at DTU in Denmark (which is actually the device developed
at Cardiff University which they bought) takes up to thirty days due to the spectral
characterization [Andersen and de Boer, 2006]. A device that can detect multiple points
simultaneously, rather than just individual points, would accelerate this process. A CCD
camera is such a device and can be calibrated for use as a multiple-point luminance meter
when used in conjunction with a light collection device. In goniophotometers that use
this approach, light from the sample is reflected or transmitted onto a projection surface
on which the multi-point luminance meter is trained. The CCD camera images this
scene, and through a camera calibration process, is able to extract directional luminance
information. This technology is much faster than scanning goniophotometers, taking
about eight hours rather than several days [Andersen and de Boer, 2006]. In addition,
this approach has an added advantage of capturing a continuous scene, allowing it to
characterize high dynamic data sets accurately. However, the reliability of this system is
effected by the accuracy of the calibration procedure of the CCD camera [Andersen, 2004]
[Andersen and de Boer, 2006].
Presently, three projection goniophotometers have been built and are in operation: one
at LBNL for computer graphics applications, one at the Universit6 de Rennes 1 in France
to improve photo-realistic rendering, and one at EPFL to test advanced fenestration
systems. All use a CCD camera to capture data, but their projection surfaces differ
[Andersen and de Boer, 2006].
At Universit6 de Rennes, Deniel created a goniophotometer with a cube as projection
mechanism (Figure 2.3) [Deniel, 2002]. A spatial calibration was required to convert the
data into hemispherical coordinates needed for BT(R)DF characterization.
Andersen at Ecole Polytechnique F6derale de Lausanne (EPFL) in Lausanne, Switzer22
Light
source
I
I
I
-
/
01
yr
a
in
I
'4
'4
'4
N
,
Out
/
/
Sample
Figure 2.3: The concept of the goniophotometer at the Universit6 de Rennes 1 in France. A reflective
box fits over the sample as the projection device [Deniel, 2002].
23
Figure 2.4: The EPFL Goniophotometer uses a six-sided structure as its projection mechanism. The
CCD camera images each of these screens and combines the results to characterize all incident angles
[Andersen et al., 2005a].
land devised a goniophotometer that required only six data collections (Figure 2.4)
[Andersen, 2004] [Andersen et al., 2005a]. Reflection and transmission of light from the
sample traveled to a triangular, diffuse screen, six of which surrounded the sample. A
CCD camera imaged each screen, and after six 60' movements of the screen and camera
system, data from all incident angles was collected. As with the device at Universit6
de Rennes, a spatial calibration was necessary to convert the data into hemispherical
coordinates (In fact, this is the case for all projection goniophotometers.). Although this
method was more than two orders of magnitude faster than using a standard scanning
goniophotometer, it still took around eight hours to collect data from a sample. It was
recognized that a system which allows the camera to image all incident angles at once
would accelerate the process even further.
In the early 1990s, Ward created a goniophotometer that was able to reduce the number
of data acquisitions to one (Figure 2.5) [Ward, 1992]. A half-mirrored hemisphere was
attached to the goniophotometer and the light source shone through the hemisphere to
the sample, placed approximately along the central axis of the hemisphere. Light would
pass through the hemisphere to the sample and reflect off or transmit through the sample
into the interior of the hemisphere. The reflected or transmitted light would be collected
by the mirrored hemisphere and then, due to geometry of a hemisphere, be reflected from
the sample to a fisheye lens of the CCD camera, which was also placed roughly at the
central axis of the hemisphere. The CCD camera with the aid of the fisheye lens would
collect all of the data in one acquisition.
In order to obtain the most precise measurements, the camera and sample would both
be located exactly along the central axis of the hemisphere, a physical impossibility.
24
Front View
metal
Side View
cunerweight
stand/base
assmiblv
Source support
sample
=
Collimlated
liuht
source
~~
Sample/target
holder
---CCD camera
(
with fisheve lens
source
calcra
tal f-silvered
optical hemisphere
c isae
Half-silvered
plastic hemisphere
Metal stand
Figure 2.5: The LBNL goniophotometer uses a half-mirrored, hollow hemisphere as its projection device.
The CCD camera images the interior of the hemisphere and is able to record data for all incident angles
in one data collection [Ward, 1992].
Therefore, a complicated spatial calibration was performed to extract meaningful data.
On the other hand, if the hemisphere were replaced by a semi-ellipsoid, this shortcoming
could be eliminated, allowing for more precise data acquisition without the aid of spatial
conversion.
All of this information has been summarized and compiled in Table 2.1, which is an
excerpt from [Andersen and de Boer, 20061.
2.2
Shortcomings of existing goniophotometers
Current goniophotometers have several shortcomings. Data collection using scanning goniophotometers is a very slow process and only allows for discrete data measurement,
affecting its ability to measure highly dynamic data sets. Projection goniophotometers
allow for more rapid data acquisition and characterization of highly dynamic data sets,
but current models can still be improved to improve the time of data acquisition, allowing
measurement of much more data. Furthermore, except for the goniophotometer at Cardiff
University [Breitenbach and Rosenfeld, 1998], all current technology lack the capability
to provide information about how the color spectrum of light is redirected. However, the
time needed for this device to measure spectral information is too long to make the measurements practical and extensive. Even so, this information allows the characterization
of spectrally selective materials, an area that that should be explored further experimentally [Andersen and de Boer, 2006]. Additionally, when light enters a building through
a window, it carries heat with it. This is mainly due to near-infrared wavelengths of
light traveling through the window. No method exists to directly determine the solar
gains through a given material in a directional and time efficient manner. Because different materials reflect visible light and near-infrared wavelengths in different ways, it is
important to consider these materials's infrared transmissivity and the subsequent heat
gains when designing buildings that utilize these light-redirecting elements. Without this
consideration, it is possible to spend more on cooling a building than the savings from
25
Institute
BTDF
BRDF
Coverage
Direc
kcp
LBNL, USA
/
-
discrete
sample
ISE, Germany
/
/
discrete
BTF
pabRopto,
Germany
/
/
discrete
sample
TUB, Germany
/
-
discrete
sample
TNO, The
Netherlands
/
/
discrete
not publ.
-
-4 days
fsphere
DTU, Denmark
/
-
discrete
±0.50
/
-30 days
fsphcrc, g-val., an.
-
/
Time
Validation
~4 days
g-val. comp. (10%)
'-4 days
old: fsphcre (5%-61%);
new: tinder devlpmt
-4 days
not published
- 4 days
under devlpmt
Replica
(10%-20%)
-
model (-11%)
UTS, Australia
/
/
discrete
sample
EPFL,
/
/
full
sample
-
-~4 days
not published
8 hours
fsphere, an.model,
Switzerland
MIT, USA
ray-trac., BTDF
comp. (2%-14%)
/
/
full
sample
/
< 10 min
under devlpmt
Table 2.1: A comparison of current goniophotometers in terms of their features how they experimentally
assess fagade components and fenestration systems [Andersen and de Boer, 2006].
having a natural lighting system [Scartezzini, 2003]. The latter two improvements were
proposed by Andersen in her PhD thesis [Andersen, 2004].
2.3
Existing Heliodons for Scale Models Solar Simulation
When designing a daylighting strategy, it is also important to consider how a building's
structure interacts with sunlight throughout the day. Architects use certain generalizations (i.e. south facing windows are desirable for sunlight exposure, etc.) but these
provide insufficient information to predict how a building will be affected by sunlight.
Tools have been developed to give a designer a more accurate idea of how his or her
building will behave when exposed to sunlight.
A heliodon is a machine that allows sunlight simulation on scale models for a given
time of day and year at a certain place on earth. Often for beginning architecture courses
at educational institutions (for which the MIT device will be used), scale models of ones
building designs are built before CAD models are even approached. Although modern
light simulation software permits a similar study of a building design, heliodons allow
a quick investigation of scale models that could take hours of input (building the CAD
model, assigning materials, etc.) to a computer program. However, once architecture
courses start to require the construction of a CAD model for ones building design, light
simulation software can be a faster approach than heliodon studies. Additionally, CAD
models are more precise than scale models, so the information obtained from simulations
which use them can be more accurate. However, heliodons still provide the designer with
a physical, in-person view of the shadows on their scale model, a more intuitive approach
26
-
Figure 2.6: The heliodon built by the PEC group of UC Berkeley at the PG&E Energy Center in San
Francisco and a scale model being tested with the heliodon [UC Berkeley, 2006].
than computer programs. Taken together, simulations with a CAD model and shadow
tests of a scale model with a heliodon can provide more insight than either approach
alone.
In a heliodon, a "sun" is attached in a fixed position relative to the heliodon and
the heliodon is rotated to simulate the desired orientation relative to the sun. Using a
heliodon, a designer can determine how far natural light will likely penetrate into his
or her building at a certain time of year over the course of a day, given the building's
location.
2.3.1
Current Heliodons
One such example of a heliodon was built at the PG&E Energy Center in San Francisco
by the PEC group of UC Berkeley (Figure 2.6). It consists of an articulated table which
can rotate appropriately relative to a spotlight (the "sun") that is mounted on the ceiling
30 feet from the table. The spotlight is used in order to approximate collimated light (the
rays exiting the spotlight are quasi-parallel), a characteristic that the actual sun exhibits.
A camera is mounted to the moving platform to capture the change in shadows on the
exterior of the scale model through the heliodon's range of motion [UC Berkeley, 2006].
Heliodons can either be automated or manual; the heliodon at PG&E is automated.
Manual versions can be much less expensive because they do not require a complicated
control system or an advanced lighting system to mimic the sun. They can often be
used outdoors where the real sun, or even a cloudy day, can be characterized. However,
automated versions are often far more practical because they can simulate the Earth's
motion in a continuous fashion, providing a more complete characterization.
27
2.4
Development of an ellipsoidal goniophotometer and heliodon
Because the information gathered from goniophotometers and solar simulators are both
relevant to daylight design, it would be valuable to combine their functions into one
machine while also improving on the shortcomings of existing goniophotometers. This
project proposes to design and build a goniophotometer that will improve upon current
designs in the following ways: it will allow for rapid data collection, be able to detect
color information from the light emitted from the sample in order to characterize spectrally selective materials, and be able to measure heat information from the near infrared
spectrum to aid in balancing daylight considerations with their associated solar gains. In
addition, the structural design of this goniophotometer allows it to be used as a heliodon
as well. This goniophotometer/heliodon will be useful for conducting detailed assessments of the light redirecting properties of materials such as window coatings and novel
blind systems and also performing sunlight simulations on scale models.
The research described here focuses on several aspects of the overall project. A dual
illumination system is described in Chapter 3, Section 3.2, with one light for the heliodon
and one for the goniophotometer. For the goniophotometer application, the light collected
must be focused at the CCD camera. As discussed in Chapter 4, Section 4.1, an acrylic
semi-ellipsoid with a half-mirrored coating has been designed and fabricated for this
purpose. The light collected must then be analyzed. The procedure for this is described
in Chapter 4, Section 4.2. A color CCD camera is calibrated such that it has the same
spectral sensitivity over the visible range as the human eye and is able to produce a
luminance map of the captured scene. This calibration will permit the use of the camera
as a multi-point luminance meter and will allow it to be used eventually for the collection
of the light redirecting properties (or BT(R)DFs) of various materials of interest.
28
Chapter 3
Design and Development of the
architecture of the
goniophotometer/heliodon system
3.1
Principle of goniophotometer/helidon operation
To compile BT(R)DFs for use in effective daylighting design, it is necessary to know
the directional reflectance and transmittance of light for a given material for different
orientations of an advanced fenestration system. For architects to use a heliodon for
shadow simulation on their scale model, the heliodon must be able to simulate all locations
on Earth at different times of the day and the year. The goniophotometer/heliodon aims
to produce these various conditions through mechanical means.
This goniophotometer/heliodon simulates different illumination conditions via its rotation about two axes. As shown in Figure 3.1, the altitude axis runs through the side
supports of the machine and the azimuthal axis is normal to it. These two axes allow
the goniophotometer to simulate any location on Earth at a given time of year and at
a certain time of day. A computer interface has been developed that allows the user
of the goniophotometer/heliodon to automatically move the machine to the proper location for the given study. Further information about the detailed design, fabrication
and implementation of the structure of the machine can be found in [Ljubicic, 2005] and
[Clifford, 2006].
3.2
The dual illumination system
The ideal lighting design system for the goniophotometer/heliodon would have only one
light source, simplifying the experimental setup. However, practically speaking, it is
impossible to have one source for both applications. The beam produced by the light
source for the heliodon would be far too big for the goniophotometer application, flooding
the entire goniophotometer/heliodon surface (desirable for the heliodon application, but
not for the goniophotometer application). If the same light source were used for both
applications, an awkward system of diaphragms would have to be implemented to shape
29
Figure 3.1: The structure of the goniophotometer/heliodon [Ljubicic, 2005].
the beam down to the size necessary for the goniophotometer application. An attempt
was made to do this, but it proved to be unfeasible due to size constraints. In addition to
requiring a smaller beam, the goniophotometer requires a more uniform source (in terms
of light levels across the sample surface), a requirement easily met using a small beam
size. However, it proved impossible to achieve sufficient uniformity of the light using the
beam size necessary for the heliodon. Because of these difficulties, it was decided that
two separate light sources would be used, one for the heliodon application and one for
the goniophotometer application, would be used.
3.2.1
Placement of light sources
The next step was to determine the placement of the sources within the lab space. There
were several constraints for the placement of the larger light source for the heliodon.
The heliodon light source must provide nearly collimated light, so the degree of spread
must not be more than 50 which allows for approximation of the light as collimated
(justification for this assertion is discussed in Section 3.2.2), but the source itself must
be as small as possible to reduce costs. Therefore, the light must travel a long distance
in order to diverge enough to flood the entire surface of the heliodon. In addition, the
large light source must be carefully placed to prevent the heliodon from blocking the
light emitted from it when the heliodon is in a fully vertical position. Also, the space
available for the experimental setup was limited to a 12' by 14' room that had already
been constructed, so all of these constraints had to be met within a small space.
It was realized that a mirror could be used to increase the virtual beam distance of
the larger light source to the heliodon, and that the light source could be mounted on the
ceiling to prevent the heliodon blocking its light path. A Microsoft Excel optimization
with the solver function was performed to calculate the optimal location for the light
source and the mirror. It was found that the optimal location of the source was on the
ceiling above the heliodon and that the mirror would be most useful mounted in the corner
of the room, near the ceiling, opposite the spotlight, and rotated to a slight angle (60).
The goniophotometer/heliodon was placed in one corner of the room, 56 inches diagonally
30
Figure 3.2: The final room layout
from the corner, with enough space around it so that a person could still walk behind
it. The normal position of the goniophotometer to the spotlight is 21.930 off the vertical.
The layout of the room was modeled using Solidworks to verify that the proposed solution
fit within the room's space constraints (Figure 3.2). The goniophotometer light source
was not included in this optimization because the anticipated size was small. A mirror
system was not needed to produce the beam diameter necessary for goniophotometer
tests. Therefore, the goniophotometer source could be placed in the corner of the room
with its light beam perpendicular to the goniophotometer is in a vertical position.
3.2.2
The heliodon light source
Each light source had to be carefully selected by taking into consideration specific constraints in terms of color temperature, collimation, illuminated area, and spectral profile
such that its properties resembled the characteristics of the sun as closely as possible.
The light source for the heliodon had to be selected symbiotically with the room layout
design. If the source were too small, the degree of spread required to reach the necessary
spot size at the heliodon's surface would be greater than 5'. If it were too big, it would be
too expensive. If it were too big or mounted incorrectly, the heliodon itself might block
31
Figure 3.3: If the light source is too big or mounted too low, the heliodon itself might block its light
path. The light rays from the spotlight will hit the back side of the heliodon instead of the mirror. If
this is the case, then the light will never reach its intended destination on the front side of the heliodon.
its light path (Figure 3.3).
After researching available options, HMI (Hydrargyrum medium-arc iodide) spotlights
were selected as an appropriate choice for this application. HMIs are a mercury-halide
discharge short-arc lamps that exhibit a spectrum and color temperature similar to natural light. Using the Excel optimization to test available spotlight sizes for feasibility in
the overall design, it was found that a light source with an eighteen inch diameter would
work. The HMI source selected was available from Mole Richardson (Figure 3.4). The
spotlight has a minimum 5' spread, 5600 K color temperature and was rated at 6200 lux
at 25 feet (more than adequate for this project- since the light must travel less than 25
feet).
Once the spotlight was purchased and installed, a validation of the expected behavior
was performed. By using a lux meter, the uniformity of the light source on the surface of
the heliodon was found to vary by 7.4%, ranging from 17,000 to 22,880 lux (Figure 3.5) if
a small non-uniformity is not considered). This small non-uniformity is produced by the
HMI bulb itself in the spotlight; it blocks light emerging from the back of the spotlight.
Therefore, there is a dark area in the spotlight (the trough in Figure 3.5) which is the
32
Figure 3.4: Mole Richardson Mole Beam
x 10
Illuminance map of the Molebeam
2.5
X le
3,
2.5,
.2
1.5
1.5
E
0.5
0
40
0.5
10
0
00
20
310
Inches
40
0
Figure 3.5: An illuminance map of the Molebeam. The dip in intensity in the center is due to the HMI
bulb in the spotlight
vertical length of the spot and approximately five inches in width, meaning that it is only
about 4% of the total area of the spot. This area has a low value of 10,000 lux. Since the
heliodon is used mostly for qualitative measurements, this non-uniformity is undesirable,
but acceptable.
This illuminance map describes the illuminance of the Molebeam on the heliodon's
surface when the light beam is perfectly perpendicular to the heliodon surface. When the
heliodon is at it's most extreme position (its surface is parallel to the light beam), it is
estimated (by performing a calculation based on the behavior of light) that the variance
in light over the surface will be as much as 30%. However, this variance will only occur in
this extreme case for very grazing angles. Therefore, the light farthest from the spotlight
will not actually be generating shadows because the scale model itself will block this light
from reaching the farthest surface of the heliodon. The scale model will be experiencing
the sun as if it were sunrise, so it will actually only be subject to light from the front half
33
of the heliodon surface. Therefore, the variance of the light the scale model is subjected
to is actually only about 16% for this situation. At these light levels (10,000 lux and
above), humans are not able to perceive a difference between something that is a fifth as
bright or even a third as bright [AIM Digital Imaging, 2006a]. Since the heliodon is to
be used for qualitative measurements, these contrasts are considered acceptable.
A shadow test was performed to confirm that shadows produced by the Mole-Richardson
spotlight are similar to those produced by sunlight. Outside at noon on a sunny day, a
transparent ruler with dark gradations was held at three inches and then one foot above
the ground. The same ruler was held at three inches and then one foot above the surface
of the heliodon in its noon configuration. As shown in Figure 3.6, the shadows produced
by the Molebeam are less sharp than those produced by the sun, but they are comparable.
This qualitative assessment verifies that the light exiting the Molebeam is will produce
shadows similar to the sun.
Finally, the intensity of the light source as a function of wavelength (its spectrum)
was tested with the aid of an Ocean Optics spectrometer (Figure 3.7) which can measure
spectra up to 880 nm. The spectrum was found to be similar to other HMI lamps, and
although it is not an exact match to natural sunlight, it is recognized as a good substitute
for natural light. As mentioned previously, the Molebeam has a color temperature of 5600
Kelvin (compared to 6000 Kelvin from the sun) and the shape of its spectral curve roughly
matches that of natural light. However, it does lack wavelengths in the red range of the
visible spectrum when compared to natural light.
3.2.3
The goniophotometer light source
As with the light source for the heliodon, the light source for the goniophotometer must
mimic sunlight as closely as possible. It must be collimated (up to 5' of spread- justification for this value is discussed later in this section), have a color temperature similar
to the sun (approximately 6,000 Kelvin), have a spectral output similar to the sun, and
be uniform in its emission such that all areas are illuminated equally. These constraints
are even more important for the goniophotometer light source since it will be used for
quantitative measurements rather than the mostly qualitative measurements taken with
the heliodon source.
Although HMI lamps offer characteristics similar to sunlight, xenon lamps are superior.
They emit a more even spectrum which is closer to natural light (Figure 3.8) [xen, 2006].
They were much too expensive to use for the large heliodon source, but it was thought that
they could be utilized in the goniophotometer application. However, integrated systems
were cost prohibitive. Hence, it was decided that a light source would be designed and
built.
The design for the small source included a paraboloid with an aluminum reflective
coating. An L2274 150 watt xenon bulb (Figure 3.9) was purchased from Hammamatsu
which would approximate a point source. The design principle specified the point source
would sit at the paraboloid focus so that the paraboloid would collimate all exiting light
rays exiting the paraboloid (Figure 3.10).
This design was fabricated, assembled and tested (Figure 3.11). Initially, the results
were quite poor. Because the paraboloid was focusing the light and not completely
34
Figure 3.6: The shadow test- a transparent ruler was held three inches and one foot off the ground in
sunlight, outdoors, at noon and with the Molebeam on the heliodon. a) outdoors, 1', b) outdoors, 3", c)
Molebeam, 1', d) Molebeam, 3"
35
200
400 600 800
Wavelength (nm)
1000
Figure 3.7: Spectrum of natural sunlight from [Wikipedia, 2006f] and the spectrum of the Molebeam
(the jagged curve) measured with an Ocean Optics spectrometer
0100--
60-
... 20
300
-600 750
Wavelength (nm)
450
Figure 3.8: The emission spectrum of a xenon lamp compared to natural light [xen, 2006]. The dotted
line is natural light and the solid line is the xenon spectrum.
36
0
MAETA BASE
Figure 3.9: The xenon lamps available from Hamamatsu and the specifications of the L2274 150 watt
lamp we purchased [AIM Digital Imaging, 2006b].
Figure 3.10: Principle of point source and paraboloid
37
Figure 3.11: The xenon lamp
collimating it, it was clear that the bulb was not placed precisely at the focal point of
the paraboloid. Although the design allowed for fine adjustments, the exact position
of the xenon bulb's arc point was unknown due to manufacturing variations in xenon
bulbs. Also, because the cathode and anode (between which the arc point is created)
are encased in the bulb, measurements to determine this position were not completely
accurate. Additionally, the arc point is not a perfect point source. Therefore, it was
impossible to exactly place the arc point of the xenon lamp at the focal point of the
paraboloid even with trial and error, which was a requirement of the system. The behavior
improved, but not to the degree required by the system.
Furthermore, the xenon bulb itself exhibited strange diffraction properties through its
glass. With the paraboloid's focusing, these diffractions were quite noticeable and caused
a large variation in intensity of the spot. With methods available, it was impossible to
correct for these diffractions. Unfortunately, due to non-uniformities in the bulb and
adjustments needed to place the bulb with respect to the paraboloid, an acceptable
uniformity and collimation was never achieved.
After this setback, research began to find an integrated HMI source with the desired
optics. Even though the spectral characteristics of xenon bulbs are superior in terms of
mimicking natural light, integrated systems were still too expensive to consider. Therefore, HMI sources were reconsidered. The 400 watt Dedolight was found to meet the
requirements of the system (Figure 3.12). The Dedolight has a similar color temperature
and spectrum to sunlight due to its HMI bulb. It also produces a uniform beam with a
small degree of spread.
With data provided by Dedolight's manufacturer, it was confirmed that the Dedolight
has a spectrum similar to other HMI sources, which is considered to be a sufficient
substitute of natural light (Figure 3.13). As the figure shows, the Dedolight does not
have as much intensity in the red and near-infrared range. Because of this characteristic,
the light source does not produce as much heat as natural light. Also, the Dedolight has
a color temperature of 5600 Kelvin, another similarity to natural light which has a color
38
Figure 3.12: 400 watt Dedolight spotlight
temperature of 6000 Kelvin at noon [Wikipedia, 2006c].
With focusing, it is able to achieve a high level of uniformity. Over a 15 cm diameter,
the illuminance of the source was determined (with the aid of a LMT Pocket-Lux photometer) to vary by 8%, which is commonly considered acceptable in optics applications.
Figure 3.14 shows how the illuminance varies over the entire diameter. For extreme angles of the goniophotometer, this variance rises to as much as 30%. However, the only
way to improve this error is to have a spotlight with a smaller degree of spread. For HMI
spotlights, 50 is the smallest degree of spread available, so this is the best situation for
our given financial constraints.
A shadow test was also performed on the Dedolight. Just as with the Molebeam for the
heliodon, shadows of a transparent ruler were compared for sunlight and the Dedolight.
The images in Figure 3.15 confirm that they were similar. The gradations on the ruler
are easily viewed in the shadow when the ruler is held at three inches. At a foot, the
ruler becomes blurry in both cases. However, the behavior of the ruler in the sunlight is
superior to the Dedolight, as is expected since the sun only has a 0.25' spread, while the
Dedolight has a 10' spread.
3.2.4
Beam Shaper
For the goniophotometer application, it is important that the apparent beam on the
material being characterized is the correct size. When the goniophotometer rotates on its
altitude axis, the apparent beam will expand, becoming ellipsoidal. Without intervention,
the apparent beam will flood the surface of the goniophotometer when it is rotated to an
extreme angle (nearly parallel to the ground). The apparent beam will be the wrong size
39
200
1000
400 600 800
Wavelength (n)
Figure 3.13: Spectrum of the 400 watt Dedolight (the jagged curve) compared to the spectrum of sunlight.
Illuminance map
of Dedolight
x 10 4
1.34
1.32
-1.3
1.4
1.28
a 1.31
1.26
CD
1.2,
1.24
5
4
4
3
2
2
Figure 3.14: Illuminance map of the Dedolight
40
Figure 3.15: The shadow test- a transparent ruler was held three inches and one foot off the ground
under sunlight, at noon, outdoors and with the Dedolight on the goniophotometer. a) outdoors, 1', b)
outdoors, 3", c) Dedolight, 1', d) Dedolight, 3"
41
15 cm
15 cm
t
r'
Figure 3.16: The apparent beam on the goniophotometer with the beam shaper
unless something is put in its path to stop this excess light from reaching the table.
To stop this occurrence, a "beam shaper" has been developed (Figure 3.17). It consists
of an aluminum diaphragm that is rotated to the same angle as the altitude axis of the
table. From this action, the light beam from the Dedolight is "shaped" into the correct
size (Figure 3.16). The hinged flaps shown in the figure and in Figure 3.17 are intended
to block light when the beam shaper is rotated to grazing angles. Without these flaps,
the beam shaper diaphragm would have to be several feet long to block these grazing
angles. They are a passive system which can be flipped depending which way the beam
shaper is faced so that they always block unwanted light.
The rotation of the beam shaper is controlled with a stepper motor and a fabricated
motion control chip (Figure 3.18) which allows the motor to find a "home" position. Code
written in Visual Basic specifies the angle to which the beam shaper travels. The degree
of accuracy of the angle to which it can travel is determined by the amount of "steps"
to which the motor can discretely travel. For this particular motor, there are 1720 steps,
which means that there are approximately 0.21 degrees per step (1720p,). Therefore,
this fraction is the maximum resolution of this rotation.
The beam shaper must mimic the altitude axis of the goniophotometer in angle in
order to "shape" the beam.
The beam shaper is placed in the path of the Dedolight to the goniophotometer. It
is as close to the goniophotometer as possible. However, when using the goniophotometer/heliodon as a heliodon, the beam shaper must be moved out of position so that it
does not block the light path of the heliodon light source bouncing off the mirror. In the
future, a mechanism will be built which will allow the beam shaper to be moved easily
and replaced reliably in its correct position. Until that mechanism has been developed,
instructions for replacing the beam shaper can be found in Appendix 6.2.
42
Figure 3.17: The beam shaper "shapes" the light from the Dedolight so that the apparent beam on the
goniophotometer is always circular.
Sc
R2<
out
U,
RL
C1
L§7RI
*
_
Figure 3.18: The motion control chip circuit
43
R3
Also, when using the beam shaper, the spotlight must be focused on the beam shaper
rather than the goniophotometer surface in order to image the beam shaper. Otherwise,
the beam shaper will not work. Because of this constraint, beam uniformity is compromised. It is quite uniform at the beam shaper, but it has spread again once it reaches the
goniophotometer. To maintain acceptable beam uniformity, the actual size of the beam
on the goniophotometer's surface must be larger than the 15 cm originally specified. It
is large enough so that the center of the beam has sufficient uniformity, while the outer
edges decrease in intensity. This outer edge will be dealt with by applying a black velvet
diaphragm around samples. This diaphragm will absorb the light in this outer ring, effectively making the beam diameter a uniform spot of 15 cm. The optimal configuration was
determined by manually moving the spotlight relative to the beam shaper and focusing
it. From this trial and error procedure, the best setup was found to be a inner uniform
spot of 8% uniformity (as discussed previously) and an outer diameter of approximately
25 cm.
Validation
In this configuration, the beam shaper was able to shape light so that the apparent beam
on the goniophotometer is circular. However, future work must be completed to integrate
the beam shaper into the goniophotometer system. The code which controls the motion
of the beam shaper must be coupled to the code used to control the altitude axis of the
goniophotometer. Also, it was noticed that for extreme angles (near horizontal position),
the beam shaper did not make the apparent beam completely circular. A calibration
procedure must be performed to account for this irregularity.
Additionally, there are many features of the beam shaper which introduce an error
into it moving to the right position. For example, the mechanism for finding "home"
consists of a light diode, which when the light beam of the diode is broken, the circuit
is broken and the beam shaper can "know" where it is. A tab exists on the edge of the
beam shaper which breaks this diode beam. However, this tab is not negligible in size; it
is 0.03125" in thickness. Therefore, there is a slight discrepancy in its use because when
entering the diode beam, it cuts the beam with its front surface. When exiting, it exits
last with its back surface. Therefore, there is a slight delay in the signal which would
not occur if this tab were of negligible thickness. This thickness causes a discrepancy of
approximately 1'. However, it can be accounted for in the code.
Furthermore, the resolution of the angle, as discussed earlier, can cause a build-up
of errors over time. At this time, it is programmed to move to one angle and then
another with each calculation of degrees to steps occurring independently for each angle.
Therefore, if it is given the command to go to 600 and goes to 59.80', and then is given
the command to go to 1200, it will go to 119.60'. However, if the user "homes" the beam
shaper before each experiment, the errors will not accumulate. This homing is possible
within the user interface of the beam shaper and is recommended for its proper use.
Alternatively, the code could be altered to relate the degrees to steps for each final angle
rather than the intermediate ones.
Finally, until the beam shaper can be reliably placed in the exact position every time
after it has been removed for heliodon tests, there is will be an error associated with
44
replacing it. It might not be in the exact same position every time. In fact, this error
is expected to be much more significant than all other errors in the system. Once a
mechanism must be has been installed which can replace the beam shaper in the same
position every time, this error will be eliminated thereby vastly reducing this major source
of error.
Additionally, the beam on the goniophotometer need not be perfectly circular. The
primary purpose of the beam shaper is to prevent the light from entirely flooding the
goniophotometer's surface. Even with these errors, the beam shaper performs this task.
A more accurate measure of the beam shaper's performance its to measure the beam
exiting it onto the goniophotometer. This test was performed in Section 3.2.3 and it was
confirmed that the uniformity of the beam was 8%, which is considered acceptable for
optical applications.
45
Chapter 4
Data collection
To improve upon previous goniophotometer designs, we wanted to reduce the data acquisition time as well as provide a way to evaluate the spectral selectivity of various
materials. Most goniophotometers use a scanning principle to create a three-dimensional
profile of the light reflected or transmitted through a sample. At EPFL and LBNL, researchers were able to reduce the data acquisition time by projecting the light reflected
off of or transmitted through the sample onto a screen or a shell and measuring the luminance on these projection surfaces with a multi-point luminance meter [Ward, 1992]
[Andersen, 2004]. At LBNL, they used a CCD camera fitted with a fisheye lens so that
simultaneous imaging of the entire interior of the shell was possible. Like Ward and Andersen, we also decided to use a projection device (a hollow semi-ellipsoid) and a CCD
camera with a fisheye lens. For CCD cameras to provide accurate information about
BT(R)DFs, it they must be calibrated so that pixel intensity levels can be associated
with physical properties of light. Several methods have been devised by researchers to relate the pixel value a camera records with a luminance value. Once applied, the resulting
three-dimensional luminance map, can be used to calculate BT(R)DFs can be derived .
In this project, the camera setup captures an entire panorama panorama of either
reflected or transmitted light through a sample. This is possible with the aid of a light
collection device (a hollow semi-ellipsoid with a half-mirrored coating) which focuses the
light from the sample to the camera. The design, fabrication and installation of the semiellipsoid and the calibration procedure for the CCD camera are described in the following
chapter.
4.1
4.1.1
Development of a Semi-ellipsoid
Principle
An ellipse is an algebraic curve where the sum of the distances from any point on the
curve to two fixed points (the "foci") is constant. Optically, any light generated at one
focus of an ellipse or ellipsoid (an ellipse spun around it's major axis or minor axis so
that it is a three-dimensional object) will emit in all directions and be focused reflected
back onto the other focus of the ellipse if the interior of the ellipse is mirrored (Figure
4.1).
46
Figure 4.1: Optical principle of ellipses. Light from one focus is emitted, reflected off the mirrored
ellipsoid surface, and focused back to the other focus.
For this project, the sample is placed at the center of the rotating platform of the
goniophotometer. A CCD camera fitted with a fish-eye lens is embedded in the table
(Figure 4.3). Attached above the table is a semi-ellipsoid shell, which is mirrored on
the inside to behave as a one-way mirror. In reflectance measurements, light passes
through the shell to the sample where it is then reflected into the interior of the shell.
In transmission experiments, light travels through the sample and is reflected within the
shell. Because the sample is located at one focus of the semi-ellipsoid light reflects off
the mirrored surface to the other focus, where the camera lies (Figure 4.2). The sample
and camera can be approximated as point sources as required by the ellipsoid principle
because they are small relative to the overall size of the ellipsoid. The fisheye lens permits
full-scene imaging.
4.1.2
Design
Several factors were considered when determining the geometry of the semi-ellipsoid. The
semi-ellipsoid needed to be small enough so that it would not interfere with goniophotometer rotation, while big enough to allow one to approximate the non-finite sample as
a point source. The following characteristics describe these constraints more specifically.
" The major axis was held constant at 1150 mm since this was the maximum allowable
dimension for it to fit on the existing goniophotometer.
" Two constraints limited the size of the minor axis.
1. It must be sufficiently small such that the goniophotometer can rotate through
a full 360' with the goniophotometer attached. If the semi-ellipsoid were too
tall, it would interfere with this rotation.
2. It must be large enough such that the difference between the major axis and
focal distance is at least 10 times the sample diameter. This is needed so that the
sample will approximate a point source which is required to satisfy the optical
principle of using the semi-ellipsoid.
47
Camera
Sample
Figure 4.2: The sample and CCD camera with fisheye lens are located at the foci of the semi-ellipsoid,
along the major axis of the base.
Figure 4.3: The holding mechanism for the CCD camera. This setup allows the camera's fisheye lens to
be flush with the table at the focus of the semi-ellipsoid
48
Figure 4.4: The final dimensions of the semi-ellipsoid
o The focal distance must be maximized so that light emitted directly from the sample
to the camera is minimized.
These factors were optimized in a Microsoft Excel spreadsheet. The final design specified a major axis of 1150 mm and a minor axis of 1122.50 mm (Figures 4.4 and 4.5).
Additionally, the ellipsoid was selected to be made of acrylic since it is the most
transmissive plastic and is non-polarizing [Wikipedia, 2006a]. Acrylic transmits at least
90% of light throughout the UV, visible, and infrared spectrum (the typical value for
transparent acrylic). Also, to minimize distortions within the plastic, a refractive index
no greater than 1.5 was specified. Again, this value was the best for acrylic which would
minimize distortions. Finally, the thickness of the part needed to be kept as thin and
uniform as possible to maintain the refractive index of 1.5. An exact specification for the
thickness was not given, as it was informed by the manufacturing process.
4.1.3
Development
Due to the large size of the designed semi-ellipsoid, it was necessary to have an outside
company fabricate the part. An aluminum tool for thermoforming was machined by
American Tooling and Engineering, Inc. Using this tool, six semi-ellipsoids were made
by Spartech PDC by pulling melted acrylic sheets over the tool surface. An analysis of
these parts is discussed in Section 4.1.4.
To be able to accurately position the semi-ellipsoid on the goniophotometer- (which
is a vital step in data collection), -four holes were drilled into the lip of the semi-ellipsoid
(Figure 4.6). Two of the holes serve to place the semi-ellipsoid in an accurate position
on the goniophotometer (within 0.01"). They are fitted onto pegs protruding from the
goniophotometer's rotating surface. The pegs on the goniophotometer and holes in the
semi-ellipsoid have been very accurately placed to guarantee that the semi-ellipsoid's
49
2F~ (~.
0.395 i). thr
ra
~
____
WLle
i)HI - 5C
M
Hollow Semi-ellipsold
Courtney Browne
MIT
2 0.395 'n. thr hoe
Dirnensions irtn m niess
Figure 4.5: The data sheet sent to the fabrication company.
50
othemise
rreed
Securini hole
Stile hp
Plac e-findei
hole in flaiie
Figure 4.6: Four holes were placed on the edge of the semi-ellipsoid to accurately position it on the
goniophotometer. Two are located on the flange between the semi-ellipsoid and the lip and serve to
accurately position the semi-ellipsoid. The other two holes are found on the lip and their purpose is to
attach the semi-ellipsoid to the goniophotometer.
position with respect to the goniophotometer is correct.
The other two holes on the lip of the semi-ellipsoid are used to attach it to the goniophotometer. These holes correspond in position to two other holes on the rotating surface
of the ellipsoid. Bolts travel through both sets of holes and are secured by a nut. From
these two sets of holes, the semi-ellipsoid is placed accurately on the goniophotometer
and secured.
After the semi-ellipsoid was fabricated, it was sent to Tanury Industries to apply a
reflective coating. Again, this part of the project was outsourced because MIT did not
contain facilities to complete the coating.
Aluminum was chosen as the coating material because it reflects the complete spectrum
of white light in a uniform manner (around 90% reflection for an opaque coating-, Figure
4.7 [Griot, 2006]). We considered other coating materials such as inconel which gives a
more uniform reflection and transmission through the spectrum. However, it has a lower
overall reflection and transmission than aluminum (because it has a higher absorption).
Also, and inconel coating was cost prohibitive, so we decided to use aluminum.
If the percentage of reflection for a coating is 90%, the absorption and transmission
percentages sum to only 10% (the total of reflection, transmission and absorption must
equal 100%). For reflection experiments where light travels through the semi-ellipsoid to
the sample and then reflects to the lens (Figure 4.8), transmission and reflection had to
be balanced so that as much light as possible would reach the camera. The light is first
reduced when it passes through the ellipsoid (transmission). It is reduced again when
it bounces off the sample into the interior of the semi-ellipsoid and is reflected off the
coating.
Light at the camera
=
(%transmission)(%reflection)(original light level of Dedolight)
Therefore, reflection and transmission percentages would both ideally be 50%. How51
-
typkaf reNctame curwos
100
normal Inddence
~90
400
1000
800
600
WAVELENGTh IN NANOMETERS
Figure 4.7: The general reflection of aluminum as a function of wavelength [Griot, 2006].
Camera
Sample
Sample
b)
a)
Camera
Figure 4.8: In reflection experiments (a), light travels through the semi-ellipsoid, reflects off the sample,
and is focused with the semi-ellipsoid to the camera. In transmission experiments (b), light travels
through the sample and is focused with the semi-ellipsoid to the camera.
52
----------
ever, this is not possible in practice. There must be some absorption, and the reflection
and transmission percentages can rarely be perfectly balanced. Tanury Industries is still
working to determine the exact ratio they can achieve.
In transmission experiments, however, light will not need to travel through the semiellipsoid; light will only be reflected within the semi-ellipsoid itself (Figure 4.8). Therefore,
it is useful to maximize reflection and minimize transmission. Since Spartech was able
to make more than one useful shell for the project, we decided that two shells would
be coated-: one to be used for reflection experiments (balancing transmission and reflection as much as possible) and one to be used for transmission experiments (maximizing
transmission).
4.1.4
Validation
Because thermoforming is not a controlled process, the semi-ellipsoid is expected to have
imperfections. These imperfections will impact the effectiveness of the part. If the semiellipsoid is slightly the wrong size, light that is expected to travel from the sample to
the camera might not reach the camera because the semi-ellipsoid reflects it elsewhere.
Additionally, when air gets caught between the tool and plastic during the thermoforming
process, small bubbles can occur. These small surface imperfections will cause the same
effect as an ellipsoid that is the wrong size; - light will not be reflected back to the other
focus of the ellipse. We believe that the primary source of error in the semi-ellipsoid's
ability to reflect will be due to these small surface imperfections. This hypothesis was
confirmed by an inspection of the semi-ellipsoids after they were fabricated. They were
uniform in shape, but there were many minor surface imperfections. The manufacturer
used both cast and extruded acrylic in fabrication, and the extruded acrylic exhibited
less distortion than the cast versions. In the best two parts, we estimated these surface
imperfections to be less than 1% of the total surface area of the semi-ellipsoid. In a
worst-case scenario, the light reflected off of these surface imperfections will not reach
the camera at all. Therefore, less than 1% of light from the sample does not reach the
camera due to these surface imperfections. Even with this worst-case scenario, we believe
this percentage to be negligible. Therefore, the part should perform effectively.
A more exact calculation of the error can be performed after the semi-ellipsoids have
been aluminum coated. Unfortunately, this part of the project has not been completed
at this time. Once coated, we propose the following procedure for this error calculation
after the ellipsoids have been coated.
" A perfect reflector is used as a sample on the goniophotometer.
" The camera records, with the fisheye lens, the grey levels (0-255) of the light reflected
off the sample to the semi-ellipsoid to the camera.
" This signal is averaged.
" If the semi-ellipsoid, perfect reflector and light source were all completely uniform,
every pixel recorded by the camera would have the same level. Therefore, the perfect
reflector's uniformity or error (O-reflector) must be characterized. One might utilize
a standardized reflector in which this has been characterized by the manufacturer.
53
The light source's uniformity (Ulight source) has been characterized and can be found
in Section ??. Since these error sources multiply to give the total error, the total
error for a given position for the system is as follows.
=mtotal 1
where mtotal,
components.
mreflector,
reflector )2
7total
reflector
mlght
source,
+
( ('light source )2
mlight source
±
(,semi-ellipsoid
)2
msemi--ellipsoid
and msemi-ellipsoid are the averages of the given
" The standard deviation from the average of the pixels imaged by the camera can
be set equal to the total error (-total). The error associated with the semi-ellipsoid
(csemi-ellipsoid) can be solved for in this equation.
" This signal
noise ratio for the semi-ellipsoid will inform the calibration procedure for the
semi-ellipsoid. Spatial calibration procedures must be developed to account for this
error.
4.2
Calibration of a color CCD camera to behave as a multipoint luminance meter
For BT(R)DFs to be extracted from the image that the CCD camera captures when viewing the interior of the ellipse, the camera must be calibrated to behave as a multipoint
luminance meter. To do so, the picture is converted, pixel by pixel, into luminance values.
Calibrating a CCD camera to behave as a multi-point luminance meter has been performed in many other projects. More specifically, this process has been performed many
times for black-and-white cameras, with some examples described in [Andersen, 2004]
[Bellia et al., 2002]. The general principle is to calibrate the camera so that it has the
same sensitivity to light as the human eye, and then find the relation between the pixel
values of this calibrated camera to luminance values to be applied to images the camera
captures in the future.
Conversion of a color camera to behave as a multi-point luminance meter has been
the focus of more recent projects. For example, at LBNL, Inanici and Galvin used a color
HDR camera to capture light levels of a high dynamic range [Inanici and Galvin, 20043.
This procedure is possible because a HDR camera takes images for several integration
times and combines them into one image so one can see light levels of a larger range than
with regular cameras. At Ecole Nationale des Travaux Publics de l'Etat, Dumortier used a
Nikon Coolpix camera in which the integration time could be adjusted to capture this high
dynamic range [Dumortier et al., 2004]. Because luminance can be directly derived from
RGB values via a conversion to XYZ tristimulus values (another system for describing
colors in which the Y value is the measure of luminance [Wikipedia, 2006b]) both groups
used the camera's output to directly determine the luminance of a given pixel rather than
experimentally determining this relationship. They actually did not perform the spectral
calibration which would give the camera the same sensitivity to light as the human
eye. Therefore, their measurements ranged in error of 10-20% [Inanici and Galvin, 20043
[Dumortier et al., 20043.
54
For this project, we want to be able to evaluate the spectral selectivity of a sample.
Therefore, we used a color camera- a Kappa DX20 color CCD camera. This camera
allows adjustment in integration time to capture a large range of light levels. However,
since we want to deduce spectral selectivity, the camera needed to be calibrated so that
it had the same sensitivity to light as the human eye. From this calibration, the camera
will be able to evaluate the spectrum of light after it has interacted with a sample. After
this step has been completed, the same method used by Inanici and Galvin can be used
to convert RGB values into luminance for a given pixel [Inanici and Galvin, 2004].
4.2.1
Spectral Calibration
Often a camera does not have the same sensitivity to colors as the human eye. It will
be more efficient at imaging some wavelengths and less efficient at others. Therefore, to
determine the relationship between luminance and pixel values required to convert the
CCD camera into a multi-point luminance meter, the camera must be calibrated such
that it has the same sensitivity to colors as the human eye. When the color of light is
taken into consideration, this spectral sensitivity is specified by the CIE Standard Colorimetric Observer functions (Figure 4.9), which have been experimentally derived from
physiological measurements of how a human "standard observer" perceives monochromatic light [Wikipedia, 2006d]. It has been normalized to show the relative sensitivity
of a human to different wavelengths of monochromatic light. If one can make a camera
have the same sensitivity to this light as shown in this figure, the camera can be used
to determine luminance values a human observer would perceive directly from the RGB
values it captures, via a conversion to XYZ tristimulus values in which the Y value is
actually luminance. For this project, the calibration of the camera so that it has the
same sensitivity to colors as the human eye has been termed the "spectral calibration."
In previous works, researchers completed this task with the aid of physical filters
[Andersen, 2004], an impossibility for this project because, unlike previous projects, our
goal is to characterize light not just in intensity, but in color as well. This innovation
will allow the description of spectrally selective materials and light source variation. For
the spectral calibration, other projects only had to match one sensitivity curve (the V(A)
curve shown in Figure 4.10) for gray levels instead of three. Physical band-pass filters
(the types of filters used in applications such as these) often block all wavelengths above
or below a prescribed wavelength. Therefore, if physical filters were used, by matching
one curve, the other wavelengths would be filtered out. To circumvent this problem, a set
of three filters (one for each of the CIE Standard Colorimetric Observer functions- Figure
4.9) could have been used. We explored the possibility of coating three fish-eye lenses with
very thin filters to achieve this goal and found that the coatings were possible. However,
this approach would increase the data acquisition time because the fisheye lenses would
have to be changed in the middle of an experiment to obtain information about how a
sample behaves to the full visible spectrum of light.
Other projects using color cameras have ignored this spectral calibration and determined the degree of reliability one can expect from the measurements with an uncalibrated camera. As mentioned before, the measurements can be off by as much as 20%
[Inanici and Galvin, 2004] [Ward 1992].
55
CIE Standard Observer Curves
2
Red
Green
- - - Blue
I'
I
C)
U)
I
I
I
1.5
I
-0
I
0
CO)
=3
72
I
1
I
I
I
I
I
I
I
I:
E
co
-I.
w)
C,-
0.5
I
I
'I
2
I
0
400
450
500
550
600
650
Wavelength (nm)
700
750
Figure 4.9: The CIE Standard Colorimetric Observer functions: how a human perceives the color of
monochromatic light. These curves indicate how sensitive the human eye's receptor's are for red, green,
and blue for the visible spectrum of light [Wikipedia, 2006d].
Photometric sensitivity of the human eye
1
0.8
0.6
E
cz 0.4
0.2
0
400
500
600
Wavelength (nm)
700
Figure 4.10: The composite V(A) curve gives the overall sensitivity of the human eye to radiation at
different wavelengths [Wikipedia, 2006e). It is not divided into three colors as with Figure 4.9.
56
Figure 4.11: Spectral calibration experimental setup. Light passes from a Labsphere tungsten-halogen
calibrated light source through a Spectral Products CM110 Im monochromator to a Labsphere SRS-99010 pure white reflectance standard. Reflected light is measured with a Minolta LS-110 luminance meter
and a Kappa DX20 color CCD camera.
Ignoring the spectral calibration or using physical filters were not seen as an appropriate solutions because we wanted a very accurate measurement device that is quick in
its data collection (After much research, it was determined that our method is actually
not much faster than other methods. However, it will be able to characterize spectrally
selective materials, an innovation in this field. The time of data collection will be discussed later in this paper). Therefore, we decided that a new approach would be used.
We proposed that a computer program could take the place of physical filters by altering
the information the camera captured such that the final data were the same as if physical
filters were present. This approach was termed a "virtual filter." In this approach, the
"virtual filter" will alter the RGB values that the camera captures so that the camera
has the correct response to the light levels present.
To complete the spectral calibration, the relationship between the RGB values the
camera observes and the RGB values that a human would perceive must be determined.
To accomplish this task, an experiment is set up in which a Kappa DX20 color
CCD camera and a Minolta LS-110 luminance meter are pointed at a Labsphere SRS99-010 pure white reflectance standard with an average reflectance factor of 98% over
the visible spectrum at approximately the same distance and angle. A Labsphere calibrated tungsten-halogen light source is passed through a Spectral Products CM110 }m
monochromator (which can divide white light into monochromatic spectra) and hits the
reflectance standard. The camera images and the luminance meter takes a reading of
the monochromatic light (Figure 4.11). In order to calibrate what the camera observes
of the monochromatic light (in terms of RGB values) to what light the luminance meter
measures, the following experimental procedure was performed. First, what the camera
observes must be determined. Then, what the camera should observe is calculated and
57
compared.
Determining what the camera observes
" An image is captured with the Kappa CCD camera of wavelengths at intervals of
5 nm in visible light range (380-780 nm) as divided by monochromator. This 5 nm
division is appropriate to characterize the spectrum; anything greater would leave
out information, while anything finer would not add new information
[American Society for Testing and Materials, 20011. The settings for the camera are
as follows:
- 2.2 second integration time
- 0 gain
- 4095 HI
- 1200 LO: By altering HI and LO, one can change the range of levels which the
camera images. With this setting, a level of "0" corresponds to no light.
- 100 gamma (input'Y= output: -y is a percentage that describes the contrast. For
a -y of 100, the output pixel value is directly proportional to the input.)
- 1000 for R, G, and B
" The image file produced by the camera is imported into MATLAB with "imread.m,"
a function in the image processing toolbox. This function returns a matrix of intensity values for the three color channels (red (R), green (G), and blue (B)). Since we
are using an 8-bit camera, the intensity values range from 0 to 255. To manipulate
these pixel values, they must be converted to signed integers. This procedure has
been automated in the "opentif.m" file (Appendix 6.3.6).
" A single light intensity value is extracted from each image for each color channel
(R, G, and B) by means of an averaging algorithm implemented in MATLAB (see
Appendix 6.3.4).
This algorithm averages a prescribed range of pixel values. This range was determined by examining the spot created by the monochromator. Its size and uniformity
informed the choice of range; it appeared as though there were more than 10 pixels
of similar levels, so the range was chosen as 10. Additionally, the twelfth through
second pixels were averaged, throwing out the top two pixels to discard aberrantly
high, possibly false signal. The curve of these RGB values with their associated
errors due to this averaging can be found in Figure 4.12. As is shown in the figure,
the error bars are quite large (as much as 30% for some wavelengths) in relation to
the overall data set. This data seems to indicate that the uniformity of this spot is
not very good or that there were less than 10 pixels of acceptable uniformity. Examination of the spots imaged by the camera confirmed that the spot did decrease
in intensity at the outer edges such that there were less than 10 pixels of uniform
intensity. For future experiments, these errors can be reduced by having a larger
spot size with more pixels from which to sample so that this decrease in intensity
58
Camera Calibration for Threshold of 5
A
200
C,
0
Red
Green
Blue
100.0
coA
400
-
200 ---
--
450
500
650
700
750
650
600
550
Wavelength (nm)
700
750
600
550
R
Green
-Blue
100
0
400
450
500
Figure 4.12: The RGB sensitivity of the CCD camera (top) and its associated errors in averaging pixels
(bottom).
does not effect the results.
In addition, the Kappa DX20 color CCD camera appeared to have many pixels
that exhibited signal even when darkness was imaged (over 100 in some extreme
cases). We learned this was inherent camera behavior for long integration times
and the camera rising in temperature during operation. We deemed these false signals invalid and discarded them via the MATLAB functions turnOffBadPixels.m
and compare.m (see Appendices 6.3.3 and 6.3.5 for more information about how
these algorithms work). In the final experiment shown in Figure 4.12, this algorithm
eliminated 0.4% of total image pixels which were consistently exhibiting false levels
among many dark images.
o To be able to compare colors from light of different intensity levels, each RGB value is
converted to XYZ tristimulus values in which the Y value is a measure of luminance
[Wikipedia, 2006b]. XYZ values scale linearly with light intensity, while RGB values
do not [Rea, 2000]. Without this conversion, light comprised of the same wavelengths
but different intensity would have completely different RGB values and one would
not be able to compare light from different light sources that have the same color but
different intensity. For this situation, the XYZ values will be the same, just scaled
to reflect the light intensity. This conversion allows the recognition of two colors of
different intensities as the same color, while an RGB comparison would not. More
about this recognition is discussed at the end of this section and in Section 4.2.2.
The conversion is completed with the following equations [Inanici and Galvin, 2004]
59
Relative Intensity of the Labsphere tungsten-halogen lamp
0.15C 0.10-.1
0
0.050
400
450
500
550
600
650
Wavelength (nm)
700
750
Figure 4.13: The Labsphere tungsten-halogen source does not emit light in wavelengths close to the UV.
[Rea, 2000] [American Society for Testing and Materials, 2001].
X = k(0.412424R + 0.357579G + 0.180464B)
Y = k(0.212656R + 0.715158G + 0.072186B)
Z
=
k(0.019332R + 0.119193G + 0.950444B)
These equations have been automated in the MATLAB code rgb2XYZ.m which can
be found in Appendix 6.3.11.
In Figure 4.12, it appears as though the camera is only sensitive starting around 430
nm. However, this is not the case. Its sensitivity actually starts around 380 nm, as does
the human eye. This discrepancy is due to the light source used in these experiments; it
produces very little light in the wavelengths near the UV range (Figure 4.13). To verify
this assumption, we used the xenon source (spectrum found in Figure 3.8) discussed in
section 3.2.3 to perform the same experiment outlined above. The xenon source generates
wavelengths from UV to infrared, including all of the visible range. From this experiment,
the camera's sensitivity was determined for the entire visible range (380-780 nm) (Figure
4.14). As is evident in the figure, the error bars (due to the averaging function) for the
xenon lamp are much smaller (between 1-2%). This is due to the larger and more even
spot the xenon lamp was able to produce through the monochromator. However, it was
not as intense as the spot for the tungsten-halogen experiment, so longer integration time
was needed (8.1 seconds), which increased the number of bad pixels thrown out.
In order to characterize the error of the camera, in future work, this experiment should
be repeated with the xenon source at least three times to improve the probability that
the errors are accurately characterized. Because great care has been taken to eliminate
systematic errors (false signal due to the camera heating up, camera settings, etc.), the
variability between the results of these experiments can best be attributed to noise.
60
Camera Calibration for Threshold of 5
250
-
200 -
150 -
0
260
400
450
500
400
450
500
550
600
650
700
750
550
600
650
700
750
-
200 150 100S0
50
Wavelength (nm)
Figure 4.14: The RGB sensitivity of the CCD camera (top) and its associated errors in averaging pixels
(bottom) when the xenon source is used instead of the tungsten-halogen source. Note that the signal
starts at 380 nm instead of 430 nm (as in Figure 4.12) because the xenon source emits these wavelengths
and the tungsten-halogen source does not. In future work, the xenon source should be used for camera
calibration in order to obtain an accurate portrayal of camera behavior for these wavelengths.
61
This noise should be noted to characterize the final reliability of the camera within the
goniophotometer system.
Additionally, the errors due to the averaging function should be characterized as a
function of wavelength and intensity. We do not have enough experimental data at this
point to determine this relation. However, since the error in the averaging function is
related to spot uniformity, we expect that the percent error will be related to intensity
since it affects spot uniformity. Also, the errors could be greater because the intensity
of the signal is of similar magnitude to the noise in the system. This error will help
determine how reliably one can assign future XYZ values to the measured value from this
experiment. How this assigning occurs is discussed in the next section.
Determining what the camera should observe
In order to find the X, Y and Z tristimulus values the camera should observe to compare
with the derived XYZ values the camera captured, the amount of light (radiance) that
reaches the camera for each wavelength in 5 nm intervals from 380-780 nm must be
determined. Originally, this quantity was to be determined with a luminance meter placed
at an angle and distance to the reflectance standard similar to the angle and distance of
the camera to the reflectance standard. However, it was difficult to get accurate readings
with the luminance meter because it is difficult to keep ones hand steady to take the
reading. Measurements varied by more than 100% for very low light levels. Also, the
Labsphere tungsten-halogen calibrated light source produced very little light in the blue
range. Since the luminance meter is calibrated to have the same sensitivity as the human
eye (Figure 4.9), it was not very sensitive in this range, which compounded the problem.
Therefore, a spectrometer which can be kept steady easily was used to collect radiance
as a function of wavelength. This task can be accomplished if the luminance for white
light is measured, which is possible with the luminance meter because the intensity of
white light in much greater than the intensity of blue light generated by the Lapsphere
light source. The fluctuations in the values for the white light are around 5%. Our
spectrometer measures irradiance, which is a radiometric instead of photometric value.
Therefore, the data the luminance meter collected needed to be converted into radiometric
units for proper comparison. The relation between photometric and radiometric units are
related by the composite V(A) curve (Figure 4.10).
780nm
#photo
=
>
683 * V(A) * #(A),rdioA(A)
A=380
where
#photo
=
flux [lumens] and ca/rdi= flux [watts]
This relation can be used to convert any radiometric quantity (radiance, irradiance, etc.)
into its corresponding photometric quantity (luminance, illuminance, etc.).
Furthermore, the spectrometer must be placed along the axis of the monochromator
and light source so that it is directly in its path (allowing it to detect sufficient light
62
Figure 4.15: New spectral calibration experimental setup with the spectrometer instead of the luminance
meter
levels) rather than in the same position as the camera, as was the case with the luminance
meter (Figure 4.15). Placing the spectrometer in the path of the monochromator allows
the determination of the relative light flux exiting the spectrometer; provided that the
spectrometer is measuring some light, the size of the relative curve, and thus, the position
of the spectrometer relative to the monochromator, is not important.
A conversion is needed to relate the light detected by the spectrometer to the light
it would detect if it were in the same position as the camera or luminance meter. This
conversion is accomplished by relating the total amount of light recorded by the luminance
meter at the center of the spot produced by the monochromator to the amount of light
measured by the spectrometer, as a function of wavelength. This measurement can be
done with the luminance meter for bright, white light because it is around 20 1 for
white light instead of 0.1 - for blue light. In the case of the blue light, the error was
larger than the measured value, which meant the data is unreliable. Once the relationship
between the luminance reading and the spectrometer reading has been determined, all
light measured by the spectrometer in the path of the light source can be converted to
light the camera sees. This experiment is limited by the reliability of the luminance
reading for white light. Although a reliable reading can be taken (with an error of less
than 10%), better equipment which reduces this error would improve the reliability of
the experiment. The following mathematical procedure was performed to complete this
task.
e Variables:
total luminance at the luminance meter =
-
Lphoto,Ium =
-
L(A)photo,lum=
lumens 2
_
steradiai*m2]
luminance at the luminance meter, as a function of wavelength
steradian*m
-
L(A)radiolum =
[watts
radiance at the luminance meter, as a function of wavelength
63
=
-
E(A)radio,spec =
length
(Eradio =
relative irradiance at the spectrometer, as a function of wave[a"s], but since this is a relative value, units are not important.)
Note that these values exist for white and monochromatic light.
* Measured:
- Lphoto,lum
of white light
-
E(A)radio,spec of
-
E(A)radio,spec
white light
of monochromatic light
* Determine a for white light
Lradio,um
Eradio,spec
which defines the relationship between the light present at the luminance meter and
the light present at the spectrometer.
- E(A)radio,spec must be integrated over all wavelengths.
/780nm
E(A)radio,specdA
Eradio,spec =
A=380
- LphotoIum must be converted to L(A)radio,lum (photo2radio.m- Appendix 6.3.8)
780nm
Lphoto,Ium =
683 * V(A) * L(A)radio,lumA(A)
A=380
Because of the summation, it is difficult to isolate L(A)radio,lum. However, this
is possible because the relative radiance at the luminance meter is known. It is
the same as the relative radiance at the spectrometer; distance and angle will
not affect the relative radiance. However, to find the absolute radiance, which is
needed for determining ce, some algebraic manipulation is required. L(A)radio,lum
can be isolated from the equation by determining the wavelength at which it at
a maximum. Then each term of the summation includes L(A)radio,lum,max and
the fraction, (3), of the maximum it is at that wavelength.
L(A)photolum = L(A)radio,lum,max[683V(A) 3 8 0 3380 + ... + 683V(A) 78 o3 78 0]
L(A)radio,um,ma
L(A)photo,lum
=683V(A) 38 0 038 0 +
L(A)radio,lum,380 =
...
+ 683V(A) 78 0Q 78 0
3 8 0L(A)radiomax
L(A)radio,lum,385 =/ 3 385L(A)radio,max...
64
L(A)radio,um,780 =
3
780L(A)radio,max
From this procedure, the absolute radiance at the luminance meter (or camera)
can be determined. The above procedure has been automated in Appendix
6.3.8. Next, the total radiance of white light at the luminance meter must be
found by summing under the found curve to determine a.
780nm
Lradio,lum
L(A)radio,IumdA
-=
A=380
Finally, ce can be determined.
Lradio,lum
Eradio,spec
* This a must be applied to monochromatic light for each wavelength in 5 nm intervals from 380-780 nm measured by the spectrometer to determine the radiance of
monochromatic light that the camera observes.
L(A)radio,lum
-
aE(A)radio,spec
* Now the absolute radiance of monochromatic light that the camera observes has been
determined. The color of this light (X, Y, and Z) can be derived with the following
procedure (can be calculated with MATLAB file found in Appendix 6.3.10):
780nm
X
=
k EL(A)radio,lumT(A)AA
A=380nm
780nm
Y = k EL(A)radiolumg(A)AA
A=380nm
780nm
Z
=
k
L(A)radioIum7Z(A)AA
A=380nm
Where
- X, Y, Z = tristimulus values
-
spectral radiance distribution of monochromatic light from the
source, derived from previous procedure
L(A)radio,lum =
- !(A), Y(A), i (A) = spectral tristimulus value given for CIE standard observer in
The IESNA Lighting Handbook (Figure 4.9) [Rea, 2000]
65
Determining K Value
0)
cTz
01
400
450
500
550
600
650
700
750
400
450
500
550
600
650
700
750
550 600
650
Wavelength (nm)
700
750
2-
0
IiI-
I
C1
400
450
500
Figure 4.16: k is determined by dividing the luminance for the monochromatic light at a given wavelength
by the Y value determined in the above procedure. This k value can now scale the X, Y and Z values so
that they will compare to what the camera imaged.
- k
=
normalizing factor which makes Y equal the luminance for a given wave-
length ([Inanici and Galvin, 2004]).
Figure 4.16 shows how k was determined.
L(A)radio,spec -
The
L(A)phOt,Spc
(derived from
procedure outlined below) for monochromatic light was divided
by Yk.1 from the above summation. From this, the constant which relates
L(A)photo,spec and Yk=1 was found (k), which can be used to convert X, Y, and Z
into normalized values based on the amount of luminance present in the system.
To derive L(A)pho,,, 8 pc from L(A)radio,sp,,ec, complete the following summation for
each monochromatic spectra.
780nm
L(A)photo,Ium
S1
683 * V(A) * L(A)radio,lumA(A)
A=380
e Each XYZ value the camera observes can be compared to a value that the human eye
would render, based on the amount of light present, as determined in the procedure
above (Figure 4.17). This plot was rendered with SpecIntegrate.m (see Appendix
6.3.1).
Now that this relationship has been determined, there must be a way to apply this
relationship to images the camera captures. Ideally, the camera would capture an image,
and based on the pixel values, new pixel values would be assigned to account for the way
66
Camera XYZ and Expected XYZ values
based on radiance present
A M.
M.
M.
E.
E.
x E.
2-
I
:3
> 1.5 N
CO
Red
Green
Blue
Red
Green
Blue
1
0.5
400
450
500
650
550
600
Wavelength (nm)
700
750
Figure 4.17: The smaller curve (based on the RGBs the camera captured) must be mapped onto the
larger curve (the XYZs derived from the amount of light present)
the human eye would image the scene based on the amount of light present, rather than
how the camera imaged the scene. However, the camera does not capture the exact same
pixel values from acquisition to acquisition (due to slight variations in photons interacting
with the camera's sensor, slight variation in light levels, etc.). The average due to the
averaging function will be similar, but the exact pixel values will be different. This
anomaly is the main reason for the averaging function. Therefore, the exact relationship
between XYZs the camera captures to expected XYZs based on light levels derived in
the procedure above is unlikely to occur more than once. However, this relationship map
can be created, allowing the conversion of XYZ values the camera captures to proper
XYZ values based on the light present. The following procedure is a proposal for how
this assignment would occur.
" The ratio between the X, Y, and Z values is determined as a unique ratio for each
wavelength. This unique ratio is associated with the expected XYZs based on the
amount of light present (Figure 4.17). There will be a ratio for each 5 nm increment
from 380-780 nm.
" Ranges of the unique ratios are created for each of the monochromatic light experiments. The size of these ranges is statistically determined from the errors associated
with the averaging function to ensure correct placement of the experimental XYZ
values in the correct range (Section 4.2.1).
" Monochromatic light is imaged with the camera. The ratio between these experi67
mental X, Y, and Z values is found and they are placed in the appropriate range.
The experimental XYZ values will be a multiple of the original XYZ values from
which the unique ratios were determined. New XYZ values can be assigned after
multiplying the expected XYZ values based on the amount of light present by the
multiple the experimental XYZs are of the original ratios.
This approach is limited by the certainty of placing experimental XYZ values into
these ranges. There will be an error associated with this assignment. Additionally, there
will be an error associated with scaling the XYZ values with the appropriate multiple.
These errors must be determined experimentally after the characterization is complete
to determine the overall reliability of this assigning method. However, we expect these
errors to be negligible. Additionally, this approach allows for spectral characterization
of camera RGB values (via conversion to XYZ tristimulus values), which has not been
previously accomplished.
4.2.2
Spectral calibration for continuous spectra
Although this procedure is a novel approach, it does have a shortcoming. It only "maps"
light the camera sees to what the human eye would see for monochromatic light. In order
to make the procedure valid for all combinations of light, one must be able to identify
what amount of monochromatic light is a sum of the total continuous spectrum.
This separation is not specifically covered by this project. However, if a mechanism
were available to filter continuous spectra into monochromatic light, the computer based
spectral calibration would work. The information about how monochromatic light behaves with a sample could be combined to determine the response of the entire spectrum.
However, the filter system need not filter as finely as the spectral calibration required.
It must only filter to a degree which allows classification of the quasi-monochromatic
light into the original monochromatic divisions. We propose that this division would be
most appropriate at the peaks of the CIE Standard Observer function curves (Figure
4.18) because the XYZs in these ranges should be unique. From this point, the XYZ
values of the quasi-monochromatic light will be a sum of XYZs for the monochromatic
light within that quasi-monochromatic range. It has been proposed that the intensity
of these wavelengths can be determined by guessing what combination of the XYZs for
monochromatic light add up to the XYZs for quasi-monochromatic light. The range of
XYZs to try in this guessing are determined from the ranges mentioned in the previous
section. For instance,
XYZquasi-mono(550-560nm) = XYZmno(550nm)+XYZmono(555nm)+XYZmono(560nm)
If this summation can be determined, then the monochromatic content of the quasimonochromatic light can be determined. From there, the multiple of the ratios can be
applied to the XYZs based on the amount of light present.
For example, if the quasi-monochromatic light produced by the filter system ranged
from 550-570 nm, and the original XYZ values from the spectral calibration were as
68
CIE Standard Observer Curves
Red
Green
- - - Blue
2
a)
a)a,
0,
1.5
-0
0
1
II%
-a E
2 a)
0.5
----
0
400
450
500
650
600
550
Wavelength (nm)
700
750
Figure 4.18: The proposal of how the CIE Standard Observer function curves should be divided by the
filter system. Within these ranges, the XYZ values will be unique, allowing monochromatic light with
the same XYZs to be assigned to a certain wavelength. The XYZs of quasi-monochromatic light should
be the summation of the XYZs of monochromatic light from these ranges.
69
follows
550 nm: X =2, Y =5, Z =4
555 nm: X
=
1, Y =4, Z
=
6
560 nm: X
=
5, Y
=
2
565 nm: X
=
3, Y
570 nm: X = 2, Y
3, Z
=
5, Z =3
3, Z =-1
=
then the unique ratios of these monochromatic light experiments would be
550 nm: X:Y = 1, Y:Z = , X:Z = 1
555 nm: X:Y-=1, Y:Z =Z, X:Z =
560 nm: X:Y =,
Y:Z =-1, X:Z<= 5
565 nm: X:Y= , Y:Z =j,
570 unm: X:Y
=
j,
33
Y:Z
=
X:Z-= 3
3, X:Z= 2
The camera would be used experimentally to determine the spectral selectivity of a given
sample in the 550-570 nm range. If the sample transmits light that the camera images
as XYZ values of 3, 9, and 10 respectively. The ratios would be
X:Y =1, Y:Z=-, X:Z =These ratios do not fit in any of the ranges. However, it is clear that the XYZ values
from 550 and 555 nm sum to equal these new XYZ values.
XYZ 5 50 5-
70 nm
= XYZ5 50 lnm + XYZ 5 5 5 nm
X:3=2+1
Y:9=5+4
Z: 10= 4 + 6
Therefore, it is likely that the material is blocking 560-570 nm wavelengths and per-
mitting 550-555 nm.
Although the errors for this procedure have not yet been determined, it is recognized
that they can be minimized by selecting ranges which permit the reliable determination of these summations. However, smaller ranges mean longer data acquisition times.
Therefore, the error should be balanced with the desired data acquisition time.
70
4.2.3
Photometric Calibration
After the spectral calibration has been performed, the camera has the same sensitivity to
light as the human eye when one can subdivide the spectrum into quasi-monochromatic
spectra with the filtering system. Next, a calibration must be performed which relates
the RGB values the camera captures to luminance (the "photometric calibration"). The
experimental setup is the same as in the spectral calibration. The RGB levels of white
light are determined as a function of luminance level for a given integration time. The luminance level can be altered by varying the distance of the light source to the reflectance
standard or with filters. The range of integration times for this experiment is governed
by what integration time the camera needs to image the lowest light level the luminance
meter can measure and what integration time the camera needs to image (without saturating) the light the Dedolight produces at the particular distance of the setup and
after a reflection off the semi-ellipsoid. Once this relationship has been determined for all
integration times within this range, a pixel value can be used to determine the luminance
at that pixel for a given integration time.
Luminance can be derived from RGB values, if one knows the relationship between
luminance and integration time for the camera [Inanici and Galvin, 2004].
Y
=
k(0.212656R + 0.715158G + 0.072186B)
The Y value (of X, Y and Z) is equal to luminance.
Lphoto
The k is determined experimentally from the photometric calibration. It relates the
luminance the camera measures for a given integration time to the luminance that is
present and is unique for each combination of luminance and integration time (k*).
k*
=
Lpresent
Ldetermined with the camera
In order to determine the k for each integration time, the luminance for the white light
at a given distance should be determined, along with the Yk.1 value, based on the RGBs
the camera images at that distance. Then their quotient should be taken according to the
above relation and the k* value can be determined for a given luminance and integration
time.
*=
Lhoto
0.212656R + 0.715758G + 0.072186B
This procedure should be repeated for all relevant integration times and luminance
levels. After this constant has been determined for each integration time, luminances can
be derived directly from RGB values the camera captures. Code was written in MATLAB
(Appendix 6.3.11) which converts RGB into XYZ values. In future work, the k values for
different integration times must be integrated into this code.
71
Presently, the k value is a constant in the code (in rad2XYZ2.m - Appendix 6.3.10)
which is inputted when the function is called. It must be rewritten to allow the assignment
of a k based on an integration time. By listing k as a function of integration time in
an array in the code, and changing the input to be integration time, one can call the
appropriate k by inputting the integration time.
This experiment should also be repeated three times to determine the errors in measurement. By averaging the results from the three experiments, the error associated with
relating luminance to RGB values can be determined. The error arises from discrepancies in the RGB values the camera captures and the luminance values measured with
the luminance meter. When the results from three consecutive experiments are averaged, the mean (m) and standard deviation (a) of the luminance and RGB values can be
determined. The mean (m) and standard deviation (7) of the RGB values are as follows.
2
a2
URGB
2
MiRGB = 0. 2 12 6 5 6MR + 0. 7 151 5 8 MG + 0.0 7 2 18 6 mB
There will be a different CRGB and MRGB for each k*. Additionally, there will be a
different, corresponding ML and cL for each k*. The total error, O'k-, will be
ik* = nk*
(
GL
_L)2 +
mL
(
RGB
)2
mRGB
Once the photometric calibration has been completed, luminance values can be directly
derived, pixel by pixel from RGB values the camera captures with the aid of the filtering
system discussed in section 4.2.2. Within the entire goniophotometer system, the camera
will image the interior of the semi-ellipsoid and through a spatial conversion, be able to
characterize the BT(R)DFs which emitted from the sample. However, the speed of the
data acquisition is limited by this filtering system because a measurement must be taken
within each filtered range (Figure 4.18) to determine the spectral selectivity of a material.
The exact errors expected from this system have yet to be determined, but we expect
they will arise from the range chosen for the filters which will directly effect the certainty
of XYZ assignment as discussed in section 4.2.2 and the errors associated with the relation
between luminance and RGB values as discussed previously in this section.
72
Chapter 5
Conclusion
The work described in this thesis has focused on selecting and installing the light sources
for the two applications of our device and setting up the CCD camera used for collection
and analysis of the data produced by the goniophotometer application. A light source for
each application (heliodon and goniophotometer) was successfully selected and installed
in the room layout and they were found to exhibit a uniformity over their beam diameter of 7.4% (for the heliodon source) and 8% (for the goniophotometer source), a color
temperature similar to sunlight (5600 Kelvin) and a spectrum that is considered to be a
suitable simulator of sunlight.
In addition, this thesis features the creation of a projection device used in data collection. The developed projection device is an acrylic semi-ellipsoid, which fits over the
rotating surface of the goniophotometer and functions to focus light emitted from the
sample to the CCD camera such that the light can be characterized as a BT(R)DF. The
acrylic semi-ellipsoid has been fabricated, but due to delays in the outsourcer's schedule,
the coating of it has not yet been completed. In future work, the semi-ellipsoid must
be tested with the goniophotometer to characterize the system and coatings. A rough
estimation of the error associated with the semi-ellipsoid has been approximated to be
less than 1%. When it is coated, a more accurate characterization will be completed to
determine its reliability within the goniophotometer system.
Finally, the thesis discusses the calibration of a color CCD camera to behave as a multipoint luminance meter. This portion is the most innovative; although other researchers
have used color CCD cameras as multi-point luminance meters, they have neglected the
spectral calibration which gives the camera the same sensitivity to light as the human
eye. Without this calibration, the measurements the camera takes can be off by 10-20%
[Dumortier et al., 2004] [Inanici and Galvin, 2004]. A new approach has been proposed
in this thesis that will allow the spectral calibration involving a system which filters the
light into quasi-monochromatic spectra to permit the extraction of spectral data from
RGB values, and eventually a calibration which relates pixel value to luminance. The
errors associated with this calibration have yet to be determined. However, we expect
that there will be errors associated with the entire system to will be less than the 10%
accomplished by [Dumortier et al., 20041 and (Inanici and Galvin, 2004] since we will be
completing a spectral calibration.
Once complete, the goniophotometer will be the only one of its kind. In addition to
73
performing BT(R)DF characterization as the existing goniophotometers do, it will be
able to characterize spectrally selective materials and directional near infrared transmission through materials. With this new system, databanks of BF(R)DF information can
be quickly and easily created. This data will aid in advanced daylighting design; the
BT(R)DFs can be inputted into computer simulation programs to inform a building's
design. Generalized properties of these materials can be better determined so that building managers are more willing to use the materials. This information will foster better
understanding of novel materials and thus, promote sustainability in future design and
construction.
There is still much to complete on this project. The performance of the semi-ellipsoid
must be determined with the procedure outlined in Section 4.1.4. The camera calibration
must be finished, but the framework is in place to collect the data and quickly analyze
it (Section 4.2). An near infrared camera must undergo a similar calibration so that the
goniophotometer can evaluate near infrared transmission and reflection through these
materials of interest. Finally, all of these systems must be integrated to foster use amongst
building designers.
74
Chapter 6
Appendix
6.1
Positioning the Dedolight
The Dedolight is placed along the axis of the room with its light rays parallel to the
floor and perpendicular to the goniophotometer in its vertical position. Because the
Dedolight's proper position is in the middle of the lab, it is forseeable that it will be
bumped slightly out of position on occassion. Therefore, the following steps should be
performed occassionally to ensure that the Dedolight is in the proper position:
" Spotlight stand must be placed on painted markers in lab room.
" Check that the center of the lens is 36" off the ground. If not, loosen the knob on
the spotlight right above the stand and adjust the height. Retighten the knob.
" With the goniophotometer in a horizontal position, shine the Dedolight to the corner
of the room. Make sure that the light passes the goniophotometer and hits the corner
of the room in a centered way.
" Check that the the spotlight is set to have a 100 of spread (which is ok, because the
iris selects the central rays of the spotlight, meaning that those outputed do have 50
spread.
* The condenser lens should be retracted to its full extent.
6.2
Positioning the beam shaper
The following procedure should be performed when the beam shaper is returned to its
proper position for the goniophotometer application:
" Place the beam shaper within the marks painted on the floor.
" Make sure that the Dedolight has been focused properly by confirming that the lens
from the Dedolight is imaged on the beam shaper (it should have a textured look,
as the lens in the Dedolight is slightly textured).
75
SpecInteg rate6.m
photo2radio.m
radio2photo.m
rad2XYZ2.m
integrate.m
calcam.m
turnOffBadPixels.m
datamean.m
compare. m
opentif.m
Figure 6.1: The commands written for the camera calibration and how they relate. SpecIntegrate is the
root command that calls all of the other functions.
"
Adjust the iris so that the beam of light from the Dedolight is only slightly bigger
than the opening on the beam shaper. There is a piece of tape on it currently to
show where the iris should be adjusted.
" Adjust the height of the Dedolight so that the beam is centered on the goniophotometer. It should just graze the top opening of the beam shaper.
6.3
Matlab Code
This Appendix contains the MATLAB code created for the camera calibration.
ordered by the root command and the files that that command calls (Figure 6.1).
6.3.1
It is
Speclntegrate.m
SpecIntegrate determines XYZ values a human eye would render, based on the amount
of light present.
clear all;
close all;
cd .. ;
cd 22;
thres = 5;
A = calcam(thres);
sizeA = size(A);
plotxyz-result =
for i = 1:sizeA(1)
temp = cat(2,A(i,1),rgb2xyz([A(i,2) A(i,3) A(i,4)]));
plotxyz-result = cat(1,plotxyz_resulttemp);
end;
76
cd .. ;
cd CameraCalibration;
rho = 0.987173;
luminancemeter_wl = 21.5;
darkfilename = 'black.txt';
[A B] = textread(dark-filename,'Xf %f',
dark-spectrum
[A B];
clear A;
[A B];
for i = 1:2048
spectrum(i)
end;
clear A;
%Read in two column vectors:
% wavelength, relative irradiance
clear B;
white-filename = 'O.txt';
[A B]= textread(white-filename,'Xf %f',
spectrum =
2048);
clear B;
2048);
%Read in two column vectors:
% wavelength, relative
% irradiance
XCreat e matrix C by combining A and B
= spectrum(i)-darkspectrum(i,2);
radio-output = photo2radio(luminance-meter-wl,spectrum);
radianceluminance-meter = radio-output(:,1:2)
%Find radiance with
% photo2radio
% Take wavelength and radiance
radiance-spectrometer(:,1) = spectrum(:,1);
radiance-spectrometer(:,2) = spectrum(:,2) * rho / pi;
alpha = integrate (radiance luminancemeter)/integrate (radiance-spectrometer);
figure(5);
plot(radiance-spectrometer(:,1),radiancespectrometer(:,2),"'r');
hold on;
%Loop to Process Files
for i = 1:81
wavelength = 375 + 5 * i;
%Create File Names
filename = [num2str(wavelength) '.txt'];
[A B]= textread(filename,'f .f', 2048);
C = [A B];
clear A;
clear B;
%Create matrix C
for
j
= 1:2048
C(j,2) = C(j,2) - dark-spectrum(j,2);
end;
C_irr = C;
luminance(i,1) = wavelength;
luminance(i,2:3) = alpha * radio2photo(C-irr);
C(:,2)
=
C(:,2)*rho/pi;
%convert
77
to radiance
plot(C(:,1),C(:,2));
C(:,2) = C(:,2) * alpha;
XYZ(i,2:4)=rad2XYZ2(C(:,1:2),1);
XYZ(i,1)=wavelength;
k = luminance(i,2) / XYZ(i,3);
XYZ(i,2:4)=rad2XYZ2(C(:,1:2),k);
end;
figure(1);
plot(XYZ(:,1),
hold on;
plot(XYZ(:,1),
plot(XYZ(:,1),
XYZ(:,2),
'r');
XYZ(:,3),
'g');
XYZ(:,4),
'b');
xlabel('Wavelength (nm)');
ylabel('tristimulus (XYZ) values');
title('Expected XYZ values based on radiance present');
figure(2);
subplot(2,1,1);
hold off;
plot(plotxyz-result(:,1),plotxyzresult(:,2),r)
axis([350 800 0 1]);
hold on;
plot(plotxyz-result(:,1) ,plotxyzresult(:,3),'g');
plot (plotxyz-result(:,1) ,plotxyz-result(:,4),'b');
ylabel('Measured/Converted Chromaticity Coordinates');
title(['Camera Calibration for Threshold of ' num2str(thres)]);
x_bar
=
y-bar =
z_bar
=
[ 0.0002 0.0007 0.0024 0.0072 0.0191 0.0434 0.0487 0.1406 0.2045 0.2647
0.3147 0.3577 0.3837 0.3867
0.0805 0.0411 0.0162 0.0051
0.2365 0.3042 0.3768 0.4516
1.0142 1.0743 1.1185 1.1343
0.6475 0.5351 0.4316 0.3437
0.0409 0.0286 0.0199 0.0138
0.0010 0.0007 0.0005 0.0004
0.0000 1 ;
[ 0.0000 0.0001 0.0003 0.0008
0.0387 0.0496 0.0621 0.0747
0.2536 0.2977 0.3391 0.3954
0.8752 0.9238 0.9620 0.9822
0.8689 0.8256 0.7774 0.7204
0.2835 0.2283 0.1798 0.1402
0.0159 0.0111 0.0077 0.0054
0.0004 0.0003 0.0002 0.0001
0.0000 1];
[ 0.0007 0.0029 0.0105 0.0323
1.5535 1.7985 1.9673 2.0273
0.3707
0.0038
0.5298
1.1240
0.2683
0.0096
0.0003
0.3430
0.0154
0.6161
1.0891
0.2043
0.0066
0.0002
0.3023
0.0375
0.7052
1.0305
0.1526
0.0046
0.0020
0.0895
0.4608
0.9918
0.6583
0.1076
0.0037
0.0001
0.0045
0.1063
0.5314
0.9991
0.5939
0.0812
0.0026
0.0001
0.0088 0.0145
0.1282 0.1528
0.6067 0.6857
0.9973 0.9824
0.5280 0.4618
0.0603 0.0441
0.0018 0.0012
0.0000 0.0000
0.2541 0.1956 0.1323
0.0714 0.1177 0.1730
0.7938 0.8787 0.9512
0.9597 0.8563 0.7549
0.1122 0.0813 0.0579
0.0031 0.0022 0.0015
0.0001 0.0001 0.0001 0.0000
0.0214 0.0295
0.1852 0.2199
0.7618
0.9556
0.3981
0.0318
0.0008
0.0000
0.8233
0.9152
0.3396
0.0226
0.0006
0.0000
0.0860 0.1971 0.3894 0.6568 0.9725 1.2825
1.9948 1.9007 1.7454 1.5549 1.3176 1.0302
78
0.7721 0.5701 0.4153 0.3024 0.2185 0.1592 0.1120 0.0822 0.0607 0.0431
0.0305
0.0000
0.0000
0.0000
0.0000
0.0000
wavelength =
0.0206
0.0000
0.0000
0.0000
0.0000
];
0.0137
0.0000
0.0000
0.0000
0.0000
0.0079
0.0000
0.0000
0.0000
0.0000
0.0040
0.0000
0.0000
0.0000
0.0000
0.0011
0.0000
0.0000
0.0000
0.0000
% CIE standard observer data
[380 385
460 465
540 545
620 625
700 705
780];
390
470
550
630
710
395
475
555
635
715
400
480
560
640
720
405 410 415
485 490 495
565 570 575
645 650 655
725 730 735
420
500
580
660
740
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
(tristimulus values)
425
505
585
665
745
430
510
590
670
750
435
515
595
675
755
440
520
600
680
760
445
525
605
685
765
subplot(2,1,2);
hold off;
plot(wavelengthx_bar,'r--');
axis([350 800 0 2.5]);
hold on;
plot(wavelength,ybar,'g--');
plot(wavelength,z-bar,'b--');
xlabel('Wavelength (nm)');
ylabel('CIE Standard Observer Tristimulus Values');
figure(3);
subplot(1,1,1);
plot(plotxyz-result(:,1),plotxyz-result(:,2) ,'r');
axis([350 800 0 2.5]);
hold on;
plot(plotxyz-result(: ,1),plotxyz-result(:,3),'g');
plot (plotxyz-result(: , 1),plotxyz.result(:,4),'b');
ylabel('Camera XYZ/Light Source XYZ');
plot(XYZ(:,1), XYZ(:,2), 'r--');
plot(XYZ(:,1), XYZ(:,3), 'g--');
plot(XYZ(:,1), XYZ(:,4), 'b--');
xlabel('Wavelength (nm)');
ylabel('tristimulus (XYZ) values');
title('Camera XYZ and Expected XYZ values based on radiance present');
figure(4);
subplot(3,1,1);
plot(luminance(:,1),luminance(:,2));
ylabel('Luminance');
subplot(3,1,2);
plot(XYZ(:,1),XYZ(:,3));
hold on;
plot(540,0.02,'*r');
ylabel('Y');
subplot(3,1,3);
plot(XYZ(:,1),XYZ(:,3)./luminance(:,2));
ylabel('Y/Luminance')
79
0.0000
0.0000
0.0000
0.0000
0.0000
450
530
610
690
770
455
535
615
695
775
for i = 1:81
k(i,1)=luminance(i,1);
k(i,2)=luminance(i,2)/plotxyz-result(i,3);
end;
figure(6);
subplot(3,1,1);
plot(k(:,1),luminance(:,2));
ylabel('luminance');
subplot(3,1,2);
plot(k(:,1),plotxyz-result(:,3));
ylabel('Y');
subplot(3,1,3);
plot(k(:,1),k(:,2));
ylabel('k');
6.3.2
calcam.m
Calcam.m plots camera sensitivity.
function calresult = calcam(thres)
%calcan(thres) returns vector
[wavelength r g b rSD gSD bSD]
% thres is a noise threshold for discarding bad pixels.
%Calculates calibration curve for a given directory of data
/Assumes noise files are 'noisel.tif' and 'noise2.tif'
IAssumes image files are '420.tif' '480.tif' etc.
%Outputs output.txt with columns of wavelength rgb and rgbSD
i = 1;
noiseArray
while(1)
tempName = ['noise' num2str(i) '.tif'];
try imread(tempName);
catch break;
end;
noiseArray = strvcat(noiseArray, tempName);
i = i + 1;
end;
calresult = El;
for i = 1:81
wavelength = i*5 + 375;
imagel = [num2str(wavelength) '.tif'];
try calresult = cat(1,calresult,
[wavelength turnOffBadPixels(imagei,noiseArray,thres)]);
end;
end;
try
dlmwrite('output.txt', calresult,
end;
'delimiter',
80
'\t',
'newline', 'pc');
sizeCal = size(calresult);
subplot(2,1,1);
hold off;
plot(calresult(:,1),calresult(:,2),'r');
axis([calresult(1,1) calresult(sizeCal(1),1) 0 260]);
hold on;
plot(calresult(:,1),calresult(:,3),'g');
plot(calresult(:,1),calresult(:,4),'b');
ylabel('8-bit Intensity');
title(['Camera Calibration for Threshold of ' num2str(thres)]);
subplot (2,1,2);
hold off;
errorbar(calresult(: , 1) ,calresult(: ,2) ,calresult(: ,5), 'r');
axis([calresult(1,1) calresult(sizeCal(1),1) 0 260]);
hold on;
errorbar (calresult (: ,1) , calresult (: ,3), calresult (: , 6), 'g');
errorbar(calresult(:,1),calresult(:,4),calresult(:,7),'b');
xlabel('Wavelength (nm)');
ylabel('8-bit Intensity');
6.3.3
turnOffBadPixels.m
turnOffBadPixels.m takes information from compare.m to determine if pixels are "bad"
(see section 6.3.5 for how this determination is performed). If it is a bad pixel, and it
throws it out.
function output = turnOffBadPixels(tifname,noiseArray,thres)
imagel = opentif(tifname);
badpix = compare(noiseArray,thres);
numbad = size(badpix);
for i = 1 : numbad(1)
imagel(badpix(i,1),badpix(i,2),badpix(i,3)) = 0; % turns off bad pixels
end;
image-uint8 = uint8(imagel);
% image(image-uint8); % images the matrix
outputfile = ['minusNoise' tifname];
imwrite(image-uint8,outputfile,'TIFF');
output = datamean(outputfile);
6.3.4
datamean.m
Datamean.m selects the twelfth through second highest pixel value in an image for each
channel and averages them to determine an overall value for the pixel. The twelfth and
second pixel were selected as the range to reflect the amount of pixels the camera imaged
81
of monochromatic light. The second highest pixel value was to throw out any pixel
that might be falsely too high. This range can be altered in the code to reflect other
experimental situations. It also determines the standard deviation for the average of the
pixels selected.
function doutput = datamean (tifname);
imagel = opentif(tifname);
sizeOfTIF = size(imagel);
%array of matrix dimensions (row, col)
totalSize = sizeffTIF(1) * sizeOfTIF(2);
R = reshape (image(: , :,1) , 1,sizeOf TIF(1)*size0f TIF (2));
G = reshape (imagel(: , ,2),1,sizeOf TIF(1)*size0f TIF(2)) ;
B = reshape (imagel(:,
,3) ,1,sizeOf TIF(1)*size0f TIF(2));
Rorder = sort(R);
Gorder = sort(G);
Border = sort(B);
%reshape into one long row
%sort
%control pixel selection criteria here
IMPORTANT!!!!!H!
% pick out desired data points between j and k
% this lets you set how many pixels you want the program to average
Rordertrunc = Rorder(totalSize-12:totalSize-2);
Gordertrunc = Gorder(totalSize-12totalSize-2);
Bordertrunc = Border(totalSize-12:totalSize-2);
M_R = mean(Rordertrunc);
M_G = mean(Gordertrunc);
M_B = mean(Bordertrunc);
M = [MR MG MB];
stdR = std(Rordertrunc);
stdG = std(Gordertrunc);
stdB = std(Bordertrunc);
STD =
[stdR stdG std_B];
MMaxR = max(Rordertrunc);
MMaxG = max(Gordertrunc);
MMaxB = max(Bordertrunc);
MMax =
[MMaxR MMaxG MMaxB];
doutput =
6.3.5
[M STD MMax];
compare.m
compare.m takes dark images (images taken without light present, named by the user
noisel.tif, noise2.tif, etc.) and compares the pixels to see if there is a signal. If all of the
dark images have signal at the same pixel above a certain threshold (set as "thres" by the
82
user), turnOffBadPixel.m (see next section) assumes it is a "bad" pixel, and it throws it
out.
function badPixels = compare(noiseArray, thres)
%compare(image1,image2,threshold)
%Shows bad pixels for two noise images above a threshold
%Returns bad pixels
size(noiseArray);
sizeArray
noiseImage = [1;
for i = 1:sizeArray(i)
noiseImage(:,:,:,i) = opentif(noiseArray(i,:));
end;
%compares Rdatasinti and Rdatasint2 to see if the bogus pixels are in the
%same position
sizeOfTIF = size(noiseImage);
badPixels = [];
% shows the i,j position of the bad pixel
in (RGB)
% k is what channel it's
for i = 1 : sizeOfTIF(1)
sizeOfTIF(2)
for j = 1
for k= 1 :3
for m = 1:sizeArray(1)
if noiseImage(i,j,k,m) <= thres
break;
end;
if m == sizeArray(1)
badPixels = cat(1,badPixels,[i
j
k]);
end;
end;
end;
end;
end;
numBad = size(badPixels);
numberBad = numBad(1);
percentBadPixels = numBad(i) / (sizeOfTIF(i)
%determines percentage of bad pixels
6.3.6
* size~fTIF(2)
* 3);
opentif.m
opentif.m automates the opening of image files. It also changes the unsigned integers
outputed from imread.m to signed integers needed for data manipulation.
function tifimage = opentif(tif)
%returns row x col x rgb double matrix of tif file
83
A = imread(tif);
tifimage = double(A);
integrate.m
6.3.7
integrate.m performs a numerical integration.
function result = integrate(func)
%input is x, y
%output is integrated result
integral = 0;
for i = 1:2048
if func(i,1) >= 380
if func(i,1) <= 780
integral = integral + func(i,2)
*
(func(i+1,1)-func(i-1,1))
/
2;
end;
end;
end;
result = integral;
6.3.8
photo2radio.m
photo2radio converts photometric units to radiometric units.
function result = photo2radio(photo, spectrum)
%input is luminance and spectrum of light (wavelength,
%output is wavelength and radiance and irradiance
V
=
irradiance or radiance)
[ 0.0000 0.0001 0.0001 0.0002 0.0004 0.0006 0.0012 0.0022 0.0040 0.0073
0.0116 0.0168
0.1390 0.1693
0.8620 0.9149
0.8700 0.8163
0.2650 0.2170
0.0170 0.0119
0.0005 0.0004
0.0000];
0.0230
0.2080
0.9540
0.7570
0.1750
0.0082
0.0002
0.0298
0.2586
0.9803
0.6949
0.1382
0.0057
0.0002
0.0380
0.3230
0.9950
0.6310
0.1070
0.0041
0.0001
0.0480
0.4073
1.0000
0.5668
0.0816
0.0029
0.0001
0.0600
0.5030
0.9950
0.5030
0.0610
0.0021
0.0001
0.0739
0.6082
0.9786
0.4412
0.0446
0.0015
0.0000
0.0910
0.7100
0.9520
0.3810
0.0320
0.0010
0.0000
0.1126
0.7932
0.9154
0.3210
0.0232
0.0007
0.0000
rho = 0.987173;
multiplier = 683;
integral = 0;
for i = 1:2048
if spectrum(i,1) >= 380
if spectrum(i,1) <= 780
integral = integral +
(V(floor(spectrum(i,1)/5)-75) +
(V(floor(spectrum(i,1)/5)-75+1)-V(floor(spectrum(i,1)/5)-75)) *
mod(spectrum(i,1),5) / 5 ) *
spectrum(i,2) * (spectrum(i+1,1)-spectrum(i-1,1)) / 2;
84
end;
end;
end;
integral = integral * multiplier;
correctionfactor = photo /
integral;
result(:,1) = spectrum(:,1);
result(:,2) = correction-factor * spectrum(:,2);
result(:,3) = correction-factor * spectrum(:,2) * pi / rho;
6.3.9
radio2photo.m
radio2photo.m converts radiometric units into photometric units.
function result = radio2photo(radio)
%input is irradiance
%output(1) is luminance
Xoutput(2) is illuminance
V
=
[ 0.0000 0.0001 0.0001 0.0002 0.0004 0.0006 0.0012 0.0022 0.0040 0.0073
0.0116 0.0168
0.1390 0.1693
0.8620 0.9149
0.8700 0.8163
0.2650 0.2170
0.0170 0.0119
0.0005 0.0004
0.0000];
0.0230
0.2080
0.9540
0.7570
0.1750
0.0082
0.0002
0.0298
0.2586
0.9803
0.6949
0.1382
0.0057
0.0002
0.0380
0.3230
0.9950
0.6310
0.1070
0.0041
0.0001
0.0480
0.4073
1.0000
0.5668
0.0816
0.0029
0.0001
0.0600
0.5030
0.9950
0.5030
0.0610
0.0021
0.0001
0.0739
0.6082
0.9786
0.4412
0.0446
0.0015
0.0000
0.0910
0.7100
0.9520
0.3810
0.0320
0.0010
0.0000
0.1126
0.7932
0.9154
0.3210
0.0232
0.0007
0.0000
rho = 0.987173;
multiplier = 683;
radio-integral = 0;
for i = 1:2048
if radio(i,1) >= 380
if radio(i,1) <= 780
radio-integral = radio-integral +
(V(floor(radio(i,1)/5)-75) +
(V(floor(radio(i,1)/5)-75+1)-V(floor(radio(i,1)/5)-75)) *
mod(radio(i,1),5) / 5 ) *
radio(i,2) * (radio(i+1,1)-radio(i-1,1)) / 2;
end;
end;
end;
result(1) = multiplier * rho / pi * radiointegral;
result(2) = multiplier * radiointegral;
85
6.3.10
rad2XYZ2.m
rad2XYZ2.m converts radiance to XYZ values.
function XYZ = rad2XYZ2(spectrum, k)
% CIE
standard observer data (tristimulus values)
x_bar = [ 0.0002
0.3147
0.0805
0.2365
1.0142
0.0007
0.3577
0.0411
0.3042
1.0743
0.0024
0.3837
0.0162
0.3768
0.0409
0.0010
0.0000
[ 0.0000
0.0387
0.2536
0.8752
0.8689
0.2835
0.0159
0.0004
0.0000
[ 0.0007
1.5535
0.7721
0.0305
0.0000
0.0000
0.0000
0.0000
0.0000
0.0286
0.0007
];
0.0001
0.0496
0.2977
0.9238
0.8256
0.2283
0.0111
0.0003
];
0.0029
1.7985
0.5701
0.0206
0.0000
0.0000
0.0000
0.0000
0.0199
0.0005
1.1185
0.6475 0.5351 0.4316
y-bar
z_bar
=
0.0487
0.3023
0.0375
0.7052
1.1343 1.1240 1.0891 1.0305
0.3437 0.2683 0.2043 0.1526
0.0138 0.0096 0.0066 0.0046
0.0004 0.0003 0.0002 0.0001
0.0072
0.3867
0.0051
0.4516
0.0191
0.3707
0.0038
0.5298
0.0434
0.3430
0.0154
0.6161
0.1406
0.2541
0.0714
0.7938
0.9597
0.2045 0.2647
0.1956 0.1323
0.1177 0.1730
0.8787 0.9512
0.8563 0.7549
0.1122 0.0813 0.0579
0.0031 0.0022 0.0015
0.0001 0.0001 0.0000
0.0003 0.0008 0.0020
0.0621 0.0747 0.0895
0.3391 0.3954 0.4608
0.9620 0.9822 0.9918
0.7774 0.7204 0.6583
0.1798 0.1402 0.1076
0.0077 0.0054 0.0037
0.0002 0.0001 0.0001
0.0045 0.0088 0.0145 0.0214
0.1063 0.1282 0.1528 0.1852
0.5314 0.6067 0.6857 0.7618
0.9991 0.9973 0.9824 0.9556
0.5939 0.5280 0.4618 0.3981
0.0812 0.0603 0.0441 0.0318
0.0026 0.0018 0.0012 0.0008
0.0001 0.0000 0.0000 0.0000
0.0105
1.9673
0.4153
0.0137
0.0000
0.0000
0.0000
0.0000
0.1971
1.9007
0.1592
0.0011
0.0000
0.0000
0.0000
0.0000
0.0323
2.0273
0.3024
0.0079
0.0000
0.0000
0.0000
0.0000
0.0860
1.9948
0.2185
0.0040
0.0000
0.0000
0.0000
0.0000
0.3894 0.6568
1.7454 1.5549
0.1120 0.0822
0.0000 0.0000
0.0000 0.0000
0.0000 0.0000
0.0000 0.0000
0.0000 0.0000
0.0295
0.2199
0.8233
0.9152
0.3396
0.0226
0.0006
0.0000
0.9725 1.2825
1.3176
0.0607
0.0000
0.0000
0.0000
0.0000
1.0302
0.0431
0.0000
0.0000
0.0000
0.0000
0.0000 0.0000
1;
redoutput = 0;
green-output = 0;
blueoutput = 0;
for i = 1:2048
if spectrum(i,1) >= 380
if spectrum(i,1) <= 780
red-output = redoutput +
(x-bar(floor(spectrum(i,1)/5)-75) +
(x-bar(floor(spectrum(i,1)/5)-75+1)-x-bar(floor(spectrum(i,1)/5)-75)
) * mod(spectrum(i,1),5) / 5)spectrum(i,2) * (spectrum(i+1,1)spectrum(i-1,1)) / 2;
green-output = green-output +
(y-bar(floor(spectrum(i,1)/5)-75) +
(y-bar(floor(spectrum (i,1)/5)-75+1)-y-bar(floor(spectrum(i,1) /5)-75)
) * mod(spectrum(i,1),5) / 5 ) * spectrum(i,2) * (spectrum(i+1,1)spectrum(i-1,1)) / 2;
blue-output = blueoutput +
86
(z-bar(floor(spectrum(i,1)/5)-75) +
(z-bar(floor(spectrum(i,1)/5)-75+1)-z-bar(floor(spectrum(i,1)/5)-75)
) * mod(spectrum(i,1),5) / 5 ) * spectrum(i,2) * (spectrum(i+1,1)spectrum(i-1,1)) / 2;
end;
end;
end;
X = k * redoutput; % sums the products for all wavelengths
Y = k * green-output;
Z = k * blue-output;
XYZ =
[X Y ZI;
6.3.11
rgb2XYZ.m
rgb2XYZ.m converts RGB values into XYZ values [Inanici and Galvin, 2004] [Wikipedia, 2006g].
Basically, it performs this matrix operation:
X = k(O.412424R + 0.357579G + 0.180464B)
Y = k(0.212656R + 0.715158G + 0.072186B)
Z = k(0.019332R + 0.119193G + 0.950444B)
function xyz = rgb2xyz(rgb)
%function rgb2xyz([red green blue])
returns
[X
a = 0.055;
gamma = 2.4;
rgb = rgb / 255;
for i = 1:3
if rgb(i) > 0.04045
g(i) =
((rgb(i)+a)/(1+a))^gamma;
else
g(i) = rgb(i)/12.92;
end;
end;
transformMatrix =
[0.412424 0.357579 0.180464;
0.212656 0.715158 0.072186;
0.019332 0.119193 0.950444];
xyz = transformMatrix
xyz = xyz';
* g';
87
Y Z]
6.3.12
XYZ2rgb.m
XYZ2rgb.m converts XYZ values into RGB values [Inanici and Galvin, 20041. This code
was not specifically used in this project, but it can be used in the future to post-process
images from XYZs into RGBs so they can be reimaged by MATLAB.
function output = xyz2rgb(input)
%function rgb2xyz([X Y Z]) returns
[red green blue]
a = 0.055;
gamma = 2.4;
A
[ 3.2406 -1.5372 -0.4986;
=
-0.9689
1.8758
0.0557 -0.2040
%
0.0415;
1.0570];
input = input/80;
RGB = A *
input';
RGB(RGB<0)=0;
RGB(RGB>1)=1;
for i = 1:3
if RGB(i) <= 0.0031308
RGB(i) = 12.92*RGB(i);
else
RGB(i) = (1+a) * RGB(i)^(1/gamma)
end;
-
a;
end;
output = RGB';
output=round(output*255);
output(output<O)=0;
output(output>255)=255;
88
Bibliography
[xen, 2006] (2006). Xenon Short Arc Lamps.
http://www.ushio.co.jp/products/catalog/imp-e/1036.html.
Accessed on May 27,
2006.
[AIM Digital Imaging, 2006a] AIM
Digital
Imaging
(2006a).
Digi-
tal
Home
Theater
Projectors
FAQs:
How
black
are
http://www.ausmedia.com.au/projector%20contrast%20ratio.htm.
the
blacks?
Accessed on
May 30, 2006.
[AIM Digital Imaging, 2006b] AIM Digital Imaging (2006b).
Light Sources- Xenon
Lamps.
http://sales.hamamatsu.com/en/products/electron-tube-division/lightsources/xenon-lamps.php. Accessed on May 30, 2006.
[Albrecht and Eichele, 2003] Albrecht, U. and Eichele, G. (2003). The mammalian circadian clock. Current Opinion in Genetics & Development, 13(3):271-277.
[American Society for Testing and Materials, 2001] American Society for Testing and
Materials (2001). Method E308, Standard Practicefor Computing Colors of Objects by
Using the CIE System. ASTM International, West Conshohocken, PA, astm e308-01
edition.
[Andersen, 2004] Andersen, M. (2004). Innovative Video-Goniophotometerfor Advanced
FenestrationSystems. PhD thesis, EPFL, Lausanne.
[Andersen and de Boer, 2006] Andersen, M. and de Boer, J. (2006). Goniophotometry
and assessment of bidirectional photometric properties of complex fenestration systmems. Energy and Buildings, 38:836-848.
[Andersen et al., 2005a] Andersen, M., Roecker, C., and Scartezzini, J.-L. (2005a). Design of a time-efficient video-goniophotometer combining bidirectional functions assessment for transmission and reflection. Solar Energy Materials and Solar Cells, 88(1):97-
118. 15.
[Andersen et al., 2005b] Andersen, M., Rubin, M., Powles, R., and Scartezzini, J.-L.
(2005b). Bi-directional transmission properties of venetian blinds: experimental assessment compared to ray-tracing calculations. Solar Energy, 78:187-198.
[Andersen and Scartezzini, 2003] Andersen, M. and Scartezzini, J. (2003). Including of
the specular component in a BTDF or BRDF assessment based on digital imaging. In
Proceedings CISBAT 2003, pages 223-228, Lausanne, EPFL.
89
[Apian-Bennewitz, 2006] Apian-Bennewitz, P. (2006).
http://www.pab-opto.de/. Accessed on May 20, 2006.
pab
Goniophotometer
II.
[Apian-Bennewitz and von der Hardt, 1998] Apian-Bennewitz, P. and von der Hardt, J.
(1998). Enhancing and calibrating a goniophotometer. Solar Energy Materials ard
Solar Cells, 54(1-4):309-322.
[Aydinli, 1996] Aydinli, S. (1996). Short description of the spiral goniophotometer for
bidirectional measurements. Report for IEA SHC Task 21, ECBCS Annex 29, Subtask
A, Techische Universit'ate Berlin (TUB), Berlin.
[Bellia et al., 2002] Bellia, L., Cesarano, A., Minichiello, F., and Sibilio, S. (2002). Setting
up a CCD photometer for lighting research and design.
Building and Environment,
37(11):1099-1106.
[Bj6rkst6n et al., 2005] Bj6rksten, K., Bjerregaard, P., and Kripke, D. (2005). Suicides
in the midnight sun- a study of seasonality in suicides in West Greenland. Psychiatry
Research, 133(2-3):205-213.
[Breitenbach and Rosenfeld, 1998] Breitenbach, J.
photogoniometer to measure angular dependent
International Conference on Renewable Energy
386-391, Ottawa, Canada. Solar Energy Society
and Rosenfeld, J. (1998). Design of a
optical properties. In Proceedings of
Technologies in Cold Climates, pages
of Canada Inc.
[CCOHS, 2006] CCOHS (2006). OSH Answers.
http://www.ccohs.ca/oshanswers/ergonomics/lighting-flicker.html.
Accessed on May
20, 2006.
[C.E.Ochoa and Capeluto, 2006] C.E.Ochoa and Capeluto, I. (2006). Evaluating visual
comfort and performance of three natural lighting systems for deep office buildings in
highly luminous climates. Building and Environment, 41(8):1128-1135.
[Clifford, 2006] Clifford, Z. (2006). Heliodon application hardware and software control.
Technical report, Massachusetts Institute of Technology, Cambridge.
[Deniel, 2002] Deniel, J. (2002). Modelisation des luminaires et des BRDF: realisation,
mesure et compression. PhD thesis, Universit6 de Rennes 1, France.
[Dumortier et al., 2004] Dumortier, D., Coutelier, B., Faulcon, T., and Roy, F. V.
(September 2004). PHOTOLUX: a new luminance mapping system based on Nikon
Coolpix digital cameras. In Lux Europa Proceedings, pages 308-311, Berlin.
[Franzetti et al., 2004] Franzetti, C., Fraisse, G., and Achard, G. (2004). Influence of the
coupling between daylight and artificial lighting on thermal loads in office buildings.
Energy and Buildings, 36(2):117-126.
[Granderson and Agogino, 2006] Granderson, J. and Agogino, A. (2006). Intelligent office
lighting: Demand-responsive conditioning and increased user satisfaction. LEUKOS,
2(3):185-198.
90
[Griot, 2006] Griot,
M.
(2006).
Optical
coatings.
Metallic
coatings:
http://www.mellesgriot.com/products/optics/oc_5-i.htm.
Accessed
on
May
27,
2006.
[Group, 2003a] Group, H. M. (2003a). Daylight and retail sales. Technical Report P50003-082-A-5, California Energy Commission, Fair Oaks, California.
[Group, 2003b] Group, H. M. (2003b). Daylighting in schools: Reanalysis report. Technical Report P500-03-082-A-3, California Energy Commission, Fair Oaks, California.
[Group, 2003c] Group, H. M. (2003c). Windows and offices: A study of office worker performance and the indoor environment. Technical Report P500-03-082-A-9, California
Energy Commission, Fair Oaks, California.
[Heschong et al., 20023 Heschong, L., Wright, R., and Okura, S. (2002).
impacts on retail sales performance.
Daylighting
Journal of the Illuminating Engineering Society,
31(2).
[H.F.O. Muller, 2005] H.F.O. Muller (2005). Laminated glass with light directing holograms in buildings. Glass Science and Technology, 78:124-130.
[Inanici and Galvin, 2004] Inanici, M. and Galvin, J. (2004). Evaluation of high dynamic
range photography as a luminance mapping technique. Technical Report LBNL-57545,
Lawrence Berkeley National Laboratory, University of California, University of California.
[James and Bahaj, 2005a] James, P. and Bahaj, A. (2005a).
Holographic optical ele-
ments: various principles for solar control of conservatories and sunrooms. Solar En-
ergy, 78(3):441-454.
[James and Bahaj, 2005b] James, P. and Bahaj, A. (2005b). Smart glazing solutions to
glare and solar gain: a 'sick building' case study. Energy and Buildings, 37(10):10581067.
[Krarti et al., 2005] Krarti, M., Erickson, P., and Hillman, T. (2005). A simplified method
to estimate energy savings of artificial lighting use from daylighting.
Building and
Environment, 40(6):747-754.
[Lam et al., 2006] Lam, J., Tsang, C., and Yang, L. (2006). Impacts of lighting density
on heating and cooling loads in different climates in china.
Energy Conversion and
Management, 47(13-14):1942-1953.
[Li and Lam, 2001] Li, D. and Lam, J. (2001). Evaluation of lighting performance in
office buildings with daylighting controls. Energy and Buildings, 33(8):793-803.
[Li and Lam, 20031 Li, D. and Lam, J. (2003).
An investigation of daylighting perfor-
mance and energy saving in daylit corridor. Energy and Buildings, 35(4):365-373.
[Li and Tsang, 2005] Li, D. and Tsang, E. (2005).
An analysis of measured and sim-
ulated daylight illuminance and lighting savings in a daylit corridor.
Environment, 40(7):973-982.
91
Building and
[Libby, 2003] Libby, B. (2003). Beyond the Bulbs: In Praise of Natural Light. New York
Times.
[Ljubicic, 2005] Ljubicic, D. (2005). Automated Support for Experimental Approaches
in Daylighting Performances Assessment. undergraduate, Massachusetts Institute of
Technology, Cambridge.
oikos@, 2006] oikos@(2006).
Special
features
http://oikos.com/library/eem/skylights/lightpipes.html.
of
light
Accessed
on
pipes.
May
20,
2006.
[Osterhaus, 2005] Osterhaus, W. (2005). Discomfort glare assessment and prevention for
daylight applications in office environments. Solar Energy, 79(2):140-158.
[Raymond and Adler, 2005] Raymond, J. and Adler, J. (2005). A neglected nutrient: are
Americans dying from lack of vitamin D? Newsweek.
[Rea, 2000] Rea, M. S. (2000). Lighting Handbook: reference & application. Illuminating
Society of North America, New York, N.Y., 9 edition.
[Scartezzini, 2003] Scartezzini, J. (2003). Advances in daylighting and artificial lighting.
In Proceedings of Research in Building Physics.
[Scartezzini and Courret, 2002] Scartezzini, J. and Courret, G. (2002). Anidolic daylight-
ing systems. Solar Energy, 73(2):123-135.
[Scott, 2006] Scott, A. (2006). Dan Flavin: A Retrospective of Contemporary Art.
http://flakmag.com/features/flavin.html. Accessed on May 20, 2006.
[Sullivan et al., 1998] Sullivan, R., Beltran, L., Lee, E., Rubin, M., and Selkowitz,
S. (1998).
Energy and daylight performance of angular selective glazings.
In
ASHRAE/DOE/BTECC Conference, Thermal Performance of the Exterior Envelopes
of Buildings VII, Clearwater Beach, Florida.
[Sumova et al., 2004] Sumova', A., Bendovai, Z., Sladek, M., Kova'ikov6', Z., and IllnerovA, H. (2004). Seasonal molecular timekeeping within the rat circadian clock.
Physiology Research, 53.
[Sweitzer, 1993] Sweitzer, G. (1993). Three advanced daylighting technologies for offices.
Energy, 18(2):107-114.
[Taylor and Socov, 1974] Taylor, L. and Socov, E. (1974).
The movement of people to-
ward light. J. Illum. Eng. Soc., 3(3):237-241.
[UC Berkeley, 2006] UC
Berkeley
(2006).
Building
science
http://arch.ced.berkeley.edu/resources/bldgsci/bsl/heliodon.html.
uc
berkeley.
accessed
on
May 20, 2006.
[US Department of Energy, 2006] US
Department
of
Energy,
E.
(2006).
Energy
consumption
estimates
by
source
and
end-use
sector,
2001.
http://www.eia.doe.gov/emeu/states/sep-sum/html/pdf/sum-use-all.pdf.
Accessed
on May 24, 2006.
92
live
wellness,
for
Tools
(2006).
[Verilux, 2006] Verilux
http://www.toolsforwellness.com/verilux.html. Accessed on May 20, 2006.
smarter.
[Walze et al., 2005] Walze, G., Nitz, P., Ell, J., Georg, A., Gombert, A., and Hossfeld,
W. (2005). Combination of microstructures and optically functional coatings for solar
control. Solar Energy Materials and Solar Cells, 89(2-3):233-248. 15.
[Ward, 1992] Ward, G. (1992). Measuring and modeling anisotropic reflection. Computer
Graphics, 26(2):265-272.
[Wikipedia, 2006a] Wikipedia
glass.
Acrylic
(2006a).
http://enwikipedia.org/wiki/Polymethyl-methacrylate.
Accessed
on
May
27,
2006.
[Wikipedia, 2006b] Wikipedia
Cie
(2006b).
http://en.wikipedia.org/wiki/CIE-XYZ..color-space.
[Wikipedia, 2006c] Wikipedia
(2006c).
1931
space.
color
Accessed on May 27, 2006.
Color
temperature.
http://en.wikipedia.org/wiki/Color-temperature. Accessed on May 20, 2006.
International commission on illumination.
[Wikipedia, 2006d] Wikipedia (2006d).
http://en.wikipedia.org/wiki/International-Commission-on-Illumination.
Accessed on
May 20, 2006.
[Wikipedia, 2006e] Wikipedia
(2006e).
http://en.wikipedia.org/wiki/Luminosity-function.
Luminosity
funtion.
Accessed on May 20, 2006.
Solar
(2006f).
[Wikipedia, 2006f] Wikipedia
http://en.wikipedia.org/wiki/SRGB. Accessed on May 20, 2006.
radiation.
[Wikipedia, 2006g] Wikipedia (2006g). srgb. http://en.wikipedia.org/wiki/Solarradiation.
Accessed on May 20, 2006.
[Wilkins et al., 1989] Wilkins, A., Nimmo-Smith, I., Slater, A., and Bedocs, L. (1989).
Fluorescent lighting, headaches and eye-strain.
Lighting Research and Technology,
21:11-18.
[Women's Health Weekly, 2005] Women's Health Weekly (2005). Seasonal affective disorder; sunbox offers treatment option for seasonal affective disorder sufferers. NewsRx,
page 136.
[W.S. Strouse Watt, 2006] W.S. Strouse Watt, 0. (2006). Computer vision syndrome
and computer glasses. http://www.mdsupport.org/library/cvs.html. Accessed on May
20, 2006.
[Yorks and Ginthner, 1987] Yorks, P. and Ginthner, D. (1987). Wall lighting placement:
Effect on behavior in the work environment. Light. Des. Appl., 17(7):30-37.
[Zain-Ahmed et al., 2002] Zain-Ahmed, A., Sopian, K., Othman, M., Sayigh, A., and
Surendran, P. (2002). Daylighting as a passive solar design strategy in tropical buildings: a case study of malaysia. Energy Conversion and Management, 43(13):1725-1736.
93