ArticlePDF Available

A Guided Tour of Colour Space

Authors:

Abstract and Figures

This article describes the theory of color repro-duction in video, and some of the engineering compromises necessary to make practical cameras and practical coding systems. Video processing is generally concerned with color represented in three components derived from the scene, usually red, green, and blue or components computed from these. Once red, green, and blue components of a scene are obtained, color coding systems transform these components into other forms suitable for processing, recording, and transmission. Accu-rate color reproduction depends on knowledge of exactly how the physical spectra of the orig-inal scene are transformed into these compo-nents and exactly how the color components are transformed to physical spectra at the display. I use the term luminance and the symbol Y to refer to true, CIE luminance, a linear-light quan-tity proportional to physical intensity. I use the term luma and the symbol Y' to refer to the signal that is used in video to approximate lumi-nance. I you are unfamiliar with these terms or this usage, see page 24 of my book 1 .
Content may be subject to copyright.
© 1997-08-19 Charles Poynton 1 of 14
A Guided Tour of Color Space
Charles Poynton
This article describes the theory of color repro-
duction in video, and some of the engineering
compromises necessary to make practical
cameras and practical coding systems.
Video processing is generally concerned with
color represented in three components derived
from the scene, usually red, green, and blue or
components computed from these. Once red,
green, and blue components of a scene are
obtained, color coding systems transform these
components into other forms suitable for
processing, recording, and transmission. Accu-
rate color reproduction depends on knowledge
of exactly how the physical spectra of the orig-
inal scene are transformed into these compo-
nents and exactly how the color components are
transformed to physical spectra at the display.
I use the term
luminance
and the symbol
Y
to
refer to true, CIE luminance, a linear-light quan-
tity proportional to physical intensity. I use the
term
luma
and the symbol
Y’
to refer to the
signal that is used in video to approximate lumi-
nance. I you are unfamiliar with these terms or
this usage, see page 24 of my book
1
.
Trichromaticity
Color is the perceptual result of light having
wavelength from 400 to 700 nm that is incident
upon the retina.
The human retina has three types of color
photoreceptor
cone
cells, which respond to inci-
1 Charles Poynton,
A Technical Introduction to Digital
Video
(New York: John Wiley & Sons, 1996)
dent radiation with different spectral response
curves. Because there are exactly three types of
color photoreceptors, three components are
necessary and sufficient to describe a color,
providing that appropriate spectral weighting
functions are used: Color vision is inherently
trichromatic
.
A fourth type of photoreceptor cell, the
rod
, is
also present in the retina. Rods are effective only
at extremely low light intensities. Because there
is only one type of rod cell, “night vision” cannot
perceive color. Image reproduction takes place at
light levels sufficiently high that the rod recep-
tors play no role.
Isaac Newton said, “Indeed rays, properly
expressed, are not colored.” Spectral power
distributions (SPDs) exist in the physical world,
but color exists only in the eye and the brain.
Color specification
The Commission Internationale de L’Éclairage
(CIE) has defined a system to compute a triple of
numerical components that can be considered to
be the mathematical coordinates of color space.
Their function is analogous to coordinates on a
map. Cartographers have different map projec-
tions for different functions: Some map projec-
tions preserve areas, others show latitudes and
longitudes as straight lines. No single map
projection fills all the needs of map users. Simi-
larly, no single color system fills all of the needs
of color users.
This is an edited version of a paper first published in New
Foundations for Video Technology (Proceedings of the
SMPTE Advanced Television and Electronic Imaging
Conference, San Francisco, Feb. 1995), 167- 180.
Author’s address: 56A Lawrence Avenue E,
Toronto ON M4N 1S3, Canada.
E-mail poynton@poynton.com,
www.inforamp.net/~poynton
A Guided Tour of Color Space
2
Any system for color specification must be inti-
mately related to the CIE specifications. The
systems useful for color specification are all
based on CIE
XYZ
. Numerical specification of
hue and saturation has been standardized by the
CIE, but is not useful for color specification.
A color specification system needs to be able to
represent any color with high precision. Since
few colors are handled at a time, a specification
system can be computationally complex.
An ink color can be specified by the proportions
of standard or proprietary inks that can be mixed
to make the color. Ink mixture is beyond the
scope of this paper.
The science of colorimetry forms the basis for
describing a color as three numbers. However,
classical colorimetry is intended for the specifica-
tion of color, not for color image coding.
Although an understanding of colorimetry is
necessary to achieve good color performance in
video, strict application of colorimetry is unsuit-
able.
Color image coding
A color image is represented as an array of
pixels, where each pixel contains numerical
components that define a color. Three compo-
nents are necessary and sufficient for this
purpose, although in printing it is convenient to
use a fourth (black) component.
In theory, the three numerical values for image
coding could be provided by a color specification
system. But a practical image coding system
needs to be computationally efficient; cannot
afford unlimited precision; need not be inti-
mately related to the CIE system; and generally
needs to cover only a reasonably wide range of
colors and not all of the colors. So image coding
uses different systems than color specification.
The systems useful for image coding are linear
RGB
, nonlinear
RGB
, nonlinear
CMY
, nonlinear
CMYK
, and derivatives of nonlinear
RGB
such as
Y’C
B
C
R
. Numerical values of hue and saturation
are not useful in color image coding.
If you manufacture cars, you have to match the
color of paint on the door with the color of paint
on the fender. A color specification system will
be necessary. However, to convey a picture of
the car, you need image coding. You can afford
to do quite a bit of computation in the first case
because you have only two colored elements,
the door and the fender. In the second case, the
color coding must be quite efficient because you
may have a million colored elements or more.
For a highly readable short introduction to color
image coding, consult DeMarsh and
Linear-Light
Tristimulus
(x, y)
Chromaticity
Image Coding Systems
Perceptually
Uniform
Hue-
Oriented
CIE XYZ
CIE xyY
Linear RGB
}
HLS, HSB
Nonlinear
R’G’B’
CIE L*a*b* CIE L* h
ab
c
ab
CIE L* h
uv
c
uv
CIE L*u*v* TekHVC
Nonlinear
Y’C
B
C
R
3
×
3
3
×
3
POLAR/
RECTANGULAR
SCALED POLAR/
RECTANGULAR
POLAR/
RECTANGULAR
NONLINEAR
TRANSFORM
NONLINEAR
TRANSFORM
TRANSFER
FUNCTION
PROJECTIVE
TRANSFORM
NONLINEAR
TRANSFORM
PROJECTIVE
TRANSFORM
Figure 1 Color systems can be classified into four
groups that are related by different kinds of transfor-
mations. Tristumulus systems and perceptually-
uniform systems are useful for image coding.
A Guided Tour of Color Space
3
Giorgianni
2
. For a terse, complete technical
treatment, read Schreiber
3
.
Definitions
Intensity
is a measure over some interval of the
electromagnetic spectrum of the flow of power
that is radiated from, or incident on, a surface.
Intensity is what I call a
linear-light
measure,
expressed in units such as watts per square
meter.
Brightness
is defined by the CIE as “the attribute
of a visual sensation according to which an area
appears to exhibit more or less light.” Because
brightness perception is very complex, the CIE
defined a more tractable quantity,
luminance
,
which is radiant power weighted by a spectral
sensitivity function that is characteristic of vision.
Hue
is the attribute of a color perception
denoted by blue, green, yellow, red, and so on.
Roughly speaking, if the dominant wavelength
of an SPD shifts, the hue of the associated color
will shift.
Saturation
, or
purity
, is the degree of colorful-
ness, from neutral gray through pastel to satu-
2 LeRoy E. DeMarsh and Edward J. Giorgianni, “Color
Science for Imaging Systems,” in
Physics Today
,
September 1989, 44-52.
3 W.F. Schreiber,
Fundamentals of Electronic Imaging
Systems,
Third Edition (New York: Springer-Verlag, 1993).
rated colors. Roughly speaking, the more
concentrated an SPD is at one wavelength, the
more saturated the associated color. You can
desaturate a color by adding light that contains
power at all wavelengths.
Spectral Power Distribution (SPD) and
tristimulus
Physical power (or
radiance
) is expressed in a
spectral power distribution
(SPD), often in 31
components each representing power in a
10 nm band from 400 to 700 nm. The SPD of
the CIE standard daylight illuminant CIE D
65
is
sketched at the top of Figure 2 above.
One way to describe a color is to directly repro-
duce its spectral power distribution. This is
sketched in the middle row of Figure 2, where
31 components are transmitted. It is reasonable
to use this method to describe a single color or a
few colors, but using 31 components for each
pixel is an impractical way to code an image.
Based on the trichromatic nature of vision, we
can determine suitable spectral weighting func-
tions to describe a color using just three compo-
nents, indicated at the bottom of Figure 2.
The relationship between SPD and perceived
color is the concern of the science of
colorimetry
.
In 1931, the CIE adopted standard curves for a
hypothetical
Standard Observer
. These curves
specify how an SPD can be transformed into a
triple that specifies a color.
400 500 600 700
Wavelength, nm
Spectral reproduction (31 components)
Tristimulus reproduction (3 components)
Figure 2 Spectral and tristimulus color reproduction.
A color can be described as a spectral power distribution
(SPD) of perhaps 31 numbers, each representing power
over a 10 nm band. However, if appropriate spectral
weighting functions are used, three values are sufficient.
A Guided Tour of Color Space
4
The CIE system is immediately and almost
universally applicable to self-luminous sources
and displays. However, the colors produced by
reflective systems such as photography, printing
or paint are a function not only of the colorants
but also of the SPD of the ambient illumination.
If your application has a strong dependence
upon the spectrum of the illuminant, you may
have to resort to spectral matching.
Scanner spectral constraints
The relationship between spectral distributions
and the three components of a color value is
conventionally explained starting from the
famous color matching experiment. I will instead
explain the relationship by illustrating the prac-
tical concerns of engineering the spectral filters
required by a color scanner or camera, as illus-
trated in Figure 3 above.
The top row shows the spectral sensitivity of
three wideband filters having uniform response
across each of the red, green, and blue regions
of the spectrum. In a typical filter, whether for
electrical signals or for optical power, it is gener-
ally desirable to have a response that is as
uniform as possible across the passband, to have
transition zones as narrow as possible, and to
have maximum possible attenuation in the stop-
bands. Entire textbooks are devoted to filter
design; they concentrate on optimizing filters in
that way. The example on the top row of the
illustration shows two monochromatic sources
which are seen as saturated orange and red.
Applying a “textbook” filter to that example
causes these two different wavelength distribu-
tions to report the identical
RGB
triple, [1, 0, 0].
That is a serious problem, because these SPDs
are seen as different colors!
At first sight it may seem that the problem with
the wideband filters is insufficient wavelength
discrimination. The middle row of the example
addresses the apparent lack of spectral discrimi-
nation of the filters of the top row, by using three
narrowband filters. This set solves one problem,
R
G
B
Wideband filter set
Narrowband filter set
CIE-based filter set
400 500 600 700
450
400 500 600 700
ZYX
R
G
B
400 500 600 700 400 500 600 700
400 500 600 700 400 500 600 700
540 620
Figure 3 Scanner spectral constraints. This diagram
shows the spectral constraints associated with scanners
and cameras. The top row shows the spectral sensitivity
of three wideband filters having uniform response across
the shortwave, mediumwave and longwave regions of
the spectrum. The problem with this approach is that two
monochromatic sources seen by the eye to have different
colors – in this case, saturated orange and a saturated
red – cannot be distinguished by the filter set.
The middle row of the illustration shows three narrow
-
band filters. This set solves the problem of the first set,
but creates another: many monochromatic sources “fall
between” the filters, and are seen by the scanner as
black. To see color as the eye does, the three filter
responses must be closely related to the color response of
the eye.
The bottom row shows the spectral response functions
of the CIE Standard Observer.
A Guided Tour of Color Space
5
but creates another: many monochromatic
sources “fall between” the filters, and are seen
by the scanner as black. In the example, the
orange source reports an
RGB
triple of [0, 0, 0],
identical to the result of scanning black.
Although this example is contrived, the problem
is not. Ultimately, the test of whether a camera
or scanner is successful is whether it reports
distinct
RGB
triples if and only if human vision
sees different colors. To see color as the eye
does, the three filter sensitivity curves must be
closely related to the color response of the eye.
A famous experiment, the
color matching
experiment
, was devised during the 1920s to
characterize the relationship between SPD and
color. The experiment measures mixtures of
different spectral distributions that are required
for human observers to match colors. Exploiting
this indirect method, the CIE in 1931 standard-
ized a set of spectral weighting functions that
models the perception of color. These curves,
defined numerically, are referred to as the x, y,
and z
color matching functions
(
CMFs
) for the
CIE Standard Observer
4
. They are illustrated in
the bottom row of Figure 3 opposite, and are
graphed at a larger scale in Figure 4 above.
CIE XYZ tristimulus
The CIE designed their system so that one of the
three tristimulus values – the
Y
value – has a
spectral sensitivity that corresponds to the light-
ness sensitivity of human vision. The
luminance
Y
of a source is obtained as the integral of its
SPD weighted by the
y
color matching function.
When luminance is augmented with two other
components
X
and
Z,
computed using the
x
and
z
color matching functions, the resulting
(
X
,
Y
,
Z
) components are known as
XYZ
tristim-
ulus
values (pronounced
“big-X, big-Y, big-Z”
or
“cap-X, cap-Y, cap-Z”
). These are linear-light
values that embed the spectral properties of
human color vision.
Tristimulus values are computed from contin-
uous SPDs by integrating the SPD using the
x
,
y
,
and
z
color matching functions. In discrete form,
tristimulus values can be computed by a matrix
multiplication, as shown in Figure 5 overleaf.
4 The CIE system is standardized in Publication
CIE No 15.2,
Colorimetry, Second Edition
(Vienna,
Austria: Central Bureau of the Commission Internationale
de L’Éclairage, 1986).
0.0
0.5
1.0
1.5
2.0
z
x
y
400 500 600 700
Wavelength, nm
Response
Figure 4 CIE Color Matching Functions. A camera must
have these spectral response curves, or linear combina-
tions of these, in order to capture all colors. However,
practical considerations make this difficult.
A Guided Tour of Color Space
6
Human color vision follows a principle of super-
position, first elaborated by Grassman and now
known as Grassman’s Law: The tristimulus
values that result after summing a set of SPDs is
identical to the sum of the tristimulus values of
each of the SPDs. Due to this linearity of additive
color mixture, any set of three components that
is a nontrivial linear combination of
X, Y, and Z is
also a set of tristimulus values.
The CIE system is based on the description of
color as a luminance component Y, as described
above, and two additional components X and Z.
The spectral weighting curves of X and Z have
been standardized by the CIE based on statistics
from experiments involving human observers.
XYZ tristimulus values can describe any color.
(RGB tristimulus values will be described later.)
The magnitudes of the XYZ components are
proportional to physical power, but their spectral
composition corresponds to the color matching
characteristics of human vision.
CIE x, y chromaticity
It is convenient, for both conceptual under-
standing and computation, to have a represen-
tation of “pure” color in the absence of
luminance. The CIE standardized a procedure for
normalizing XYZ tristimulus values to obtain two
chromaticity values x and y (pronounced
“little-x, little-y”). The relationships are
computed by this projective transformation:
A color plots as a point in an (x, y) chromaticity
diagram, illustrated in Figure 6 opposite. When a
narrowband SPD comprising power at just one
wavelength is swept across the range 400 to
700 nm, it traces a shark-fin shaped spectral
locus in (x, y) coordinates. The sensation of
purple cannot be produced by a single wave-
length: To produce purple requires a mixture of
shortwave and longwave light. The line of
purples joins extreme blue to extreme red. The
chromaticity coordinates of real (physical) SPDs
are bounded by the line of purples and the spec-
tral locus: All colors are contained in this region
of the chromaticity diagram.
It is common to specify a color by its chroma-
ticity and luminance, in the form of an xyY triple.
(A third chromaticity value, z, can be computed
similarly. However, z is redundant if x and y are
known, due to the identity x + y + z = 1.)
=•
82.75
91.49
93.43
86.68
104.86
117.01
117.81
114.86
115.92
108.81
109.35
107.80
104.79
107.69
104.41
104.05
100.00
96.33
95.79
88.69
90.01
89.60
87.70
83.29
83.70
80.03
80.21
82.28
78.28
69.72
71.61
400 nm
450 nm
500 nm
550 nm
600 nm
650 nm
700 nm
0.0143
0.0435
0.1344
0.2839
0.3483
0.3362
0.2908
0.1954
0.0956
0.0320
0.0049
0.0093
0.0633
0.1655
0.2904
0.4334
0.5945
0.7621
0.9163
1.0263
1.0622
1.0026
0.8544
0.6424
0.4479
0.2835
0.1649
0.0874
0.0468
0.0227
0.0114
0.0004
0.0012
0.0040
0.0116
0.0230
0.0380
0.0600
0.0910
0.1390
0.2080
0.3230
0.5030
0.7100
0.8620
0.9540
0.9950
0.9950
0.9520
0.8700
0.7570
0.6310
0.5030
0.3810
0.2650
0.1750
0.1070
0.0610
0.0320
0.0170
0.0082
0.0041
0.0679
0.2074
0.6456
1.3856
1.7471
1.7721
1.6692
1.2876
0.8130
0.4652
0.2720
0.1582
0.0782
0.0422
0.0203
0.0087
0.0039
0.0021
0.0017
0.0011
0.0008
0.0003
0.0002
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
0.0000
X
Y
Z
T
Figure 5 Calculation of tristimulus values by matrix
multiplication. The column vector at the right is a
discrete version of CIE Illuminant D65. The 31-by-3
matrix is a discrete version of the set of CIE x, y and z
color matching functions, here represented in 31 wave-
lengths at 10 nm intervals across the spectrum. The result
of performing the matrix multiplication is a set of XYZ
tristimulus values.
xX
XYZ yY
XYZ
=++ =++
Eq 1
A Guided Tour of Color Space
7
To recover X and Z tristimulus values from chro-
maticities and luminance, use the inverse of Eq 1:
The projective transformation used to compute x
and y is such that any linear combination or two
spectra, or two tristimulus values, plots on a
straight line in the (x, y) plane. However, the
transformation does not preserve distances, so
chromaticity values do not combine linearly.
There is no unique physical or perceptual defini-
tion of white. An SPD that is considered white
will have CIE (x, y) coordinates roughly in the
central area of the chromaticity diagram, in the
region of (13, 13).
Additive mixture (RGB)
The simplest way to reproduce a wide range of
colors is to mix light from three lights of different
colors, usually red, green, and blue. Figure 7
overleaf illustrates additive reproduction. In
physical terms, the spectra from each of the
lights add together wavelength by wavelength
to form the spectrum of the mixture.
As a consequence of the principle of superposi-
tion, the color of an additive mixture is a strict
function of the colors of the primaries and the
fraction of each primary that is mixed.
Additive reproduction is employed directly in a
video projector, where the spectra from a red
beam, a green beam, and a blue beam are phys-
ically summed at the surface of the projection
screen.
Additive reproduction is also employed in a
direct-view color CRT, but through slightly indi-
rect means. The screen of a CRT comprises small
dots that produce red, green, and blue light.
When the screen is viewed from a sufficient
distance, the spectra of these dots add at the
observer’s retina.
0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8
x
0.0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
y
Comparison of SMPTE 240M,
EBU Tech. 3213-E
and ITU-R Rec. BT.709 Primaries
400
440
460
480
500
520
540
580
600
620
640 700
560
SMPTE
GREEN
RED
BLUE
CIE D65
EBU
Rec. 709
Figure 6 The (x, y) chromaticity
diagram. The spectral locus, an
inverted U-shaped path, is swept out as
a monochromatic source runs from 400
to 700 nm. The locus of all colors is
closed by the line of purples, which
traces SPDs that combine longwave
and shortwave power but have no
mediumwave contribution.
On this chromaticity diagram, I plot
the chromaticity of the D65 white point,
and the chromaticities of the red, green,
and blue primaries of contemporary
SMPTE, EBU, and ITU-R standards for
video primaries. See A Technical Intro-
duction to Digital Video for details.
Xx
yYZ
xy
yY==
−−1
Eq 2
A Guided Tour of Color Space
8
White
In additive image reproduction, the white point
is the chromaticity of the color reproduced by
equal red, green, and blue components. White
point is a function of the ratio (or balance) of
power among the primaries.
It is often convenient for purposes of calculation
to define white as a uniform SPD. This white
reference is known as the equal-power illumi-
nant, or CIE Illuminant E.
A more realistic reference that approximates
daylight has been specified numerically by the
CIE as Illuminant D65. You should use this unless
you have a good reason to use something else.
The print industry commonly uses D50, and
photography commonly uses D55. These repre-
sent compromises between the conditions of
indoor (tungsten) and daylight viewing.
Human vision adapts to white in the viewing
environment. An image viewed in isolation –
such as a slide projected in a dark room – creates
its own white reference, and a viewer will be
quite tolerant of errors in the white point.
However, if the same image is viewed in the
presence of an external white reference or a
second image, then differences in white point
can be objectionable.
Planck determined that the SPD radiated from a
hot object – a black body radiator – is a function
of the temperature to which the object is heated.
A typical source of illumination has a heated
object at its core, so it is often useful to charac-
terize an illuminant by specifying the tempera-
ture (in units of kelvin, K) of a black body
radiator that appears to have the same color.
Although an illuminant can be specified infor-
mally by its color temperature, a more complete
specification is provided by the chromaticity
coordinates of the SPD of the source. Figure 8
shows the SPDs of several standard illuminants.
Camera white reference
There is an implicit assumption in television that
the camera operates as if the scene were illumi-
nated by a source having the chromaticity of
CIE D65. In practice, the scene illumination is
often deficient in the shortwave (blue) region of
the spectrum. Television studio lighting is often
accomplished by tungsten lamps, which are very
yellow. This situation is accommodated by white
balancing, that is, by adjusting the gain of the
red, green, and blue components of the scene so
that a white object reports the values that would
be reported if the scene were illuminated by D65.
R
G
B
400
Wavelength, nm
500 600 700
Figure 7 Additive mixture. This diagram illustrates the
physical process underlying additive color mixture, as is
used in color television. Each colorant has an indepen-
dent, direct path to the image. The spectral power of the
image is, at each wavelength, the sum of the spectra of
the colorants. The colors (chromaticities) of the mixtures
are completely determined by the colors of the primaries,
and analysis and prediction of mixtures is reasonably
simple. The spectral curves shown here are those of an
actual Sony Trinitron™ monitor.
A Guided Tour of Color Space
9
Monitor white reference
In an additive mixture the illumination of the
reproduced image is generated entirely by the
display device. In particular, reproduced white is
determined by the characteristics of the display,
and is not dependent on the environment in
which the display is viewed. In a completely dark
viewing environment such as a cinema theater,
this is desirable. However, in an environment
where the viewer’s field of view encompasses
objects other than the display, the viewer’s
notion of “white” is likely to be influenced or
even dominated by what he perceives as
“white” in the ambient. To avoid subjective
mismatches, the chromaticity of white repro-
duced by the display and the chromaticity of
white in the ambient should be reasonably close.
Modern blue CRT phosphors are more efficient
with respect to human vision than red or green.
In a quest for brightness at the expense of color
accuracy, it is common for a computer display to
have excessive blue content, about twice as blue
as daylight, with white at about 9300 K. This
situation can be corrected by calibrating the
monitor to a white reference with a lower color
temperature.
Characterization of RGB primaries
Additive reproduction is based on physical
devices that produce all-positive SPDs for each
primary. Physically and mathematically, the
spectra add. The largest range of colors will be
produced with primaries that appear red, green,
and blue. Human color vision obeys the principle
of superposition, so the color produced by any
additive mixture of three primary spectra can be
predicted by adding the corresponding fractions
of the XYZ components of the primaries: the
colors that can be mixed from a particular set of
RGB primaries are completely determined by the
colors of the primaries by themselves. Subtrac-
tive reproduction is much more complicated: the
colors of mixtures are determined by the prima-
ries and by the colors of their combinations.
An additive RGB system is specified by the chro-
maticities of its primaries and its white point. The
extent (gamut) of the colors that can be mixed
from a given set of RGB primaries is given in the
(x, y) chromaticity diagram by a triangle whose
vertices are the chromaticities of the primaries.
In computing, there are no standard primaries
and there is no standard white point. If you have
an RGB image but have no information about its
450400350 500 550 600 650 700 750 800
0.5
1
1.5
2
2.5
D
75
D
65
D
55
D
50
C
A
Figure 8 CIE illuminants are graphed here. Illuminant A
is now obsolete; it is representative of tungsten illumina-
tion. Illuminant C was an early standard for daylight; it
too is obsolete. The family of D Illuminants represent
daylight at several color temperatures.
A Guided Tour of Color Space
10
chromaticities, you cannot accurately determine
the colors represented by the image data.
The NTSC in 1953 specified a set of primaries
that were representative of phosphors used in
color CRTs of that era. However, phosphors
changed over the years, primarily in response to
market pressures for brighter receivers, and by
the time of the first videotape recorder the
primaries in use were quite different than those
on the books.” So although you may see the
NTSC primary chromaticities documented, they
are of no practical use today.
Contemporary studio monitors have slightly
different standards in North America, Europe,
and Japan. However, international agreement5
has been obtained on primaries for high-defini-
tion television (HDTV), and these primaries are
closely representative of contemporary monitors
in studio video, computing, and computer
graphics. The standard is formally denoted
ITU-R Recommendation BT.709 (formerly CCIR
Rec. 709). I’ll call it Rec. 709. The Rec. 709
primaries and its D65 white point are these:
For a discussion of nonlinear RGB in computer
graphics, read Lindbloom’s SIGGRAPH paper6.
For technical details on monitor calibration,
consult Cowan7.
Video standards specify abstract R’G’B’ systems
that are closely matched to the characteristics of
5 Basic Parameter Values for the HDTV Standard for the
Studio and for International Programme Exchange
(Geneva, Switzerland: ITU, 1990).
6 Bruce Lindbloom, “Accurate Color Reproduction for
Computer Graphics Applications,” in Computer
Graphics, Vol. 23, No. 3 (July 1989), 117-126.
7 William B. Cowan, “An Inexpensive Scheme for Cali-
bration of a Colour Monitor in terms of CIE Standard
Coordinates,” in Computer Graphics, Vol. 17, No. 3
(July 1983), 315-321.
real monitors. Physical devices that produce
additive color involve tolerances and uncertain-
ties, but if you have a monitor that conforms to
Rec. 709 within some tolerance, you can
consider the monitor to be device-independent.
The importance Rec. 709 as an interchange
standard in studio video, broadcast television,
and high definition television, and the percep-
tual basis of the standard, assures that its param-
eters will be used even by devices such as flat-
panel displays that do not have the same physics
as CRTs.
Non-realizable spectra
In the x, y, and z functions, the contribution of
each weighting function to each tristimulus
value is positive at every wavelength. Light
cannot convey negative power; if a color
matching function had a negative excursion, its
filter could not be physically realized.
If you examine the x, y, and z functions, you will
see that – except for the trivial case of an identi-
cally zero spectrum – it is not possible to have a
physical spectrum that produces zero luminance:
Every wavelength has some positive contribu-
tion to Y. There is no way to generate nonzero X
or Z values without also generating nonzero Y.
Thus, a “pure” X tristimulus triple [1, 0, 0] can
never be obtained from a physical (all-positive)
spectral distribution. However, setting aside for
the moment the physical constraint, a pure X
tristimulus value would correspond mathemati-
cally to a spectrum that has negative power at
some wavelengths, in order to generate unity X
and zero luminance Y.
The SPDs corresponding to the CIE x, y, and z
color matching functions correspond to physi-
cally realizable spectra, SPDs having these
shapes could in theory be used as the primaries
in a display device such as a CRT. However, using
XYZ tristimulus values to drive a display having
x, y, and z primaries would not correctly repro-
duce color. This is fundamentally because the
analysis of the source SPD into X, Y, and Z
components proceeds independently during
scanning, but in an additive display device, the
SPDs of the primaries interact when they
Table 1 Rec. 709 primaries and white reference
RGBwhite
x0.640 0.300 0.150 0.3127
y0.330 0.600 0.060 0.3290
z0.030 0.100 0.790 0.3582
A Guided Tour of Color Space
11
combine at the display. If power from the source
could be “steered” into three components at the
source – if power contributing to one compo-
nent could be subtracted from the other two –
then the component values at the scanner could
be directly used at the display.
Unfortunately this “steering” operation involves
subtraction, which in the case of the x, y, and z
functions leads to spectral sensitivity functions
that are negative at some wavelengths, and
therefore not realizable in practice.
Alternatively, the display could be arranged so
that the X tristimulus component produced a
contribution having a spectral power distribution
x-- y -- z (and similarly for Y and Z). Unfortunately
this proposal suffers from nonrealizable prima-
ries at the display. So we have a dilemma: to
utilize XYZ tristimulus components directly to
achieve color reproduction, we must have either
a nonrealizable scanner or a nonrealizable
display!
The solution to this dilemma involves trans-
forming the tristimulus components into a math-
ematical domain where physical realizability is
not a constraint. The transformation depends on
the fact that the principle of superposition
applies to human color vision. We process the
XYZ components through a 3×3 linear (matrix)
multiplication, to produce a new set of compo-
nents appropriate for physical RGB display
primaries.
Transformations between CIE XYZ and RGB
RGB values in a particular set of primaries can be
transformed to and from CIE XYZ by a 3×3
matrix transform. These transforms involve
tristimulus values, that is, sets of three linear-
light components that conform to the CIE color
matching functions. SMPTE has published
details on how to compute these transforma-
tions8. CIE XYZ is a special case of tristimulus
values. In XYZ, any color is represented by a
positive set of values.
8 SMPTE RP 177-1993, Derivation of Basic Television
Color Equations.
To transform from CIE XYZ into Rec. 709 RGB
(with its D65 white point), use this transform:
This matrix has some negative coefficients: XYZ
colors that are out of gamut for a particular RGB
transform to RGB where one or more RGB
components is negative or greater than unity.
Here’s the inverse transform. Because white is
normalized to unity, the middle row sums to
unity:
To recover primary chromaticities from such a
matrix, compute little x and y for each RGB
column vector. To recover the white point, trans-
form RGB = [1, 1, 1] to XYZ, then compute x
and y.
Transforms among RGB systems
RGB tristimulus values in a system employing
one set of primaries can be transformed into
another set by a 3×3 matrix multiplication.
Generally these matrices are normalized for a
white point luminance of unity. For details, see
Television Engineering Handbook9.
As an example, here is the transform from
SMPTE 240M (or SMPTE RP-145) RGB to
Rec. 709:
This matrix is very close to the identity matrix. In
a case like this, if the transform is computed in
the nonlinear (gamma-corrected) R’G’B’
domain the resulting errors will be insignificant.
9 Television Engineering Handbook, Featuring HDTV
Systems, Revised Edition by K. Blair Benson, revised by
Jerry C. Whitaker (New York: McGraw-Hill, 1992). This
supersedes the Second Edition.
R
G
B
X
Y
Z
709
709
709
3 240479 1 537150 0 498535
0 969256 1 875992 0 041556
0 055648 0 204043 1 057311
=−−
...
...
...
X
Y
Z
R
G
B
=
0 412453 0 357580 0 180423
0 212671 0 715160 0 072169
0 019334 0 119193 0 950227
709
709
709
...
...
...
R
G
B
R
G
B
709
709
709
145
145
145
0 939555 0 050173 0 010272
0 017775 0 965795 0 016430
0 001622 0 004371 1 005993
=−−
...
...
...
A Guided Tour of Color Space
12
Here’s another example. To transform EBU 3213
RGB to Rec. 709:
Transforming among RGB systems may lead to
an out of gamut result where one or more RGB
components is negative or greater than unity.
Scanning colored media
The previous section discusses the analysis filters
necessary to obtain three tristimulus values from
incident light from a colored object. However, if
the incident light has itself been formed by an
imaging system based on three components, the
spectral composition of light from the original
scene has already been analyzed. To perform a
spectral analysis based on the CIE color
matching functions is unnecessary at best, and is
likely to introduce degraded color information.
A much more direct, practical, and accurate
approach is to perform an analysis using filters
tuned to the spectrum of the colorants of the
reproduced image, and thereby extract the three
color “records” present in the reproduction
being scanned.
In practical terms, if an image to be scanned has
already been recorded on color film, or printed
by color photography or offset printing, then
better results will be obtained by using three
narrowband filters designed to extract the color
components that have already been recorded on
the media. A densitometer is an instrument that
performs this task (but reporting logarithmic
density values instead of linear-light tristimulus
values). Several spectral sensitivity functions
have been standardized for densitometers for
use in different applications.
Practical camera spectral response
The previous section explained how the CIE x, y,
and z functions are representative of the spectral
response of the three color photoreceptor cells
of human vision, and explained how XYZ tristim-
ulus components from such a scanner are trans-
formed for use by a display device having
realizable primaries. Although an ideal scanner
or a camera would use these spectral functions,
many practical difficulties preclude their use in
real cameras and scanners.
The CIE x, y, and z functions capture all colors.
However, the range of colors commonly
encountered in the natural or man-made world
is a fraction of total color space. Although it is
useful for an instrument such as a colorimeter to
measure all colors, in an imaging system we are
generally concerned with the subset of colors
that occur frequently. If a scanner is designed to
capture all colors, its complexity is necessarily
higher and its performance is necessarily worse
than a camera designed to capture a smaller
range of colors.
Furthermore, in a system that reproduces
images, such as television, any economical
display device will be limited in the range of
colors that it can reproduce. There are usually
good engineering reasons to capture a range of
colors that is not excessively larger than the
range of colors that can be reproduced. For
much of the history of color television, cameras
were designed to incorporate assumptions about
the color reproduction capabilities of color CRTs.
However, nowadays, video production equip-
ment is being used to originate images for a
wider range of applications than just television
broadcast. The desire to make video cameras
suitable for originating images for these wider
applications has led to cameras with increased
color range.
“Subtractive” mixture (CMY)
In contrast to additive mixture described earlier,
another way to produce a range of color
mixtures is to selectively remove portions of the
spectrum from a relatively broadband illuminant.
This is illustrated in Figure 9 opposite.
The illuminant at the left of the sketch produces
light over most or all of the visible spectrum, and
each successive filter transmits some portion of
the band and attenuates other portions. The
overall attenuation of each filter is controlled: In
the case of color photography, this is accom-
plished by varying the thickness of each dye
R
G
B
R
G
B
EBU
EBU
EBU
709
709
709
1 044036 0 044036 0
010
0 0 011797 0 988203
=
...
...
...
A Guided Tour of Color Space
13
layer that functions as a filter, the shape of each
curve remaining constant.
In physical terms, the spectrum of the mixture is
the wavelength by wavelength product of the
spectrum of the illuminant and the spectral
transmission curves of each of the colorants. The
spectral transmission curves of the colorants
multiply, so this method of color reproduction
should really be called “multiplicative.” Photo-
graphers and printers have for decades
measured transmission in base-10 logarithmic
density units, where transmission of unity corre-
sponds to a density of 0, transmission of 0.1
corresponds to a density of 1, transmission of
0.01 corresponds to a density of 2, and so on.
When a printer or photographer computes the
effect of filters in tandem, he subtracts density
values instead of multiplying transmission
values, so he calls the system subtractive.
To achieve a wide range of colors in a subtractive
system requires filters that appear colored cyan,
yellow, and magenta (CMY). Figure 9 illustrates
the situation of a transmissive display such as in
the projection of color photographic film.
RGB information can be used as the basis for
subtractive image reproduction. If the color to be
reproduced has a blue component of zero, then
the yellow filter must attenuate the shortwave
components of the spectrum as much as
possible. As the amount of blue to be repro-
duced increases, the attenuation of the yellow
filter should decrease. This reasoning leads to the
“one-minus-RGB” relationships:
Cyan in tandem with magenta produces blue,
cyan with yellow produces green, and magenta
with yellow produces red. Here is a memory aid:
Additive primaries are at the top, subtractive at
the bottom. On the left, magenta and yellow
filters combine subtractively to produce red. On
the right, red and green sources combine addi-
tively to produce yellow.
Subtractive details
Although the “one-minus-RGB” relationships
are a good starting point, they are not useful in
practice. One problem is that the relationships
assume RGB to represent linear-light tristimulus
values and CMY to represent linear-light trans-
400
Wavelength, nm
500 600 700
Yl
Illuminant
Mg
Cy
Figure 9 Subtractive mixture as employed in color
photography and color offset printing, is illustrated in this
diagram. The colorants act in succession to remove spec-
tral power from the illuminant. In physical terms, the
spectral power of the mixture is, at each wavelength, the
product of the spectrum of the illuminant and the trans-
mission of the colorants: the mixture could be called
multiplicative. If the amount of each colorant is repre-
sented in the form of density – the negative of the
base-10 logarithm of transmission – then color mixtures
can be determined by subtraction.
Color mixtures in subtractive systems are complex
because the colorants have absorption not only in the
intended region of the spectrum but also unwanted
absorptions in other regions. The unwanted absorptions
cause interaction among the channels, and therefore
nonlinearity in the mixtures. The illuminant shown here is
CIE D65; the colorant absorption curves are those of the
dyes of a commercial color photographic print film.
RGB RGB
Cy Mg Yl Cy Mg Yl
Cy 1R=
Mg 1 G=
Yl 1B=
A Guided Tour of Color Space
14
mission values. In computing and in offset
printing, neither the RGB nor the CMY quanti-
ties are directly proportional to intensity.
Another problem is that the “one-minus-RGB
relationships do not incorporate any about infor-
mation about tristimulus values of the assumed
RGB primaries or the assumed CMY colorants!
Obviously the spectra of the colorants has some
influence over reproduced color, but the model
embeds no information about the spectra!
The most serious practical problem with
“subtractive” color mixture is that any overlap
among the absorption spectra of the colorants
results in “unwanted absorption” in the mixture.
For example, the magenta dye is intended to
absorb only medium wavelengths, roughly
between 500 and 600 nm. However, for the
photographic film dyes shown in Figure 9, the
magenta dye also has substantial absorption in
the shortwave region of the spectrum. Conse-
quently, the amount of shortwave light that
contributes to any mixture is not only a function
of the amount of attenuation introduced by the
yellow filter, but also the incidental attenuation
introduced by the magenta filter. There is no
simple way to compensate these interactions,
because the multiplication of spectra is not
mathematically a linear operation.
In practice, the conversion from RGB to CMY is
accomplished using either fairly complicated
polynomial arithmetic, or by three-dimensional
interpolation of lookup tables.
In a subtractive mixture, reproduced white is
determined by characteristics of the colorants
and by the spectrum of the illuminant used at
the display. In a reproduction such as a color
photograph that is illuminated by the ambient
light in the viewer’s environment, mismatch
between the white reference in the scene and
the white reference in the viewing environment
is eliminated. In a subtractive reproduction using
a device that generates its own illumination, the
subjective effect of white mismatch may need to
be considered.
Image synthesis
Once light is on its way to the eye, any tristim-
ulus-based system can accurately represent
color. However, the interaction of light and
objects involves SPDs and spectral absorbtion
curves, not tristimulus values. In synthetic
computer graphics, the calculations are actually
simulating sampled SPDs, even if only three
components are used. Details concerning the
resultant errors are found in Hall10.
10 Roy Hall, Illumination and Color in Computer Gener-
ated Imagery (New York: Springer-Verlag, 1989)
... In this work, the CIE XYZ color space (43) was utilized to analyze the chromaticity of the infected and healthy chicken combs. The extracted RGB data were normalized and converted to CIE XYZ color space using the linear matrix transformation as shown in Equation 1. ...
... Therefore, we can conclude that any small change in the x chromaticity variable would significantly contribute to the classification of the chicken. Since the classifier coefficient of x chromaticity variable results in a positive sign, the increments of x value would increase the value of the sigmoid input function; thus, the results of the sigmoid function would converge to 1. Theoretically, the chroma or actual perceived color was indicated by the x and z values (43). The x chromaticity value can be approximately described as green to red part. ...
... Since the coefficient carries a negative sign (yz model), increasing the z chromaticity value would encourage the classifier model to predict the chicken to be infected. The z chromaticity value can be approximately described as a yellow to blue part for any increment in value (43). Therefore, we conclude that the infected chickens converged more to the blue region according to Figures 7A,B. ...
Article
Full-text available
Bacteria- or virus-infected chicken is conventionally detected by manual observation and confirmed by a laboratory test, which may lead to late detection, significant economic loss, and threaten human health. This paper reports on the development of an innovative technique to detect bacteria- or virus-infected chickens based on the optical chromaticity of the chicken comb. The chromaticity of the infected and healthy chicken comb was extracted and analyzed with International Commission on Illumination (CIE) XYZ color space. Logistic Regression, Support Vector Machines (SVMs), K-Nearest Neighbors (KNN), and Decision Trees have been developed to detect infected chickens using the chromaticity data. Based on the X and Z chromaticity data from the chromaticity analysis, the color of the infected chicken’s comb converged from red to green and yellow to blue. The development of the algorithms shows that Logistic Regression, SVM with Linear and Polynomial kernels performed the best with 95% accuracy, followed by SVM-RBF kernel, and KNN with 93% accuracy, Decision Tree with 90% accuracy, and lastly, SVM-Sigmoidal kernel with 83% accuracy. The iteration of the probability threshold parameter for Logistic Regression models has shown that the model can detect all infected chickens with 100% sensitivity and 95% accuracy at the probability threshold of 0.54. These works have shown that, despite using only the optical chromaticity of the chicken comb as the input data, the developed models (95% accuracy) have performed exceptionally well, compared to other reported results (99.469% accuracy) which utilize more sophisticated input data such as morphological and mobility features. This work has demonstrated a new feature for bacteria- or virus-infected chicken detection and contributes to the development of modern technology in agriculture applications.
... The CIE XYZ colour space described in the literature [14] will be used in the current study to examine the chromaticity of the diseased and healthy chicken comb. As indicated in equation (1), the extracted RGB values were normalized and transformed to CIE XYZ colour space using a linear matrix transformation. ...
... According to Figure 5, the healthy chicken was scattered above 0.4 of x chromaticity value, while the infected chicken was primarily scattered below. Theoretically, the chroma or actual perceived colour of green to red was indicated by the x chromaticity value [14]. So, infected chicken colour combs were converging from red to green region, which can be concluded because infected chickens were mostly scattered to the negative side. ...
... The y chromaticity values of both class chickens were overlapped, and no particular threshold value could be assigned based on the y chromaticity variable. The y chromaticity value represents the brightness, luminosity, or intensity as presented in literature [14]. Therefore, a possible explanation for the y chromaticity results is that our data comes from different sources with different illuminants. ...
... Absorption spectra of (a) EBT solution and (b) calcium EBT complex and spectral ranges covered by channels B and R of the RGB system [18,24]. Inset: photos of the measurement tube containing (I) EBT solution and (II) Ca or Mg EBT complexes. ...
... The determination of calcium and magnesium was based on the discoloration of a blue EBT solution because of complex formation, which is thermodynamically favored: (1) The RGB color system, widely used in digital image colorimetry, was investigated due to the simplicity of data processing and successful previous applications [17,18,22,23]. The absorption spectra of the EBT solution (Fig. 1, graph a) and calcium/magnesium complex correspond to the spectral ranges covered by the R and B channels of the RGB system [18,24], respectively (see e.g. the calcium complex spectra in Fig. 1. graph b). In principle, both channels can be used to monitor complex formation, which turns the blue EBT solution pink (inset Fig. 1). ...
Article
A microanalytical procedure using smartphone-based digital image colorimetry is proposed for determining calcium and magnesium concentrations in biodiesel without analyte extraction. The analytical method relies on the discoloration of an alkaline Eriochrome Black T (EBT) solution because of complex formation with the analytes. Ethanol was used as a mediator solvent for dissolving both biodiesel and EBT. The analytical response was based on the measurement of reflected radiation from digital images captured using a smartphone camera. Photometric measurements were based on the RGB color system, taken R channel values as the analytical signal because of complementarity with the color of the EBT solution. The coefficient of variation (n = 10) and limit of detection were 1.0% and 3 µmol L–1, respectively. A linear response was observed in the 10–75 µmol L–1 range, described by the equation R = 0.612C + 93.1 (r = 0.999). The reagent amounts consumed per determination were as little as 25 µg of EBT and 120 µg of NaOH, with generation of only 935 µL of waste. The procedure was selective for calcium and magnesium, without interference of metal ions (Na+, K+, Zn2+, Cu2+, Ni2+, Fe2+, and Fe3+) as well as other sample components (glycerol, methanol, and water) in concentrations higher than those expected in biodiesel. Matrix effects for biodiesel obtained from different raw materials were negligible (recoveries from 90 to 104%), making feasible external standard calibration with aqueous/ethanolic reference solutions. The results agreed with those obtained by ICP OES after sample preparation by emulsion breaking, demonstrating the reliability of the proposed approach.
... Entropy (i.e., the randomness/complexity of an image) was calculated using Shannon's (1948) method. CIELAB colour space values were calculated as an average for each channel for each image (Poynton, 2015). CIELAB values are based on the opponent-process theory of colour vision, providing a luminance (L*) value, as well as two chromatic measures. ...
Article
Full-text available
The growing interest in harnessing natural environments to enhance mental health, including cognitive functioning and mood, has yielded encouraging results in initial studies. Given that images of nature have demonstrated similar benefits, they are frequently employed as proxies for real-world environments. To ensure precision and control, researchers often manipulate images of natural environments. The effectiveness of this approach relies on standardization of imagery, and therefore, inconsistency in methods and stimuli has limited the synthesis of research findings in the area. Responding to these limitations, the current paper introduces the Salford Nature Environments Database (SNED), a standardized database of natural images created to support ongoing research into the benefits of nature exposure. The SNED currently exists as the most comprehensive nature image database available, comprising 500 high-quality, standardized photographs capturing a variety of possible natural environments across the seasons. It also includes normative scores for user-rated (801 participants) characteristics of fascination, refuge and prospect, compatibility, preference, valence, arousal, and approach–avoidance, as well as data on physical properties of the images, specifically luminance, contrast, entropy, CIELAB colour space parameter values, and fractal dimensions. All image ratings and content detail, along with participant details, are freely available online. Researchers are encouraged to use this open-access database in accordance with the specific aims and design of their study. The SNED represents a valuable resource for continued research in areas such as nature-based therapy, social prescribing, and experimental approaches investigating underlying mechanisms that help explain how natural environments improve mental health and wellbeing.
... Maximum times RGB is taken as a default color space in common sense. RGB defines a color as the percentage of red, green and blue hues mixed together (Poynton, 1995) ("Color Management: Understanding Color Spaces",2017).A color space can be either device dependent or device independent. Device dependent color space is somehow restricted to the parameters and device used for display. ...
Preprint
Color image segmentation is a very emerging research topic in the area of color image analysis and pattern recognition. Many state-of-the-art algorithms have been developed for this purpose. But, often the segmentation results of these algorithms seem to be suffering from miss-classifications and over-segmentation. The reasons behind these are the degradation of image quality during the acquisition, transmission and color space conversion. So, here arises the need of an efficient image enhancement technique which can remove the redundant pixels or noises from the color image before proceeding for final segmentation. In this paper, an effort has been made to study and analyze different image enhancement techniques and thereby finding out the better one for color image segmentation. Also, this comparative study is done on two well-known color spaces HSV and LAB separately to find out which color space supports segmentation task more efficiently with respect to those enhancement techniques.
... The RGB colour space is an additive colour model used to represent colour images, reflecting the sensitivity of three types of cone cells in the retina to specific spectra. In this model, RGB stands for red, green, and blue [36]. In this system, the intensity of each colour can be represented by an integer value ranging from 0 to 255, as illustrated in Figure 1. ...
Article
Full-text available
In the field of machine vision, target detection models have experienced rapid development and have been practically applied in various domains. In agriculture, target detection models are commonly used to identify various types of fruits. However, when it comes to recognizing berries, such as raspberries, the fruits nearing ripeness exhibit highly similar colours, posing a challenge for existing target detection models to accurately identify raspberries in this stage. Addressing this issue, a raspberry detection method called HSA‐YOLOv5 (HSV self‐adaption YOLOv5) is proposed. This method detects immature, nearly ripe, and ripe raspberries. The approach involves transforming the RGB colour space of the original dataset images into an improved HSV colour space. By adjusting corresponding parameters and enhancing the contrast of similar colours while retaining the maximum features of the original image, the method strengthens data features. Adaptive selection of HSV parameters is performed based on data captured under different weather conditions, applying homogeneous preprocessing to the dataset. The improved model is compared with the original YOLOv5 model using a self‐constructed dataset. Experimental results demonstrate that the improved model achieves a mean average precision (mAP) of 0.97, a 6.42 percentage point increase compared to the baseline YOLOv5 model. In terms of immature, nearly ripe, and ripe raspberries, there are improvements of 6, 4, and 7 percentage points, respectively, validating the effectiveness of the proposed model.
... sRGB, Adobe RGB, ProPhotoRGB, etc.), but do so mostly for historical reasons relating to technology. There are a variety of other tristimulus color models that attempt to capture more of the color space detectable by the average human (Poynton, 1995;Stone, 2015), many of which derive from the CIE 1931 RGB and XYZ color space definitions (Wright, 1929). However, it is unclear which model is most beneficial for machine learning algorithms processing sky images. ...
Article
Whole sky spectral radiance distribution measurements are difficult and expensive to obtain, yet important for real-time applications of radiative transfer, building performance, physically based rendering, and photovoltaic panel alignment. This work presents a validated machine learning approach to predicting spectral radiance distributions (350–1780 nm) across the entire hemispherical sky, using regression models trained on high dynamic range (HDR) imagery and spectroradiometer measurements. First, we present and evaluate measured, engineered, and computed machine learning features used to train regression models. Next, we perform experiments comparing regular and HDR imagery, sky sample color models, and spectral resolution. Finally, we present a tool that reconstructs a spectral radiance distribution for every single point of a hemispherical clear sky image given only a photograph of the sky and its capture timestamp. We recommend this tool for building performance and spectral rendering pipelines. The spectral radiance of 81 sample points per test sky is estimated to within 7.5% RMSD overall at 1 nm resolution. Spectral radiance distributions are validated against libRadtran and spectroradiometer measurements. Our entire sky dataset and processing software is open source and freely available on our project website.
... Then, the actual colour of the materials is identified through their relative values. Yet, another important parameter that can be extracted from colour photographs is HSV, which is a linear colour code [54]. In this, H is the angular representation of colour on a scale of 0 • -360 • . ...
Article
Pollution of aquatic bodies with Lead and Mercury is a serious environmental issue, and there exists an unmet need to develop easy and portable detection techniques. Herein, we report a silver nanoparticle (AgNP) based system exhibiting high selectivity and distinct spectrophotometric behaviour towards these metal ions in the aqueous medium. Though many of the metal ions did not influence the characteristic plasmon spectral profile of AgNPs, Mercury caused the disappearance of absorbance at ~400 nm. Consequently, the yellow solution became colourless at around 7.6 ppm of Hg 2+ ions. While the presence of Pb 2+ ions induced a decrease in the initial plasmon intensity with the concomitant emergence of a new red-shifted band above 500 nm. As a result, the colour of the solution turned from yellow to red. These colour changes are found to fall well within the spectral responses of R, G and B filters of the digital camera. By making use of this, an android mobile application was developed for digitizing the colour values and for doing further quantitative analysis. The estimated concentration value by this method was found to be in good match with those obtained from absorbance measurement. The achieved lower detection limits were ~ 0.8 ppm for Hg 2+ and ~ 1.5 ppm for Pb 2+ ions. This illustrates the potential of the proposed combination of smartphone enabled system with AgNPs for easy discrimination and quantification of these two metal ions from the aqueous medium.
Chapter
Voucher image classification is a complex process that can Please confirm if the corresponding author is correctly identified. Amend if necessary. be influenced by many factors. Currently, people use their hands to classify voucher images. However, there are many problems such as processing time, efficiency, and accuracy of classification. When the data is small and their distributions are not uniform, the machine learning techniques used to classify them are not high accuracy. Therefore, the Siamese model is proposed to classify images with little data without equal between classes in this paper. The results show that the model achieves an accuracy of 99.57% for both forward and backward images when the number of epochs for training is only 100. This proves that the proposed method has a promising application in practice.KeywordsImage classificationMachine learningObject detectionSiamese modelDeep learning
Article
Full-text available
Tone reproduction curves (TRCs) have long been the basis for color adjustment in printing systems. Moreover, new algorithms for their processing and calculation, including tone value increase (TVI)‐based ones, are being constantly proposed. However, due to the fact that TRCs are just one‐dimensional curves covering only one colorant, they are not able to produce an exact match of all colors. Color adjustment with the help of the TRC is unstable, that is, a slight change in one color channel has a dramatic effect on the color reproduction in output. The use of gradation trajectories (GT) as an analytical model within the CIE Lab color space describing the behavior of the entire color gamut of the reproduction device makes it possible to more accurately adjust the individual color channel. Describing GT using the parametric Bézier curves further simplifies this process. In this work, we use GT to fit the TRC/TVI with the parameters recommended by ISO 12647‐2:2013.
Conference Paper
A method is presented for accurate color reproduction among a wide variety of display devices. The method is very general, in that it may be applied to virtually any color display device. Its generality has been demonstrated by application to color monitors, film recorders, electronic pre-press systems and color hardcopy devices. The algorithm has been used to accurately translate between device dependent and device independent color specifications and to translate from one device dependent color specification to another.The method separates the color reproduction process into two distinct components: device characterization, which accounts for the colorimetric properties of each class of display device, and device calibration, which accounts for local variations from one instance of a device to another.A companded RGB color space is introduced, which is used with trivariate parametric polynomial volumes (i.e. hyperpatches) to perform accurate color transformations.A color separation algorithm is presented which converts companded RGB to and from the subtractive printing colors (cyan, magenta, yellow, black) using gray component replacement.
Article
The Commission Internationale d'Eclairage system of colorimetry is a method of measuring colours that has been standardized, and is widely used by industries involved with colour. Knowing the CIE coordinates of a colour allows it to be reproduced easily and exactly in many different media. For this reason graphics installations which utilize colour extensively ought to have the capability of knowing the CIE coordinates of displayed colours, and of displaying colours of given CIE coordinates. Such a capability requires a function which transforms video monitor gun voltages (RGB colour space) into CIE coordinates (XYZ colour space), and vice versa. The function incorporates certain monitor parameters. The purpose of this paper is to demonstrate the form that such a function takes, and to show how the necessary monitor parameters can be measured using little more than a simple light meter. Because space is limited, and because each user is likely to implement the calibration differently, few technical details are given, but principles and methods are discussed in sufficient depth to allow the full use of the system. In addition, several visual checks which can be used for quick verification of the integrity of the calibration are described. The paper begins with an overview of the CIE system of colorimetry. It continues with a general discussion of transformations from RGB colour space to XYZ colour space, after which a detailed step-by-step procedure for monitor calibration is presented.
Book
This is a discussion of the physics of illumination and the associated techniques for modeling global and local illumination in computer generated imagery. It was state-of-the-art in 1988, but is now rather outdated. It does include discussions of physics and color theory basics that have not changed, and a discussion of illumination models through ray tracing models using various specular reflectance functions and including Fresnel effects. This text is currently out of print. However, we still receive numerous requests for an electronic version of the source code in the book.