Color Image Processing and Applications

22 downloads 0 Views 569KB Size Report
Feb 18, 2000 - 7.4 Lossless Waveform-based Image compression Techniques . . . . 286 .... 1.15 The opponent color stage of the human visual system. . . . . . . . . . . ...... Cyan. Magenta. White. I=1. Black I=0. Intensity. P. Fig. 1.10. The HSI. Color ...
K.N. Plataniotis and A.N. Venetsanopoulos

Color Image Processing and Applications Engineering { Monograph (English)

February 18, 2000

Springer-Verlag Berlin Heidelberg New York London Paris Tokyo Hong Kong Barcelona Budapest

Preface

The perception of color is of paramount importance to humans since they routinely use color features to sense the environment, recognize objects and convey information. Color image processing and analysis is concerned with the manipulation of digital color images on a computer utilizing digital signal processing techniques. Like most advanced signal processing techniques, it was, until recently, con ned to academic institutions and research laboratories that could a ord the expensive image processing hardware needed to handle the processing overhead required to process large numbers of color images. However, with the advent of powerful desktop computers and the proliferation of image collection devices, such as digital cameras and scanners, color image processing techniques are now within the grasp of the general public. This book is aimed at researchers and practitioners that work in the area of color image processing. Its purpose is to ll an existing gap in scienti c literature by presenting the state of the art research in the area. It is written at a level which can be easily understood by a graduate student in an Electrical and Computer Engineering or Computer Science program. Therefore, it can be used as a textbook that covers part of a modern graduate course in digital image processing or multimedia systems. It can also be used as a textbook for a graduate course on digital signal processing since it contains algorithms, design criteria and architectures for processing and analysis systems. The book is structured into four parts. The rst, Chapter 1, deals with color principles and is aimed at readers who have very little prior knowledge of color science. Readers interested in color image processing may read the second part of the book (Chapters 2-5). It covers the major, although somewhat mature, elds of color image processing. Color image processing is characterized by a large number of algorithms that are speci c solutions to speci c problems, for example vector median lters have been developed to remove impulsive noise from images. Some of them are mathematical or content independent operations that are applied to each and every pixel, such as morphological operators. Others are algorithmic in nature, in the sense that a recursive strategy may be necessary to nd edge pixels in an image.

The third part of the book, Chapters 6-7, deals with color image analysis and coding techniques. The ultimate goal of color image analysis is to enhance human-computer interaction. Recent applications of image analysis includes compression of color images either for transmission across the internetwork or coding of video images for video conferencing. Finally, the fourth part (Chapter 8) covers emerging applications of color image processing. Color is useful for accessing multimedia databases. Local color information, for example in the form of color histograms, can be used to index and retrieve images from the database. Color features can also be used to identify objects of interest, such as human faces and hand areas, for applications ranging from video conferencing, to perceptual interfaces and virtual environments. Because of the dual nature of this investigation, processing and analysis, the logical dependence of the chapters is somewhat unusual. The following diagram can help the reader chart the course.

Logical dependence between chapters

IX Acknowledgment

We acknowledge a number of individuals who have contributed in di erent ways to the preparation of this book. In particular, we wish to extend our appreciation to Prof. M. Zervakis for contributing the image restoration section, and to Dr. N. Herodotou for his informative inputs and valuable suggestions in the emerging applications chapter. Three graduate students of ours also merit special thanks. Shu Yu Zhu for her input and high quality gures included in the color edge detection chapter, Ido Rabinovitch for his contribution to the color image coding section and Nicolaos Ikonomakis for his valuable contribution in the color segmentation chapter. We also thank Nicolaos for reviewing the chapters of the book and helping with the Latex formating of the manuscript. We also grateful to Terri Vlassopoulos for proofreading the manuscript, and Frank Holzwarth of Springer Verlag for his help during the preparation of the book. Finally, we are indebted to Peter Androutsos who helped us tremendously on the development of the companion software.

X

Contents

1. Color Spaces : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 1.1 1.2 1.3 1.4 1.5

1.6 1.7 1.8 1.9

1.10 1.11 1.12 1.13 1.14

Basics of Color Vision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The CIE Chromaticity-based Models . . . . . . . . . . . . . . . . . . . . . . The CIE-RGB Color Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Gamma Correction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Linear and Non-linear RGB Color Spaces . . . . . . . . . . . . . . . . . . 1.5.1 Linear RGB Color Space . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5.2 Non-linear RGB Color Space . . . . . . . . . . . . . . . . . . . . . . . Color Spaces Linearly Related to the RGB . . . . . . . . . . . . . . . . . The YIQ Color Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The HSI Family of Color Models . . . . . . . . . . . . . . . . . . . . . . . . . Perceptually Uniform Color Spaces . . . . . . . . . . . . . . . . . . . . . . . 1.9.1 The CIE L u v Color Space . . . . . . . . . . . . . . . . . . . . . . 1.9.2 The CIE L a b Color Space . . . . . . . . . . . . . . . . . . . . . . 1.9.3 Cylindrical L u v and L a b Color Space . . . . . . . . . . 1.9.4 Applications of L u v and L a b spaces . . . . . . . . . . . The Munsell Color Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The Opponent Color Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . New Trends . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Color Images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

1 1 4 9 13 16 16 17 20 23 25 32 33 35 37 37 39 41 42 45 45

2. Color Image Filtering : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 51 2.1 2.2 2.3 2.4 2.5

Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Color Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Modeling Sensor Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Modeling Transmission Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . Multivariate Data Ordering Schemes . . . . . . . . . . . . . . . . . . . . . . 2.5.1 Marginal Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.2 Conditional Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.3 Partial Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.4 Reduced Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.6 A Practical Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7 Vector Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

51 52 53 55 58 59 62 62 63 67 69

XII

2.8 2.9 2.10 2.11 2.12 2.13 2.14 2.15

The Distance Measures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The Similarity Measures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Filters Based on Marginal Ordering . . . . . . . . . . . . . . . . . . . . . . . Filters Based on Reduced Ordering . . . . . . . . . . . . . . . . . . . . . . . Filters Based on Vector Ordering . . . . . . . . . . . . . . . . . . . . . . . . . Directional-based Filters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Computational Complexity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

70 72 77 81 89 92 98 100

: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 107 3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107 3.2 The Adaptive Fuzzy System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109 3.2.1 Determining the Parameters . . . . . . . . . . . . . . . . . . . . . . . 112 3.2.2 The Membership Function . . . . . . . . . . . . . . . . . . . . . . . . . 113 3.2.3 The Generalized Membership Function . . . . . . . . . . . . . . 115 3.2.4 Members of the Adaptive Fuzzy Filter Family . . . . . . . . 116 3.2.5 A Combined Fuzzy Directional and Fuzzy Median Filter122 3.2.6 Comments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125 3.2.7 Application to 1-D Signals . . . . . . . . . . . . . . . . . . . . . . . . . 128 3.3 The Bayesian Parametric Approach . . . . . . . . . . . . . . . . . . . . . . . 131 3.4 The Non-parametric Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . 137 3.5 Adaptive Morphological Filters . . . . . . . . . . . . . . . . . . . . . . . . . . . 146 3.5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146 3.5.2 Computation of the NOP and the NCP . . . . . . . . . . . . . 152 3.5.3 Computational Complexity and Fast Algorithms . . . . . 154 3.6 Simulation Studies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157 3.7 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 173

3. Adaptive Image Filters

4. Color Edge Detection : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 179

4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 179 4.2 Overview Of Color Edge Detection Methodology . . . . . . . . . . . 181 4.2.1 Techniques Extended From Monochrome Edge Detection181 4.2.2 Vector Space Approaches . . . . . . . . . . . . . . . . . . . . . . . . . . 183 4.3 Vector Order Statistic Edge Operators . . . . . . . . . . . . . . . . . . . . 189 4.4 Di erence Vector Operators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 194 4.5 Evaluation Procedures and Results . . . . . . . . . . . . . . . . . . . . . . . 197 4.5.1 Probabilistic Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . 198 4.5.2 Noise Performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200 4.5.3 Subjective Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201 4.6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203

XIII

5. Color Image Enhancement and Restoration : : : : : : : : : : : : : : : 209 5.1 5.2 5.3 5.4 5.5

Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Histogram Equalization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Color Image Restoration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Restoration Algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Algorithm Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.5.1 De nitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.5.2 Direct Algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.5.3 Robust Algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.6 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

209 210 214 217 220 220 223 227 229

6. Color Image Segmentation : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 237 6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2 Pixel-based Techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.1 Histogram Thresholding . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.2 Clustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3 Region-based Techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.1 Region Growing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.2 Split and Merge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4 Edge-based Techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.5 Model-based Techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.5.1 The Maximum A-posteriori Method . . . . . . . . . . . . . . . . 6.5.2 The Adaptive MAP Method . . . . . . . . . . . . . . . . . . . . . . . 6.6 Physics-based Techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.7 Hybrid Techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8 Application . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.1 Pixel Classi cation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.2 Seed Determination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.3 Region Growing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.4 Region Merging . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.5 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.9 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

7. Color Image Compression 7.1 7.2 7.3 7.4

237 239 239 242 247 248 250 252 253 254 255 256 257 260 260 262 267 269 271 273

: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 279

Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Image Compression Comparison Terminology . . . . . . . . . . . . . . Image Representation for Compression Applications . . . . . . . . Lossless Waveform-based Image Compression Techniques . . . . 7.4.1 Entropy Coding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.2 Lossless Compression Using Spatial Redundancy . . . . . 7.5 Lossy Waveform-based Image Compression Techniques . . . . . . 7.5.1 Spatial Domain Methodologies . . . . . . . . . . . . . . . . . . . . . 7.5.2 Transform Domain Methodologies . . . . . . . . . . . . . . . . . . 7.6 Second Generation Image Compression Techniques . . . . . . . . . 7.7 Perceptually Motivated Compression Techniques . . . . . . . . . . .

279 282 285 286 286 288 290 290 292 304 307

XIV

7.7.1 Modeling the Human Visual System . . . . . . . . . . . . . . . . 7.7.2 Perceptually Motivated DCT Image Coding . . . . . . . . . 7.7.3 Perceptually Motivated Wavelet-based Coding . . . . . . . 7.7.4 Perceptually Motivated Region-based Coding . . . . . . . . 7.8 Color Video Compression . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.9 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

307 311 313 317 319 324

8. Emerging Applications : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 329 8.1 Input Analysis Using Color Information . . . . . . . . . . . . . . . . . . . 8.2 Shape and Color Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.1 Fuzzy Membership Functions . . . . . . . . . . . . . . . . . . . . . . 8.2.2 Aggregation Operators . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

A. Companion Image Processing Software A.1 A.2 A.3 A.4

331 337 338 340 343 345

: : : : : : : : : : : : : : : : : : : 349

Image Filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Image Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Image Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Noise Generation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

350 350 351 351

Index : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 353

List of Figures

1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 1.10 1.11 1.12 1.13 1.14 1.15 1.16

The visible light spectrum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The CIE XYZ color matching functions . . . . . . . . . . . . . . . . . . . . . . . The CIE RGB color matching functions . . . . . . . . . . . . . . . . . . . . . . . The chromaticity diagram . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The Maxwell triangle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The RGB color model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Linear to Non-linear Light Transformation . . . . . . . . . . . . . . . . . . . . . Non-linear to linear Light Transformation . . . . . . . . . . . . . . . . . . . . . Transformation of Intensities from Image Capture to Image Display The HSI Color Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The HLS Color Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The HSV Color Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The L u v Color Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The Munsell color system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The Opponent color stage of the human visual system. . . . . . . . . . . A taxonomy of color models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

1 7 7 9 10 11 18 19 19 26 31 31 34 40 42 46

3.1 3.2 3.3 3.4 3.5 3.6 3.7 3.8 3.9 3.10

Simulation I: Filter outputs (1st component) . . . . . . . . . . . . . . . . . . . Simulation I: Filter outputs (2nd component) . . . . . . . . . . . . . . . . . . Simulation II: Actual signal and noisy input (1st component) . . . . Simulation II: Actual signal and noisy input (2nd component) . . . . Simulation II: Filter outputs (1st component) . . . . . . . . . . . . . . . . . . Simulation II: Filter outputs (2nd component) . . . . . . . . . . . . . . . . . . A owchart of the NOP research algorithm . . . . . . . . . . . . . . . . . . . . The adaptive morphological lter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . `Peppers' corrupted by 4% impulsive noise . . . . . . . . . . . . . . . . . . . . `Lenna' corrupted with Gaussian noise  = 15 mixed with 2% impulsive noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . V MF of (3.9) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . BV DF of (3.9) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . . HF of (3.9) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . AHF of (3.9) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . FV DF of (3.9) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . . ANNMF of (3.9) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . CANNMF of (3.9) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . .

129 129 130 131 132 132 155 157 169

3.11 3.12 3.13 3.14 3.15 3.16 3.17

169 170 170 170 170 170 170 170

XVI

3.18 BFMA of (3.9) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.19 V MF of (3.10) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.20 BV DF of (3.10) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.21 HF of (3.10) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.22 AHF of (3.10) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.23 FV DF of (3.10) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.24 ANNMF of (3.10) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . 3.25 CANNMF of (3.10) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . 3.26 BFMA of (3.10) using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . 3.27 `Mandrill' - 10% impulsive noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.28 NOP-NCP ltering results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.29 V MF using 3x3 window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.30 Mutistage Close-opening ltering results . . . . . . . . . . . . . . . . . . . . . . .

170 171 171 171 171 171 171 171 171 173 173 173 173

4.1 4.2 4.3 4.4 4.5 4.6 4.7 4.8 4.9 4.10 4.11 4.12 4.13 4.14 4.15 4.16 4.17

180 195 202 202 202 203 203 203 203 204 204 204 204 205 205 205 205

Edge detection by derivative operators . . . . . . . . . . . . . . . . . . . . . . . . Sub-window Con gurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Test color image `ellipse' . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Test color image ` ower' . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Test color image `Lenna' . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of `ellipse': Sobel detector . . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of `ellipse': VR detector . . . . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of `ellipse': DV detector . . . . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of `ellipse': DV hv detector . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of ` ower': Sobel detector . . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of ` ower': VR detector . . . . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of ` ower': DV detector . . . . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of ` ower': DVadap detector . . . . . . . . . . . . . . . . . . . . . . . . Edge map of `Lenna': Sobel detector . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of `Lenna': VR detector . . . . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of `Lenna': DV detector . . . . . . . . . . . . . . . . . . . . . . . . . . . . Edge map of `Lenna': DVadap detector . . . . . . . . . . . . . . . . . . . . . . . .

5.1 The original color image `mountain' . . . . . . . . . . . . . . . . . . . . . . . . . . . 215 5.2 The histogram equalized color output . . . . . . . . . . . . . . . . . . . . . . . . . 215 6.1 6.2 6.3 6.4 6.5 6.6 6.7 6.8 6.9 6.10

Partitioned image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Corresponding quad-tree . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The HSI cone with achromatic region in yellow . . . . . . . . . . . . . . . . . Original image. Achromatic pixels: intensity < 10, > 90 . . . . . . . . . Saturation < 5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Saturation < 10 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Saturation< 15 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Original image. Achromatic pixels: saturation < 10, intensity> 90 Intensity < 5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Intensity < 10 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

250 250 261 262 262 262 262 263 263 263

XVII

6.11 6.12 6.13 6.14 6.15 6.16 6.17 6.18 6.19 6.20 6.21 6.22 6.23 6.24 6.25 6.26 6.27 6.28 6.29 6.30 6.31 6.32 6.33 7.1 7.2 7.3 7.4 7.5 7.6 7.7 7.8 7.9 7.10 7.11 7.12 7.13 7.14 7.15

Intensity < 15 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Original image. Achromatic pixels: saturation< 10, intensity< 10 . Intensity > 85 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Intensity > 90 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Intensity > 95 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Original image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Pixel classi cation with chromatic pixels in red and achromatic pixels in the original color . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Original image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Pixel classi cation with chromatic pixels in tan and achromatic pixels in the original color . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Arti cial image with level 1, 2, and 3 seeds. . . . . . . . . . . . . . . . . . . . . The region growing algorithm. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Original 'Claire' image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 'Claire' image showing seeds with V AR = 0:2 . . . . . . . . . . . . . . . . . . Segmented 'Claire' image (before merging), Tchrom = 0:15 . . . . . . . Segmented 'Claire' image (after merging), Tchrom = 0:15 and Tmerge = 0:2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Original 'Carphone' image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 'Carphone' image showing seeds with V AR = 0:2 . . . . . . . . . . . . . . . Segmented 'Carphone' image (before merging), Tchrom = 0:15 . . . . Segmented 'Carphone' image (after merging), Tchrom = 0:15 and Tmerge = 0:2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Original 'Mother-Daughter' image . . . . . . . . . . . . . . . . . . . . . . . . . . . . 'Mother-Daughter' image showing seeds with V AR = 0:2 . . . . . . . . Segmented 'Mother-Daughter' image (before merging), Tchrom = 0:15 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Segmented 'Mother-Daughter' image (after merging), Tchrom = 0:15 and Tmerge = 0:2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

263 264 264 264 264 265

The zig-zag scan . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . DCT based coding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Original color image `Peppers' . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Image coded at a compression ratio 5 : 1 . . . . . . . . . . . . . . . . . . . . . . . Image coded at a compression ratio 6 : 1 . . . . . . . . . . . . . . . . . . . . . . . Image coded at a compression ratio 6:3 : 1 . . . . . . . . . . . . . . . . . . . . . Image coded at a compression ratio 6:35 : 1 . . . . . . . . . . . . . . . . . . . . Image coded at a compression ratio 6:75 : 1 . . . . . . . . . . . . . . . . . . . . Subband coding scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Relationship between di erent scale subspaces . . . . . . . . . . . . . . . . . . Multiresolution analysis decomposition . . . . . . . . . . . . . . . . . . . . . . . . The wavelet-based scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Second generation coding schemes . . . . . . . . . . . . . . . . . . . . . . . . . . . . The human visual system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Overall operation of the processing module . . . . . . . . . . . . . . . . . . . .

297 298 299 299 299 299 299 299 301 302 303 304 304 307 318

265 265 265 266 267 270 270 270 270 271 271 271 271 272 272 272 272

XVIII

7.16 MPEG-1: Coding module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 322 7.17 MPEG-1: Decoding module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 322 8.1 8.2 8.3 8.4 8.5 8.6 8.7 8.8 8.9 8.10 8.11 8.12

Skin and Lip Clusters in the RGB color space . . . . . . . . . . . . . . . . . . Skin and Lip Clusters in the L a b color space . . . . . . . . . . . . . . . . Skin and Lip hue Distributions in the HSV color space . . . . . . . . . . Overall scheme to extract the facial regions within a scene . . . . . . . Template for hair color classi cation = R1 + R2 + R3 . . . . . . . . . . . Carphone: Frame 80 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Segmented frame . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Frames 20-95 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Miss America: Frame 20 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Frames 20-120 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Akiyo: Frame 20 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Frames 20-110 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

333 333 334 337 342 344 344 344 345 345 345 345

A.1 A.2 A.3 A.4

Screenshot of the main CIPAView window at startup. . . . . . . . . . . . Screenshot of Di erence Vector Mean edge detector being applied Gray scale image quantized to 4 levels . . . . . . . . . . . . . . . . . . . . . . . . . Screenshot of an image being corrupted by Impulsive Noise.. . . . . .

350 351 352 352

List of Tables

1.1 EBU Tech 3213 Primaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 1.2 EBU Tech 3213 Primaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 1.3 Color Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 2.1 Computational Complexity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100 3.1 3.2 3.3 3.4 3.5 3.6 3.7 3.8 3.9 3.10 3.11 3.12 3.13 3.14

Noise Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Filters Compared . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Subjective Image Evaluation Guidelines . . . . . . . . . . . . . . . . . . . . . . . Figure of Merit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . NMSE(x10 2 ) for the RGB `Lenna' image, 33 window . . . . . . . . . NMSE(x10 2 ) for the RGB `Lenna' image, 55 window . . . . . . . . . NMSE(x10 2 ) for the RGB `peppers' image, 33 window . . . . . . . NMSE(x10 2 ) for the RGB `peppers' image, 55 window . . . . . . . NCD for the RGB `Lenna' image, 33 window . . . . . . . . . . . . . . . . . NCD for the RGB `Lenna' image, 55 window . . . . . . . . . . . . . . . . . NCD for the RGB `peppers' image, 33 window . . . . . . . . . . . . . . . . NCD for the RGB `peppers' image, 55 window . . . . . . . . . . . . . . . . Subjective Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Performance measures for the image Mandrill . . . . . . . . . . . . . . . . . .

158 159 161 162 164 165 165 166 166 167 167 168 168 172

4.1 4.2 4.3 4.4

Vector Order Statistic Operators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Di erence Vector Operators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Numerical Evaluation with Synthetic Images . . . . . . . . . . . . . . . . . . . Noise Performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

198 199 199 201

6.1 Comparison of Chromatic Distance Measures . . . . . . . . . . . . . . . . . . 269 6.2 Color Image Segmentation Techniques . . . . . . . . . . . . . . . . . . . . . . . . 273 7.1 Storage requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 280 7.2 A taxonomy of image compression methodologies: First Generation283 7.3 A taxonomy of image compression methodologies: Second Generation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 283 7.4 Quantization table for the luminance component . . . . . . . . . . . . . . . 296 7.5 Quantization table for the chrominance components . . . . . . . . . . . . 296

XX

7.6 The JPEG suggested quantization table . . . . . . . . . . . . . . . . . . . . . . . 312 7.7 Quantization matrix based on the contrast sensitivity function for 1.0 min/pixel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 312 8.1 Miss America (WidthHeight=360288):Shape & Color Analysis. 343

1. Color Spaces

1.1 Basics of Color Vision

Color is a sensation created in response to excitation of our visual system by electromagnetic radiation known as light [1], [2], [3]. More speci c, color is the perceptual result of light in the visible region of the electromagnetic spectrum, having wavelengths in the region of 400nm to 700nm, incident upon the retina of the human eye. Physical power or radiance of the incident light is in a spectral power distribution (SPD), often divided into 31 components each representing a 10nm band [4]-[13].

1.1. The visible light spectrum Fig.

The human retina has three types of color photo-receptor cells, called cones , which respond to radiation with somewhat di erent spectral response curves [4]-[5]. A fourth type of photo-receptor cells, called roads , are also present in the retina. These are e ective only at extremely low light levels, for example during night vision. Although rods are important for vision, they play no role in image reproduction [14], [15]. The branch of color science concerned with the appropriate description and speci cation of a color is called colorimetry [5], [10]. Since there are exactly three types of color photo-receptor cone cells, three numerical components are necessary and suÆcient to describe a color, providing that appropriate spectral weighting functions are used. Therefore, a color can be speci ed by a tri-component vector. The set of all colors form a vector space called color space or color model. The three components of a color can be de ned in many di erent ways leading to various color spaces [5], [9]. Before proceeding with color speci cation systems (color spaces), it is appropriate to de ne a few terms: Intensity (usually denoted I ), brightness

2

(Br), luminance (Y ), lightness (L ), hue (H ) and saturation (S ), which are often confused or misused in the literature. The intensity (I ) is a measure, over some interval of the electromagnetic spectrum, of the ow of power that is radiated from, or incident on a surface and expressed in units of watts per square meter [4], [18], [16]. The intensity (I ) is often called a linear light measure and thus is expressed in units, such as watts per square meter [4], [5]. The brightness (Br) is de ned as the attribute of a visual sensation according to which an area appears to emit more or less light [5]. Since brightness perception is very complex, the Commission Internationale de L'Eclairage (CIE) de ned another quantity luminance (Y ) which is radiant power weighted by a spectral sensitivity function that is characteristic of human vision [5]. Human vision has a nonlinear perceptual response to luminance which is called lightness (L ). The nonlinearity is roughly logarithmic [4]. Humans interpret a color based on its lightness (L ), hue (H ) and saturation (S ) [5]. Hue is a color attribute associated with the dominant wavelength in a mixture of light waves. Thus hue represents the dominant color as perceived by an observer; when an object is said to be red, orange, or yellow the hue is being speci ed. In other words, it is the attribute of a visual sensation according to which an area appears to be similar to one of the perceived colors: red, yellow, green and blue, or a combination of two of them [4], [5]. Saturation refers to the relative purity or the amount of white light mixed with a hue. The pure spectrum colors are fully saturated and contain no white light. Colors such as pink (red and white) and lavender (violet and white) are less saturated, with the degree of saturation being inversely proportional to the amount of white light added [1]. A color can be de-saturated by adding white light that contains power at all wavelengths [4]. Hue and saturation together describe the chrominance. The perception of color is basically determined by luminance and chrominance [1]. To utilize color as a visual cue in multimedia, image processing, graphics and computer vision applications, an appropriate method for representing the color signal is needed. The di erent color speci cation systems or color models (color spaces or solids) address this need. Color spaces provide a rational method to specify, order, manipulate and e ectively display the object colors taken into consideration. A well chosen representation preserves essential information and provides insight to the visual operation needed. Thus, the selected color model should be well suited to address the problem's statement and solution. The process of selecting the best color representation involves knowing how color signals are generated and what information is needed from these signals. Although color spaces impose constraints on color perception and representation they also help humans perform important tasks. In particular, the color models may be used to de ne colors, discriminate between colors, judge similarity between color and identify color categories for a number of applications [12], [13].

3

Color model literature can be found in the domain of modern sciences, such as physics, engineering, arti cial intelligence, computer science, psychology and philosophy. In the literature four basic color model families can be distinguished [14]: 1. Colorimetric color models, which are based on physical measurements of spectral re ectance. Three primary color lters and a photo-meter, such as the CIE chromaticity diagram usually serve as the initial points for such models. 2. Psychophysical color models, which are based on the human perception of color. Such models are either based on subjective observation criteria and comparative references (e.g. Munsell color model) or are built through experimentation to comply with the human perception of color (e.g. Hue, Saturation and Lightness model). 3. Physiologically inspired color models, which are based on the three primaries, the three types of cones in the human retina. The Red-GreenBlue (RGB) color space used in computer hardware is the best known example of a physiologically inspired color model. 4. Opponent color models, which are based on perception experiments, utilizing mainly pairwise opponent primary colors, such as the YellowBlue and Red-Green color pairs. In image processing applications, color models can alternatively be divided into three categories. Namely: 1. Device-oriented color models, which are associated with input, processing and output signal devices. Such spaces are of paramount importance in modern applications, where there is a need to specify color in a way that is compatible with the hardware tools used to provide, manipulate or receive the color signals. 2. User-oriented color models, which are utilized as a bridge between the human operators and the hardware used to manipulate the color information. Such models allow the user to specify color in terms of perceptual attributes and they can be considered an experimental approximation of the human perception of color. 3. Device-independent color models, which are used to specify color signals independently of the characteristics of a given device or application. Such models are of importance in applications, where color comparisons and transmission of visual information over networks connecting di erent hardware platforms are required. In 1931, the Commission Internationale de L'Eclairage (CIE) adopted standard color curves for a hypothetical standard observer. These color curves specify how a speci c spectral power distribution (SPD) of an external stimulus (visible radiant light incident on the eye) can be transformed into a set of three numbers that specify the color. The CIE color speci cation system

4

is based on the description of color as the luminance component Y and two additional components X and Z [5]. The spectral weighting curves of X and Z have been standardized by the CIE based on statistics from experiments involving human observers [5]. The CIE XYZ tristimulus values can be used to describe any color. The corresponding color space is called the CIE XYZ color space. The XYZ model is a device independent color space that is useful in applications where consistent color representation across devices with di erent characteristics is important. Thus, it is exceptionally useful for color management purposes. The CIE XYZ space is perceptually highly non uniform [4]. Therefore, it is not appropriate for quantitative manipulations involving color perception and is seldom used in image processing applications [4], [10]. Traditionally, color images have been speci ed by the non-linear red (R0 ), green (G0 ) and blue (B 0 ) tristimulus values where color image storage, processing and analysis is done in this non-linear RGB (R0 G0 B0 ) color space. The red, green and blue components are called the primary colors . In general, hardware devices such as video cameras, color image scanners and computer monitors process the color information based on these primary colors. Other popular color spaces in image processing are the YIQ (North American TV standard), the HSI (Hue, Saturation and Intensity), and the HSV (Hue, Saturation, Value) color spaces used in computer graphics. Although XYZ is used only indirectly it has a signi cant role in image processing since other color spaces can be derived from it through mathematical transforms. For example, the linear RGB color space can be transformed to and from the CIE XYZ color space using a simple linear three-by-three matrix transform. Similarly, other color spaces, such as non-linear RGB, YIQ and HSI can be transformed to and from the CIE XYZ space, but might require complex and non-linear computations. The CIE have also derived and standardized two other color spaces, called L u v and L a b , from the CIE XYZ color space which are perceptually uniform [5]. The rest of this chapter is devoted to the analysis of the di erent color spaces in use today. The di erent color representation models are discussed and analyzed in detail with emphasis placed on motivation and design characteristics. 1.2 The CIE Chromaticity-based Models

Over the years, the CIE committee has sponsored the research of color perception. This has lead to a class of widely used mathematical color models. The derivation of these models has been based on a number of color matching experiments, where an observer judges whether two parts of a visual stimulus match in appearance. Since the colorimetry experiments are based on a matching procedure in which the human observer judges the visual similarity of two areas the theoretical model predicts only matching and not perceived

5

colors. Through these experiments it was found that light of almost any spectral composition can be matched by mixtures of only three primaries (lights of a single wavelength). The CIE had de ned a number of standard observer color matching functions by compiling experiments with di erent observers, di erent light sources and with various power and spectral compositions. Based on the experiments performed by CIE early in this century, it was determined that these three primary colors can be broadly chosen, provided that they are independent. The CIE's experimental matching laws allow for the representation of colors as vectors in a three-dimensional space de ned by the three primary colors. In this way, changes between color spaces can be accomplished easily. The next few paragraphs will brie y outline how such a task can be accomplished. According to experiments conducted by Thomas Young in the nineteenth century [19], and later validated by other researchers [20], there are three di erent types of cones in the human retina, each with di erent absorption spectra: S1 (), S2 (), S3 (), where 380780 (nm). These approximately peak in the yellow-green, green and blue regions of the electromagnetic spectrum with signi cant overlap between S1 and S2 . For each wavelength the absorption spectra provides the weight with which light of a given spectral distribution (SPD) contributes to the cone's output. Based on Young's theory, the color sensation that is produced by a light having SPD C () can be de ned as:

i (C ) =

Z 2 1

Si ()C () d

(1.1)

for i = 1; 2; 3. According to (1.1) any two colors C1 (), C2 () such that i (C1 ) = i (C2 ) , i = 1; 2; 3 will be perceived to be identical even if C1() and C2 () are di erent. This well known phenomenon of spectrally di erent

stimuli that are indistinguishable to a human observer is called metamers [14] and constitutes a rather dramatic illustration of the perceptual nature of color and the limitations of the color modeling process. Assume that three primary colors Ck , k = 1; 2; 3 with SPD Ck () are available and let Z

Ck () d = 1 (1.2) To match a color C with spectral energy distribution C (), the three primaries are mixed in proportions of k , k = 1; 2; 3. Their linear combination P3 k=1 k Ck () should be perceived as C (). Substituting this into (1.1) leads to:

i (C ) = for i = 1; 2; 3.

Z

(

3 X k=1

k Ck ())Si () d =

3 X k=1

Z

k Si ()Ck () d

(1.3)

6 R

The quantity Si ()Ck () d can be interpreted as the ith , cone response generated by one unit of the k th primary color: Z

i;k = i (Ck ) = Si()Ck () d

i = 1; 2; 3 (1.4)

Therefore, the color matching equations are: 3 X k=1

Z

k i;k = i (C ) = Si ()C () d

(1.5)

assuming a certain set of primary colors Ck () and spectral sensitivity curves Si (). For a given arbitrary color, k can be found by simply solving (1.4) and (1.5). Following the same approach wk can be de ned as the amount of the kth primary required to match the reference white, providing that there is available a reference white light source with known energy distribution w(). In such a case, the values obtained through

Tk (C ) = w k (1.6) k for k = 1; 2; 3 are called tristimulus values of the color C , and determine the relative amounts of primitives required to match that color. The tristimulus values of any given color C () can be obtained given the spectral tristimulus values Tk (), which are de ned as the tristimulus values of unit energy spectral color at wavelength  . The spectral tristimulus Tk () provide the so-called spectral matching curves which are obtained by setting C () = Æ( ) in (1.5).

The spectral matching curves for a particular choice of color primaries with an approximately red, green and blue appearance were de ned in the CIE 1931 standard [9]. A set of pure monochromatic primaries are used, blue (435:8nm), green (546:1nm) and red (700nm). In Figures 1.2 and 1.3 the Yaxis indicates the relative amount of each primary needed to match a stimulus of the wavelength reported on the X-axis. It can be seen that some of the values are negative. Negative numbers require that the primary in question be added to the opposite side of the original stimulus. Since negative sources are not physically realizable it can be concluded that the arbitrary set of three primary sources cannot match all the visible colors. However, for any given color a suitable set of three primary colors can be found. Based on the assumption that the human visual system behaves linearly, the CIE had de ned spectral matching curves in terms of virtual primaries. This constitutes a linear transformation such that the spectral matching curves are all positive and thus immediately applicable for a range of practical situations. The end results are referred to as the CIE 1931 standard observer matching curves and the individual curves (functions) are labeled

7

x, y, z respectively. In the CIE 1931 standard the matching curves were selected so that y was proportional to the human luminosity function, which was an experimentally determined measure of the perceived brightness of monochromatic light. CIE 1964 XYZ color matching functions 2.5

2

1.5

1

0.5

0 0

50

100

150

200 250 300 Wavelength, nm

350

The CIE XYZ color matching functions

400

450

500

Fig. 1.2.

400

450

500

Fig. 1.3.

Color Matching Functions −:r −−: g

Tristimulus value

−.: b

1

0 0

50

100

150

200 250 300 Wavelength, nm

350

The CIE RGB color matching functions

If the spectral energy distribution C () of a stimulus is given, then the chromaticity coordinates can be determined in two stages. First, the tristimulus values X , Y , Z are calculated as follows:

8 Z

X = x()C () d Y

Z

(1.7)

y()C () d

(1.8)

Z = z()C () d

(1.9)

=

Z

The new set of primaries must satisfy the following conditions: 1. The XYZ components for all visible colors should be non-negative. 2. Two of the primaries should have zero luminance. 3. As many spectral colors as possible should have at least one zero XYZ component. Secondly, normalized tristimulus values, called chromaticity coordinates, are calculated based on the primaries as follows:

x = X + XY + Z (1.10) y = X + YY + Z (1.11) (1.12) z = X + ZY + Z Clearly z = 1 (x + y ) and hence only two coordinates are necessary to describe a color match. Therefore, the chromaticity coordinates project the 3 D color solid on a plane, and they are usually plotted as a parametric x y plot with z implicitly evaluated as z = 1 (x + y ). This diagram is known as the chromaticity diagram and has a number of interesting properties that are used extensively in image processing. In particular,

1. The chromaticity coordinates (x; y ) jointly represent the chrominance components of a given color. 2. The entire color space can be represented by the coordinates (x; y; T ), in which T = constant is a given chrominance plane. 3. The chromaticity diagram represents every physically realizable color as a point within a well de ned boundary. The boundary represents the primary sources. The boundary vertices have coordinates de ned by the chromaticities of the primaries. 4. A white point is located in the center of the chromaticity diagram. More saturated colors radiate outwards from white. Complementary pure colors can easily be determined from the diagram. 5. In the chromaticity diagram, the color perception obtained through the superposition of light coming from two di erent sources, lies on a straight line between the points representing the component lights in the diagram.

9

6. Since the chromaticity diagram reveals the range of all colors which can be produced by means of the three primaries (gamut), it can be used to guide the selection of primaries subject to design constraints and technical speci cations. 7. The chromaticity diagram can be utilized to determine the hue and saturation of a given color since it represents chrominance by eliminating luminance. Based on the initial objectives set out by CIE, two of the primaries, X and Z , have zero luminance while the primary Y is the luminance indicator determined by the light-eÆciency function V () at the spectral matching curve y. Thus, in the chromaticity diagram the dominant wavelength (hue) can be de ned as the intersection between a line drawn from the reference white through the given color to the boundaries of the diagram. Once the hue has been determined, then the purity wc of the line segments of a given color can be found as the ratio r = wp that connect the reference white with the color (wc) to the line segment between the reference white and the dominant wavelength/hue (wp).

Fig. 1.4.

gram

The chromaticity dia-

1.3 The CIE-RGB Color Model

The fundamental assumption behind modern colorimetry theory, as it applies to image processing tasks, is that the initial basis for color vision lies in the di erent excitation of three classes of photo-receptor cones in the retina. These include the red, green and blue receptors, which de ne a trichromatic

10

space whose basis of primaries are pure colors in the short, medium and high portions of the visible spectrum [4], [5], [10]. As a result of the assumed linear nature of light, and due to the principle of superposition, the colors of a mixture are a function of the primaries and the fraction of each primary that is mixed. Throughout this analysis, the primaries need not be known, just their tristimulus values. This principle is called additive reproduction. It is employed in image and video devices used today where the color spectra from red, green and blue light beams are physically summed at the surface of the projection screen. Direct view color CRT's (cathode ray tube) also utilize additive reproduction. In particular, the CRT's screen consists of small dots which produce red, green and blue light. When the screen is viewed from a distance the spectra of these dots add up in the retina of the observer. In practice, it is possible to reproduce a large number of colors by additive reproduction using the three primaries: red, green and blue. The colors that result from additive reproduction are completely determined by the three primaries. The video projectors and the color CRT's in use today utilize a color space collectively known under the name RGB, which is based on the red, green and blue primaries and a white reference point. To uniquely specify a color space based on the three primary colors the chromaticity values of each primary color and a white reference point need to be speci ed. The gamut of colors which can be mixed from the set of the RGB primaries is given in the (x; y ) chromaticity diagram by a triangle whose vertices are the chromaticities of the primaries (Maxwell triangle) [5], [20]. This is shown in Figure 1.5. P 3

1

C

C’

1 P

2

1

P 1

Fig. 1.5.

triangle

The Maxwell

11

Blue(0,0,B)

Cyan(0,G,B)

White (R,G,B)

Magenta (R,0,B)

Grey-scale line

Green(0,G,0) Black(0,0,0)

Red(R,0,0)

Yellow(R,G,0)

Fig. 1.6.

model

The RGB color

In the red, green and blue system the color solid generated is a bounded subset of the space generated by each primary. Using an appropriate scale along each primary axis, the space can normalized, so that the maximum is 1. Therefore, as can be seen in Figure 1.6 the RGB color solid is a cube, called the RGB cube. The origin of the cube, de ned as (0; 0; 0) corresponds to black and the point with coordinates (1; 1; 1) corresponds to the system's brightest white. In image processing, computer graphics and multimedia systems the RGB representation is the most often used. A digital color image is represented by a two dimensional array of three variate vectors which are comprised of the pixel's red, green and blue values. However, these pixel values are relative to the three primary colors which form the color space. As it was mentioned earlier, to uniquely de ne a color space, the chromaticities of the three primary colors and the reference white must be speci ed. If these are not speci ed within the chromaticity diagram, the pixel values which are used in the digital representation of the color image are meaningless [16]. In practice, although a number of RGB space variants have been de ned and are in use today, their exact speci cations are usually not available to the end-user. Multimedia users assume that all digital images are represented in the same RGB space and thus use, compare or manipulate them directly no matter where these images are from. If a color digital image is represented in the RGB system and no information about its chromaticity characteristics is available, the user cannot accurately reproduce or manipulate the image. Although in computing and multimedia systems there are no standard primaries or white point chromaticities, a number of color space standards

12

have been de ned and used in the television industry. Among them are the Federal Communication Commission of America (FCC) 1953 primaries, the Society of Motion Picture and Television Engineers (SMPTE) `C' primaries, the European Broadcasting Union (EBU) primaries and the ITU-R BT.709 standard (formerly known as CCIR Rec. 709) [24]. Most of these standards use a white reference point known as CIE D65 but other reference points, such as the CIE illuminant E are also be used [4]. In additive color mixtures the white point is de ned as the one with equal red, green and blue components. However, there is no unique physical or perceptual de nition of white, so the characteristics of the white reference point should be de ned prior to its utilization in the color space de nition. In the CIE illuminant E, or equal-energy illuminant, white is de ned as the point whose spectral power distribution is uniform throughout the visible spectrum. A more realistic reference white, which approximates daylight has been speci ed numerically by the CIE as illuminant D65. The D65 reference white is the one most often used for color interchange and the reference point used throughout this work. The appropriate red, green and blue chromaticities are determined by the technology employed, such as the sensors in the cameras, the phosphors within the CTR's and the illuminants used. The standards are an attempt to quantify the industry's practice. For example, in the FCC-NTSC standard, the set of primaries and speci ed white reference point were representative of the phosphors used in color CRTs of a certain era. Although the sensor technology has changed over the years in response to market demands for brighter television receivers, the standards remain the same. To alleviate this problem, the European Broadcasting Union (EBU) has established a new standard (EBU Tech 3213). It is de ned in Table 1.1. Table 1.1.

EBU Tech 3213 Primaries

Colorimetry x y z

Red 0.640 0.330 0.030

Green 0.290 0.600 0.110

Blue 0.150 0.060 0.790

White D65 0.3127 0.3290 0.3582

Recently, an international agreement has nally been reached on the primaries for the High De nition Television (HDTV) speci cation. These primaries are representative of contemporary monitors in computing, computer graphics and studio video production. The standard is known as ITU-R BT.709 and its primaries along with the D65 reference white are de ned in Table 1.2. The di erent RGB systems can be converted amongst each other using a linear transformation assuming that the white references values being used are known. As an example, if it is assumed that the D65 is used in both

13 Table 1.2.

EBU Tech 3213 Primaries

Colorimetry x y z

Red 0.640 0.330 0.030

Green 0.300 0.600 0.100

Blue 0.150 0.060 0.790

White D65 0.3127 0.3290 0.3582

systems, then the conversion between the ITU-R BT.709 and SMPTE `C' primaries is de ned by the following matrix transformation: 2

R709 3 2 0:939555 0:050173 0:010272 3 2 Rc 3 4 G709 5 = 4 0:017775 0:9655795 0:016430 5 4 Gc 5 (1.13) B709 0:001622 0:004371 1:005993 Bc where R709 , G709 , B709 are the linear red, green and blue components of the ITU-R BT.709 and Rc , Gc , Bc are the linear components in the SMPTE `C'

system. The conversion should be carried out in the linear voltage domain, where the pixel values must rst be converted into linear voltages. This is achieved by applying the so-called gamma correction. 1.4 Gamma Correction

In image processing, computer graphics, digital video and photography, the symbol represents a numerical parameter which describes the nonlinearity of the intensity reproduction. The cathode-ray tube (CRT) employed in modern computing systems is nonlinear in the sense that the intensity of light reproduced at the screen of a CRT monitor is a nonlinear function of the voltage input. A CRT has a power law response to applied voltage. The light intensity produced on the display is proportional to the applied voltage raised to a power denoted by [4], [16], [17]. Thus, the produced intensity by the CRT and the voltage applied on the CRT have the following relationship:

Iint = (v0 )

(1.14)

The relationship which is called the ` ve-halves' power law is dictated by the physics of the CRT electron gun. The above function applies to a single electron gun of a gray-scale CRT or each of the three red, green and blue electron guns of a color CRT. The functions associated with the three guns on a color CRT are very similar to each other but not necessarily identical. The actual value of for a particular CRT may range from about 2.3 to 2.6 although most practitioners frequently claim values lower than 2.2 for video monitors. The process of pre-computing for the nonlinearity by computing a voltage signal from an intensity value is called gamma correction. The function required is approximately a 0:45 power function. In image processing applications, gamma correction is accomplished by analog circuits at the camera.

14

In computer graphics, gamma correction is usually accomplished by incorporating the function into a frame bu er lookup table. Although in image processing systems gamma was originally used to refer to the nonlinearity of the CRT, it is generalized to refer to the nonlinearity of an entire image processing system. The value of an image or an image processing system can be calculated by multiplying the 's of its individual components from the image capture stage to the display. The model used in (1.14) can cause wide variability in the value of gamma mainly due to the black level errors since it forces the zero voltage to map to zero intensity for any value of gamma. A slightly di erent model can be used in order to resolve the black level error. The modi ed model is given as:

Iint = (voltage + )2:5

(1.15)

By xing the exponent of the power function at 2.5 and using the single parameter to accommodate black level errors the modi ed model ts the observed nonlinearity much better than the variable gamma model in (1.14). The voltage-to-intensity function de ned in (1.15) is nearly the inverse of the luminance-to-brightness relationship of human vision. Human vision de nes luminance as a weighted mixture of the spectral energy where the weights are determined by the characteristics of the human retina. The CIE has standardized a weighting function which relates spectral power to luminance. In this standardized function, the perceived luminance by humans relates to the physical luminance (proportional to intensity) by the following equation: (

1

Y 3 L = 116( Yn )

16 if Y if 903:3( Yn ) 1 3

Y Yn Y Yn

> 0:008856 0:008856

(1.16)

where Yn is the luminance of the reference white, usually normalized either to 1.0 or 100. Thus, the lightness perceived by humans is, approximately, the cubic root of the luminance. The lightness sensation can be computed as intensity raised, approximately to the third power. Thus, the entire image processing system can be considered linear or almost linear. To compensate for the nonlinearity of the display (CRT), gamma correction with a power of ( 1 ) can be used so that the overall system is approximately 1. In a video system, the gamma correction is applied to the camera for precomputing the nonlinearity of the display. The gamma correction performs the following transfer function:

voltage0 = (voltage) 1

(1.17)

where voltage is the voltage generated by the camera sensors. The gamma corrected value is the reciprocal of the gamma resulting in a transfer function with unit power exponent.

15

To achieve subjectively pleasing images, the end-to-end power function of the overall imaging system should be around 1.1 or 1.2 instead of the mathematically correct linear system. The REC 709 speci es a power exponent of 0.45 at the camera which, in conjunction with the 2.5 exponent at the display, results in an overall exponent value of about 1.13. If the value is greater than 1, the image appears sharper but the scene contrast range, which can be reproduced, is reduced. On the other hand, reducing the value has a tendency to make the image appear soft and washed out. For color images, the linear values R; G, and B values should be converted into nonlinear voltages R0 , G0 and B 0 through the application of the gamma correction process. The color CRT will then convert R0 , G0 and B 0 into linear red, green and blue light to reproduce the original color. The ITU-R BT.709 standard recommends a gamma exponent value of 0.45 for the High De nition Television. In practical systems, such as TV cameras, certain modi cations are required to ensure proper operation near the dark regions of an image, where the slope of a pure power function is in nite at zero. The red tristimulus (linear light) component may be gamma-corrected at the camera by applying the following convention:

R0

709

=



4:5R 1:099R0:45

if R0:018 0:099 if 0:018 < R

(1.18)

0 the resulting gamma corrected with R denoting the linear light and R709 value. The computations are identical for the G and B components. The linear R; G, and B are normally in the range [0; 1] when color images are used in digital form. The software library translates these oating point values to 8-bit integers in the range of 0 to 255 for use by the graphics hardware. Thus, the gamma corrected value should be:

R0 = 255R 1

(1.19)

The constant 255 in (1.19) is added during the A/D process. However, gamma correction is usually performed in cameras, and thus, pixel values are in most cases nonlinear voltages. Thus, intensity values stored in the framebu er of the computing device are gamma corrected on-the- y by hardware look up tables on their way to the computer monitor display. Modern image processing systems utilize a wide variety of sources of color images, such as images captured by digital cameras, scanned images, digitized video frames and computer generated images. Digitized video frames usually have a gamma correction value between 0.5 and 0.45. Digital scanners assume an output gamma in the range of 1.4 to 2.2 and they perform their gamma correction accordingly. For computer generated images the gamma correction value is usually unknown. In the absence of the actual gamma value the recommended gamma correction is 0.45.

16

In summary, pixel values alone cannot specify the actual color. The gamma correction value used for capturing or generating the color image is needed. Thus, two images which have been captured with two cameras operating under di erent gamma correction values will represent colors differently even if the same primaries and the same white reference point are used. 1.5 Linear and Non-linear RGB Color Spaces

The image processing literature rarely discriminates between linear RGB and non-linear (R0 G0 B0 ) gamma corrected values. For example, in the JPEG and MPEG standards and in image ltering, non-linear RGB (R0 G0 B0 ) color values are implicit. Unacceptable results are obtained when JPEG or MPEG schemes are applied to linear RGB image data [4]. On the other hand, in computer graphics, linear RGB values are implicitly used [4]. Therefore, it is very important to understand the di erence between linear and non-linear RGB values and be aware of which values are used in an image processing application. Hereafter, the notation R0 G0 B0 will be used for non-linear RGB values so that they can be clearly distinguished from the linear RGB values.

1.5.1 Linear RGB Color Space As mentioned earlier, intensity is a measure, over some interval of the electromagnetic spectrum, of the ow of power that is radiated from an object. Intensity is often called a linear light measure. The linear R value is proportional to the intensity of the physical power that is radiated from an object around the 700 nm band of the visible spectrum. Similarly, a linear G value corresponds to the 546:1 nm band and a linear B value corresponds to the 435:8 nm band. As a result the linear RGB space is device independent and used in some color management systems to achieve color consistency across diverse devices. The linear RGB values in the range [0, 1] can be converted to the corresponding CIE XYZ values in the range [0, 1] using the following matrix transformation [4]: 2

X 3 2 0:4125 0:3576 0:1804 3 2 R 3 4 Y 5 = 4 0:2127 0:7152 0:0722 5 4 G 5 Z 0:0193 0:1192 0:9502 B

(1.20)

The transformation from CIE XYZ values in the range [0, 1] to RGB values in the range [0, 1] is de ned by: 2

R 3 2 3:2405 1:5372 0:4985 3 2 X 3 4 G 5 = 4 0:9693 1:8760 0:0416 5 4 Y 5 B 0:0556 0:2040 1:0573 Z

(1.21)

17

Alternatively, tristimulus XYZ values can be obtained from the linear RGB values through the following matrix [5]: 2

X 3 2 0:490 0:310 0:200 3 2 R 3 4 Y 5 = 4 0:117 0:812 0:011 5 4 G 5 Z 0:000 0:010 0:990 B

(1.22)

The linear RGB values are a physical representation of the chromatic light radiated from an object. However, the perceptual response of the human visual system to radiate red, green, and blue intensities is non-linear and more complex. The linear RGB space is, perceptually, highly non-uniform and not suitable for numerical analysis of the perceptual attributes. Thus, the linear RGB values are very rarely used to represent an image. On the contrary, non-linear R0 G0 B0 values are traditionally used in image processing applications such as ltering.

1.5.2 Non-linear RGB Color Space When an image acquisition system, e.g. a video camera, is used to capture the image of an object, the camera is exposed to the linear light radiated from the object. The linear RGB intensities incident on the camera are transformed to non-linear RGB signals using gamma correction. The transformation to non-linear R0 G0 B0 values in the range [0, 1] from linear RGB values in the range [0, 1] is de ned by:

R0

(

=

G0 = B0 =

( (

4:5R; 1 1:099R C

4:5G; 1 1:099G C

4:5B; 1 1:099B C

0:099; 0:099; 0:099;

if R  0:018 otherwise

if G  0:018 otherwise

if B  0:018 otherwise

(1.23)

where C is known as the gamma factor of the camera or the acquisition device. The value of C that is commonly used in video cameras is 0:145 (' 2:22) [4]. The above transformation is graphically depicted in Figure 1.7. The linear segment near low intensities minimizes the e ect of sensor noise in practical cameras and scanners. Thus, the digital values of the image pixels acquired from the object and stored within a camera or a scanner are the R0 G0 B0 values usually converted to the range of 0 to 255. Three bytes are then required to represent the three components, R0 , G0 , and B 0 of a color image pixel with one byte for each component. It is these non-linear R0 G0 B0 values that are stored as image data les in computers and are used in image processing applications. The RGB symbol used in image processing literature usually refers to the R0 G0 B0

18 1

0.8

NMSE x 100

0.6

0.4

0.2

0

Linear to Nonlinear Light Transformation Fig. 1.7.

−0.2 0

0.1

0.2

0.3

0.4

0.5 0.6 BVOSF #

0.7

0.8

0.9

1

values and, therefore, care must be taken in color space conversions and other relevant calculations. Suppose the acquired image of an object needs to be displayed in a display device such as a computer monitor. Ideally, a user would like to see (perceive) the exact reproduction of the object. As pointed out, the image data is in R0 G0 B0 values. Signals (usually voltage) proportional to the R0 G0 B0 values will be applied to the red, green, and blue guns of the CRT (Cathode Ray Tube) respectively. The intensity of the red, green, and blue lights generated by the CRT is a non-linear function of the applied signal. The non-linearity of the CRT is a function of the electrostatics of the cathode and the grid of the electron gun. In order to achieve correct reproduction of intensities, an ideal monitor should invert the transformation at the acquisition device (camera) so that the intensities generated are identical to the linear RGB intensities that were radiated from the object and incident in the acquisition device. Only then will the perception of the displayed image be identical to the perceived object. A conventional CRT has a power-law response, as depicted in Figure 1.8. This power-law response, which inverts the non-linear (R0 G0 B0 ) values in the range [0, 1] back to linear RGB values in the range [0, 1], is de ned by the following power function [4]:

R G

= =

( R0 4:5 

;

R0

( G0 4:5  0

;

+ 0:099 1:099

G + 0:099 1:099

 D

if R0  0:018 otherwise

 D

if G0  0:018 otherwise

(1.24)

19

B

=

( B0 4:5  0

;

B + 0:099 1:099

 D

if B 0  0:018 otherwise

The value of the power function, D , is known as the gamma factor of the display device or CRT. Normal display devices have D in the range of 2:2 to 2:45. For exact reproduction of the intensities, gamma factor of the display device must be equal to the gamma factor of the acquisition device ( C = D ). Therefore, a CRT with a gamma factor of 2:2 should correctly reproduce the intensities. 1 0.9

Linear Light Intensities (R, G, B)

0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0

Non-linear to linear Light Transformation Fig. 1.8.

0.1

0.2

0.3 0.4 0.5 0.6 0.7 Non−linear Light Intensties (R’, G’, B’)

0.8

0.9

1

The transformations that take place throughout the process of image acquisition to image display and perception are illustrated in Figure 1.9. R Object

G B

Fig. 1.9.

Digital Video Camera

R’ G’ B’

Storage

G’ B’

Perceived Intensities

R

R’ CRT

G

HVS

B

R’ G’ B’

Transformation of Intensities from Image Capture to Image Display

It is obvious from the above discussion that the R0 G0 B0 space is a device dependent space. Suppose a color image, represented in the R0 G0 B0 space, is displayed on two computer monitors having di erent gamma factors. The red, green, and blue intensities produced by the monitors will not be identical and the displayed images might have di erent appearances. Device dependent spaces cannot be used if color consistency across various devices, such as display devices, printers, etc., is of primary concern. However, similar devices

20

(e.g. two computer monitors) usually have similar gamma factors and in such cases device dependency might not be an important issue. As mentioned before, the human visual system has a non-linear perceptual response to intensity, which is roughly logarithmic and is, approximately, the inverse of a conventional CRT's non-linearity [4]. In other words, the perceived red, green, and blue intensities are approximately related to the R0 G0 B0 values. Due to this fact, computations involving R0 G0 B0 values have an approximate relation to the human color perception and the R0 G0 B0 space is less perceptually non-uniform relative to the CIE XYZ and linear RGB spaces [4]. Hence, distance measures de ned between the R0 G0 B0 values of two color vectors provide a computationally simple estimation of the error between them. This is very useful for real-time applications and systems in which computational resources are at premium. However, the R0 G0 B0 space is not adequately uniform, and it cannot be used for accurate perceptual computations. In such instances, perceptually uniform color spaces (e.g. L u v and L a b ) that are derived based on the attributes of human color perception are more desirable than the R0 G0 B0 space [4]. 1.6 Color Spaces Linearly Related to the RGB

In transmitting color images through a computer-centric network, all three primaries should be transmitted. Thus, storage or transmission of a color image using RGB components requires a channel capacity three times that of gray scale images. To reduce these requirements and to boost bandwidth utilization, the properties of the human visual system must be taken into consideration. There is strong evidence that the human visual system forms an achromatic channel and two chromatic color-di erence channels in the retina. Consequently, a color image can be represented as a wide band component corresponding to brightness, and two narrow band color components with considerably less data rate than that allocated to brightness. Since the large percentage (around 60%) of brightness is attributed to the green primary, then it is advantageous to base the color components on the other two primaries. The simplest way to form the two color components is to remove them by subtraction, (e.g. the brightness from the blue and red primaries). In this way the unit RGB color cube is transformed into the luminance Y and two color di erence components B Y and R Y [33], [34]. Once these color di erence components have been formed, they can be sub-sampled to reduce the bandwidth or data capacity without any visible degradation in performance. The color di erence components are calculated from nonlinear gamma corrected values R0 ,G0 ,B 0 rather than the tristimulus (linear voltage) R; G; B primary components. According to the CIE standards the color imaging system should operate similarly to a gray scale system, with a CIE luminance component Y formed

21

as a weighted sum of RGB tristimulus values. The coeÆcients in the weighted sum correspond to the sensitivity of the human visual system to each of the RGB primaries. The coeÆcients are also a function of the chromaticity of the white reference point used. International agreement on the REC. 709 standard provides a value for the luminance component based on the REC. 0 luminance equation is: 709 primaries [24]. Thus, the Y709

0 = 0:2125R0 + 0:7154G0 + 0:0721B 0 Y709 (1.25) 709 709 709 0 , B709 0 and G0709 are the gamma-corrected (nonlinear) values of where R709 0 0 and the three primaries. The two color di erence components B709 Y709 0 0 R709 Y709 can be formed on the basis of the above equation. Various scale factors are applied to the basic color di erence components for di erent applications. For example, the Y 0 PR PB is used for component analog video, such as BetaCam, and Y 0 CB CR for component digital video,

such as studio video, JPEG and MPEG. Kodak's YCC (PhotoCD model) uses scale factors optimized for the gamut of lm colors [31]. All these systems 0 ; B709 0 Y709 0 ; R709 0 0 ) which are scaled utilize di erent versions of the (Y709 Y709 to place the extrema of the component signals at more convenient values. In particular, the Y 0 PR PB system used in component analog equipment is de ned by the following set: 2

32 03 0 3 2 Y601 0:299 0:587 0:114 R 4 PB 5 = 4 0:168736 0:331264 0:5 5 4 G0 5 PR 0:5 0:418686 0:081312 B0

and

(1.26)

2

32 0 3 R0 3 2 1: 0: 1:402 Y601 4 G0 5 = 4 1: 0:344136 0:714136 5 4 PB 5 B0 1: 1:772 0: PR

(1.27)

The rst row comprises the luminance coeÆcients which sum to unity. For each of the other two columns the coeÆcients sum to zero, a necessity for color di erence formulas. The 0.5 weights re ect the maximum excursion of PB and PR for the blue and the red primaries. The Y0 CB CR is the Rec ITU-R BT. 601-4 international standard for studio quality component digital video. The luminance signal is coded in 8 bits. The Y 0 has an excursion of 219 with an o set of 16, with the black point coded at 16 and the white at code 235. Color di erences are also coded in 8-bit forms with excursions of 112 and o set of 128 for a range of 16 through 240 inclusive. To compute Y0 CB CR from nonlinear R0 G0 B0 in the range of [0; 1] the following set should be used: 2

0 3 2 16 3 2 Y601 4 CB 5 = 4 128 5 + 4 CR 128

32

3

65:481 128:553 24:966 R0 37:797 74:203 112:0 5 4 G0 5 112:0 93:786 18:214 B0

(1.28)

22

with the inverse transform 2 03 2 3 R0 0:00456821 0:0 0:00625893 4 G 5 = 4 0:00456621 0:00153632 0:00318811 5  B0 0:00456621 0:00791071 0:0 02

0 3 Y601 @4 PB 5 PR

2

31

16 4 128 5A 128

(1.29)

When 8-bit R0 G0 B0 are used, black is coded at 0 and white is at 255. To encode Y0 CB CR from R0 G0 B0 in the range of [0; 255] using 8-bit binary arithmetic the transformation matrix should be scaled by 256 255 . The resulting transformation pair is as follows: 2 0 3 2 3 2 32 0 3 Y601 16 65:481 128:553 24:966 R255 1 4 PB 5 = 4 128 5 + 4 37:797 74:203 112:0 5 4 G0255 5 (1.30) 256 0 PR 128 112:0 93:786 18:214 B255

0 is the gamma-corrected value, using a gamma-correction lookup where R255 table for 1 . This yields the RGB intensity values with integer components between 0 and 255 which are gamma-corrected by the hardware. To obtain R0 G0 B0 values in the range [0; 255] from Y0 CB CR using 8-bit arithmetic the following transformation should be used: 3 0:0 0:00625893 R0 3 1 2 0:00456821 4 0:00456621 0:00153632 0:00318811 5 4 G0 5 = 256 0 0:00456621 0:00791071 0:0 B 02 0 3 2 31 Y601 16 @4 PB 5 4 128 5A PR 128 2

(1.31)

1 may be larger than unity and, Some of the coeÆcients when scaled by 256 thus some clipping may be required so that they fall within the acceptable RGB range. The Kodak YCC color space is another example of a predistorted color space, which has been designed for the storage of still color images on the Photo-CD. It is derived from the predistorted (gamma-corrected) R0 G0 B0 values using the ITU-R BT.709 recommended white reference point, primaries, and gamma correction values. The YCC space is similar to the Y0 CB CR discussed, although scaling of B 0 Y 0 and R0 Y 0 is asymmetrical in order to accommodate a wide color gamut, similar to that of a photographic lm. In particular the following relationship holds for Photo-CD compressed formats: 0 (1.32) Y 0 = 1255 :402 Y601 C1 = 156 + 111:40(B0 Y 0) (1.33)

23

C2 = 137 + 135:64(R0 Y 0)

(1.34)

The two chrominance components are compressed by factors of 2 both horizontally and vertically. To reproduce predistorted R0 G0 B0 values in the range of [0; 1] from integer PhotoYCC components the following transform is applied: 2

3 0:0 0:0051681 R00 3 1 2 0:00549804 4G 5 = 4 0:00549804 0:0015446 0:0026325 5 256 0 0:00549804 0:0079533 0:0 B 02 0 3 2 31 Y 0 @4 C1 5 4 156 5A (1.35) C2 137 The B 0 Y 0 and R0 Y 0 components can be converted into polar coordinates

to represent the perceptual attributes of hue and saturation. The values can be computed using the following formulas [34]:

0 0 H = tan 1( BR0 YY 0 ) S = ((B0 Y 0 )2 + (R0 Y 0)2) 21

(1.36) (1.37)

where the saturation S is the length of the vector from the origin of the chromatic plane to the speci c color and the hue H is the angle between the R0 Y 0 axis and the saturation component [33]. 1.7 The YIQ Color Space

The YIQ color speci cation system, used in commercial color TV broadcasting and video systems, is based upon the color television standard that was adopted in the 1950s by the National Television Standard committee (NTSC) [10], [1], [27], [28]. Basically, YIQ is a recoding of non-linear R0 G0 B0 for transmission eÆciency and for maintaining compatibility with monochrome TV standards [1], [4]. In fact, the Y component of the YIQ system provides all the video information required by a monochrome television system. The YIQ model was designed to take advantage of the human visual system's greater sensitivity to change in luminance than to changes in hue or saturation [1]. Due to these characteristics of the human visual system, it is useful in a video system to specify a color with a component representative of luminance Y and two other components: the in-phase I , an orange-cyan axis, and the quadrature Q component, the magenta-green axis. The two chrominance components are used to jointly represent hue and saturation .

24

With this model, it is possible to convey the component representative of luminance Y in such a way that noise (or quantization) introduced in transmission, processing and storage is minimal and has a perceptually similar e ect across the entire tone scale from black to white [4]. This is done by allowing more bandwidth (bits) to code the luminance (Y ) and less bandwidth (bits) to code the chrominance (I and Q) for eÆcient transmission and storage purposes without introducing large perceptual errors due to quantization [1]. Another implication is that the luminance (Y ) component of an image can be processed without a ecting its chrominance (color content). For instance, histogram equalization to a color image represented in YIQ format can be done simply by applying histogram equalization to its Y component [1]. The relative colors in the image are not a ected by this process. The ideal way to accomplish these goals would be to form a luminance component (Y ) by applying a matrix transform to the linear RGB components and then subjecting the luminance (Y ) to a non-linear transfer function to achieve a component similar to lightness L . However, there are practical reasons in a video system why these operations are performed in the opposite order [4]. First, gamma correction is applied to each of the linear RGB. Then, a weighted sum of the nonlinear components is computed to form a component representative of luminance Y . The resulting component (luma) is related to luminance but is not the same as the CIE luminance Y although the same symbol is used for both of them. The nonlinear RGB to YIQ conversion is de ned by the following matrix transformation [4], [1]: 2

Y 3 2 0:299 0:587 0:114 3 2 R00 3 4 I 5 = 4 0:596 0:275 0:321 5 4 G 5 Q 0:212 0:523 0:311 B0

(1.38)

As can be seen from the above transformation, the blue component has a small contribution to the brightness sensation (luma Y ) despite the fact that human vision has extraordinarily good color discrimination capability in the blue color [4]. The inverse matrix transformation is performed to convert YIQ to nonlinear R0 G0 B0 . Introducing a cylindrical coordinate transformation, numerical values for hue and saturation can be calculated as follows:

HY IQ = tan 1( QI ) 1 SY IQ = (I 2 + Q2) 2

(1.39) (1.40)

As described it, the YIQ model is developed from a perceptual point of view and provides several advantages in image coding and communications applications by decoupling the luma (Y ) and chrominance components (I and Q). Nevertheless, YIQ is a perceptually non-uniform color space and thus not appropriate for perceptual color di erence quanti cation. For example,

25

the Euclidean distance is not capable of accurately measuring the perceptual color distance in the perceptually non-uniform YIQ color space. Therefore, YIQ is not the best color space for quantitative computations involving human color perception. 1.8 The HSI Family of Color Models

In image processing systems, it is often convenient to specify colors in a way that is compatible with the hardware used. The di erent variants of the RGB monitor model address that need. Although these systems are computationally practical, they are not useful for user speci cation and recognition of colors. The user cannot easily specify a desired color in the RGB model. On the other hand, perceptual features, such as perceived luminance (intensity), saturation and hue correlate well with the human perception of color . Therefore, a color model in which these color attributes form the basis of the space is preferable from the users point of view. Models based on lightness, hue and saturation are considered to be better suited for human interaction. The analysis of the user-oriented color spaces starts by introducing the family of intensity, hue and saturation (HSI) models [28], [29]. This family of models is used primarily in computer graphics to specify colors using the artistic notion of tints, shades and tones. However, all the HSI models are derived from the RGB color space by coordinate transformations. In a computer centered image processing system, it is necessary to transform the color coordinates to RGB for display and vice versa for color manipulation within the selected space. The HSI family of color models use approximately cylindrical coordinates. The saturation (S ) is proportional to radial distance, and the hue (H ) is a function of the angle in the polar coordinate system. The intensity (I ) or lightness (L) is the distance along the axis perpendicular to the polar coordinate plane. The dominant factor in selecting a particular HSI model is the de nition of the lightness, which determines the constant-lightness surfaces, and thus, the shape of the color solid that represents the model. In the cylindrical models, the set of color pixels in the RGB cube which are assigned a common lightness value (L) form a constant-lightness surface. Any line parallel to the main diagonal of the color RGB cube meets the constantlightness surface at most in one point. The HSI color space was developed to specify, numerically, the values of hue, saturation, and intensity of a color [4]. The HSI color model is depicted in Figure 1.10. The hue (H ) is measured by the angle around the vertical axis and has a range of values between 0 and 360 degrees beginning with red at 0Æ . It gives a measure of the spectral composition of a color. The saturation (S ) is a ratio that ranges from 0 (i.e. on the I axis), extending radially outwards to a maximum value of 1 on the surface of the cone. This component refers to the proportion of pure light of the dominant wavelength and indicates how

26

far a color is from a gray of equal brightness. The intensity (I ) also ranges between 0 and 1 and is a measure of the relative brightness. At the top and bottom of the cone, where I = 0 and 1 respectively, H and S are unde ned and meaningless. At any point along the I axis the Saturation component is zero and the hue is unde ned. This singularity occurs whenever R = G = B . I=1

White

gray-scale

purely saturated Red

Yellow

P

H Magenta

S

Intensity

Green

Cyan

Blue

Black

I=0

Fig.

1.10.

Color Space

The HSI

The HSI color model owes its usefulness to two principal facts [1], [28]. First, like in the YIQ model, the intensity component I is decoupled from the chrominance information represented as hue H and saturation S . Second, the hue (H ) and saturation (S ) components are intimately related to the way in which humans perceive chrominance [1]. Hence, these features make the HSI an ideal color model for image processing applications where the chrominance is of importance rather than the overall color perception (which is determined by both luminance and chrominance). One example of the usefulness of the

27

HSI model is in the design of imaging systems that automatically determine the ripeness of fruits and vegetables [1]. Another application is color image histogram equalization performed in the HSI space to avoid undesirable shifts in image hue [10]. The simplest way to choose constant-lightness surfaces is to de ne them as planes. A simpli ed 0de nition of the perceived lightness in terms of the 0 0 R,G,B values is L = R +G3 +B , where the normalization is used to control the range of lightness values. The di erent constant-lightness surfaces are perpendicular to the main diagonal of the RGB cube and parallel to each other. The shape of a constant lightness surface is a triangle for 0L M3 and 2M 3 LM with L2[0; M ] and where M is a given lightness threshold. The theory underlying the derivation of conversion formulas between the RGB space and HSI space is described in detail in [1], [28]. The image processing literature on HSI does not clearly indicate whether the linear or the non-linear RGB is used in these conversions [4]. Thus the non-linear (R0 G0 B0 ), which is implicit in traditional image processing, shall be used. But this ambiguity must be noted. The conversion from R0 G0 B0 (range [0, 1]) to HSI (range [0, 1]) is highly nonlinear and considerably complicated: "

#

R0 G0) + (R0 B0 )] H = cos (1.41) [(R0 G0 )2 + (R0 B0 )(G0 B0)] 21 (1.42) S = 1 (R0 + G30 + B0 ) [min(R0; G0 ; B0 )] (1.43) I = 13 (R0 + G0 + B0) where H = 360Æ H , if (B 0 =I ) > (G0 =I ). Hue is normalized to the range [0, 1] by letting H = H=360Æ. Hue (H ) is not de ned when the saturation (S ) is zero. Similarly, saturation (S ) is unde ned if intensity (I ) is zero. To transform the HSI values (range [0, 1]) back to the R0 G0 B0 values (range [0, 1]), then the H values in [0, 1] range must rst be converted back to the un-normalized [0o , 360o] range by letting H = 360Æ (H ). For the R0 G0 (red and green) sector (0Æ < H  120Æ), the conversion is: B0 = I (1 S ) (1.44)   H R0 = I 1 + cos S(60cos (1.45) Æ H) G0 = 3I (R0 + B0) (1.46) The conversion for the G0 B 0 (green and blue) sector (120Æ < H  240Æ ) 1

is given by:

H

=

H

120Æ

1 2 [(

(1.47)

28

R0

=

I (1

G0

=

I

B0

= 3I

(R0 +

H G0

=

H = I (1

240Æ

B0

=

R0

= 3I

S)



1 +

S

cos H cos (60Æ H)

(1.49)

G0 )

(1.50)

Finally, for the B 0 R0 (blue and red) sector (240Æ corresponding equations are:

I

1 +

S cos H cos (60Æ H)

(G0 +

< H 

360Æ ), the (1.51)

S)



(1.48)



B0)



(1.52) (1.53) (1.54)

Fast versions of the transformation, containing fewer multiplications and avoiding square roots, are often used in hue calculations. Also, formulas without trigonometric functions can be used. For example, hue can be evaluated using the following formula [44]: 1. If B 0 = min(R0 ;

G0 ; B0 ) then 0 0 H = 3(R0G+ G0B 2B0 2. If R0 = min(R0 ; G0 ; B 0 ) then 0 0 H = R0 +B G0 R 2B0 + 31 3. If G0 = min(R0 ; G0 ; B 0 ) then 0 0 H = R0 +B G0 R 2B0 + 32

(1.55)

(1.56)

(1.57)

Although the HSI model is useful in some image processing applications, the formulation of it is awed with respect to the properties of color vision. The usual formulation makes no clear reference to the linearity or nonlinearity of the underlying RGB and to the lightness perception of human vision [4]. It computes the brightness as (R0 + G0 + B 0 ) =3 and assigns the name intensity I . Recall that the brightness perception is related to luminance Y . Thus, this computation con icts with the properties of color vision [4]. In addition to this, there is a discontinuity in the hue at 360o and thus, the formulation introduces visible discontinuities in the color space. Another major disadvantage of the HSI space is that it is not perceptually uniform.

29

Consequently, the HSI model is not very useful for perceptual image computation and for conveyance of accurate color information. As such, distance measures, such as the Euclidean distance, cannot estimate adequately the perceptual color distance in this space. The model discussed above is not the only member of the family. In particular, the double hexcone HLS model can be de ned by simply modifying the constant-lightness surface. It is depicted in Figure 1.11. In the HLS model the lightness is de ned as:

L = max (R ; G ; B ) +2 min (R ; G ; B ) 0

0

0

0

0

0

(1.58)

If the maximum and the minimum value coincide then S = 0 and the hue is unde ned. Otherwise based on the lightness value, saturation is de ned as follows: Max Min) 1. If L0:5 then S = ((Max +Min) Min) 2. If L > 0:5 then S = (2(Max Max Min)

where Max = max (R0 ; G0 ; B 0 ) and Min = min (R0 ; G0 ; B 0 ) respectively. Similarly, hue is calculated according to: 1. If R0 = Max then

G0 B0 H = Max Min 0 2. If G = Max then B 0 R0 H = Max Min 0 3. If B = Max then R0 G0 H = 4 + Max Min

(1.59)

(1.60)

(1.61)

The backward transform starts by rescaling the hue angles into the range [0; 6]. Then, the following cases are considered:

1. If S = 0, hue is unde ned and (R0 ; G0 ; B 0 ) = (L; L; L) 2. Otherwise, i = Floor(H ) (the Floor(X ) function returns the largest integer which is not greater than X ), in which i is the sector number of the hue and f = H i is the hue value in each sector. The following cases are considered:

 if LLcritical = 255 2 then

Max = L(1 + S ) Mid1 = L(2fS + 1 S )

(1.62) (1.63)

30

Mid2 = L(2(1 f )S + 1 S ) Min = L(1 S )  if L > Lcritical = 255 2 then Max = L(1 S ) + 255S Mid1 = 2((1 f )S (0:5 f )Max) Mid2 = 2(fL (f 0:5)Max) Min = L(1 + S ) 255S

(1.64) (1.65) (1.66) (1.67) (1.68) (1.69)

Based on these intermediate values the following assignments should be made: 1. 2. 3. 4. 5. 6.

if i = 0 then (R0 ; G0 ; B 0 ) = (Max; Mid1; Min) if i = 1 then (R0 ; G0 ; B 0 ) = (Mid2; Max; Min) if i = 2 then (R0 ; G0 ; B 0 ) = (Min; Max; Mid1) if i = 3 then (R0 ; G0 ; B 0 ) = (Min; Mid2; Max) if i = 4 then (R0 ; G0 ; B 0 ) = (Mid1; Min; Max) if i = 5 then (R0 ; G0 ; B 0 ) = (Max; Min; Mid2)

The HSV (hue, saturation, value) color model also belongs to this group of hue-oriented color coordinate systems which correspond more closely to the human perception of color. This user-oriented color space is based on the intuitive appeal of the artist's tint, shade, and tone. The HSV coordinate system, proposed originally in Smith [36], is cylindrical and is conveniently represented by the hexcone model shown in Figure 1.12 [23], [27]. The set of equations below can be used to transform a point in the RGB coordinate system to the appropriate value in the HSV space. 1 H1 = cos 1f p(R2 [(RG)2 +G)(R+ (RB)(BG)] B) g H = H1 ; if B  G H = 360Æ H1 ; if B > G

G; B) min(R; G; B) S = max(R; max( R; G; B) V = max(R; G; B)

(1.70) (1.71) (1.72) (1.73)

(1.74) 255 Here the RGB values are between 0 and 255. A fast algorithm used here to convert the set of RGB values to the HSV color space is provided in [23]. The important advantages of the HSI family of color spaces over other color spaces are:

31 White L=0

Green

Yellow S

P H

Cyan

Red Lightness (L) Magenta

Blue

Fig.

1.11.

The HLS

Fig.

1.12.

The HSV

Color Space

Black L=1

V Green

Yellow S

P H

Cyan

Blue

Red

White V=0

Magenta Value (V)

Black V=1

Color Space

32

  

Good compatibility with human intuition. Separability of chromatic values from achromatic values. The possibility of using one color feature, hue, only for segmentation purposes. Many image segmentation approaches take advantage of this. Segmentation is usually performed in one color feature (hue) instead of three, allowing the use of much faster algorithms.

as:

  

However, hue-oriented color spaces have some signi cant drawbacks, such

singularities in the transform, e.g. unde ned hue for achromatic points sensitivity to small deviations of RGB values near singular points numerical instability when operating on hue due to the angular nature of the feature.

1.9 Perceptually Uniform Color Spaces

Visual sensitivity to small di erences among colors is of paramount importance in color perception and speci cation experiments. A color system that is to be used for color speci cation should be able to represent any color with high precision. All systems currently available for such tasks are based on the CIE XYZ color model. In image processing, it is of particular interest in a perceptually uniform color space where a small perturbation in a component value is approximately equally perceptible across the range of that value. The color speci cation systems discussed until now, such as the XYZ or RGB tristimulus values and the various RGB hardware oriented systems are far from uniform. Recalling the discussion of YIQ space earlier in this chapter, the ideal way to compute the perceptual components representative of luminance and chrominance is to appropriately form the matrix of linear RGB components and then subject them to nonlinear transfer functions based on the color sensing properties of the human visual system. A similar procedure is used by CIE to formulate the L u v and L a b spaces. The linear RGB components are rst transformed to CIE XYZ components using the appropriate matrix. Finding a transformation of XYZ which transforms this color space into a reasonably perceptually uniform color space consumed a decade or more at the CIE and in the end, no single system could be agreed upon [4], [5]. Finally, in 1976, CIE standardized two spaces, L u v and L a b , as perceptually uniform. They are slightly di erent because of the di erent approaches to their formulation [4], [5], [25], [30]. Nevertheless, both spaces are equally good in perceptual uniformity and provide very good estimates of color di erence (distance) between two color vectors. Both systems are based on the perceived lightness L and a set of opponent color axes, approximately red-green versus yellow-blue. According to

33

the CIE 1976 standard, the perceived lightness of a standard observer is assumed to follow the physical luminance (a quantity proportional to intensity) according to a cubic root law. Therefore, the lightness L is de ned by the CIE as: (

1

Y 3 L = 116( Yn )

16 if Y 903:3( Yn ) if 1 3

Y Yn Y Yn

> 0:008856 0:008856

(1.75)

where Yn is the physical luminance of the white reference point . The range of values for L is from 0 to 100 representing a black and a reference white respectively. A di erence of unity between two L values, the so-called L is the threshold of discrimination. This standard function relates perceived lightness to linear light luminance. Luminance can be computed as a weighted sum of red, green and blue components. If three sources appear red, green and blue and have the same power in the visible spectrum, the green will appear the brightest of the three because the luminous eÆciency function peaks in the green region of the spectrum. Thus, the coeÆcients that correspond to contemporary CRT displays (ITU-R BT. 709 recommendation) [24] re ect that fact, when using the following equation for the calculation of the luminance:

Y709 = 0:2125R + 0:7154G + 0:0721B (1.76) The u and v  components in L u v space and the the a and b compo-

nents in L a b space are representative of chrominance. In addition, both are device independent color spaces. Both these color spaces are, however, computationally intensive to transform to and from the linear as well as nonlinear RGB spaces. This is a disadvantage if real-time processing is required or if computational resources are at a premium.

1.9.1 The CIE L u v Color Space The rst uniform color space standardized by CIE is the L u v illustrated in Figure 1.13. It is derived based on the CIE XYZ space and white reference point [4], [5]. The white reference point [Xn ; Yn ; Zn ] is the linear RGB = [1; 1; 1] values converted to the XYZ values using the following transformation: 2

Xn 3 2 0:4125 0:3576 0:1804 3 2 1 3 4 Yn 5 = 4 0:2127 0:7152 0:0722 5 4 1 5 Zn 0:0193 0:1192 0:9502 1

(1.77)

Alternatively, white reference points can be de ned based on the Federal Communications Commission (FCC) or the European Broadcasting Union (EBU) RGB values using the following transformations respectively [35]:

34 2

Xn 3 2 0:607 0:174 0:200 3 2 1 3 4 Yn 5 = 4 0:299 0:587 0:114 5 4 1 5 Zn 0:000 0:066 1:116 1 2 3 2 32 3 Xn 0:430 0:342 0:178 1 4 Yn 5 = 4 0:222 0:702 0:071 5 4 1 5 Zn 0:020 0:130 0:939 1

(1.78) (1.79)

Fig. 1.13.

Color Space

The L u v

The lightness component L is de ned by the CIE as a modi ed cube root of luminance Y [4], [31], [37], [32]:

L

=

8  1 < 116 Y 3 Yn  : 903 3 Y

:

Yn

16

if

Y Yn

> 0:008856

E otherwise

(1.80)

The CIE de nition of L applies a linear segment near black for (Y=Yn )  0:008856. This linear segment is unimportant for practical purposes [4]. L has a range [0, 100], and a L of unity is roughly the threshold of visibility [4]. Computation of u and v  involves intermediate u0 , v 0 , u0n , and vn0 quantities de ned as: u0 = X + 154XY + 3Z v0 = X + 159YY + 3Z (1.81) (1.82) u0n = X + 154XYn + 3Z vn0 = X + 159YYn + 3Z n n n n n n with the CIE XYZ values computed through (1.20) and (1.21). Finally, u and v  are computed as: u = 13L(u0 u0n) (1.83)

v

= 13L (v 0

vn0 )

(1.84)

Conversion from L u v to XYZ is accomplished by ignoring the linear segment of L . In particular, the linear segment can be ignored if the luminance variable Y is represented with eight bits of precision or less.

35

Then, the luminance Y is given by:

Y

=



L

To compute X



+ 16 3 Yn 116 and Z , rst compute u0 and v 0 as:

(1.85)

u + u0 v0 = v + v0 n n 13L 13L Finally, X and Z are given by:   1 u0 (9:0 15:0 v 0 ) Y 0 X = 4 + 15:0 u Y v0   1 (9:0 15:0 v 0 ) Y X Z = 3 v0 u0

=

(1.86) (1.87) (1.88)

Consider two color vectors xL u v and yL u v in the L u v space represented as:

xL u v = [xL ;

xu  ; xv  ]T

and yL u v = [yL ;

yu ; yv ]T (1.89)

The perceptual color distance in the L u v space, called the total color  in [5], is de ned as the Euclidean distance (L2 norm) between di erence Euv the two color vectors xL u v and yL u v :

Euv = jjxLuv

yL u v jjL2

h

Euv = (xL yL )2 + (xu yu )2 + (xv yv )2

i 21

(1.90)

It should be mentioned that in a perceptually uniform space, the Euclidean distance is an accurate measure of the perceptual color di erence [5]. As such,  is widely used for the evaluation of color the color di erence formula Euv reproduction quality in an image processing system, such as color coding systems.

1.9.2 The CIE L a b Color Space The L a b color space is the second uniform color space standardized by

CIE. It is also derived based on the CIE XYZ space and white reference point [5], [37]. The lightness L component is the same as in the L u v space. The L ,  a and b components are given by:

Y  13 Yn "  31 X 500 Xn

L = 116 a =



16 

(1.91)

Y  13 Yn

#

(1.92)

36 "

#



Z  13 (1.93) Zn with the constraint that XXn ; YYn ; ZZn > 0:01. This constraint will be satis ed b = 200

Y  13 Yn

for most practical purposes [4]. Hence, the modi ed formulae described in [5] for cases that do not not satisfy this constraint can be ignored in practice [4], [10]. The back conversion to the XYZ space from the L a b space is done by rst computing the luminance Y , as described in the back conversion of L u v , followed by the computation of X and Z :

Y X Z

= = =



L

 + 16 3 116

a

+

500

b

+

200



Yn

Y  13 Yn



(1.94) !3

Y  13 Yn

Xn

!3

(1.95)

Zn

(1.96)

The perceptual color distance in the L a b is similar to the one in the L u v . The two color vectors xL a b and yL a b in the L a b space can be represented as:

xL ab = [xL ;

xa ; xb ]T

and

yL ab = [yL ;

ya ; yb ]T (1.97)

The perceptual color distance (or total color di erence) in the L a b space, Eab , between two color vectors xLuv and yLuv is given by the Euclidean distance (L2 norm):

Eab = jjxLab

yL a b jjL2

h

i1

= (xL

yL )2 + (xa ya )2 + (xb yb )2 2 (1.98)  is applicable to the observing conditions The color di erence formula Eab  . However, this simple normally found in practice, as in the case of Euv

di erence formula values color di erences too strongly when compared to experimental results. To correct the problem a new di erence formula was recommended in 1994 by CIE [25], [31]. The new formula is as follows: 1

2 2 2 2 Eab 94 = [ KLSyLL ) + (xaKcSyca ) + (xbK H SyHb ) ] (1.99) where the factors KL , Kc , KH are factors to match the perception of the background conditions, and SL , Sc , SH are linear functions of the di erences  have been in chroma. Standard reference values for the calculation for Eab 94

(xL

37

speci ed by the CIE. Namely, the values most often in use are KL = Kc = KH = 1, SL = 1, Sc = 1 + 0:045((xa ya ) and SH = 1 + 0:015((xb yb ) respectively. The parametric values may be modi ed to correspond to typical experimental conditions. As an example, for the textile industry, the KL factor should be 2, and the Kc and KH factors should be 1. For all other applications a value of 1 is recommended for all parametric factors [38].

1.9.3 Cylindrical L u v and L a b Color Space Any color expressed in the rectangular coordinate system of axes L u v or L a b can also be expressed in terms of cylindrical coordinates with the perceived lightness L and the psychometric correlates of chroma and hue  and that in the [37]. The chroma in the L u v space is denoted as Cuv     L a b space Cab . They are de ned as [5]:

Cuv Cab

=

(u )2 + (v  )2

=

(a )2 + (b )2

 21

 12

(1.100) (1.101)

The hue angles are useful quantities in specifying hue numerically [5], [37]. Hue angle huv in the L u v space and hab in the L a b space are de ned as [5]: 

v  u   hab = arctan ab The saturation suv in the L u v space is given by:  suv = CLuv huv

= arctan

(1.102) (1.103)

(1.104)

1.9.4 Applications of L u v and L a b spaces The L u v and L a b spaces are very useful in applications where precise

quanti cation of perceptual distance between two colors is necessary [5]. For example in the realization of perceptual based vector order statistics lters. If a degraded color image has to be ltered so that it closely resembles, in perception, the un-degraded original image, then a good criterion to optimize is the perceptual error between the output image and the un-degraded original image. Also, they are very useful for evaluation of perceptual closeness or perceptual error between two color images [4]. Precise evaluation of perceptual closeness between two colors is also essential in color matching systems used in various applications such as multimedia products, image arts, entertainment, and advertisements [6], [14], [22].

38

L u v and L a b color spaces are extremely useful in imaging systems where exact perceptual reproduction of color images (color consistency) across the entire system is of primary concern rather than real-time or simple computing. Applications include advertising, graphic arts, digitized or animated paintings etc. Suppose, an imaging system consists of various color devices, for example video camera/digital scanner, display device, and printer. A painting has to be digitized, displayed, and printed. The displayed and printed versions of the painting must appear as close as possible to the original image. L u v and L a b color spaces are the best to work with in such cases. Both these systems have been successfully applied to image coding for printing [4], [16]. Color calibration is another important process related to color consistency. It basically equalizes an image to be viewed under di erent illumination or viewing conditions. For instance, an image of a target object can only be taken under a speci c lighting condition in a laboratory. But the appearance of this target object under normal viewing conditions, say in ambient light, has to be known. Suppose, there is a sample object whose image under ambient light is available. Then the solution is to obtain the image of the sample object under the same speci c lighting condition in the laboratory. Then a correction formula can be formulated based on the images of the sample object obtained and these can be used to correct the target object for the ambient light [14]. Perceptual based color spaces, such as L a b , are very useful for computations in such problems [31], [37]. An instance, where such calibration techniques have great potential, is medical imaging in dentistry. Perceptually uniform color spaces, with the Euclidean metric to quantify color distances, are particularly useful in color image segmentation of natural scenes using histogram-based or clustering techniques. A method of detecting clusters by tting to them some circular-cylindrical decision elements in the L a b uniform color coordinate system was proposed in [39], [40]. The method estimates the clusters' color distributions without imposing any constraints on their forms. Boundaries of the decision elements are formed with constant lightness and constant chromaticity loci. Each boundary is obtained using only 1-D histograms of the L HÆ C cylindrical coordinates of the image data. The cylindrical coordinates L HÆ C [30] of the L a b color space known as lightness, hue, and chroma, are given by:

L = L H Æ = arctan(b=a) C  = (a2 + b2)1=2

(1.105) (1.106) (1.107)

The L a b space is often used in color management systems (CMS). A color management system handles the color calibration and color consistency issues. It is a layer of software resident on a computer that negotiates color reproduction between the application and color devices. Color management systems perform the color transformations necessary to exchange accurate

39

color between diverse devices [4], [43]. A uniform, based on CIE L u v , color space named TekHVC was proposed by Tektronix as part of its commercially available CMS [45]. 1.10 The Munsell Color Space

The Munsell color space represents the earliest attempt to organize color perception into a color space [5], [14], [46]. The Munsell space is de ned as a comparative reference for artists. Its general shape is that of a cylindrical representation with three dimensions roughly corresponding to the perceived lightness, hue and saturation. However, contrary to the HSV or HSI color models where the color solids were parameterized by hue, saturation and perceived lightness, the Munsell space uses the method of the color atlas, where the perception attributes are used for sampling. The fundamental principle behind the Munsell color space is that of equality of visual spacing between each of the three attributes. Hue is scaled according to some uniquely identi able color. It is represented by a circular band divided into ten sections. The sections are de ned as red, yellow-red, yellow, green-yellow, green, blue-green, blue, purple-blue, purple and red-purple. Each section can be further divided into ten subsections if ner divisions of hue are necessary. A chromatic hue is described according to its resemblance to one or two adjacent hues. Value in the Munsell color space refers to a color's lightness or darkness and is divided into eleven sections numbered zero to ten. Value zero represents black while a value of ten represent white. The chroma de nes the color's strength. It is measured in numbered steps starting at one with weak colors having low chroma values. The maximum possible chroma depends on the hue and the value being used. As can be seen in Fig. (1.14), the vertical axis of the Munsell color solid is the line of V values ranging from black to white. Hue changes along each of the circles perpendicular to the vertical axis. Finally, chroma starts at zero on the V axis and changes along the radius of each circle. The Munsell space is comprised of a set of 1200 color chips each assigned a unique hue, value and chroma component. These chips are grouped in such a way that they form a three dimensional solid, which resembles a warped sphere [5]. There are di erent editions of the basic Munsell book of colors, with di erent nishes (glossy or matte), di erent sample sizes and a di erent number of samples. The glossy nish collection displays color point chips arranged on 40 constant-hue charts. On each constant-hue chart the chips are arranged in rows and columns. In this edition the colors progress from light at the top of each chart to very dark at the bottom by steps which are intended to be perceptually equal. They also progress from achromatic colors, such as white and gray at the inside edge of the chart, to chromatic colors at the outside edge of the chart by steps that are also intended to be

40

perceptually equal. All the charts together make up the color atlas, which is the color solid of the Munsell system. Value

Y

GY

Hue

G

YR Chroma R

BG

RP

B PB

P

The Munsell color system Fig. 1.14.

Although the Munsell book of colors can be used to de ne or name colors, in practice is not used directly for image processing applications. Usually stored image data, most often in RGB format, are converted to the Munsell coordinates using either lookup tables or closed formulas prior to the actual application. The conversion from the RGB components to the Munsell hue (H ), value (V ) corresponding to luminance and chroma (C ) corresponding to saturation, can be achieved by using the following mathematical algorithm [47]: x = 0:620R + 0:178G + 0:204B

y = 0:299R + 0:587G + 0:144B z = 0:056G + 0:942B

(1.108)

A nonlinear transformation is applied to the intermediate values as follows: p = f (x) f (y) (1.109)

q = 0:4(f (z) f (y)) (1.110) 1 where f (r) = 11:6r 3 1:6. Further the new variables are transformed to: s = (a + bcos())p (1.111) t = (c + dsin())q (1.112) where  = tan 1 ( pq ), a = 8:880, b = 0:966, c = 8:025 and d = 2:558. Finally,

the requested values are obtained as:

H=

s t

arctan( )

(1.113)

41

V = f (y) and

(1.114)

C = (s2 + t2) 2

1

(1.115)

Alternatively, conversion from RGB, or other color spaces, to the Munsell color space can be achieved through look-up tables and published charts [5]. In summary, the Munsell color system is an attempt to de ne color in terms of hue, chroma and lightness parameters based on subjective observations rather than direct measurements or controlled perceptual experiments. Although it has been found that the Munsell space is not as perceptually uniform as originally claimed and, despite the fact that it cannot directly integrate with additive color schemes, it is still in use today despite attempts to introduce colorimetric models for its replacement. 1.11 The Opponent Color Space

The opponent color space family is a set of physiologically motivated color spaces inspired by the physiology of the human visual system. According to the theory of color vision discussed in [48] the human vision system can be expressed in terms of opponent hues, yellow and blue on one hand and green and red on the other, which cancel each other when superimposed. In [49] an experimental procedure was developed which allowed researchers to quantitatively express the amounts of each of the basic hues present in any spectral stimulus. The color model of [50], [51], [52], [44] suggests the transformation of the RGB `cone' signals to three channels, one achromatic channel (I) and two opponent color channels (RG, YB) according to:

RG = R G Y B = 2B R G I =R+G+B

(1.116) (1.117) (1.118)

At the same time a set of e ective color features was derived by systematic experiments of region segmentation [53]. According to the segmentation procedure of [53] the color which has the deep valleys on its histogram and has the largest discriminant power to separate the color clusters in a given region need not be the R, G, and B color features. Since a feature is said to have large discriminant power if its variance is large, color features with large discriminant power were derived by utilizing the Karhunen-Loeve (KL) transformation. At every step of segmenting a region, calculation of the new color features is done for the pixels in that region by the KL transform of R, G, and B signals. Based on extensive experiments [53], it was concluded

42 Cones

Opponent Signals

R R+G+B

G

R-G

B

2B-R-G

Fig. 1.15.

The Opponent color stage of the human visual system

that three color features constitute an e ective set of features for segmenting color images, [54], [55]:

I 1 = (R + G3 + B) I 2 = (R B) I 3 = (2G 2R B)

(1.119) (1.120) (1.121)

In the opponent color space hue could be coded in a circular format ranging through blue, green, yellow, red and black to white. Saturation is de ned as distance from the hue circle making hue and saturation speciable with in color categories. Therefore, although opponent representation are often thought as a linear transforms of RGB space, the opponent representation is much more suitable for modeling perceived color than RGB is [14]. 1.12 New Trends

The plethora of color models available poses application diÆculties. Since most of them are designed to perform well in a speci c application, their performance deteriorates rapidly under di erent operating conditions. Therefore, there is a need to merge the di erent (mainly device dependent) color spaces into a single standard space. The di erences between the monitor RGB space and device independent spaces, such as the HVS and the CIE L a b spaces impose problems in applications, such as multimedia database navigation and face recognition primarily due to the complexity of the operations needed to support the transform from/to device dependent color spaces. To overcome such problems and to serve the needs of network-centric applications and WWW-based color imaging systems, a new standardized color space based on a colorimetric RGB (sRGB) space has recently been proposed [56]. The aim of the new color space is to complement the current color space

43

management strategies by providing a simple, yet eÆcient and cost e ective method of handling color in the operating systems, device drivers and the Web using a simple and robust device independent color de nition. Since most computer monitors are similar in their key color characteristics and the RGB space is the most suitable color space for the devices forming a modern computer-based imaging systems, the colorimetric RGB space seems to be the best candidate for such a standardized color space. In de ning a colorimetric color space, two factors are of paramount importance:

 

the viewing environment parameters with its dependencies on the Human Visual System the standard device space colorimetric de nitions and transformations [56]

The viewing environment descriptions contain all the necessary transforms needed to support conversions between standard and target viewing environments. On the other hand, the colorimetric de nitions provide the transforms necessary to convert between the new sRGB and the CIE-XYZ color space. The reference viewing environment parameters can be found in [56] with the sRGB tristimulus values calculated from the CIE-XYZ values according to the following transform: 2

RsRGB 3 2 3:2410 1:5374 0:4986 3 2 X 3 4 GsRGB 5 = 4 0:9692 1:8760 0:0416 5 4 Y 5 BsRGB 0:0556 0:2040 1:0570 Z

(1.122)

In practical image processing systems negative sRGB tristimulus values and sRGB values greater than 1 are not retained and typically removed by utilizing some form of clipping. In the sequence, the linear tristimulus values are transformed to nonlinear sR0 G0 B0 as follows: 1. If RsRGB ; GsRGB ; BsRGB 0:0034 then

sR0 = 12:92RsRGB sG0 = 12:92GsRGB sB0 = 12:92BsRGB 2. else if RsRGB ; GsRGB ; BsRGB > 0:0034 then sR0 = 1:055RsRGB 21::40 0:055 sG0 = 1:055GsRGB 12::04 0:055 sB0 = 1:055BsRGB 12::04 0:055

(1.123) (1.124) (1.125) (1.126) (1.127) (1.128)

44

The e ect of the above transformation is to closely t a straightforward value of 2.2 with a slight o set to allow for invertibility in integer mathematics. The nonlinear R0 G0 B0 values are then converted to digital values with a black digital count of 0 and a white digital count of 255 for 24-bit coding as follows:

sRd = 255:0sR0 sGd = 255:0sG0 sBd = 255:0sB0

(1.129) (1.130) (1.131)

The backwards transform is de ned as follows:

sR0 = sRd + 255:0 sG0 = sGd + 255:0 sB0 = sBd + 255:0

(1.132) (1.133) (1.134)

and

1. if RsRGB ; GsRGB ; BsRGB 0:03928 then

RsRGB = sR0 + 12:92 GsRGB = sG0 + 12:92 BsRGB = sB0 + 12:92 2. else if RsRGB ; GsRGB ; BsRGB > 0:03928 then 0 0:055 2:4 RsRGB = ( sR 1+:055 ) 0 0:055 2:4 GsRGB = ( sG 1+:055 ) 0 0:055 2:4 ) BsRGB = ( sB 1+:055

with

X 3 2 0:4124 0:3576 0:1805 3 2 RsRGB 3 4 Y 5 = 4 0:2126 0:7152 0:0722 5 4 GsRGB 5 Z 0:0193 0:1192 0:9505 BsRGB

(1.135) (1.136) (1.137)

(1.138) (1.139) (1.140)

2

(1.141)

The addition of a new standardized color space which supports Webbased imaging systems, device drivers, printers and monitors complementing the existing color management support can bene t producers and users alike by presenting a clear path towards an improved color management system.

45 1.13 Color Images

Color imaging systems are used to capture and reproduce the scenes that humans see. Imaging systems can be built using a variety of optical, electronic or chemical components. However, all of them perform three basic operations, namely: (i) image capture, (ii) signal processing, and (iii) image formation. Color-imaging devices exploit the trichromatic theory of color to regulate how much light from the three primary colors is absorbed or re ected to produce a desired color. There are a number of ways to acquiring and reproducing color images, including but not limited to:

 Photographic lm. The lm which is used by conventional cameras con 

 

tains three emulation layers, which are sensitive to red and blue light, which enters through the camera lens. Digital cameras. Digital cameras use a CCD to capture image information. Color information is captured by placing red, green and blue lters before the CCD and storing the response to each channel. Cathode-Ray tubes. CRTs are the display device used in televisions and computer monitors. They utilize a extremely ne array of phosphors that emit red, green and blue light at intensities governed by an electron gun, in accordance to an image signal. Due to the close proximity of the phosphors and the spatial ltering characteristics of the human eye, the emitted primary colors are mixed together producing an overall color. Image scanners. The most common method of scanning color images is the utilization of three CCD's each with a lter to capture red, green and blue light re ectance. These three images are then merged to create a copy of the scanned image. Color printers. Color printers are the most common method of attaining a printed copy of a captured color image. Although the trichromatic theory is still implemented, color in this domain is subtractive. The primaries which are used are usually cyan, magenta and yellow. The amount of the three primaries which appear on the printed media govern how much light is re ected.

1.14 Summary

In this chapter the phenomenon of color was discussed. The basic color sensing properties of the human visual system and the CIE standard color speci cation system XYZ were described in detail. The existence of three types of spectral absorption cones in the human eyes serves as the basis of the trichromatic theory of color, according to which all visible colors can be created by combining three . Thus, any color can be uniquely represented by a three dimensional vector in a color model de ned by the three primary colors.

46 Table 1.3.

Color Model

Color System RGB R0 G0 B0 XYZ YIQ YCC I1I2I3 HSV HSI HLS L u v L a b Munsell

Fig. 1.16.

Transform (from RGB) non linear linear linear linear linear non linear non linear non linear non linear non linear non linear

Component correlation highly correlated correlated uncorrelated uncorrelated correlated correlated correlated correlated correlated correlated correlated

A taxonomy of color models

Color speci cation models are of paramount importance in applications where eÆcient manipulation and communication of images and video frames are required. A number of color speci cation models are in use today. Examples include color spaces, such as the RGB, R0 G0 B0 , YIQ, HSI, HSV, HLS,L u v , and L a b . The color model is a mathematical representation of spectral colors in a nite dimensional vector space. In each one of them the actual color is reconstructed by combining the basis elements of the vector

References

47

Color Spaces Models Colorimetric Device-oriented

User-oriented Munsell

XYZ - non-uniform spaces RGB, YIQ, YCC - uniform spaces L a b , L u v HSI, HSV, HLS, I1I2I3

Applications

colorimetric calculations storage, processing, analysis coding, color TV, storage (CD-ROM) color di erence evaluation analysis, color management systems human color perception multimedia, computer graphics human visual system

spaces, the so called primary colors. By de ning di erent primary colors for the representation of the system di erent color models can be devised. One important aspect is the color transformation, the change of coordinates from one color system to another (see Table 1.3). Such a transformation associates to each color in one system a color in the other model. Each color model comes into existence for a speci c application in color image processing. Unfortunately, there is no technique for determining the optimum coordinate model for all image processing applications. For a speci c application the choice of a color model depends on the properties of the model and the design characteristics of the application. Table 1.14 summarizes the most popular color systems and some of their applications. References

1. Gonzalez, R., Woods, R.E. (1992): Digital Image Processing. Addisson Wesley, Reading MA. 2. Robertson, P., Schonhut, J. (1999): Color in computer graphics. IEEE Computer Graphics and Applications, 19(4), 18-19. 3. MacDonald, L.W. (1999): Using color e ectively in computer graphics. IEEE Computer Graphics and Applications, 19(4), 20-35. 4. Poynton, C.A. (1996): A Technical Introduction to Digital Video. Prentice Hall, Toronto, also available at http://www.inforamp.net/poynton/Poynton{ Digital-Video.html . 5. Wyszecki, G., Stiles, W.S. (1982): Color Science, Concepts and Methods, Quantitative Data and Formulas. John Wiley, N.Y. , 2nd Edition. 6. Hall, R.A. (1981): Illumination and Color in Computer Generated Imagery. Springer Verlag, New York, N.Y. 7. Hurlbert, A. (1989): The Computation of Color. Ph.D Dissertation, Massachusetts Institute of Technology. 8. Hurvich, Leo M. (1981): Color Vision. Sinauer Associates, Sunderland MA. 9. Boynton, R.M. (1990): Human Color Vision. Halt, Rinehart and Winston. 10. Gomes, J., Velho, L. (1997): Image Processing for Computer Graphics. Springer Verlag, New York, N.Y., also available at http://www.springerny.com/catalog/np/mar97np/DATA/0-387-94854-6.html .

48 11. Fairchild, M.D. (1998): Color Appearance Models. Addison-Wesley, Readings, MA. 12. Sharma, G., Yrzel, M.J., Trussel, H.J. (1998): Color imaging for multimedia. Proceedings of the IEEE, 86(6): 1088{1108. 13. Sharma, G., Trussel, H.J. (1997): Digital color processing. IEEE Trans. on Image Processing, 6(7): 901-932. 14. Lammens, J.M.G. (1994): A Computational Model for Color Perception and Color Naming. Ph.D Dissertation, State University of New York at Bu alo, Bu alo, New York. 15. Johnson, G.M., Fairchild, M.D. (1999): Full spectral color calculations in realistic image synthesis. IEEE Computer Graphics and Applications, 19(4), 47-53. 16. Lu, Guoyun (1996): Communication and Computing for Distributed Multimedia Systems. Artech House Publishers, Boston, MA. 17. Kubinger, W., Vincze, M., Ayromlou, M. (1998): The role of gamma correction in colour image processing. in Proceedings of the European Signal Processing Conference, 2: 1041{1044. 18. Luong, Q.T. (1993): Color in computer vision. in Handbook of Pattern Recognition and Computer Vision, Word Scienti c Publishing Company): 311{368. 19. Young, T. (1802): On the theory of light and colors. Philosophical Transactions of the Royal Society of London, 92: 20{71. 20. Maxwell, J.C. (1890): On the theory of three primary colors. Science Papers 1, Cambridge University Press: 445{450. 21. Padgham, C.A., Saunders, J.E. (1975): The Perception of Light and Color. Academic Press, New York, N.Y. 22. Judd, D.B., Wyszecki, G. (1975): Color in Business, Science and Industry. John Wiley, New York, N.Y. 23. Foley, J.D., vanDam, A., Feiner, S.K., Hughes, J.F. (1990): Fundamentals of Interactive Computer Graphics. Addison Wesley, Reading, MA. 24. CCIR (1990): CCIR Recommendation 709. Basic parameter values for the HDTV standard for studio and for international program exchange. Geneva, Switcherland. 25. CIE (1995): CIE Publication 116. Industrial color-di erence evaluation. Vienna, Austria. 26. Poynton, C.A. (1993): Gamma and its disguises. The nonlinear mappings of intensity in perception, CRTs, lm and video. SMPTE Journal: 1099{1108. 27. Kasson M.J., Ploa e, W. (1992): An analysis of selected computer interchange color spaces. ACM Transaction of Graphics, 11(4): 373-405. 28. Shih, Tian-Yuan (1995): The reversibility of six geometric color spaces. Photogrammetric Engineering and Remote Sensing, 61(10): 1223{1232. 29. Levkowitz H., Herman, G.T. (1993): GLHS: a generalized lightness, hue and saturation color model. Graphical Models and Image Processing, CVGIP-55(4): 271{285. 30. McLaren, K. (1976): The development of the CIE L a b uniform color space. J. Soc. Dyers Colour, 338{341. 31. Hill, B., Roer, T., Vorhayen, F.W. (1997): Comparative analysis of the quantization of color spaces on the basis of the CIE-Lab color di erence formula. ACM Transaction of Graphics, 16(1): 110{154. 32. Hall, R. (1999): Comparing spectral color computation methods. IEEE Computer Graphics and Applications, 19(4), 36-44. 33. Hague, G.E., Weeks, A.R., Myler, H.R. (1995): Histogram equalization of 24 bit color images in the color di erence color space. Journal of Electronic Imaging, 4(1), 15-23.

References

49

34. Weeks, A.R. (1996): Fundamentals of Electronic Image Processing. SPIE Press, Piscataway, New Jersey. 35. Benson, K. B. (1992): Television Engineering Handbook. McGraw-Hill, London, U.K. 36. Smith, A.R. (1978): Color gamut transform pairs. Computer Graphics (SIGGRAPH'78 Proceedings), 12(3): 12{19. 37. Healey, C.G., Enns, J.T. (1995): A perceptual color segmentation algorithm. Technical Report, Department of Computer Science, University of British Columbia, Vancouver. 38. Luo, M. R. (1998): Color science. in Sangwine, S.J., Horne, R.E.N. (eds.), The Colour Image Processing Handbook, 26{52, Chapman & Hall, Cambridge, Great Britain. 39. Celenk, M. (1988): A recursive clustering technique for color picture segmentation. Proceedings of the Int. Conf. on Computer Vision and Pattern Recognition, 1: 437{444. 40. Celenk, M. (1990): A color clustering technique for image segmentation. Computer Vision, Graphics, and Image Processing, 52: 145{170. 41. Cong, Y. (1998): Intelligent Image Databases. Kluwer Academic Publishers, Boston, Ma. 42. Ikeda, M. (1980): Fundamentals of Color Technology. Asakura Publishing, Tokyo, Japan. 43. Rhodes, P. A. (1998): Colour management for the textile industry. in Sangwine, S.J., Horne, R.E.N. (eds.), The Colour Image Processing Handbook, 307-328, Chapman & Hall, Cambridge, Great Britain. 44. Palus, H. (1998): Colour spaces. in Sangwine, S.J., Horne, R.E.N. (eds.), The Colour Image Processing Handbook, 67{89, Chapman & Hall, Cambridge, Great Britain. 45. Tektronix (1990): TekColor Color Management System: System Implementers Manual. Tektronix Inc. 46. Birren, F. (1969): Munsell: A Grammar of Color. Van Nostrand Reinhold, New York, N.Y. 47. Miyahara, M., Yoshida, Y. (1988): Mathematical transforms of (R,G,B) colour data to Munsell (H,V,C) colour data. Visual Communications and Image Processing, 1001, 650{657. 48. Hering, E. (1978): Zur Lehe vom Lichtsinne. C. Gerond's Sohn, Vienna, Austria. 49. Jameson, D., Hurvich, L.M. (1968): Opponent-response functions related to measured cone photo pigments. Journal of the Optical Society of America, 58: 429{430. 50. de Valois, R.L., De Valois, K.K. (1975): Neural coding of color. in Carterette, E.C., Friedman, M.P. (eds.), Handbook of Perception. Volume 5, Chapter 5, 117{166, Academic Press, New York, N.Y. 51. de Valois, R.L., De Valois, K.K. (1993): A multistage color model. Vision Research 33(8): 1053{1065. 52. Holla, K. (1982): Opponent colors as a 2-dimensional feature within a model of the rst stages of the human visual system. Proceedings of the 6th Int. Conf. on Pattern Recognition, 1: 161{163. 53. Ohta, Y., Kanade, T., Sakai, T. (1980): Color information for region segmentation. Computer Graphics and Image Processing, 13: 222{241. 54. von Stein, H.D., Reimers, W. (1983): Segmentation of color pictures with the aid of color information and spatial neighborhoods. Signal Processing II: Theories and Applications, 1: 271{273. 55. Tominaga S. (1986): Color image segmentation using three perceptual attributes. Proceedings of CVPR'86, 1: 628-630.

50 56. Stockes, M., Anderson, M., Chandrasekar, Sri., Motta, Ricardo (1997): A standard default color space for the Internet sRGB. International Color Consortium (ICC), contributed document electronic reprint (http://www.color.org).

Suggest Documents