AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (1.3 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Open Access

Convolutional Neural Network Image Classification Based on Different Color Spaces

School of Computer Science and Engineering, Macau University of Science and Technology, Macao 999078, China
School of Computer Science, University of Nottingham Ningbo China, Ningbo 315100, China
Show Author Information

Abstract

Although Convolutional Neural Networks (CNNs) have achieved remarkable success in image classification, most CNNs use image datasets in the Red-Green-Blue (RGB) color space (one of the most commonly used color spaces). The existing literature regarding the influence of color space use on the performance of CNNs is limited. This paper explores the impact of different color spaces on image classification using CNNs. We compare the performance of five CNN models with different convolution operations and numbers of layers on four image datasets, each converted to nine color spaces. We find that color space selection can significantly affect classification accuracy, and that some classes are more sensitive to color space changes than others. Different color spaces may have different expression abilities for different image features, such as brightness, saturation, hue, etc. To leverage the complementary information from different color spaces, we propose a pseudo-Siamese network that fuses two color spaces without modifying the network architecture. Our experiments show that our proposed model can outperform the single-color-space models on most datasets. We also find that our method is simple, flexible, and compatible with any CNN and image dataset.

References

[1]
D. Chai and A. Bouzerdoum, A Bayesian approach to skin color classification in YCbCr color space, in Proc. of Intelligent Systems and Technologies for the New Millennium, Kuala Lumpur, Malaysia, 2000, pp. 421–424.
[2]

B. Chen, S. Shi, J. Sun, B. Chen, K. Guo, L. Du, J. Yang, Q. Xu, S. Song, and W. Gong, Using HSI color space to improve the multispectral lidar classification error caused by measurement geometry, IEEE Trans. Geosci. Remote Sens., vol. 59, no. 4, pp. 3567–3579, 2021.

[3]
D. S. Y. Kartika and D. Herumurti, Koi fish classification based on HSV color space, in Proc. Int. Conf. Information & Communication Technology and Systems (ICTS), Surabaya, Indonesia, 2016, pp. 96–100.
[4]
E. A. Khan and E. Reinhard, Evaluation of color spaces for edge classification in outdoor scenes, in Proc. IEEE Int. Conf. Image Processing 2005, Genova, Italy, 2005, pp. 952–955.
[5]
O. R. Indriani, E. J. Kusuma, C. A. Sari, E. H. Rachmawanto, and D. R. I. M. Setiadi, Tomatoes classification using K-NN based on GLCM and HSV color space, in Proc. Int. Conf. Innovative and Creative Information Technology (ICITech), Salatiga, Indonesia, 2017, pp. 1–6.
[6]
S. Sural, G. Qian, and S. Pramanik, Segmentation and histogram generation using the HSV color space for image retrieval, in Proc. Proceedings. Int. Conf. Image Processing, Rochester, NY, USA, 2002, pp. 589–592.
[7]
P. Ganesan, V. Rajini, and R. I. Rajkumar, Segmentation and edge detection of color images using CIELAB color space and edge detectors, in Proc. INTERACT-2010, Chennai, India, 2010, pp. 393–397.
[8]
N. M. Kwok, Q. P. Ha, and G. Fang, Effect of color space on color image segmentation, in Proc. 2nd Int. Congress on Image and Signal Processing, Tianjin, China, 2009, pp. 1–5.
[9]

A. B. A. Hassanat, M. Alkasassbeh, M. Al-Awadi, and E. A. A. Alhasanat, Color-based object segmentation method using artificial neural network, Simul. Model. Pract. Theory, vol. 64, pp. 3–17, 2016.

[10]

J. Liu and X. Zhong, An object tracking method based on mean shift algorithm with HSV color space and texture features, Clust. Comput., vol. 22, no. 3, pp. 6079–6090, 2019.

[11]

P. Hidayatullah and M. Zuhdi, Color-texture based object tracking using HSV color space and local binary pattern, Int. J. Electr. Eng. Inform., vol. 7, no. 2, pp. 161–174, 2015.

[12]
S. Saravanakumar, A. Vadivel, and C. G. Saneem Ahmed, Multiple object tracking using HSV color space, in Proc. 2011 Int. Conf. Communication, Computing & Security, Rourkela, India, 2011, pp. 247–252.
[13]

R. Azhar, D. Tuwohingide, D. Kamudi, Sarimuddin, and N. Suciati, Batik image classification using SIFT feature extraction, bag of features and support vector machine, Procedia Comput. Sci., vol. 72, pp. 24–30, 2015.

[14]
Q. Li and X. Wang, Image classification based on SIFT and SVM, in Proc. IEEE/ACIS 17th Int. Conf. Computer and Information Science (ICIS), Singapore, Singapore, 2018, pp. 762–765.
[15]
Z. Xian, M. Azam, and N. Bouguila, Statistical modeling using bounded asymmetric Gaussian mixtures: Application to human action and gender recognition, in Proc. IEEE 22nd Int. Conf. Information Reuse and Integration for Data Science (IRI), Las Vegas, NV, USA, 2021, pp. 41–48.
[16]
Z. Xian, M. Azam, M. Amayri, W. Fan, and N. Bouguila, Bounded asymmetric Gaussian mixture-based hidden Markov models, in Hidden Markov Models and Applications, N. Bouguila, W. Fan, and M. Amayri, eds. Cham, Switzerland: Springer International Publishing, 2022, pp. 33–58.
[17]
S. L. Phung, A. Bouzerdoum, and D. Chai, A novel skin color model in YCbCr color space and its application to human face detection, in Proc. Int. Conf. Image Processing, Rochester, NY, USA, 2002, pp. 289–292.
[18]

I. Philipp and T. Rath, Improving plant discrimination in image processing by use of different colour space transformations, Comput. Electron. Agric., vol. 35, no. 1, pp. 1–15, 2002.

[19]
P. A. Herrault, D. Sheeren, M. Fauvel, and M. Paegelow, Automatic extraction of forests from historical maps based on unsupervised classification in the CIELab color space, in Geographic Information Science at the Heart of Europe. Cham, Switzerland: Springer International Publishing, 2013, pp. 95–112.
[20]
J. Chen and J. Lei, Research on color image classification based on HSV color space, in Proc. Second Int. Conf. Instrumentation, Measurement, Computer, Communication and Control, Harbin, China, 2012, pp. 944–947.
[21]

H. K. Kim, J. H. Park, and H. Y. Jung, An efficient color space for deep-learning based traffic light recognition, J. Adv. Transp., vol. 2018, p. 2365414, 2018.

[22]
R. Girshick, J. Donahue, T. Darrell, and J. Malik, Rich feature hierarchies for accurate object detection and semantic segmentation, in Proc. IEEE Conf. Computer Vision and Pattern Recognition, Columbus, OH, USA, 2014, pp. 580–587.
[23]
S. Ren, K. He, R. Girshick, and J. Sun, Faster R-CNN: Towards real-time object detection with region proposal networks, in Proc. 28th Int. Conf. Neural Information Processing Systems, Montreal, Canada, 2015, pp. 91–99.
[24]
J. Dai, Y. Li, K. He, and J. Sun, RFCN:Object detection via region-based fully convolutional networks, in Proc. 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 2016, pp. 379–387.
[25]
Y. Yuan, G. Zeng, Z. Chen, and Y. Gao, Color image quality assessment with multi deep convolutional networks, in Proc. IEEE 4th Int. Conf. Signal and Image Processing (ICSIP), Wuxi, China, 2019, pp. 934–941.
[26]

S. Khan and M. Narvekar, Novel fusion of color balancing and superpixel based approach for detection of tomato plant diseases in natural complex environment, J. King Saud Univ. Comput. Inf. Sci., vol. 34, no. 6, pp. 3506–3516, 2022.

[27]

M. Mehmood, N. Alshammari, S. A. Alanazi, A. Basharat, F. Ahmad, M. Sajjad, and K. Junaid, Improved colorization and classification of intracranial tumor expanse in MRI images via hybrid scheme of Pix2Pix-cGANs and NASNet-large, J. King Saud Univ. Comput. Inf. Sci., vol. 34, no. 7, pp. 4358–4374, 2022.

[28]
S. N. Gowda and C. Yuan, ColorNet: investigating the importance of color spaces for image classification, in Proc. Computer Vision—ACCV 2018, Cham, Switzerland, 2019, pp. 581–596.
[29]

S. Bianco, C. Cusano, P. Napoletano, and R. Schettini, Improving CNN-based texture classification by color balancing, J. Imaging, vol. 3, no. 3, p. 33, 2017.

[30]
Y. LeCun, L. D. Jackel, L. Bottou, A. Brunot, C. Cortes, J. S. Denker, H. Drucker, I. R. Subramanian, U. Muller, E. Sackinger, P. Y. Simard, and V. N. Vapnik, Comparison of learning algorithms for hand written digit recognition, in Proc. 5th Int. Conf. on Artificial Neural Networks, Perth, Australia, 1995, pp. 53–60.
[31]

A. Krizhevsky, I. Sutskever, and G. E. Hinton, ImageNet classification with deep convolutional neural networks, Commun. ACM, vol. 60, no. 6, pp. 84–90, 2017.

[32]
J. Deng, W. Dong, R. Socher, L. J. Li, K. Li, and F. F. Li, ImageNet: A large-scale hierarchical image database, in Proc. IEEE Conf. Computer Vision and Pattern Recognition. Miami, FL, USA, 2009, pp. 248–255.
[33]
F. Mamalet and C. Garcia, Simplifying ConvNets for fast learning, in Artificial Neural Networks and Machine Learning. Berlin, Germany: Springer, 2012, pp. 58–65.
[34]
L. Sifre and S. Mallat, Rigid-motion scattering for texture classification, arXiv preprint arXiv: 1403.1687, 2014.
[35]
A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, Mobilenets: Efficient convolutional neural networks for mobile vision applications, arXiv preprint arXiv: 1704.04861, 2017.
[36]
S. N. Gowda, Human activity recognition using combinatorial deep belief networks, in Proc. IEEE Conf. Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 2017, pp. 1–6.
[37]
K. Simonyan and A. Zisserman, Very deep convolutional networks for large-scale image recognition, in Proc. of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 2015, pp. 1–14.
[38]
C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, Going deeper with convolutions, in Proc. IEEE Conf. Computer Vision and Pattern Recognition, Boston, MA, USA, 2015, pp. 1–9.
[39]

S. Hochreiter, The vanishing gradient problem during learning recurrent neural nets and problem solutions, Int. J. Uncertain. Fuzziness Knowl.-Based Syst., vol. 6, no. 2, pp. 107–116, 1998.

[40]
S. Ioffe and C. Szegedy, Batch normalization: Accelerating deep network training by reducing internal covariate shift, in Proc. 32nd Int. Conf. Int. Conf. Machine Learning, Lille, France, 2015, pp. 448–456.
[41]
K. He, X. Zhang, S. Ren, and J. Sun, Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification, in Proc. IEEE Int. Conf. Computer Vision (ICCV), Santiago, Chile, 2015, pp. 1–9.
[42]
A. M. Saxe, J. L. McClelland, and S. Ganguli, Exact solutions to the nonlinear dynamics of learning in deep linear neural networks, in Proc. of the 2nd Int. Conf. Learning Representations, 2014, pp. 1–22.
[43]

Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel, Backpropagation applied to handwritten zip code recognition, Neural Comput., vol. 1, no. 4, pp. 541–551, 1989.

[44]
K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, in Proc. IEEE Conf. Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 2016, pp. 770–778.
[45]
G. Huang, Z. Liu, L. Van Der Maaten, and K. Q. Weinberger, Densely connected convolutional networks, in Proc. IEEE Conf. Computer Vision and Pattern Recognition, Honolulu, HI, USA, 2017, pp. 4700–4708.
[46]

G. Bebis and M. Georgiopoulos, Feed-forward neural networks, IEEE Potentials, vol. 13, no. 4, pp. 27–31, 1994.

[47]
N. Vandenbroucke, L. Macaire, and J. G. Postaire, Color pixels classification in an hybrid color space, in Proc. Int. Conf. Image Processing, Chicago, IL, USA, 1998, pp. 176–180.
[48]
B. D. Zarit, B. J. Super, and F. K. H. Quek, Comparison of five color models in skin pixel classification, in Proc. Int. Workshop on Recognition, Analysis, and Tracking of Faces and Gestures in Real-Time Systems, Corfu, Greece, 1999.
[49]
A. Krizhevsky and G. Hinton, Learning multiple layers of features from tiny images, Tech. Rep. 001, University of Toronto, Toronto, Canada, 2009.
[50]
Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, and A. Y. Ng, Reading digits in natural images with unsupervised feature learning, in Proc. of the NIPS Workshop on Deep Learning and Unsupervised Feature Learning, Sierra Nevada, Spain, 2011, pp. 1–9.
[51]
A. Coates, A. Ng, and H. Lee, An analysis of single-layer networks in unsupervised feature learning, in Proc. 14th Int. Conf. Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 2011, pp. 215–223.
[52]

N. A. Ibraheem, M. M. Hasan, R. Z. Khan, and P. K. Mishra, Understanding color models: A review, ARPN J. Sci. Technol., vol. 2, no. 3, pp. 265–275, 2012.

[53]

S. I. Nin, J. M. Kasson, and W. Plouffe, Printing CIELAB images on a CMYK printer using tri-linear interpolation, in Proc. SPIE/IS&T 1992 Symp. Electronic Imaging : Science and Technology, 1992, San Jose, CA, USA, pp. 316–324.

[54]

D. J. Sawicki and W. Miziolek, Human colour skin detection in CMYK colour space, IET Image Process., vol. 9, no. 9, pp. 751–757, 2015.

[55]
K. Plataniotis and A. N. Venetsanopoulos, Color Image Processing and Applications. Berlin, Germany: Springer Science & Business Media, 2000.
[56]

T. Smith and J. Guild, The C.I.E. colorimetric standards and their use, Trans. Opt. Soc., vol. 33, no. 3, pp. 73–134, 1932.

[57]

C. Wyman, P. P. Sloan, and P. Shirley, Simple analytic approximations to the CIE xyz colormatching functions, J. Comput. Graph. Tech., vol. 2, no. 2, p. 11, 2013.

[58]

H. S. Fairman, M. H. Brill, and H. Hemmendinger, How the CIE 1931 color-matching functions were derived from Wright-Guild data, 3.0.CO;2-7">Color Res. Appl., vol. 22, no. 1, p. 11–23, 1997.

[59]
C. Poynton, YUV and luminance considered harmful, in The Morgan Kaufmann Series in Computer Graphics, M. Gross and H. Pfister, eds. San Francisco, CA, USA: Morgan Kaufmann, 2003, pp. 595–600.
[60]
A. Ford and A. Roberts, Colour Space Conversions, London, UK: Westminster University, 1998.
[61]
K. Jack, Color spaces, in Video Demystified, Fifth Edition, K. Jack, ed. Burlington, VT, USA: Newnes, 2007, pp. 15–36.
[62]
L. A. Jeni, J. F. Cohn, and F. De La Torre, Facing imbalanced data: Recommendations for the use of performance metrics, in Proc. Humaine Association Conf. Affective Computing and Intelligent Interaction, Geneva, Switzerland, 2013, pp. 245–251.
[63]
D. P. Kingma and J. Ba, Adam: A method for stochastic optimization, arXiv preprint arXiv: 1412.6980, 2014.
Tsinghua Science and Technology
Pages 402-417
Cite this article:
Xian Z, Huang R, Towey D, et al. Convolutional Neural Network Image Classification Based on Different Color Spaces. Tsinghua Science and Technology, 2025, 30(1): 402-417. https://doi.org/10.26599/TST.2024.9010001

397

Views

103

Downloads

0

Crossref

0

Web of Science

0

Scopus

0

CSCD

Altmetrics

Received: 07 November 2023
Revised: 29 December 2023
Accepted: 02 January 2024
Published: 11 September 2024
© The Author(s) 2025.

The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).

Return