t-SNE最大化的自适应彩色图像灰度化方法
Adaptive decolorization method based on t-SNE maximization
- 2024年29卷第8期 页码:2333-2349
纸质出版日期: 2024-08-16
DOI: 10.11834/jig.230392
移动端阅览
浏览全部资源
扫码关注微信
纸质出版日期: 2024-08-16 ,
移动端阅览
谢斌, 徐燕, 王冠超, 杨舒敏, 李燕伟. 2024. t-SNE最大化的自适应彩色图像灰度化方法. 中国图象图形学报, 29(08):2333-2349
Xie Bin, Xu Yan, Wang Guanchao, Yang Shumin, Li Yanwei. 2024. Adaptive decolorization method based on t-SNE maximization. Journal of Image and Graphics, 29(08):2333-2349
目的
2
彩色图像的灰度化是计算机视觉领域的研究热点。针对传统彩色图像灰度化方法得到的灰度图像存在对比度保持不足、细节模糊及层次感欠缺等问题,本文结合t分布随机邻域嵌入(t-distributed stochastic neighbor embedding,t-SNE)提出了一种更加简单、高效的彩色图像灰度化新方法。
方法
2
首先,将t-SNE降维思想引入到彩色图像的灰度化过程中,设计了一种基于t-SNE最大化的彩色图像灰度化新模型,通过最大化能量函数使原始彩色图像中对比度较小的区域在灰度化后其对比度能够适当地变大或保持,让灰度图像更好地保持原始彩色图像的对比度特征和层次感。其次,在新模型中设计了一种自适应的对比度保持策略,根据颜色对比度信息来自适应地调节原始彩色图像不同区域的灰度化力度,更好地保留原始彩色图像的细节和对比度信息。最后,采用了一种高效的离散搜索方法以快速求解所提新模型。
结果
2
基于Cadik、CSDD(complex scene decolorization dataset)和Color250数据集的大量实验结果表明,与传统方法相比,本文方法得到的灰度图像具有更好的表现,在颜色对比度保持率(color contrast preserving ratio, CCPR)指标上,本文方法在上述3个数据集上的平均CCPR值最高,分别为0.874、0.862和0.864。另外,在相同硬件上测试不同灰度化方法的运行效率时,本文方法的运行时间最短。
结论
2
相较于传统灰度化方法,本文方法不仅能够更好地保持原始彩色图像的对比度、细节特征和层次感,而且在主观评价和客观评价方面均有更好的表现。
Objective
2
Image decolorization converts a color image into a grayscale one. It aims to maintain important features of the original color image as much as possible. It is normally used in the preprocessing stage of pattern recognition and image processing systems, such as extracting fingerprint features, face recognition, and medical imaging. Given that image decolorization maps color images in three dimensions into grayscale images in one dimension, it is a dimensionality reduction process essentially. Thus, the loss of information is inevitable. Accordingly, the major task for researchers can be expressed as preserving important features such as contrast, detail information, and hierarchy of the original color image as much as possible. In the process of acquiring grayscale images, the most basic method is to extract the luminance channel of color images directly. However, acquiring only the luminance information tends to lose the contrast details and appearance features of the input image. In recent years, researchers have focused more on whether grayscale images match the perception of color of the human visual system, and many classical image decolorization algorithms have been proposed. These color-to-gray conversion methods generally fall into two main categories: local mapping methods and global mapping methods. In the local mapping method, the mapping function is spatially varying, that is, the mapping functions of pixels at different spatial locations often differ. The grayscale image obtained by the local mapping method can keep part of the image structure information and local contrast information. However, such methods tend to lose the overall information of the image. They can also produce certain phenomena such as halo and noise enhancement. Some researchers have proposed the global mapping method to address the shortcomings of the abovementioned methods. In the global mapping method, the same mapping function is used for the entire image, and the grayscale value is only related to the color of the original color image. However, many of these methods focus excessively on pixels with large amplitude values, which tend to produce visual overfitting and cause difficulty in retaining information in areas with small contrast. A novel image decolorization method is proposed, in combination with t-distributed stochastic neighbor embedding (t-SNE), to address the problems of insufficient contrast maintenance, blurred details, and lack of hierarchy in grayscale images obtained by the traditional color-to-gray conversion method. This method is simpler and more efficient.
Method
2
In this study, considering that t-SNE can better preserve the important features and intrinsic connections of the original high-dimensional spatial data points, we introduce the idea of t-SNE dimensionality reduction into the process of color-to-gray conversion. A new model of color image grayscale is also designed based on t-SNE maximization. In this model, the similarity between any pixel in the original color image is represented by the joint Gaussian probability, while the similarity between any pixel in the corresponding grayscale image is represented by the joint Gaussian probability. The energy function is maximized such that the contrast areas of the original color image with less contrast can be appropriately enlarged or maintained after grayscale. Accordingly, the grayscale image can better maintain the contrast characteristics and the sense of hierarchy of the original color image. In addition, an adaptive contrast retention strategy is designed in this study and introduced into the new model to better preserve the details and contrast information of the original color image. The strategy defines a contrast retention coefficient that can adaptively adjust the graying intensity of different regions of the original color image based on the color contrast information. Given that the proposed maximization model is highly nonlinear, the solution using traditional methods such as gradient descent and iterative method will be particularly complex and time consuming. For this reason, we adopt an efficient discrete search strategy to solve the model. It transforms the continuous parameter optimization problem into a discrete parameter space search to quickly obtain grayscale images, which greatly improves the overall efficiency of the model.
Result
2
To demonstrate the effectiveness of the proposed method, we conducted an experimental comparison with various other traditional color-to-gray conversion methods on three different types of datasets, namely, Cadik, complex scene decolorization dataset (CSDD), and Color250. Experimental results show that the proposed method can better maintain the contrast, detail features, and hierarchy of the original color image. For the color contrast preserving ratio (CCPR) metric, the average CCPR values of the proposed method are 0.874, 0.862, and 0.864 for the three datasets when the contrast threshold values range from 1 to 15, respectively, which are better than the average CCPR values of the conventional grayscale method. Comparing the efficiency of different grayscale methods on the same hardware shows that the method has the shortest running time when inputting a raw color image of size 396 × 386 pixels, which is real-time speed. In addition, volunteers were randomly invited to complete preference experiments and accuracy experiments to more comprehensively evaluate the visual effects of grayscale images obtained by different grayscale methods. The experiments showed that the volunteers preferred the grayscale images obtained by the proposed method in this study. Therefore, the proposed method in this study was more consistent with the visual perception of human eyes.
Conclusion
2
In this study, we analyze the problems of traditional color-to-gray conversion methods and propose an adaptive decolorization method based on t-SNE maximization. The experimental data based on three datasets show that the proposed method in this study not only maintains the contrast, detail characteristics, and hierarchy of the original color image better but also has better performance in subjective and objective evaluations.
彩色图像灰度化t分布随机邻域嵌入(t-SNE)对比度保持离散搜索细节保持
color-to-gray conversiont-distributed stochastic neighbor embedding (t-SNE)contrast preservingdiscrete searchdetail preserving
Ancuti C O, Ancuti C and Bekaert P. 2010. Decolorizing images for robust matching//Proceedings of 2010 IEEE International Conference on Image Processing. Hong Kong, China: IEEE: 149-152 [DOI: 10.1109/ICIP.2010.5652328http://dx.doi.org/10.1109/ICIP.2010.5652328]
Ancuti C O, Ancuti C and Bekaert P. 2011. Enhancing by saliency-guided decolorization//Proceedings of 2011 IEEE Conference on Computer Vision and Pattern Recognition. Colorado Springs, USA: IEEE: 257-264 [DOI: 10.1109/CVPR.2011.5995414http://dx.doi.org/10.1109/CVPR.2011.5995414]
Bala R and Eschbach R. 2004. Spatial color-to-grayscale transform preserving chrominance edge information//Proceedings of 2004 Color and Imaging Conference. Scottsdale, USA: the Society for Imaging Science and Technology: 82-86
Ĉadík M. 2008. Perceptual evaluation of color-to-grayscale image conversions. Computer Graphics Forum, 27(7): 1745-1754 [DOI: 10.1111/j.1467-8659.2008.01319.xhttp://dx.doi.org/10.1111/j.1467-8659.2008.01319.x]
Du H, He S F, Sheng B, Ma L Z and Lau R W H. 2015. Saliency-guided color-to-gray conversion using region-based optimization. IEEE Transactions on Image Processing, 24(1): 434-443 [DOI: 10.1109/TIP.2014.2380172http://dx.doi.org/10.1109/TIP.2014.2380172]
Gooch A A, Olsen S C, Tumblin J and Gooch B. 2005. Color2Gray: salience-preserving color removal. ACM Transactions on Graphics, 24(3): 634-639 [DOI: 10.1145/1073204.1073241http://dx.doi.org/10.1145/1073204.1073241]
Guo J C, Yue H H, Zhang Y, Liu D, Liu X W and Zheng S D. 2022. The analysis of image enhancement on salient object detection. Journal of Image and Graphics, 27(7): 2129-2147
郭继昌, 岳惠惠, 张怡, 刘迪, 刘晓雯, 郑司达. 2022. 图像增强对显著性目标检测的影响研究. 中国图象图形学报, 27(7): 2129-2147 [DOI: 10.11834/ jig.200735http://dx.doi.org/10.11834/jig.200735]
Han Y, Luo L, Xie B and Xu C. 2020. Nonparametric histogram segmentation-based automatic detection of yarns. Textile Research Journal, 90(11/12): 1326-1341 [DOI: 10.1177/0040517519890212http://dx.doi.org/10.1177/0040517519890212]
Han Y, Xu C, Baciu G, Li M and Islam M R. 2017. Cartoon and texture decomposition-based color transfer for fabric images. IEEE Transactions on Multimedia, 19(1): 80-92 [DOI: 10.1109/TMM.2016.2608000http://dx.doi.org/10.1109/TMM.2016.2608000]
Jiang W, Xu Y H, Chen Z, Zhang N and Zhou J Z. 2022. Fault diagnosis for rolling bearing using a hybrid hierarchical method based on scale-variable dispersion entropy and parametric t-SNE algorithm. Measurement, 191: #110843 [DOI: 10.1016/J.MEASUREMENT.2022.110843http://dx.doi.org/10.1016/J.MEASUREMENT.2022.110843]
Jin Z M and Ng M K. 2015. A contrast maximization method for color-to-grayscale conversion. Multidimensional Systems and Signal Processing, 26(3): 869-877 [DOI: 10.1007/s11045-014-0295-2http://dx.doi.org/10.1007/s11045-014-0295-2]
Kim Y, Jang C, Demouth J and Lee S. 2009. Robust color-to-gray via nonlinear global mapping. ACM Transactions on Graphics, 28(5): 1-4 [DOI: 10.1145/1618452.1618507http://dx.doi.org/10.1145/1618452.1618507]
Liu Q G, Liu P X, Xie W S, Wang Y H and Liang D. 2015. GcsDecolor: gradient correlation similarity for efficient contrast preserving decolorization. IEEE Transactions on Image Processing, 24(9): 2889-2904 [DOI: 10.1109/TIP.2015.2423615http://dx.doi.org/10.1109/TIP.2015.2423615]
Liu Q G, Liu P X, Wang Y H and Leung H. 2017a. Semiparametric decolorization with Laplacian-based perceptual quality metric. IEEE Transactions on Circuits and Systems for Video Technology, 27(9): 1856-1868 [DOI: 10.1109/TCSVT.2016.2555779http://dx.doi.org/10.1109/TCSVT.2016.2555779]
Liu Q G, Shao G P, Wang Y H, Gao J B and Leung H. 2017b. Log-Euclidean metrics for contrast preserving decolorization. IEEE Transactions on Image Processing, 26(12): 5772-5783 [DOI: 10.1109/TIP.2017.2745104http://dx.doi.org/10.1109/TIP.2017.2745104]
Lu C W, Xu L and Jia J Y. 2012a. Contrast preserving decolorization//Proceedings of 2012 IEEE International Conference on Computational Photography. Seattle, USA: IEEE: 1-7 [DOI: 10.1109/ICCPhot.2012.6215215http://dx.doi.org/10.1109/ICCPhot.2012.6215215]
Lu C W, Xu L and Jia J Y. 2012b. Real-time contrast preserving decolorization//Proceedings of the SIGGRAPH Asia 2012 Technical Briefs. Singapore, Singapore: ACM: #34 [DOI: 10.1145/2407746.2407780http://dx.doi.org/10.1145/2407746.2407780]
Lu C W, Xu L and Jia J Y. 2014. Contrast preserving decolorization with perception-based quality metrics. International Journal of Computer Vision, 110(2): 222-239 [DOI: 10.1007/s11263-014-0732-6http://dx.doi.org/10.1007/s11263-014-0732-6]
Ma K D, Zhao T S, Zeng K and Wang Z. 2015. Objective quality assessment for color-to-gray image conversion. IEEE Transactions on Image Processing, 24(12): 4673-4685 [DOI: 10.1109/TIP.2015.2460015http://dx.doi.org/10.1109/TIP.2015.2460015]
Nafchi H Z, Shahkolaei A, Hedjam R and Cheriet M. 2017. CorrC2G: color to gray conversion by correlation. IEEE Signal Processing Letters, 24(11): 1651-1655 [DOI: 10.1109/LSP.2017.2755077http://dx.doi.org/10.1109/LSP.2017.2755077]
Neumann L, Ĉadík M and Nemcsics A. 2007. An efficient perception-based adaptive color to gray transformation//Proceedings of the 3rd Eurographics Conference on Computational Aesthetics in Graphics, Visualization and Imaging. Alberta, Canada: Eurographics Association: 20-22
Smith K, Landes P E, Thollot J and Myszkowski K. 2008. Apparent greyscale: a simple and fast conversion to perceptually accurate images and video. Computer Graphics Forum, 27(2): 193-200 [DOI: 10.1111/j.1467-8659.2008.01116.xhttp://dx.doi.org/10.1111/j.1467-8659.2008.01116.x]
Sowmya V, Govind D and Soman K P. 2017. Significance of incorporating chrominance information for effective color-to-grayscale image conversion. Signal, Image and Video Processing, 11(1): 129-136 [DOI: 10.1007/s11760-016-0911-8http://dx.doi.org/10.1007/s11760-016-0911-8]
Van der Maaten L and Hinton G. 2008. Visualizing data using t-SNE. Journal of Machine Learning Research, 9(86): 2579-2605
Wang L Y, Tian T, Xu H L and Tong H M. 2022. Short-term power load forecasting model based on t-SNE dimension reduction visualization analysis, VMD and LSSVM improved with chaotic sparrow search algorithm optimization. Journal of Electrical Engineering and Technology, 17(5): 2675-2691 [DOI: 10.1007/s42835-022-01101-7http://dx.doi.org/10.1007/s42835-022-01101-7]
Wang W, Li Z G and Wu S Q. 2018. Color contrast-preserving decolorization. IEEE Transactions on Image Processing, 27(11): 5464-5474 [DOI: 10.1109/TIP.2018.2855424http://dx.doi.org/10.1109/TIP.2018.2855424]
Wang X, Luo T T, Liu C X and Peng H Y. 2015. Decolorization with contrast, color consistency and grayscale pixel preservation. Journal of Image and Graphics, 20(3): 408-417
王勋, 罗婷婷, 刘春晓, 彭浩宇. 2015. 对比度、颜色一致性和灰度像素保持的消色算法. 中国图象图形学报, 20(3): 408-417 [DOI: 10.11834/jig.20150312http://dx.doi.org/10.11834/jig.20150312]
Wang Z, Bovik A C, Sheikh H R and Simoncelli E P. 2004. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4): 600-612 [DOI: 10.1109/TIP.2003.819861http://dx.doi.org/10.1109/TIP.2003.819861]
Xiong J J, Lu H Y, Liu Q G and Xu X L. 2018. Parametric ratio-based method for efficient contrast-preserving decolorization. Multimedia Tools and Applications, 77(12): 15721-15745 [DOI: 10.1007/s11042-017-5143-6http://dx.doi.org/10.1007/s11042-017-5143-6]
Yu J, Li F and Lyu X G. 2021. Contrast preserving decolorization based on the weighted normalized L1 norm. Multimedia Tools and Applications, 80(21/23): 31753-31782 [DOI: 10.1007/s11042-021-11172-9http://dx.doi.org/10.1007/s11042-021-11172-9]
Yu N N, Li J J and Hua Z. 2022. Decolorization algorithm based on contrast pyramid transform fusion. Multimedia Tools and Applications, 81(11): 15017-15039 [DOI: 10.1007/s11042-022-12189-4http://dx.doi.org/10.1007/s11042-022-12189-4]
Zhao H L, Zhang H N and Jin X G. 2018. Efficient image decolorization with a multimodal contrast-preserving measure. Computers and Graphics, 70: 251-260 [DOI: 10.1016/j.cag.2017.07.009http://dx.doi.org/10.1016/j.cag.2017.07.009]
Zhou M Q, Sheng B and Ma L Z. 2014. Saliency preserving decolorization//Proceedings of 2014 IEEE International Conference on Multimedia and Expo. Chengdu, China: IEEE: 1-6 [DOI: 10.1109/ICME.2014.6890153http://dx.doi.org/10.1109/ICME.2014.6890153]
相关文章
相关作者
相关机构