Grayscale image colorization is known as an ill-posed problem because of the imbalanced matching between intensity and color values. Even given prior hints about the original color image, existing colorization methods cannot recover the original color image from grayscale faithfully. In this paper, we propose to embed color information into an invertible grayscale, such that it can be easily recovered to the original color. However, a vanilla encoding-decoding network cannot produce rich representations of color information and thus the reconstruction quality is limited. Moreover, due to the neglect of the discrimination of color information, it cannot embed color information into visually inconspicuous patterns located in the grayscale. In this paper, we propose a novel color-encoding schema, dual features ensemble network (DFENet), for the effective embedding and faithfully reconstruction. In particular, we complement the residual representations with dense representations, to integrate the ability of local residual learning and local feature fusion. Furthermore, we propose an element-wise self-attention mechanism that highlights the discriminative features and suppresses the redundant ones generated from the dual path module. Extensive experiments demonstrate the proposed method outperforms state-of-the-art methods in terms of reconstruction quality as well as the similarity between the generated invertible grayscale and its groundtruth. INDEX TERMS Decolorization, colorization, dual features ensemble, convolutional neural network.
Color dimensionality reduction is believed as a non-invertible process, as re-colorization results in perceptually noticeable and unrecoverable distortion. In this article, we propose to convert a color image into a grayscale image that can fully recover its original colors, and more importantly, the encoded information is discriminative and sparse, which saves storage capacity. Particularly, we design an invertible deep neural network for color encoding and decoding purposes. This network learns to generate a residual image that encodes color information, and it is then combined with a base grayscale image for color recovering. In this way, the non-differentiable compression process (e.g., JPEG) of the base grayscale image can be integrated into the network in an end-to-end manner. To further reduce the size of the residual image, we present a specific layer to enhance Sparsity Enforcing Priors (SEP), thus leading to negligible storage space. The proposed method allows color embedding on a sparse residual image while keeping a high, 35dB PSNR on average. Extensive experiments demonstrate that the proposed method outperforms state-of-the-arts in terms of image quality and tolerability to compression.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.