Image fusion process consolidates data and information from various images of same sight into a solitary image. Each of the source images might speak to a fractional perspective of the scene, and contains both "pertinent" and "immaterial" information. In this study, a new image fusion method is proposed utilizing the Discrete Cosine Transform (DCT) to join the source image into a solitary minimized image containing more exact depiction of the sight than any of the individual source images. In addition, the fused image comes out with most ideal quality image without bending appearance or loss of data. DCT algorithm is considered efficient in image fusion. The proposed scheme is performed in five steps: (1) RGB colour image (input image) is split into three channels R, G, and B for source images. (2) DCT algorithm is applied to each channel (R, G, and B). (3) The variance values are computed for the corresponding 8 × 8 blocks of each channel. (4) Each block of R of source images is compared with each other based on the variance value and then the block with maximum variance value is selected to be the block in the new image. This process is repeated for all channels of source images. (5) Inverse discrete cosine transform is applied on each fused channel to convert coefficient values to pixel values, and then combined all the channels to generate the fused image. The proposed technique can potentially solve the problem of unwanted side effects such as blurring or blocking artifacts by reducing the quality of the subsequent image in image fusion process. The proposed approach is evaluated using three measurement units: the average of Q(abf), standard deviation, and peak Signal Noise Rate. The experimental results of this proposed technique have shown good results as compared with older techniques.
The computer vision (CV) is an emerging area with sundry promises. This communication encompasses the past development, recent trends and future directions of the CV in the context of deep learning (DL) algorithms-based object detections and localizations techniques. To identify the object location inside an image and recognize it by a computer program as fast as the human brain the machine learning and DL techniques have been evolved. However, the main limitations of the machine are related to the prolonged time consumption to handle vast amount of data to perform the same task as the human brain. To overcome these shortcomings, the convolution neural networks (NNs)-based deep NN has been developed, which detects and classifies the object with high precision. To train the deep NNs, massive amount of data (in the form of images and videos) and time is needed, making the computational cost of the CV very high. Thus, transfer learning techniques have been proposed wherein a model trained on one task can be reused on another linked task, thereby producing excellent outcomes. In this spirit, diverse DL-based algorithms have been introduced to detect and classify the object. These algorithms include the region-based convolutional NN (R-CNN), fast R-CNN, Faster R-CNN, mask E-CNN and You Only Look Once. A comparative evaluation among these techniques has been made to reveal their merits and demerits in the CV.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.