Face recognition is one of a complex biometrics in the field of pattern recognition due to the constraints imposed by variation in the appearance of facial images. These changes in appearance are affected by variation in illumination, expression or occlusions etc. Illumination can be considered a complex problem in both indoor and outdoor pattern matching. Literature studies have revealed that two problems of textural based illumination handling in face recognition seem to be very common. Firstly, textural values are changed during illumination normalization due to increase in the contrast that changes the original pixels of face. Secondly, it minimizes the distance between interclasses which increases the false acceptance rates. This paper addresses these issues and proposes a robust algorithm that overcomes these limitations. The limitations are resolved through transforming pixels from nonillumination side to illuminated side. It has been revealed that proposed algorithm produced better results as compared to existing related algorithms.
The online retrieval of images related to clothes is a crucial task because finding the exact items like the query image from a large amount of data is extremely challenging. However, large variations in clothes images degrade the retrieval accuracy of visual searches. Another problem with retrieval accuracy is high dimensions of feature vectors obtained from pre-trained deep CNN models. This research is an effort to enhance the training and test accuracy of clothes retrieval by using two different means. Initially, features are extracted using the modified AlexNet (M-AlexNet) with little modification in which ReLU activation function is replaced with a self-regularized Mish activation function because of its non-monotonic nature. The M-AlexNet with Mish is trained on CIFAR-10 dataset using SoftMax classifier. Another contribution is to reduce the dimensions of feature vectors obtained from M-AlexNet. The dimensions of features are reduced by selecting the top k ranked features and removing some of the dissimilar features using the proposed Joint Shannon's Entropy Pearson Correlation Coefficient (JSE-PCC) technique to enhance the clothes retrieval performance. To calculate the efficacy of suggested methods, the comparison is performed with other deep CNN models such as baseline AlexNet, VGG-16, VGG-19, and ResNet50 on DeepFashion2, MVC, and the proposed Clothes Image Dataset (CID). Extensive experiments indicate that AlexNet with Mish attains 85.15%, 82.04%, and 83.65% accuracy on DeepFashion2, MVC, and 83.65% on CID datasets respectively. Hence, M-AlexNet and the proposed feature selection technique surpassed the results with a margin of 5.11% on DeepFashion2, 1.95% on MVC, and 3.51% CID datasets.
Visual analysis of fashion images gain much attention in the fashion industry due to its commercial and social importance. In recent years, deep learning techniques offer overwhelming progress in improving the accuracy of fine‐grained apparel segmentation with accurate bounding box prediction. The baseline pixel‐based masking techniques show excellent performance in object detection and segmentation but sometimes ignores the boundary of objects, resulting in uneven and complicated segmentation masks. Moreover, it is time taking to generate a multi‐scale feature map against each anchor box. To remedy this problem, a more accurate, faster, and suitable deep learning architecture is proposed that accurately detects, classify, and performs fine‐grained segmentation of cloth products in a single platform. In this paper, initially, an Object Class Head Detector model is proposed in which the baseline Mask‐RCNN model is used as a reference model. Here, we replace the Region Proposal Network with the proposed modified YoloV2 model to locate apparel products with its class prediction. The modified YoloV2 model has more capability to detect tiny objects because of local and high‐level feature fusion. The goal of this step is to accurately locate the objects in minimum time intervals. Furthermore, the predicted bounding box is converted to object shape offsets using deep snake architecture that tightly fits onto the apparel shape. It can improve the accuracy of cloth shape segmentation by preserving object contours. The proposed architecture is empirically validated on various existing fashion image datasets. The experimental results illustrate that the proposed architecture performs better on the Deepfashion2 dataset with mAP of 86.86%, as compared to other state‐of‐the‐art deep learning models.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.