Infectious diseases are highly contagious due to rapid transmission and very challenging to diagnose in the early stage. Artificial Intelligence and Machine Learning now become a strategic weapon in assisting infectious disease prevention, rapid-response in diagnosis, surveillance, and management. In this paper, a bifold COVID_SCREENET architecture is introduced for providing COVID-19 screening solutions using Chest Radiography (CR) images. Transfer learning using nine pre-trained ImageNet models to extract the features of Normal, Pneumonia, and COVID-19 images is adapted in the first fold and classified using baseline Convolutional Neural Network (CNN). A Modified Stacked Ensemble Learning (MSEL) is proposed in the second fold by stacking the top five pre-trained models, and then the predictions resulted. Experimentation is carried out in two folds: In first fold, open-source samples are considered and in second fold 2216 real-time samples collected from Tamilnadu Government Hospitals, India, and the screening results for COVID data is 100% accurate in both the cases. The proposed approach is also validated and blind reviewed with the help of two radiologists at Thanjavur Medical College & Hospitals by collecting 2216 chest X-ray images between the month of April and May. Based on the reports, the measures are calculated for COVID_SCREENET and it showed 100% accuracy in performing multi-class classification.
The recent developments in deep learning techniques evolved to new heights in various domains and applications. The recognition, translation, and video generation of Sign Language (SL) still face huge challenges from the development perspective. Although numerous advancements have been made in earlier approaches, the model performance still lacks recognition accuracy and visual quality. In this paper, we introduce novel approaches for developing the complete framework for handling SL recognition, translation, and production tasks in real-time cases. To achieve higher recognition accuracy, we use the MediaPipe library and a hybrid Convolutional Neural Network + Bi-directional Long Short Term Memory (CNN + Bi-LSTM) model for pose details extraction and text generation. On the other hand, the production of sign gesture videos for given spoken sentences is implemented using a hybrid Neural Machine Translation (NMT) + MediaPipe + Dynamic Generative Adversarial Network (GAN) model. The proposed model addresses the various complexities present in the existing approaches and achieves above 95% classification accuracy. In addition to that, the model performance is tested in various phases of development, and the evaluation metrics show noticeable improvements in our model. The model has been experimented with using different multilingual benchmark sign corpus and produces greater results in terms of recognition accuracy and visual quality. The proposed model has secured a 38.06 average Bilingual Evaluation Understudy (BLEU) score, remarkable human evaluation scores, 3.46 average Fréchet Inception Distance to videos (FID2vid) score, 0.921 average Structural Similarity Index Measure (SSIM) values, 8.4 average Inception Score, 29.73 average Peak Signal-to-Noise Ratio (PSNR) score, 14.06 average Fréchet Inception Distance (FID) score, and an average 0.715 Temporal Consistency Metric (TCM) Score which is evidence of the proposed work.
The Editor-in-Chief and the publisher have retracted this article. This article was submitted to be part of a guestedited issue. An investigation concluded that the editorial process of this guest-edited issue was compromised by a third party and that the peer review process has been manipulated. Based on the investigation's findings the Editor-in-Chief therefore no longer has confidence in the results and conclusions of this article.The author disagrees with this retraction.Publisher's Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The speech and hearing-impaired community use sign language as the primary means of communication. It is quite challenging for the general population to interpret or learn sign language completely. A sign language recognition system must be designed and developed to address this communication barrier. Most current sign language recognition systems rely on wearable sensors, keeping the recognition system unaffordable for most individuals. Moreover, the existing vision-based sign recognition frameworks do not consider all of the spatial and temporal information required for accurate recognition. A novel vison-based hybrid deep neural net methodology is proposed in this study for recognizing Indian and Russian sign gestures. The proposed framework is aimed to establish a single framework for tracking and extracting multisemantic properties, such as non-manual components and manual co-articulations. Furthermore, spatial feature extraction from the sign gestures is deployed using a 3D deep neural net with atrous convolutions. The temporal and sequential feature extraction is carried out by employing attention-based Bi-LSTM. In addition, the distinguished abstract feature extraction is done using the modified autoencoders. The discriminative feature extraction for differentiating the sign gestures from unwanted transition gestures is done by leveraging the hybrid attention module. The experimentation of the proposed model has been carried out on the novel multi-signer Indo-Russian sign language dataset. The proposed sign language recognition framework with hybrid neural net yields better results than other state-of-the-art frameworks.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.