People can make use of credit card for online transactions as it provides efficient and easy-to-use facility. With the increase in usage of credit cards, the capacity of credit card misuse has also enhanced. Credit card frauds cause significant financial losses for both credit card holders and financial companies. In this research study, the main aim is to detect such frauds which include the accessibility of public data, high-class imbalance data and fraud nature can be changed and the false alarm is in high rates. The relevant literature presents a number of machines learning based approaches for credit card detection.Such as Extreme Learning Method, Decision Tree, Random Forest, Support Vector Machine, Logistic Regression and XG Boost. But due to low accuracy, there is still need to apply state of the art deep learning algorithms to reduce the fraud losses.The main focus has been to apply the recent development of deep learning algorithms for this purpose. Comparative analysis of both machine learning and deep learning algorithms was performed to find efficient outcomes. The detail empirical analysis is carried out using European card benchmark dataset for fraud detection. Machine learning algorithm is first applied on the data set which showed improvement in the accuracy of detection of the frauds to some extent. Later, three architectures based on convolutional neural network are applied to improve fraud detection performance. Further addition of layers further increased the accuracy of detection. A comprehensive empirical analysis has been carried out by applying variations in number of hidden layers, epochs and applying the latest models. The evaluation of research work shows the improved results achieved such as accuracy, f1-score, precision and AUC Curves having optimized values 99.9%,85.71%,93%,98% respectively. The purposed model outperforms over state of art machine learning and deep learning algorithms for credit card detection problems.In addition, we have performed experiments by balancing the data and applying deep learning algorithms to minimize the false negative rate. The proposed approaches can be implemented effectively for the real-world detection of credit card frauds.
Arabic is one of the official languages recognized by the United Nations (UN) and is widely used in the middle east, and parts of Asia, Africa, and other countries. Social media activity currently dominates the textual communication on the Internet and potentially represents people’s views about specific issues. Opinion mining is an important task for understanding public opinion polarity towards an issue. Understanding public opinion leads to better decisions in many fields, such as public services and business. Language background plays a vital role in understanding opinion polarity. Variation is not only due to the vocabulary but also cultural background. The sentence is a time series signal; therefore, sequence gives a significant correlation to the meaning of the text. A recurrent neural network (RNN) is a variant of deep learning where the sequence is considered. Long short-term memory (LSTM) is an implementation of RNN with a particular gate to keep or ignore specific word signals during a sequence of inputs. Text is unstructured data, and it cannot be processed further by a machine unless an algorithm transforms the representation into a readable machine learning format as a vector of numerical values. Transformation algorithms range from the Term Frequency–Inverse Document Frequency (TF-IDF) transform to advanced word embedding. Word embedding methods include GloVe, word2vec, BERT, and fastText. This research experimented with those algorithms to perform vector transformation of the Arabic text dataset. This study implements and compares the GloVe and fastText word embedding algorithms and long short-term memory (LSTM) implemented in single-, double-, and triple-layer architectures. Finally, this research compares their accuracy for opinion mining on an Arabic dataset. It evaluates the proposed algorithm with the ASAD dataset of 55,000 annotated tweets in three classes. The dataset was augmented to achieve equal proportions of positive, negative, and neutral classes. According to the evaluation results, the triple-layer LSTM with fastText word embedding achieved the best testing accuracy, at 90.9%, surpassing all other experimental scenarios.
Machine intelligence models are robust in classifying the datasets for data analytics and for predicting the insights that would assist in making clinical decisions. The models would assist in the disease prognosis and preliminary disease investigation, which is crucial for effective treatment. There is a massive demand for the interpretability and explainability of decision models in the present day. The models’ trustworthiness can be attained through deploying the ensemble classification models in the eXplainable Artificial Intelligence (XAI) framework. In the current study, the role of ensemble classifiers over the XAI framework for predicting heart disease from the cardiovascular datasets is carried out. There are 303 instances and 14 attributes in the cardiovascular dataset taken for the proposed work. The attribute characteristics in the dataset are categorical, integer, and real type and the associated task related to the dataset is classification. The classification techniques, such as the support vector machine (SVM), AdaBoost, K-nearest neighbor (KNN), bagging, logistic regression (LR), and naive Bayes, are considered for classification purposes. The experimental outcome of each of those algorithms is compared to each other and with the conventional way of implementing the classification models. The efficiency of the XAI-based classification models is reasonably fair, compared to the other state-of-the-art models, which are assessed using the various evaluation metrics, such as area under curve (AUC), receiver operating characteristic (ROC), sensitivity, specificity, and the F1-score. The performances of the XAI-driven SVM, LR, and naive Bayes are robust, with an accuracy of 89%, which is assumed to be reasonably fair, compared to the existing models.
The healthcare sector is rapidly being transformed to one that operates in new computing environments. With researchers increasingly committed to finding and expanding healthcare solutions to include the Internet of Things (IoT) and edge computing, there is a need to monitor more closely than ever the data being collected, shared, processed, and stored. The advent of cloud, IoT, and edge computing paradigms poses huge risks towards the privacy of data, especially, in the healthcare environment. However, there is a lack of comprehensive research focused on seeking efficient and effective solutions that ensure data privacy in the healthcare domain. The data being collected and processed by healthcare applications is sensitive, and its manipulation by malicious actors can have catastrophic repercussions. This paper discusses the current landscape of privacy-preservation solutions in IoT and edge healthcare applications. It describes the common techniques adopted by researchers to integrate privacy in their healthcare solutions. Furthermore, the paper discusses the limitations of these solutions in terms of their technical complexity, effectiveness, and sustainability. The paper closes with a summary and discussion of the challenges of safeguarding privacy in IoT and edge healthcare solutions which need to be resolved for future applications.
Over the past few years, a tremendous change has occurred in computer-aided diagnosis (CAD) technology. The evolution of numerous medical imaging techniques has enhanced the accuracy of the preliminary analysis of several diseases. Magnetic resonance imaging (MRI) is a prevalent technology extensively used in evaluating the progress of the spread of malignant tissues or abnormalities in the human body. This article aims to automate a computationally efficient mechanism that can accurately identify the tumor from MRI images and can analyze the impact of the tumor. The proposed model is robust enough to classify the tumors with minimal training data. The generative variational autoencoder models are efficient in reconstructing the images identical to the original images, which are used in adequately training the model. The proposed self-learning algorithm can learn from the insights from the autogenerated images and the original images. Incorporating long short-term memory (LSTM) is faster processing of the high dimensional imaging data, making the radiologist’s task and the practitioners more comfortable assessing the tumor’s progress. Self-learning models need comparatively less data for the training, and the models are more resource efficient than the various state-of-art models. The efficiency of the proposed model has been assessed using various benchmark metrics, and the obtained results have exhibited an accuracy of 89.7%. The analysis of the progress of tumor growth is presented in the current study. The obtained accuracy is not pleasing in the healthcare domain, yet the model is reasonably fair in dealing with a smaller size dataset by making use of an image generation mechanism. The study would outline the role of an autoencoder in self-learning models. Future technologies may include sturdy feature engineering models and optimized activation functions that would yield a better result.
Sentiment analysis has been one of the most active research areas in the past decade due to its vast applications. Sentiment quantification, a new research problem in this field, extends sentiment analysis from individual documents to an aggregated collection of documents. Sentiment analysis has been widely researched, but sentiment quantification has drawn less attention despite offering a greater potential to enhance current business intelligence systems. In this research, to perform sentiment quantification, a framework based on feature engineering is proposed to exploit diverse feature sets such as sentiment, content, and part of speech, as well as deep features including word2vec and GloVe. Different machine learning algorithms, including conventional, ensemble learners, and deep learning approaches, have been investigated on standard datasets of SemEval2016, SemEval2017, STS-Gold, and Sanders. The empirical-based results reveal the effectiveness of the proposed feature sets in the process of sentiment quantification when applied to machine learning algorithms. The results also reveal that the ensemble-based algorithm AdaBoost outperforms other conventional machine learning algorithms using a combination of proposed feature sets. The deep learning algorithm RNN, on the other hand, shows optimal results using word embedding-based features. This research has the potential to help diverse applications of sentiment quantification, including polling, trend analysis, automatic summarization, and rumor or fake news detection.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.