Abstract-In this paper, we present a novel automatic and unsupervised change-detection approach specifically oriented to the analysis of multitemporal single-channel single-polarization synthetic aperture radar (SAR) images. This approach is based on a closed-loop process made up of three main steps: 1) a novel preprocessing based on a controlled adaptive iterative filtering; 2) a comparison between multitemporal images carried out according to a standard log-ratio operator; and 3) a novel approach to the automatic analysis of the log-ratio image for generating the change-detection map. The first step aims at reducing the speckle noise in a controlled way in order to maximize the discrimination capability between changed and unchanged classes. In the second step, the two filtered multitemporal images are compared to generate a log-ratio image that contains explicit information on changed areas. The third step produces the change-detection map according to a thresholding procedure based on a reformulation of the Kittler-Illingworth (KI) threshold selection criterion. In particular, the modified KI criterion is derived under the generalized Gaussian assumption for modeling the distributions of changed and unchanged classes. This parametric model was chosen because it is capable of better fitting the conditional densities of classes in the log-ratio image. In order to control the filtering step and, accordingly, the effects of the filtering process on change-detection accuracy, we propose to identify automatically the optimal number of despeckling filter iterations [Step 1)] by analyzing the behavior of the modified KI criterion. This results in a completely automatic and self-consistent change-detection approach that avoids the use of empirical methods for the selection of the best number of filtering iterations. Experiments carried out on two sets of multitemporal images (characterized by different levels of speckle noise) acquired by the European Remote Sensing 2 satellite SAR sensor confirm the effectiveness of the proposed unsupervised approach, which results in change-detection accuracies very similar to those that can be achieved by a manual supervised thresholding.
The aim of this paper is twofold. First, we present a thorough experimental study to show the superiority of the generalization capability of the support vector machine (SVM) approach in the automatic classification of electrocardiogram (ECG) beats. Second, we propose a novel classification system based on particle swarm optimization (PSO) to improve the generalization performance of the SVM classifier. For this purpose, we have optimized the SVM classifier design by searching for the best value of the parameters that tune its discriminant function, and upstream by looking for the best subset of features that feed the classifier. The experiments were conducted on the basis of ECG data from the Massachusetts Institute of Technology-Beth Israel Hospital (MIT-BIH) arrhythmia database to classify five kinds of abnormal waveforms and normal beats. In particular, they were organized so as to test the sensitivity of the SVM classifier and that of two reference classifiers used for comparison, i.e., the k-nearest neighbor (kNN) classifier and the radial basis function (RBF) neural network classifier, with respect to the curse of dimensionality and the number of available training beats. The obtained results clearly confirm the superiority of the SVM approach as compared to traditional classifiers, and suggest that further substantial improvements in terms of classification accuracy can be achieved by the proposed PSO-SVM classification system. On an average, over three experiments making use of a different total number of training beats (250, 500, and 750, respectively), the PSO-SVM yielded an overall accuracy of 89.72% on 40438 test beats selected from 20 patient records against 85.98%, 83.70%, and 82.34% for the SVM, the kNN, and the RBF classifiers, respectively.
In this paper, we propose a remote-sensing scene-classification method based on vision transformers. These types of networks, which are now recognized as state-of-the-art models in natural language processing, do not rely on convolution layers as in standard convolutional neural networks (CNNs). Instead, they use multihead attention mechanisms as the main building block to derive long-range contextual relation between pixels in images. In a first step, the images under analysis are divided into patches, then converted to sequence by flattening and embedding. To keep information about the position, embedding position is added to these patches. Then, the resulting sequence is fed to several multihead attention layers for generating the final representation. At the classification stage, the first token sequence is fed to a softmax classification layer. To boost the classification performance, we explore several data augmentation strategies to generate additional data for training. Moreover, we show experimentally that we can compress the network by pruning half of the layers while keeping competing classification accuracies. Experimental results conducted on different remote-sensing image datasets demonstrate the promising capability of the model compared to state-of-the-art methods. Specifically, Vision Transformer obtains an average classification accuracy of 98.49%, 95.86%, 95.56% and 93.83% on Merced, AID, Optimal31 and NWPU datasets, respectively. While the compressed version obtained by removing half of the multihead attention layers yields 97.90%, 94.27%, 95.30% and 93.05%, respectively.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.