Hyperspectral images (HSIs) are often corrupted by a mixture of several types of noise during the acquisition process, e.g., Gaussian noise, impulse noise, dead lines, stripes, and many others. Such complex noise could degrade the quality of the acquired HSIs, limiting the precision of the subsequent processing. In this paper, we present a novel tensor-based HSI restoration approach by fully identifying the intrinsic structures of the clean HSI part and the mixed noise part respectively. Specifically, for the clean HSI part, we use tensor Tucker decomposition to describe the global correlation among all bands, and an anisotropic spatial-spectral total variation (SSTV) regularization to characterize the piecewise smooth structure in both spatial and spectral domains. For the mixed noise part, we adopt the 1 norm regularization to detect the sparse noise, including stripes, impulse noise, and dead pixels. Despite that TV regulariztion has the ability of removing Gaussian noise, the Frobenius norm term is further used to model heavy Gaussian noise for some real-world scenarios. Then, we develop an efficient algorithm for solving the resulting optimization problem by using the augmented Lagrange multiplier (ALM) method. Finally, extensive experiments on simulated and real-world noise HSIs are carried out to demonstrate the superiority of the proposed method over the existing state-of-the-art ones.Index Terms-Hyperspectral image (HSI), mixed noise, low rank tensor decomposition, total variation (TV).
Motivation Studies have shown that the accuracy of random forest (RF)-based scoring functions (SFs), such as RF-Score-v3, increases with more training samples, whereas that of classical SFs, such as X-Score, does not. Nevertheless, the impact of the similarity between training and test samples on this matter has not been studied in a systematic manner. It is therefore unclear how these SFs would perform when only trained on protein-ligand complexes that are highly dissimilar or highly similar to the test set. It is also unclear whether SFs based on machine learning algorithms other than RF can also improve accuracy with increasing training set size and to what extent they learn from dissimilar or similar training complexes. Results We present a systematic study to investigate how the accuracy of classical and machine-learning SFs varies with protein-ligand complex similarities between training and test sets. We considered three types of similarity metrics, based on the comparison of either protein structures, protein sequences or ligand structures. Regardless of the similarity metric, we found that incorporating a larger proportion of similar complexes to the training set did not make classical SFs more accurate. In contrast, RF-Score-v3 was able to outperform X-Score even when trained on just 32% of the most dissimilar complexes, showing that its superior performance owes considerably to learning from dissimilar training complexes to those in the test set. In addition, we generated the first SF employing Extreme Gradient Boosting (XGBoost), XGB-Score, and observed that it also improves with training set size while outperforming the rest of SFs. Given the continuous growth of training datasets, the development of machine-learning SFs has become very appealing. Availability and implementation https://github.com/HongjianLi/MLSF Supplementary information Supplementary data are available at Bioinformatics online.
It has recently been claimed that the outstanding performance of machine-learning scoring functions (SFs) is exclusively due to the presence of training complexes with highly similar proteins to those in the test set. Here, we revisit this question using 24 similarity-based training sets, a widely used test set, and four SFs. Three of these SFs employ machine learning instead of the classical linear regression approach of the fourth SF (X-Score which has the best test set performance out of 16 classical SFs). We have found that random forest (RF)-based RF-Score-v3 outperforms X-Score even when 68% of the most similar proteins are removed from the training set. In addition, unlike X-Score, RF-Score-v3 is able to keep learning with an increasing training set size, becoming substantially more predictive than X-Score when the full 1105 complexes are used for training. These results show that machine-learning SFs owe a substantial part of their performance to training on complexes with dissimilar proteins to those in the test set, against what has been previously concluded using the same data. Given that a growing amount of structural and interaction data will be available from academic and industrial sources, this performance gap between machine-learning SFs and classical SFs is expected to enlarge in the future.
Cerebrovascular diseases, i.e., acute stroke, are a common cause of serious long-term disability. Cerebral perfusion computed tomography (CPCT) can provide rapid, highresolution, quantitative hemodynamic maps to assess and stratify perfusion in patients with acute stroke symptoms. However, CPCT imaging typically involves a substantial radiation dose due to its repeated scanning protocol. Therefore, in this work, we present a low-dose CPCT image reconstruction method to yield high-quality CPCT images and high-precision hemodynamic maps by utilizing the great similarity information among the repeated scanned CPCT images. Specifically, a newly developed low-rank tensor decomposition with spatial-temporal total variation (LRTD-STTV) regularization is incorporated into the reconstruction model. In the LRTD-STTV regularization, the tensor Tucker decomposition is used to describe global spatialtemporal correlations hidden in the sequential CPCT images, and it is superior to the matricization model (i.e., low-rank model) that fails to fully investigate the prior knowledge of the intrinsic structures of the CPCT images after vectorizing the CPCT images. Moreover, the spatial-temporal TV regularization is used to characterize the local piecewise smooth structure in the spatial domain and the pixels' similarity with the adjacent frames in the temporal domain because the intensity at each pixel in CPCT images is similar to its neighbors. Therefore, the presented LRTD-STTV model can efficiently deliver faithful underlying information of the CPCT images and preserve the spatial structures. An efficient alternating direction method of multipliers algorithm is also developed to solve the presented LRTD-STTV model. Extensive experimental results on numerical phantom and patient data clearly demonstrated that the presented model can significantly improve the quality of CPCT images and provide accurate diagnostic features in hemodynamic maps for low-dose cases compared with the existing popular algorithms.
Perfusion computed tomography (PCT) is critical in detecting cerebral ischemic lesions. PCT examination with low-dose scans can effectively reduce radiation exposure to patients at the cost of degraded images with severe noise and artifacts. Tensor total variation (TTV) models are powerful tools that can encode the regional continuous structures underlying a PCT object. In a TTV model, the sparsity structures of the contrast-medium concentration (CMC) across PCT frames are assumed to be isotropic with identical and independent distribution. However, this assumption is inconsistent with practical PCT tasks wherein the sparsity has evident variations and correlations. Such modeling deviation hampers the performance of TTV-based PCT reconstructions. To address this issue, we developed a novel contrast-medium anisotropy-aware tensor total variation (CMAA-TTV) model to describe the intrinsic anisotropy sparsity of the CMC in PCT imaging tasks. Instead of directly on the difference matrices, the CMAA-TTV model characterizes sparsity on a low-rank subspace of the difference matrices which are calculated from the input data adaptively, thus naturally encoding the intrinsic variant and correlated anisotropy sparsity structures of the CMC. We further proposed a robust and efficient PCT reconstruction algorithm to improve low-dose PCT reconstruction performance using the CMAA-TTV model. Experimental studies using a digital brain perfusion phantom, patient data with low-dose simulation and clinical patient data were performed to validate the effectiveness of the presented algorithm. The results demonstrate that the CMAA-TTV algorithm can achieve noticeable improvements over state-of-the-art methods in low-dose PCT reconstruction tasks.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.