This paper considers blind methods for centrality estimation from graph signals. We model graph signals as the outcome of an unknown low-pass graph filter excited with influences governed by a sparse sub-graph. This model is compatible with a number of data generation process on graphs, including stock data and opinion dynamics. Based on the said graph signal model, we first prove that the folklore heuristics based on PCA of data covariance matrix may fail when the graph filter is not sufficiently low-pass. To remedy, we propose a robust blind centrality estimation method which substantially improves the centrality estimation performance. Numerical results on synthetic and real data support our findings.Index Termsnode centrality, graph signal processing, principal component analysis, robust PCA 1 The authors thank Ms. Xuan Zhang (CUHK) for collecting the dataset.
This paper is concerned with the blind identification of graph filters from graph signals. Our aim is to determine if the graph filter generating the graph signals is first-order lowpass without knowing the graph topology. Notice that lowpass graph filter is a common prerequisite for applying graph signal processing tools for sampling, denoising, and graph learning. Our method is inspired by the Perron Frobenius theorem, which observes that for first-order lowpass graph filter, the top eigenvector of output covariance would be the only eigenvector with elements of the same sign. Utilizing this observation, we develop a simple detector that answers if a given data set is produced by a first-order lowpass graph filter. We analyze the effects of finite-sample, graph size, observation noise, strength of lowpass filter, on the detector's performance. Numerical experiments on synthetic and real data support our findings.
Multi-label learning is often applied to handle complex decision tasks, and feature selection is its essential part. The relation of labels is always ignored or not enough to consider for both multi-label learning and its feature selection. To deal with the problem, F-neighborhood rough sets are employed. Different from other methods, the original approximate space is not changed, but the relation of labels is sufficient to consider. To be specific, a multi-label decision system is discomposed into a family of singlelabel decision tables with the label set(first-order strategy) at first. Secondly, calculate attribute significance in the family of single-label decision tables. Third, construct an attribute significance matrix and improved attribute significance matrices to evaluate the quality of the features, then a parallel reduct is obtained with information fusion. These processes construct F-neighborhood parallel reduction algorithm for a multi-label decision system(FNPRMS). Compared with the state-of-the-arts, experimental results show that FNPRMS is effective and efficient on 9 publicly available data sets.
Siamese network-based trackers consider tracking as features cross-correlation between the target template and the search region. Therefore, feature representation plays an important role for constructing a high-performance tracker. However, all existing Siamese networks extract the deep but low-resolution features of the entire patch, which is not robust enough to estimate the target bounding box accurately. In this work, to address this issue, we propose a novel high-resolution Siamese network, which connects the high-to-low resolution convolution streams in parallel as well as repeatedly exchanges the information across resolutions to maintain high-resolution representations. The resulting representation is semantically richer and spatially more precise by a simple yet effective multi-scale feature fusion strategy. Moreover, we exploit attention mechanisms to learn object-aware masks for adaptive feature refinement, and use deformable convolution to handle complex geometric transformations. This makes the target more discriminative against distractors and background. Without bells and whistles, extensive experiments on popular tracking benchmarks containing OTB100, UAV123, VOT2018 and LaSOT demonstrate that the proposed tracker achieves state-of-the-art performance and runs in real time, confirming its efficiency and effectiveness.
This paper considers the PageRank centrality computation problem on large graphs. We study asynchronous and distributed algorithms which are operated by aggregating information from local neighbors iteratively. Unlike prior works which rely on stochastic gradient descent (SGD) applied on a least square objective, we derive a stochastic approximation (SA) scheme for solving the PageRank problem by discretizing a linear system of ordinary differential equations. Our approach results in a family of asynchronous and distributed algorithms applicable for fixed and random topologies. Convergence rates are analyzed for both settings. In the fixed topology setting, we prove that the SA-based PageRank algorithm converges faster than the prior SGD-based method for large graphs. Numerical experiments support our findings.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.