We propose a method to identify and localize object classes in images. Instead of operating at the pixel level, we advocate the use of superpixels as the basic unit of a class segmentation or pixel localization scheme. To this end, we construct a classifier on the histogram of local features found in each superpixel. We regularize this classifier by aggregating histograms in the neighborhood of each superpixel and then refine our results further by using the classifier in a conditional random field operating on the superpixel graph. Our proposed method exceeds the previously published state-of-the-art on two challenging datasets: Graz-02 and the PASCAL VOC 2007 Segmentation Challenge.
Abstract. We present an approach to determine the category and location of objects in images. It performs very fast categorization of each pixel in an image, a brute-force approach made feasible by three key developments: First, our method reduces the size of a large generic dictionary (on the order of ten thousand words) to the low hundreds while increasing classification performance compared to k-means. This is achieved by creating a discriminative dictionary tailored to the task by following the information bottleneck principle. Second, we perform feature-based categorization efficiently on a dense grid by extending the concept of integral images to the computation of local histograms. Third, we compute SIFT descriptors densely in linear time. We compare our method to the state of the art and find that it excels in accuracy and simplicity, performing better while assuming less.
VLFeat is an open and portable library of computer vision algorithms. It aims at facilitating fast prototyping and reproducible research for computer vision scientists and students. It includes rigorous implementations of common building blocks such as feature detectors, feature extractors, (hierarchical) k-means clustering, randomized kd-tree matching, and super-pixelization. The source code and interfaces are fully documented. The library integrates directly with Matlab, a popular language for computer vision research.
Inside the mammalian nose lies a labyrinth of bony plates covered in epithelium collectively known as turbinates. Respiratory turbinates lie anteriorly and aid in heat and water conservation, while more posterior olfactory turbinates function in olfaction. Previous observations on a few carnivorans revealed that aquatic species have relatively large, complex respiratory turbinates and greatly reduced olfactory turbinates compared with terrestrial species. Body heat is lost more quickly in water than air and increased respiratory surface area likely evolved to minimize heat loss. At the same time, olfactory surface area probably diminished due to a decreased reliance on olfaction when foraging under water. To explore how widespread these adaptations are, we documented scaling of respiratory and olfactory turbinate surface area with body size in a variety of terrestrial, freshwater, and marine carnivorans, including pinnipeds, mustelids, ursids, and procyonids. Surface areas were estimated from high-resolution CT scans of dry skulls, a novel approach that enabled a greater sampling of taxa than is practical with fresh heads. Total turbinate, respiratory, and olfactory surface areas correlate well with body size (r 2 ‡ 0.7), and are relatively smaller in larger species. Relative to body mass or skull length, aquatic species have significantly less olfactory surface area than terrestrial species. Furthermore, the ratio of olfactory to respiratory surface area is associated with habitat. Using phylogenetic comparative methods, we found strong support for convergence on 1 : 3 proportions in aquatic taxa and near the inverse in terrestrial taxa, indicating that aquatic mustelids and pinnipeds independently acquired similar proportions of olfactory to respiratory turbinates. Constraints on turbinate surface area in the nasal chamber may result in a trade-off between respiratory and olfactory function in aquatic mammals.
No abstract
Abstract. The paper presents an exact GPU implementation of the quick shift image segmentation algorithm. Variants of the implementation which use global memory and texture caching are presented, and the paper shows that a method backed by texture caching can produce a 10-50X speedup for practical images, making computation of super-pixels possible at 5-10Hz on modest sized (256x256) images.
We consider the problem of relative pose regression in visual relocalization. Recently, several promising approaches have emerged in this area. We claim that even though they demonstrate on the same datasets using the same split to train and test, a faithful comparison between them was not available since on currently used evaluation metric, some approaches might perform favorably, while in reality performing worse. We reveal a tradeoff between accuracy and the 3D volume of the regressed subspace. We believe that unlike other relocalization approaches, in the case of relative pose regression, the regressed subspace 3D volume is less dependent on the scene and more affect by the method used to score the overlap, which determined how closely sampled viewpoints are. We propose three new metrics to remedy the issue mentioned above. The proposed metrics incorporate statistics about the regression subspace volume. We also propose a new pose regression network that serves as a new baseline for this task. We compare the performance of our trained model on Microsoft 7-Scenes and Cambridge Landmarks datasets both with the standard metrics and the newly proposed metrics and adjust the overlap score to reveal the tradeoff between the subspace and performance. The results show that the proposed metrics are more robust to different overlap threshold than the conventional approaches. Finally, we show that our network generalizes well, specifically, training on a single scene leads to little loss of performance on the other scenes.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.