The automatic detection of objects within complex volumetric imagery is becoming of increased interest due to the use of dual energy Computed Tomography (CT) scanners as an aviation security deterrent. These devices produce a volumetric image akin to that encountered in prior medical CT work but in this case we are dealing with a complex multi-object volumetric environment including significant noise artefacts. In this work we look at the application of the recent extension to the seminal SIFT approach to the 3D volumetric recognition of rigid objects within this complex volumetric environment. A detailed overview of the approach and results when applied to a set of exemplar CT volumetric imagery is presented.
We present a comparison of 3D feature descriptors with application to threat detection in Computed Tomography (CT) airport baggage imagery. The detectors range in complexity from a basic local density descriptor, through local region histograms and 3D extensions to both to the RIFT descriptor and the seminal SIFT feature descriptor. We show that, in the complex CT imagery domain containing a high degree of noise and imaging artefacts, an object recognition system using simpler descriptors appears to outperform a more complex RIFT/SIFT solution. Recognition rates in excess of 95% are demonstrated with minimal false positive rates for a set of exemplar 3D objects.
We introduce a novel 3D extension to the hierarchical visual cortex model used for prior work in 2D object recognition. Prior work on the use of the visual cortex standard model for the explicit task of object class recognition has solely concentrated on 2D imagery. In this paper we discuss the explicit 3D extension of each layer in this visual cortex model hierarchy for use in object recognition in 3D volumetric imagery. We apply this extended methodology to the automatic detection of a class of threat items in Computed Tomography (CT) security baggage imagery. The CT imagery suffers from poor resolution and a large number of artefacts generated through the presence of metallic objects. In our examination of recognition performance we make a comparison to a codebook approach derived from a 3D SIFT descriptor and demonstrate that the visual cortex method out-performs in this imagery. Recognition rates in excess of 95% with minimal false positive rates are demonstrated in the detection of a range of threat items.
Use policyThe full-text may be used and/or reproduced, and given to third parties in any format or medium, without prior permission or charge, for personal research or study, educational, or not-for-pro t purposes provided that:• a full bibliographic reference is made to the original source • a link is made to the metadata record in DRO • the full-text is not changed in any way The full-text must not be sold in any format or medium without the formal permission of the copyright holders.Please consult the full DRO policy for further details.
AbstractWe investigate the performance of a Bag of (Visual) Words ( Optimal performance is achieved using the DH and DGH descriptors in conjunction with an uncertainty assignment methodology. Successful detection rates in excess of 97% for handguns and 89% for bottles and false-positive rates of approximately 2-3% are achieved. We demonstrate that the underlying imaging modality and the irrelevance of illumination and scale invariance within the transmission imagery context considered here, result in the favourable performance of simpler density histogram descriptors (DH, DGH) over 3D extensions of the well-established SIFT and RIFT feature descriptor approaches.
We investigate the feasibility of a codebook approach for the automated classification of threats in pre-segmented 3D baggage Computed Tomography (CT) security imagery. We compare the performance of five codebook models, using various combinations of sampling strategies, feature encoding techniques and classifiers, to the current state-of-the-art 3D visual cortex approach [1]. We demonstrate an improvement over the state-of-the-art both in terms of accuracy as well as processing time using a codebook constructed via randomised clustering forests [2], a dense feature sampling strategy and an SVM classifier. Correct classification rates in excess of 98% and false positive rates of less than 1%, in conjunction with a reduction of several orders of magnitude in processing time, make the proposed approach an attractive option for the automated classification of threats in security screening settings.
Abstract-In this paper, we describe a Threat Image Projection (TIP) method designed for 3D Computed Tomography (CT) screening systems. The novel methodology automatically determines a valid 3D location in the passenger 3D CT baggage image into which a fictional threat 3D image can be inserted without violating the bag content. According to the scan orientation, the passenger bag content and the material of the inserted threat appropriate CT artefacts are generated using a Radon transform in order to make the insertion realistic. Densely cluttered 3D CT baggage images are used to validate our method. Experimental results confirm that our method is able to reliably insert threat items in challenging 3D images without providing any perceptible visual cue to human screeners.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.