Vehicle re-identification is key to keep track of vehicles monitored by a multicamera network with non-overlapping views. In this paper, we propose a probabilistic framework based on a two-step strategy that re-identifies vehicles in road tunnels. The first step consists of splitting the re-identification problem by connecting groups of vehicles observed in different cameras using certain motion and appearance criteria. In the second step, we build a Bayesian model that finds the optimal assignment between vehicles of connected groups. Descriptors like trace transform signatures, lane change, and motion discrepancies are used to derive our probabilistic framework. Experimental tests reveal that connected groups derived from the first step are composed of 4 vehicles on average. This allow us to constrain the number of candidate matches and increase the chances of getting the correct match. In the second step, our Bayesian model succeeds in matching vehicles among candidates with very similar appearance and under uneven illumination conditions. In general, our system reports a reidentification accuracy of 92% using a nearest-neighbor matcher, and 98% using a one-to-one matcher. These results outperform previous works and encourage us to further develop our solution for other re-identification applications.
In this paper we present an efficient way to both compute and extract salient information from trace transform signatures to perform object identification tasks. We also present a feature selection analysis of the classical trace-transform functionals, which reveals that most of them retrieve redundant information causing misleading similarity measurements. In order to overcome this problem, we propose a set of functionals based on Laguerre polynomials that return orthonormal signatures between these functionals. In this way, each signature provides salient and non-correlated information that contributes to the description of an image object. The proposed functionals were tested considering a vehicle identification problem, outperforming the classical trace transform functionals in terms of computational complexity and identification rate.
This paper proposes a generic methodology for the semi-automatic generation of reliable position annotations for evaluating multi-camera people-trackers on large video data sets. Most of the annotation data are automatically computed, by estimating a consensus tracking result from multiple existing trackers and people detectors and classifying it as either reliable or not. A small subset of the data, composed of tracks with insufficient reliability, is verified by a human using a simple binary decision task, a process faster than marking the correct person position. The proposed framework is generic and can handle additional trackers. We present results on a data set of $sim 6$ h captured by 4 cameras, featuring a person in a holiday flat, performing activities such as walking, cooking, eating, cleaning, and watching TV. When aiming for a tracking accuracy of 60 cm, 80% of all video frames are automatically annotated. The annotations for the remaining 20% of the frames were added after human verification of an automatically selected subset of data. This involved $sim 2.4$ h of manual labor. According to a subsequent comprehensive visual inspection to judge the annotation procedure, we found 99% of the automatically annotated frames to be correct. We provide guidelines on how to apply the proposed methodology to new data sets. We also provide an exploratory study for the multi-target case, applied on the existing and new benchmark video sequences.
Tracking multiple vehicles with multiple cameras is a challenging problem of great importance in tunnel surveillance. One of the main challenges is accurate vehicle matching across the cameras with non-overlapping fields of view. Since systems dedicated to this task can contain hundreds of cameras which observe dozens of vehicles each, for a real-time performance computational efficiency is essential. In this paper, we propose a low complexity, yet highly accurate method for vehicle matching using vehicle signatures composed of Radon transform like projection profiles of the vehicle image. The proposed signatures can be calculated by a simple scan-line algorithm, by the camera software itself and transmitted to the central server or to the other cameras in a smart camera environment. The amount of data is drastically reduced compared to the whole image, which relaxes the data link capacity requirements. Experiments on real vehicle images, extracted from video sequences recorded in a tunnel by two distant security cameras, validate our approach.
A novel approach to automatically detect vehicles in road tunnels is presented in this paper. Non-uniform and poor illumination conditions prevail in road tunnels making difficult to achieve robust vehicle detection. In order to cope with the illumination issues, we propose a local higher-order statistic filter to make the vehicle detection invariant to illumination changes, whereas a morphological-based background subtraction is used to generate a convex hull segmentation of the vehicles. An evaluation test comparing our approach with a benchmark object detector shows that our approach outperforms in terms of false detection rate and overlap area detection.
In this paper, a new method to compute the image\ud
histogram is presented, along with the image maximum\ud
and minimum values. It is intended for highly paral-\ud
lel architectures such as the ones found in Focal Plane\ud
Processors (FPP). This new approach exploits this par-\ud
allelism relying on the privatization technique to avoid\ud
the memory collision problem, while the bin frequency\ud
is obtained through image bitwise manipulation. Unlike\ud
traditional privatization techniques, our method exhibits\ud
a trade-o between processing time and bin size. That\ud
is, it can be adapted as a power-of-two bin size histogram\ud
and the computation time decreases exponentially as the\ud
bin size is reduced on each power of two, allowing high\ud
computational \ud
exibility.Postprint (published version
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.