2015 IEEE International Conference on Computer Vision (ICCV) 2015
DOI: 10.1109/iccv.2015.376
|View full text |Cite
|
Sign up to set email alerts
|

RGB-W: When Vision Meets Wireless

Abstract: Inspired by the recent success of RGB-D cameras, we propose the enrichment of RGB data with an additional "quasi-free" modality, namely, the wireless signal emitted by individuals' cell phones, referred to as RGB-W. The received signal strength acts as a rough proxy for depth and a reliable cue on a person's identity. Although the measured signals are noisy, we demonstrate that the combination of visual and wireless data significantly improves the localization accuracy. We introduce a novel image-driven repres… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
37
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
4

Relationship

2
6

Authors

Journals

citations
Cited by 53 publications
(39 citation statements)
references
References 34 publications
0
37
0
Order By: Relevance
“…Kundegorski and Breckon [29] achieved reasonable performances combining infrared imagery and realtime photogrammetry. Alahi et al combined monocular images with wireless signals [3] or with additional visual priors [1,2]. The seminal work of Mono3D [8] exploited deep learning to create 3D object proposals for car, pedestrian and cyclist categories but it did not evaluate 3D localization of pedestrians.…”
Section: Related Workmentioning
confidence: 99%
“…Kundegorski and Breckon [29] achieved reasonable performances combining infrared imagery and realtime photogrammetry. Alahi et al combined monocular images with wireless signals [3] or with additional visual priors [1,2]. The seminal work of Mono3D [8] exploited deep learning to create 3D object proposals for car, pedestrian and cyclist categories but it did not evaluate 3D localization of pedestrians.…”
Section: Related Workmentioning
confidence: 99%
“…Data association is widely used in radar systems, when tracking blips on a radar screen [10], as well as in object monitoring of surveillance systems [24]. When it comes to the cross-modal association, research attention is limited and all dedicated to location tracking of humans [6,49,59]. These methods heavily rely on the hypothesis that both sensor modalities are observing evolving state spaces matched precisely in the temporal domain.…”
Section: Related Workmentioning
confidence: 99%
“…Complete dataset can be found on our website. 1 In order to study how the goodness of wireless positioning will impact the visual tracking performance, we consider a BLE network with N = 32 beacons and simulate Receive Signal Strength (RSS) measurements according to the model given in Section 3 subject to Gaussian noise with zero mean and different noise variances σ 2 v . We use the Mean-Square-Error (MSE) as a measure of wireless positioning accuracy.…”
Section: Data and General Setupmentioning
confidence: 99%
“…As pointed out in [1], wireless data and visual data are complementary and should be fused for enhanced tracking performance. In the ideal cases where the above mentioned hurdles do not occur, visual tracking is in general more accurate and informative than wireless positioning.…”
Section: Introductionmentioning
confidence: 99%