2016
DOI: 10.1609/aaai.v30i1.10009
|View full text |Cite
|
Sign up to set email alerts
|

Egocentric Video Search via Physical Interactions

Abstract: Retrieving past egocentric videos about personal daily life is important to support and augment human memory. Most previous retrieval approaches have ignored the crucial feature of human-physical world interactions, which is greatly related to our memory and experience of daily activities. In this paper, we propose a gesture-based egocentric video retrieval framework, which retrieves past visual experience using body gestures as non-verbal queries. We use a probabilistic framework based on a canonical correlat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2018
2018
2018
2018

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 28 publications
0
2
0
Order By: Relevance
“…We evaluated our proposed method using datasets of ADLs, which are manually annotated with structured event descrip- tions of the ADLs that subjects performed in a house. This dataset has been used in egocentric video retrieval with gesture motions (Miyanishi et al 2016). The dataset consists of motion signals (e.g., acceleration and gyro) and first-person vision videos captured by a head-mounted wearable camera that enables us to capture various ADLs in diverse places.…”
Section: Experiments Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…We evaluated our proposed method using datasets of ADLs, which are manually annotated with structured event descrip- tions of the ADLs that subjects performed in a house. This dataset has been used in egocentric video retrieval with gesture motions (Miyanishi et al 2016). The dataset consists of motion signals (e.g., acceleration and gyro) and first-person vision videos captured by a head-mounted wearable camera that enables us to capture various ADLs in diverse places.…”
Section: Experiments Datasetsmentioning
confidence: 99%
“…Generating Semantic Concepts We made semantic concepts from the signals of wearable sensors. For feature extraction from motion signals and videos, we followed the past work (Miyanishi et al 2016). For the motion-feature extraction, we used acceleration and gyro signals and applied a short-time Fourier transform, where the window width was 75 samples (3 sec when using 25-Hz data) by shifting one sample.…”
Section: Experimental Settingsmentioning
confidence: 99%