2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2014
DOI: 10.1109/icassp.2014.6853687
|View full text |Cite
|
Sign up to set email alerts
|

Bandit framework for systematic learning in wireless video-based face recognition

Abstract: Professor Mihaela van der Schaar, Chair Video-based object or face recognition services on mobile devices have recently garnered significant attention, given that video cameras are now ubiquitous in all mobile communication devices. In one of the most typical scenarios for such services, each mobile device captures and transmits video frames over wireless to a remote computing cluster (a.k.a. "cloud" computing infrastructure) that performs the heavy-duty video feature extraction and recognition tasks for a lar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2014
2014
2018
2018

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 19 publications
0
2
0
Order By: Relevance
“…For video flow, Shot Boundary Detection is conducted first and the video is segmented into short takes with video segmentation algorithms, such as pixel algorithm, histogram algorithm, X2 histogram algorithm, X2 histogram block algorithm and contour a boundary ROC (Rate of Change) algorithm; then, the original motion trails of video object are extracted by use of moving object tracking algorithms, such as mean shift algorithm, object tracking based on Kalman filter, object tracking based on particle filter and algorithm based on modeling of moving object, with the longest trail to be processed and information extracted therefrom, including motion direction and slop of motion trail curve. At last, the said motion action will be marked by hand to extract the video verb semantic label [14][15][16][17][18][19][20][21]. Also, some other researchers proposed that semantic clews of multiple event recognitions should be fused by means of a deep-level learning strategy so that the issue of recognition would be solved by answering how to jointly analyse human actions, objects and scenes.…”
Section: Video Semantic Analysis and Relevant Researchmentioning
confidence: 99%
“…For video flow, Shot Boundary Detection is conducted first and the video is segmented into short takes with video segmentation algorithms, such as pixel algorithm, histogram algorithm, X2 histogram algorithm, X2 histogram block algorithm and contour a boundary ROC (Rate of Change) algorithm; then, the original motion trails of video object are extracted by use of moving object tracking algorithms, such as mean shift algorithm, object tracking based on Kalman filter, object tracking based on particle filter and algorithm based on modeling of moving object, with the longest trail to be processed and information extracted therefrom, including motion direction and slop of motion trail curve. At last, the said motion action will be marked by hand to extract the video verb semantic label [14][15][16][17][18][19][20][21]. Also, some other researchers proposed that semantic clews of multiple event recognitions should be fused by means of a deep-level learning strategy so that the issue of recognition would be solved by answering how to jointly analyse human actions, objects and scenes.…”
Section: Video Semantic Analysis and Relevant Researchmentioning
confidence: 99%
“…Onur Atan et al [21] amazingly launched an innovative Video-based object or face detection services on mobile devices which attracted zooming enthusiasm, in view of the fact that video cameras were then everpresent in the entire mobile communication tools. In a strikingly distinctive environment for the related services, each mobile device captured and communicated video frames over wireless to a far-flung computing cluster (a.k.a.…”
Section: Literature Surveymentioning
confidence: 99%