2012
DOI: 10.1007/s11042-012-1085-1
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal extraction of events and of information about the recording activity in user generated videos

Abstract: In this work we propose methods that exploit context sensor data modalities for the task of detecting interesting events and extracting high-level contextual information about the recording activity in user generated videos. Indeed, most camera-enabled electronic devices contain various auxiliary sensors such as accelerometers, compasses, GPS receivers, etc. Data captured by these sensors during the media acquisition have already been used to limit camera degradations such as shake and also to provide some bas… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2013
2013
2021
2021

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(27 citation statements)
references
References 33 publications
(46 reference statements)
0
10
0
Order By: Relevance
“…This information is used to annotate the generated video stream with the motion, environmental information such as the position of the device 4 . Based on this data, the detection of user behavior that leads to degraded video quality, such as camera shaking or wrong orientation of the phone, can also be observed directly on the users' device without analyzing (or transmitting) the video data itself [5]. In previous work [10], we found that degradations such as camera shakes have a huge impact on the QoE.…”
Section: Prediction Of Quality Of Expe-riencementioning
confidence: 94%
See 2 more Smart Citations
“…This information is used to annotate the generated video stream with the motion, environmental information such as the position of the device 4 . Based on this data, the detection of user behavior that leads to degraded video quality, such as camera shaking or wrong orientation of the phone, can also be observed directly on the users' device without analyzing (or transmitting) the video data itself [5]. In previous work [10], we found that degradations such as camera shakes have a huge impact on the QoE.…”
Section: Prediction Of Quality Of Expe-riencementioning
confidence: 94%
“…Also, software-based sensors indicate the current video recording properties or energy level of the device. Here, patterns can be used to identify recording degradations based on sensor readingsas shown by CriCri [5].…”
Section: Composition and Orchestration Servicementioning
confidence: 99%
See 1 more Smart Citation
“…MoviMash is an advanced approach existing for live video composition. CriCri [3] shows a first step towards replacing video analysis with mechanisms, that leverage different sensors in the recording devices to compose the video. Without respecting visual features the resulting quality of such approaches degrades [4] but the speed increases.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Common artifacts in compressing video streams such as blockiness are mitigated by the agreed video parameters (framerate, resolution and target bitrate). Thus, CrowdRecord investigates degradations commonly occurring during recording with retail smartphones, to (1) ensure the correct camera orientation, (2) identify and measure camera shaking, (3) identify the field of interest of a recording, (4) find under-/overexposed frame regions, and (5) measure frame sharpness. Streams that fail in one of these categories are not considered for composition.…”
Section: Crowdrecordmentioning
confidence: 99%