2019
DOI: 10.48550/arxiv.1909.10164
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

sZoom: A Framework for Automatic Zoom into High Resolution Surveillance Videos

Abstract: Current cameras are capable of recording high resolution video. While viewing on a mobile device, a user can manually zoom into this high resolution video to get more detailed view of objects and activities. However, manual zooming is not suitable for surveillance and monitoring. It is tiring to continuously keep zooming into various regions of the video. Also, while viewing one region, the operator may miss activities in other regions. In this paper, we propose sZoom, a framework to automatically zoom into a … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5

Relationship

5
0

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 29 publications
0
5
0
Order By: Relevance
“…In previous work [88,89,90,91,92,93] we used generative adversarial networks (GANs) for multimodal retinal image registration where the trained network outputs the registered image and the deformation field. We show that by incorporating appropriate constraints in the adversarial loss and content loss function, generative models are fairly reliable in directly generating the registered image and its corresponding deformation field.…”
Section: Related Workmentioning
confidence: 99%
“…In previous work [88,89,90,91,92,93] we used generative adversarial networks (GANs) for multimodal retinal image registration where the trained network outputs the registered image and the deformation field. We show that by incorporating appropriate constraints in the adversarial loss and content loss function, generative models are fairly reliable in directly generating the registered image and its corresponding deformation field.…”
Section: Related Workmentioning
confidence: 99%
“…Also, CNN methods capture mostly local context information and do not explore the global aspects. Zhang et al in [191,153,152,155,157,24], [22,130,151], [79,51,78,49], [42,43,11], [90,77,88], [168,8,75,76] propose a squeeze and excitation network to capture the global characteristics thus leading to improved super resolution output. However, squeeze and excitation relies on CNN features to capture global context which is not optimal.…”
Section: Introductionmentioning
confidence: 99%
“…Traditional augmentations such as image rotations or deformations have limited benefit as they do not fully represent the underlying data distribution of the training set and are sensitive to parameter choices. Recent data augmentation methods of [25], [131], [11], [113], [108], [106], [114], [129], [103], [69], [139], [39], [32], [33], [13], [73], [62], [63] use generative adversarial network (GAN), [ [23]], and show moderate success for medical image classification. However, they have limited relevance for segmentation since they do not model geometric relation between different organs and most augmentation approaches do not differentiate between normal and diseased samples.…”
Section: Introductionmentioning
confidence: 99%