2019
DOI: 10.1145/3355089.3356557
|View full text |Cite
|
Sign up to set email alerts
|

DeepFovea

Abstract: In order to provide an immersive visual experience, modern displays require head mounting, high image resolution, low latency, as well as high refresh rate. This poses a challenging computational problem. On the other hand, the human visual system can consume only a tiny fraction of this video stream due to the drastic acuity loss in the peripheral vision. Foveated rendering and compression can save computations by reducing the image quality in the peripheral vision. However, this can cause noticeable artifact… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 88 publications
(10 citation statements)
references
References 44 publications
0
10
0
Order By: Relevance
“…Their experiments reveal the effectiveness of the cloud VR system with foveated rendering. Authors in [20] conducted experimental evaluations of DeepFovea, an AI-assisted foveated rendering process and achieved more than 14 times compression on RGB video with no significant degradation in user-perceived quality. Their results indicate that deep learning (DL) based foveation would significantly reduce the rendering load.…”
Section: Cloud Vrmentioning
confidence: 99%
See 1 more Smart Citation
“…Their experiments reveal the effectiveness of the cloud VR system with foveated rendering. Authors in [20] conducted experimental evaluations of DeepFovea, an AI-assisted foveated rendering process and achieved more than 14 times compression on RGB video with no significant degradation in user-perceived quality. Their results indicate that deep learning (DL) based foveation would significantly reduce the rendering load.…”
Section: Cloud Vrmentioning
confidence: 99%
“…Its 6G vision and technology trends study are expected to be completed by 2023. The ITU-T focus group technologies for network 2030 (FG NET-2030) was established by ITU-T Study Group 13 at its meeting in Geneva, [16][17][18][19][20][21][22][23][24][25][26][27] July 2018. It intends to study the capabilities of networks for the year 2030 and beyond, when it is expected to support novel forward-looking scenarios, such as holographic type communications, extremely fast response in critical situations and high-precision communication demands of emerging market verticals.…”
Section: Introductionmentioning
confidence: 99%
“…Using eye-tracking devices, however, introduces an additional processing delay to the system and can only work with compatible client devices. Kaplanyan et al [23] proposed a neural network-based codec for 3D and AR contents. This work also assumes the availability of an eye gaze tracker at client.…”
Section: Related Workmentioning
confidence: 99%
“…DeepGame takes a learning-based approach to understand the player contextual interest within the game, predict the regions of interest (ROIs) across frames, and allocate bits to different regions based on their importance. Unlike prior works (e.g., [17,20,23]), DeepGame does not need additional feedback from players nor does it modify the existing encoders. Thus, DeepGame is easier to deploy in practice.…”
Section: Introductionmentioning
confidence: 99%
“…Focal computer vision relies on a nonhomogeneous compression of an image that maintains the pixel information at the center of fixation and strongly compresses it at the periphery, including pyramidal encoding (Kortum & Geisler, 1996;Butko & Movellan, 2010), local wavelet decomposition (Daucé, 2018) and log-polar encoding (Traver & Bernardino, 2010). A recent deep-learning-based implementation of such compression shows that in a video flow, a log-polar sampling of the image is sufficient to provide a reconstruction of the whole image (Kaplanyan et al, 2019). However, this particular algorithm lacks a system predicting the best saccadic action to perform.…”
Section: State Of the Artmentioning
confidence: 99%