The emerging MVC+D standard specifies the coding of Multiview Video plus Depth (MVD) data for enabling advanced 3D video applications. MVC+D specifications define the coding of all views of MVD at equal spatial resolution and apply a conventional MVC technique for coding the multiview texture and the depth independently. This paper presents a modified MVC+D coding scheme, where only the base view is coded at the original resolution whereas dependent views are coded at reduced resolution. To enable inter-view prediction, the base view is downsampled within the MVC coding loop to provide a relevant reference for dependent views. At the decoder side, the proposed scheme consists of a post-processing scheme which upsamples of the decoded views to their original resolution. The proposed scheme is compared against the original MVC+D scheme and an average of 4% delta bitrate reduction (dBR) in the coded views and 14.5% of dBR in the synthesized views are reported.Index Terms-3DV, MVC, asymmetric coding, spatial resolution, synthesized views INTRODUCTIONThe Moving Picture Experts Group (MPEG) has recently started 3D Video (3DV) standardization to enable support of advanced 3DV applications. The concept of advanced 3DV applications assumes that users can perceive a selected stereo-pair from numerous available views at the decoder side. Examples of such applications includes varying baseline to adjust the depth perception and multiview autostereoscopic displays (ASDs). Considering the complexity of capturing 3D scenes and the limitations in the distribution technologies, it is not possible to deliver a sufficiently large number of [4] for the independent coding of texture and depth. As a result, a forward compatibility with MVC specification is preserved, and texture views of MVC+D bitstreams can be decoded with a conventional MVC decoder. The MVC+D specification was implemented in 3DV-ATM reference software [8] and was used in this study.A possible solution to further reduce the bitrate and/or complexity of 3DV applications is to reduce the spatial resolution of a number of video views compared to the original resolution while preserving the original resolution for the remaining views. At the decoder side, views coded at the reduced resolution are upsampled to the original one using either conventional linear upsampling [9], or advanced super resolution techniques [10] that would benefit from multiview representation and the presence of depth. Being applied to texture component of MVD, this would result in a mixed-resolution texture representation and a significant bitrate reduction is hence expected.It is obvious, that a scheme with a mixed-resolution texture representation would result in decoded views (e.g. stereoscopic image-pair) with different quality, which may affect stereoscopic perception. However, this argument can be addressed with the binocular rivalry theory [11] claiming that stereoscopic vision in the human visual system (HVS) fuses the images of an asymmetric quality stereoscopic image-pair s...
Stereoscopic 3D video is becoming a reality in many application areas, ranging from high quality entertainment to mobile video services. Due to the need to process two views, the complexity of 3D video applications is significantly higher than traditional 2D counterparts. In order to enable real-time 3D video services in mobile devices, this paper proposes a novel algorithm which reduces the complexity of stereo video encoding with improvement of coding efficiency. A novel search window center prediction method is proposed that exploits the correlation between two views. Experimental results show that the average encoding time of the second view can be decreased by 80% with an increase in coding efficiency of up to 2%. The state-of-art fast motion estimation methods for stereoscopic 3D video encoding show coding efficiency decrease, whereas proposed method achieves the speed-up with increase in coding efficiency, making it suitable for high qality 3D video applications.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.