Show simple item record

dc.contributor.authorVelisavljević, Vladanen_GB
dc.contributor.authorCheung, Geneen_GB
dc.contributor.authorChakareski, Jacoben_GB
dc.date.accessioned2013-05-21T15:12:17Z
dc.date.available2013-05-21T15:12:17Z
dc.date.issued2011
dc.identifier.citationVelisavljevic, V., Cheung, G. and Chakareski, J. (2011) 'Bit allocation for multiview image compression using cubic synthesized view distortion model', IEEE International Conference on Multimedia and Expo (ICME), Barcelona, Spain, 11-15 July. Barcelona: IEEE, pp.1-6.en_GB
dc.identifier.isbn9781612843490
dc.identifier.doi10.1109/ICME.2011.6012199
dc.identifier.urihttp://hdl.handle.net/10547/292594
dc.description.abstract"Texture-plus-depth" has become a popular coding format for multiview image compression, where a decoder can synthesize images at intermediate viewpoints using encoded texture and depth maps of closest captured view locations via depth-image-based rendering (DIBR). As in other resource-constrained scenarios, limited avail able bits must be optimally distributed among captured texture and depth maps to minimize the expected signal distortion at the decoder. A specific challenge of multiview image compression for DIBR is that the encoder must allocate bits without the knowledge of how many and which specific virtual views will be synthesized at the decoder for viewing. In this paper, we derive a cubic synthesized view distortion model to describe the visual quality of an interpolated view as a function of the view's location. Given the model, one can easily find the virtual view location between two coded views where the maximum synthesized distortion occurs. Using a multi view image codec based on shape-adaptive wavelet transform, we show how optimal bit allocation can be performed to minimize the maximum view synthesis distortion at any intermediate viewpoint. Our experimental results show that the optimal bit allocation can outperform a common uniform bit allocation scheme by up to 1.0dB in coding efficiency performance, while simultaneously being competitive to a state-of-the-art H.264 codec.
dc.language.isoenen
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INCen_GB
dc.relation.urlhttp://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=6012199en_GB
dc.subjectimage codingen_GB
dc.subjectimage textureen_GB
dc.subjectrendering (computer graphics)en_GB
dc.titleBit allocation for multiview image compression using cubic synthesized view distortion modelen
dc.typeConference papers, meetings and proceedingsen
html.description.abstract"Texture-plus-depth" has become a popular coding format for multiview image compression, where a decoder can synthesize images at intermediate viewpoints using encoded texture and depth maps of closest captured view locations via depth-image-based rendering (DIBR). As in other resource-constrained scenarios, limited avail able bits must be optimally distributed among captured texture and depth maps to minimize the expected signal distortion at the decoder. A specific challenge of multiview image compression for DIBR is that the encoder must allocate bits without the knowledge of how many and which specific virtual views will be synthesized at the decoder for viewing. In this paper, we derive a cubic synthesized view distortion model to describe the visual quality of an interpolated view as a function of the view's location. Given the model, one can easily find the virtual view location between two coded views where the maximum synthesized distortion occurs. Using a multi view image codec based on shape-adaptive wavelet transform, we show how optimal bit allocation can be performed to minimize the maximum view synthesis distortion at any intermediate viewpoint. Our experimental results show that the optimal bit allocation can outperform a common uniform bit allocation scheme by up to 1.0dB in coding efficiency performance, while simultaneously being competitive to a state-of-the-art H.264 codec.


This item appears in the following Collection(s)

  • Centre for Wireless Research (CWR)
    The Centre for Wireless Research brings together expertise in the areas of mobile and wireless sensor networks. The breadth and depth of the expertise make the Centre rich with research and innovation potential.

Show simple item record