Show simple item record

dc.contributor.authorVelisavljević, Vladanen_GB
dc.contributor.authorCheung, Geneen_GB
dc.contributor.authorChakareski, Jacoben_GB
dc.date.accessioned2013-05-21T15:20:14Z
dc.date.available2013-05-21T15:20:14Z
dc.date.issued2010
dc.identifier.citationVelisavljevic, V., Cheung, G. and Chakareski, J. (2010) 'Optimal rate allocation for view synthesis along a continuous viewpoint location in multiview imaging', Picture Coding Symposium (PCS), Nagoya, Japan, 8-10 December. Nagoya: IEEE, pp.482-485.en_GB
dc.identifier.isbn9781424471348
dc.identifier.doi10.1109/PCS.2010.5702542
dc.identifier.urihttp://hdl.handle.net/10547/292597
dc.description.abstractThe authors consider the scenario of view synthesis via depth-image based rendering in multi-view imaging. We formulate a resource allocation problem of jointly assigning an optimal number of bits to compressed texture and depth images such that the maximum distortion of a synthesized view over a continuum of viewpoints between two encoded reference views is minimized, for a given bit budget. We construct simple yet accurate image models that characterize the pixel values at similar depths as first-order Gaussian auto-regressive processes. Based on our models, we derive an optimization procedure that numerically solves the formulated min-max problem using Lagrange relaxation. Through simulations we show that, for two captured views scenario, our optimization provides a significant gain (up to 2dB) in quality of the synthesized views for the same overall bit rate over a heuristic quantization that selects only two quantizers - one for the encoded texture images and the other for the depth images.
dc.language.isoenen
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INCen_GB
dc.relation.urlhttp://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=5702542en_GB
dc.subjectautoregressive processesen_GB
dc.subjectimage textureen_GB
dc.subjectrate allocationen_GB
dc.titleOptimal rate allocation for view synthesis along a continuous viewpoint location in multiview imagingen
dc.typeConference papers, meetings and proceedingsen
html.description.abstractThe authors consider the scenario of view synthesis via depth-image based rendering in multi-view imaging. We formulate a resource allocation problem of jointly assigning an optimal number of bits to compressed texture and depth images such that the maximum distortion of a synthesized view over a continuum of viewpoints between two encoded reference views is minimized, for a given bit budget. We construct simple yet accurate image models that characterize the pixel values at similar depths as first-order Gaussian auto-regressive processes. Based on our models, we derive an optimization procedure that numerically solves the formulated min-max problem using Lagrange relaxation. Through simulations we show that, for two captured views scenario, our optimization provides a significant gain (up to 2dB) in quality of the synthesized views for the same overall bit rate over a heuristic quantization that selects only two quantizers - one for the encoded texture images and the other for the depth images.


This item appears in the following Collection(s)

  • Centre for Wireless Research (CWR)
    The Centre for Wireless Research brings together expertise in the areas of mobile and wireless sensor networks. The breadth and depth of the expertise make the Centre rich with research and innovation potential.

Show simple item record