View synthesis prediction for multiview video coding

We propose a rate-distortion-optimized framework that incorporates view synthesis for improved prediction in multiview video coding. In the proposed scheme, auxiliary information, including depth data, is encoded and used at the decoder to generate the view synthesis prediction data. The proposed me...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Signal processing. Image communication 2009, Vol.24 (1), p.89-100
Hauptverfasser: Yea, Sehoon, Vetro, Anthony
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 100
container_issue 1
container_start_page 89
container_title Signal processing. Image communication
container_volume 24
creator Yea, Sehoon
Vetro, Anthony
description We propose a rate-distortion-optimized framework that incorporates view synthesis for improved prediction in multiview video coding. In the proposed scheme, auxiliary information, including depth data, is encoded and used at the decoder to generate the view synthesis prediction data. The proposed method employs optimal mode decision including view synthesis prediction, and sub-pixel reference matching to improve prediction accuracy of the view synthesis prediction. Novel variants of the skip and direct modes are also presented, which infer the depth and correction vector information from neighboring blocks in a synthesized reference picture to reduce the bits needed for the view synthesis prediction mode. We demonstrate two multiview video coding scenarios in which view synthesis prediction is employed. In the first scenario, the goal is to improve the coding efficiency of multiview video where block-based depths and correction vectors are encoded by CABAC in a lossless manner on a macroblock basis. A variable block-size depth/motion search algorithm is described. Experimental results demonstrate that view synthesis prediction does provide some coding gains when combined with disparity-compensated prediction. In the second scenario, the goal is to use view synthesis prediction for reducing rate overhead incurred by transmitting depth maps for improved support of 3DTV and free-viewpoint video applications. It is assumed that the complete depth map for each view is encoded separately from the multiview video and used at the receiver to generate intermediate views. We utilize this information for view synthesis prediction to improve overall coding efficiency. Experimental results show that the rate overhead incurred by coding depth maps of varying quality could be offset by utilizing the proposed view synthesis prediction techniques to reduce the bitrate required for coding multiview video.
doi_str_mv 10.1016/j.image.2008.10.007
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_33611204</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0923596508001173</els_id><sourcerecordid>33611204</sourcerecordid><originalsourceid>FETCH-LOGICAL-c400t-9850b05c13429004ad2e96e38cb27d94740f5533758af889900f58fd1b6fb6623</originalsourceid><addsrcrecordid>eNp9kDtPwzAUhS0EEqXwC1gysSVcP2MPDKjiJVViAVYrceziKo2LnRb13-NQZqYrHX3nSudD6BpDhQGL23XlN83KVgRA5qQCqE_QDMtalUTU9SmagSK05Erwc3SR0hoACAM1Q-zD2-8iHYbx0yafim20nTejD0PhQiw2u370-wnZ-86GwoTOD6tLdOaaPtmrvztH748Pb4vncvn69LK4X5aGAYylkhxa4AZTRhQAazpilbBUmpbUnWI1A8c5pTWXjZNSZcZx6TrcCtcKQegc3Rz_bmP42tk06o1PxvZ9M9iwS5pSgTEBlkF6BE0MKUXr9DZmI_GgMejJkF7rX0N6MjSF2VBu3R1bNm_II6NOxtvBZAPRmlF3wf_b_wH4Um7T</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>33611204</pqid></control><display><type>article</type><title>View synthesis prediction for multiview video coding</title><source>Elsevier ScienceDirect Journals Complete</source><creator>Yea, Sehoon ; Vetro, Anthony</creator><creatorcontrib>Yea, Sehoon ; Vetro, Anthony</creatorcontrib><description>We propose a rate-distortion-optimized framework that incorporates view synthesis for improved prediction in multiview video coding. In the proposed scheme, auxiliary information, including depth data, is encoded and used at the decoder to generate the view synthesis prediction data. The proposed method employs optimal mode decision including view synthesis prediction, and sub-pixel reference matching to improve prediction accuracy of the view synthesis prediction. Novel variants of the skip and direct modes are also presented, which infer the depth and correction vector information from neighboring blocks in a synthesized reference picture to reduce the bits needed for the view synthesis prediction mode. We demonstrate two multiview video coding scenarios in which view synthesis prediction is employed. In the first scenario, the goal is to improve the coding efficiency of multiview video where block-based depths and correction vectors are encoded by CABAC in a lossless manner on a macroblock basis. A variable block-size depth/motion search algorithm is described. Experimental results demonstrate that view synthesis prediction does provide some coding gains when combined with disparity-compensated prediction. In the second scenario, the goal is to use view synthesis prediction for reducing rate overhead incurred by transmitting depth maps for improved support of 3DTV and free-viewpoint video applications. It is assumed that the complete depth map for each view is encoded separately from the multiview video and used at the receiver to generate intermediate views. We utilize this information for view synthesis prediction to improve overall coding efficiency. Experimental results show that the rate overhead incurred by coding depth maps of varying quality could be offset by utilizing the proposed view synthesis prediction techniques to reduce the bitrate required for coding multiview video.</description><identifier>ISSN: 0923-5965</identifier><identifier>EISSN: 1879-2677</identifier><identifier>DOI: 10.1016/j.image.2008.10.007</identifier><language>eng</language><publisher>Elsevier B.V</publisher><subject>3DTV ; Depth ; Free-viewpoint video ; Multiview video coding ; Prediction ; View synthesis</subject><ispartof>Signal processing. Image communication, 2009, Vol.24 (1), p.89-100</ispartof><rights>2008 Elsevier B.V.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c400t-9850b05c13429004ad2e96e38cb27d94740f5533758af889900f58fd1b6fb6623</citedby><cites>FETCH-LOGICAL-c400t-9850b05c13429004ad2e96e38cb27d94740f5533758af889900f58fd1b6fb6623</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.image.2008.10.007$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,780,784,3550,4024,27923,27924,27925,45995</link.rule.ids></links><search><creatorcontrib>Yea, Sehoon</creatorcontrib><creatorcontrib>Vetro, Anthony</creatorcontrib><title>View synthesis prediction for multiview video coding</title><title>Signal processing. Image communication</title><description>We propose a rate-distortion-optimized framework that incorporates view synthesis for improved prediction in multiview video coding. In the proposed scheme, auxiliary information, including depth data, is encoded and used at the decoder to generate the view synthesis prediction data. The proposed method employs optimal mode decision including view synthesis prediction, and sub-pixel reference matching to improve prediction accuracy of the view synthesis prediction. Novel variants of the skip and direct modes are also presented, which infer the depth and correction vector information from neighboring blocks in a synthesized reference picture to reduce the bits needed for the view synthesis prediction mode. We demonstrate two multiview video coding scenarios in which view synthesis prediction is employed. In the first scenario, the goal is to improve the coding efficiency of multiview video where block-based depths and correction vectors are encoded by CABAC in a lossless manner on a macroblock basis. A variable block-size depth/motion search algorithm is described. Experimental results demonstrate that view synthesis prediction does provide some coding gains when combined with disparity-compensated prediction. In the second scenario, the goal is to use view synthesis prediction for reducing rate overhead incurred by transmitting depth maps for improved support of 3DTV and free-viewpoint video applications. It is assumed that the complete depth map for each view is encoded separately from the multiview video and used at the receiver to generate intermediate views. We utilize this information for view synthesis prediction to improve overall coding efficiency. Experimental results show that the rate overhead incurred by coding depth maps of varying quality could be offset by utilizing the proposed view synthesis prediction techniques to reduce the bitrate required for coding multiview video.</description><subject>3DTV</subject><subject>Depth</subject><subject>Free-viewpoint video</subject><subject>Multiview video coding</subject><subject>Prediction</subject><subject>View synthesis</subject><issn>0923-5965</issn><issn>1879-2677</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2009</creationdate><recordtype>article</recordtype><recordid>eNp9kDtPwzAUhS0EEqXwC1gysSVcP2MPDKjiJVViAVYrceziKo2LnRb13-NQZqYrHX3nSudD6BpDhQGL23XlN83KVgRA5qQCqE_QDMtalUTU9SmagSK05Erwc3SR0hoACAM1Q-zD2-8iHYbx0yafim20nTejD0PhQiw2u370-wnZ-86GwoTOD6tLdOaaPtmrvztH748Pb4vncvn69LK4X5aGAYylkhxa4AZTRhQAazpilbBUmpbUnWI1A8c5pTWXjZNSZcZx6TrcCtcKQegc3Rz_bmP42tk06o1PxvZ9M9iwS5pSgTEBlkF6BE0MKUXr9DZmI_GgMejJkF7rX0N6MjSF2VBu3R1bNm_II6NOxtvBZAPRmlF3wf_b_wH4Um7T</recordid><startdate>2009</startdate><enddate>2009</enddate><creator>Yea, Sehoon</creator><creator>Vetro, Anthony</creator><general>Elsevier B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>2009</creationdate><title>View synthesis prediction for multiview video coding</title><author>Yea, Sehoon ; Vetro, Anthony</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c400t-9850b05c13429004ad2e96e38cb27d94740f5533758af889900f58fd1b6fb6623</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2009</creationdate><topic>3DTV</topic><topic>Depth</topic><topic>Free-viewpoint video</topic><topic>Multiview video coding</topic><topic>Prediction</topic><topic>View synthesis</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Yea, Sehoon</creatorcontrib><creatorcontrib>Vetro, Anthony</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Signal processing. Image communication</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yea, Sehoon</au><au>Vetro, Anthony</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>View synthesis prediction for multiview video coding</atitle><jtitle>Signal processing. Image communication</jtitle><date>2009</date><risdate>2009</risdate><volume>24</volume><issue>1</issue><spage>89</spage><epage>100</epage><pages>89-100</pages><issn>0923-5965</issn><eissn>1879-2677</eissn><abstract>We propose a rate-distortion-optimized framework that incorporates view synthesis for improved prediction in multiview video coding. In the proposed scheme, auxiliary information, including depth data, is encoded and used at the decoder to generate the view synthesis prediction data. The proposed method employs optimal mode decision including view synthesis prediction, and sub-pixel reference matching to improve prediction accuracy of the view synthesis prediction. Novel variants of the skip and direct modes are also presented, which infer the depth and correction vector information from neighboring blocks in a synthesized reference picture to reduce the bits needed for the view synthesis prediction mode. We demonstrate two multiview video coding scenarios in which view synthesis prediction is employed. In the first scenario, the goal is to improve the coding efficiency of multiview video where block-based depths and correction vectors are encoded by CABAC in a lossless manner on a macroblock basis. A variable block-size depth/motion search algorithm is described. Experimental results demonstrate that view synthesis prediction does provide some coding gains when combined with disparity-compensated prediction. In the second scenario, the goal is to use view synthesis prediction for reducing rate overhead incurred by transmitting depth maps for improved support of 3DTV and free-viewpoint video applications. It is assumed that the complete depth map for each view is encoded separately from the multiview video and used at the receiver to generate intermediate views. We utilize this information for view synthesis prediction to improve overall coding efficiency. Experimental results show that the rate overhead incurred by coding depth maps of varying quality could be offset by utilizing the proposed view synthesis prediction techniques to reduce the bitrate required for coding multiview video.</abstract><pub>Elsevier B.V</pub><doi>10.1016/j.image.2008.10.007</doi><tpages>12</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0923-5965
ispartof Signal processing. Image communication, 2009, Vol.24 (1), p.89-100
issn 0923-5965
1879-2677
language eng
recordid cdi_proquest_miscellaneous_33611204
source Elsevier ScienceDirect Journals Complete
subjects 3DTV
Depth
Free-viewpoint video
Multiview video coding
Prediction
View synthesis
title View synthesis prediction for multiview video coding
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T07%3A43%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=View%20synthesis%20prediction%20for%20multiview%20video%20coding&rft.jtitle=Signal%20processing.%20Image%20communication&rft.au=Yea,%20Sehoon&rft.date=2009&rft.volume=24&rft.issue=1&rft.spage=89&rft.epage=100&rft.pages=89-100&rft.issn=0923-5965&rft.eissn=1879-2677&rft_id=info:doi/10.1016/j.image.2008.10.007&rft_dat=%3Cproquest_cross%3E33611204%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=33611204&rft_id=info:pmid/&rft_els_id=S0923596508001173&rfr_iscdi=true