Learning Cross-Video Neural Representations for High-Quality Frame Interpolation

This paper considers the problem of temporal video interpolation, where the goal is to synthesize a new video frame given its two neighbors. We propose Cross-Video Neural Representation (CURE) as the first video interpolation method based on neural fields (NF). NF refers to the recent class of metho...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Shangguan, Wentao, Sun, Yu, Gan, Weijie, Kamilov, Ulugbek S
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Shangguan, Wentao
Sun, Yu
Gan, Weijie
Kamilov, Ulugbek S
description This paper considers the problem of temporal video interpolation, where the goal is to synthesize a new video frame given its two neighbors. We propose Cross-Video Neural Representation (CURE) as the first video interpolation method based on neural fields (NF). NF refers to the recent class of methods for the neural representation of complex 3D scenes that has seen widespread success and application across computer vision. CURE represents the video as a continuous function parameterized by a coordinate-based neural network, whose inputs are the spatiotemporal coordinates and outputs are the corresponding RGB values. CURE introduces a new architecture that conditions the neural network on the input frames for imposing space-time consistency in the synthesized video. This not only improves the final interpolation quality, but also enables CURE to learn a prior across multiple videos. Experimental evaluations show that CURE achieves the state-of-the-art performance on video interpolation on several benchmark datasets.
doi_str_mv 10.48550/arxiv.2203.00137
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2203_00137</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2203_00137</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-87380a709040193210013129c840704d311842b084c74a4f9cc2f5d3bcf501c3</originalsourceid><addsrcrecordid>eNotz01OwzAUBGBvWKDCAVjhCzg8_wQ7SxRRWimiQCu20atjF0upE9kpordHDaxmMxrNR8gdh0KZsoQHTD_huxACZAHApb4mb43DFEM80DoNObPP0LmBvrpTwp5-uDG57OKEUxhipn5IdBUOX-z9hH2YznSZ8OjoOk4ujUM_t27Ilcc-u9v_XJDt8nlXr1izeVnXTw3DR62Z0dIAaqhAAa-k4Jc7XFTWKNCgOsm5UWIPRlmtUPnKWuHLTu6tL4FbuSD3f6uzqB1TOGI6txdZO8vkL4-SR9Q</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Learning Cross-Video Neural Representations for High-Quality Frame Interpolation</title><source>arXiv.org</source><creator>Shangguan, Wentao ; Sun, Yu ; Gan, Weijie ; Kamilov, Ulugbek S</creator><creatorcontrib>Shangguan, Wentao ; Sun, Yu ; Gan, Weijie ; Kamilov, Ulugbek S</creatorcontrib><description>This paper considers the problem of temporal video interpolation, where the goal is to synthesize a new video frame given its two neighbors. We propose Cross-Video Neural Representation (CURE) as the first video interpolation method based on neural fields (NF). NF refers to the recent class of methods for the neural representation of complex 3D scenes that has seen widespread success and application across computer vision. CURE represents the video as a continuous function parameterized by a coordinate-based neural network, whose inputs are the spatiotemporal coordinates and outputs are the corresponding RGB values. CURE introduces a new architecture that conditions the neural network on the input frames for imposing space-time consistency in the synthesized video. This not only improves the final interpolation quality, but also enables CURE to learn a prior across multiple videos. Experimental evaluations show that CURE achieves the state-of-the-art performance on video interpolation on several benchmark datasets.</description><identifier>DOI: 10.48550/arxiv.2203.00137</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2022-02</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2203.00137$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2203.00137$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Shangguan, Wentao</creatorcontrib><creatorcontrib>Sun, Yu</creatorcontrib><creatorcontrib>Gan, Weijie</creatorcontrib><creatorcontrib>Kamilov, Ulugbek S</creatorcontrib><title>Learning Cross-Video Neural Representations for High-Quality Frame Interpolation</title><description>This paper considers the problem of temporal video interpolation, where the goal is to synthesize a new video frame given its two neighbors. We propose Cross-Video Neural Representation (CURE) as the first video interpolation method based on neural fields (NF). NF refers to the recent class of methods for the neural representation of complex 3D scenes that has seen widespread success and application across computer vision. CURE represents the video as a continuous function parameterized by a coordinate-based neural network, whose inputs are the spatiotemporal coordinates and outputs are the corresponding RGB values. CURE introduces a new architecture that conditions the neural network on the input frames for imposing space-time consistency in the synthesized video. This not only improves the final interpolation quality, but also enables CURE to learn a prior across multiple videos. Experimental evaluations show that CURE achieves the state-of-the-art performance on video interpolation on several benchmark datasets.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz01OwzAUBGBvWKDCAVjhCzg8_wQ7SxRRWimiQCu20atjF0upE9kpordHDaxmMxrNR8gdh0KZsoQHTD_huxACZAHApb4mb43DFEM80DoNObPP0LmBvrpTwp5-uDG57OKEUxhipn5IdBUOX-z9hH2YznSZ8OjoOk4ujUM_t27Ilcc-u9v_XJDt8nlXr1izeVnXTw3DR62Z0dIAaqhAAa-k4Jc7XFTWKNCgOsm5UWIPRlmtUPnKWuHLTu6tL4FbuSD3f6uzqB1TOGI6txdZO8vkL4-SR9Q</recordid><startdate>20220228</startdate><enddate>20220228</enddate><creator>Shangguan, Wentao</creator><creator>Sun, Yu</creator><creator>Gan, Weijie</creator><creator>Kamilov, Ulugbek S</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20220228</creationdate><title>Learning Cross-Video Neural Representations for High-Quality Frame Interpolation</title><author>Shangguan, Wentao ; Sun, Yu ; Gan, Weijie ; Kamilov, Ulugbek S</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-87380a709040193210013129c840704d311842b084c74a4f9cc2f5d3bcf501c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Shangguan, Wentao</creatorcontrib><creatorcontrib>Sun, Yu</creatorcontrib><creatorcontrib>Gan, Weijie</creatorcontrib><creatorcontrib>Kamilov, Ulugbek S</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Shangguan, Wentao</au><au>Sun, Yu</au><au>Gan, Weijie</au><au>Kamilov, Ulugbek S</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Learning Cross-Video Neural Representations for High-Quality Frame Interpolation</atitle><date>2022-02-28</date><risdate>2022</risdate><abstract>This paper considers the problem of temporal video interpolation, where the goal is to synthesize a new video frame given its two neighbors. We propose Cross-Video Neural Representation (CURE) as the first video interpolation method based on neural fields (NF). NF refers to the recent class of methods for the neural representation of complex 3D scenes that has seen widespread success and application across computer vision. CURE represents the video as a continuous function parameterized by a coordinate-based neural network, whose inputs are the spatiotemporal coordinates and outputs are the corresponding RGB values. CURE introduces a new architecture that conditions the neural network on the input frames for imposing space-time consistency in the synthesized video. This not only improves the final interpolation quality, but also enables CURE to learn a prior across multiple videos. Experimental evaluations show that CURE achieves the state-of-the-art performance on video interpolation on several benchmark datasets.</abstract><doi>10.48550/arxiv.2203.00137</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2203.00137
ispartof
issn
language eng
recordid cdi_arxiv_primary_2203_00137
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
title Learning Cross-Video Neural Representations for High-Quality Frame Interpolation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-05T23%3A49%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Learning%20Cross-Video%20Neural%20Representations%20for%20High-Quality%20Frame%20Interpolation&rft.au=Shangguan,%20Wentao&rft.date=2022-02-28&rft_id=info:doi/10.48550/arxiv.2203.00137&rft_dat=%3Carxiv_GOX%3E2203_00137%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true