Learning Cross-Video Neural Representations for High-Quality Frame Interpolation
This paper considers the problem of temporal video interpolation, where the goal is to synthesize a new video frame given its two neighbors. We propose Cross-Video Neural Representation (CURE) as the first video interpolation method based on neural fields (NF). NF refers to the recent class of metho...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Shangguan, Wentao Sun, Yu Gan, Weijie Kamilov, Ulugbek S |
description | This paper considers the problem of temporal video interpolation, where the
goal is to synthesize a new video frame given its two neighbors. We propose
Cross-Video Neural Representation (CURE) as the first video interpolation
method based on neural fields (NF). NF refers to the recent class of methods
for the neural representation of complex 3D scenes that has seen widespread
success and application across computer vision. CURE represents the video as a
continuous function parameterized by a coordinate-based neural network, whose
inputs are the spatiotemporal coordinates and outputs are the corresponding RGB
values. CURE introduces a new architecture that conditions the neural network
on the input frames for imposing space-time consistency in the synthesized
video. This not only improves the final interpolation quality, but also enables
CURE to learn a prior across multiple videos. Experimental evaluations show
that CURE achieves the state-of-the-art performance on video interpolation on
several benchmark datasets. |
doi_str_mv | 10.48550/arxiv.2203.00137 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2203_00137</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2203_00137</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-87380a709040193210013129c840704d311842b084c74a4f9cc2f5d3bcf501c3</originalsourceid><addsrcrecordid>eNotz01OwzAUBGBvWKDCAVjhCzg8_wQ7SxRRWimiQCu20atjF0upE9kpordHDaxmMxrNR8gdh0KZsoQHTD_huxACZAHApb4mb43DFEM80DoNObPP0LmBvrpTwp5-uDG57OKEUxhipn5IdBUOX-z9hH2YznSZ8OjoOk4ujUM_t27Ilcc-u9v_XJDt8nlXr1izeVnXTw3DR62Z0dIAaqhAAa-k4Jc7XFTWKNCgOsm5UWIPRlmtUPnKWuHLTu6tL4FbuSD3f6uzqB1TOGI6txdZO8vkL4-SR9Q</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Learning Cross-Video Neural Representations for High-Quality Frame Interpolation</title><source>arXiv.org</source><creator>Shangguan, Wentao ; Sun, Yu ; Gan, Weijie ; Kamilov, Ulugbek S</creator><creatorcontrib>Shangguan, Wentao ; Sun, Yu ; Gan, Weijie ; Kamilov, Ulugbek S</creatorcontrib><description>This paper considers the problem of temporal video interpolation, where the
goal is to synthesize a new video frame given its two neighbors. We propose
Cross-Video Neural Representation (CURE) as the first video interpolation
method based on neural fields (NF). NF refers to the recent class of methods
for the neural representation of complex 3D scenes that has seen widespread
success and application across computer vision. CURE represents the video as a
continuous function parameterized by a coordinate-based neural network, whose
inputs are the spatiotemporal coordinates and outputs are the corresponding RGB
values. CURE introduces a new architecture that conditions the neural network
on the input frames for imposing space-time consistency in the synthesized
video. This not only improves the final interpolation quality, but also enables
CURE to learn a prior across multiple videos. Experimental evaluations show
that CURE achieves the state-of-the-art performance on video interpolation on
several benchmark datasets.</description><identifier>DOI: 10.48550/arxiv.2203.00137</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2022-02</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2203.00137$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2203.00137$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Shangguan, Wentao</creatorcontrib><creatorcontrib>Sun, Yu</creatorcontrib><creatorcontrib>Gan, Weijie</creatorcontrib><creatorcontrib>Kamilov, Ulugbek S</creatorcontrib><title>Learning Cross-Video Neural Representations for High-Quality Frame Interpolation</title><description>This paper considers the problem of temporal video interpolation, where the
goal is to synthesize a new video frame given its two neighbors. We propose
Cross-Video Neural Representation (CURE) as the first video interpolation
method based on neural fields (NF). NF refers to the recent class of methods
for the neural representation of complex 3D scenes that has seen widespread
success and application across computer vision. CURE represents the video as a
continuous function parameterized by a coordinate-based neural network, whose
inputs are the spatiotemporal coordinates and outputs are the corresponding RGB
values. CURE introduces a new architecture that conditions the neural network
on the input frames for imposing space-time consistency in the synthesized
video. This not only improves the final interpolation quality, but also enables
CURE to learn a prior across multiple videos. Experimental evaluations show
that CURE achieves the state-of-the-art performance on video interpolation on
several benchmark datasets.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz01OwzAUBGBvWKDCAVjhCzg8_wQ7SxRRWimiQCu20atjF0upE9kpordHDaxmMxrNR8gdh0KZsoQHTD_huxACZAHApb4mb43DFEM80DoNObPP0LmBvrpTwp5-uDG57OKEUxhipn5IdBUOX-z9hH2YznSZ8OjoOk4ujUM_t27Ilcc-u9v_XJDt8nlXr1izeVnXTw3DR62Z0dIAaqhAAa-k4Jc7XFTWKNCgOsm5UWIPRlmtUPnKWuHLTu6tL4FbuSD3f6uzqB1TOGI6txdZO8vkL4-SR9Q</recordid><startdate>20220228</startdate><enddate>20220228</enddate><creator>Shangguan, Wentao</creator><creator>Sun, Yu</creator><creator>Gan, Weijie</creator><creator>Kamilov, Ulugbek S</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20220228</creationdate><title>Learning Cross-Video Neural Representations for High-Quality Frame Interpolation</title><author>Shangguan, Wentao ; Sun, Yu ; Gan, Weijie ; Kamilov, Ulugbek S</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-87380a709040193210013129c840704d311842b084c74a4f9cc2f5d3bcf501c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Shangguan, Wentao</creatorcontrib><creatorcontrib>Sun, Yu</creatorcontrib><creatorcontrib>Gan, Weijie</creatorcontrib><creatorcontrib>Kamilov, Ulugbek S</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Shangguan, Wentao</au><au>Sun, Yu</au><au>Gan, Weijie</au><au>Kamilov, Ulugbek S</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Learning Cross-Video Neural Representations for High-Quality Frame Interpolation</atitle><date>2022-02-28</date><risdate>2022</risdate><abstract>This paper considers the problem of temporal video interpolation, where the
goal is to synthesize a new video frame given its two neighbors. We propose
Cross-Video Neural Representation (CURE) as the first video interpolation
method based on neural fields (NF). NF refers to the recent class of methods
for the neural representation of complex 3D scenes that has seen widespread
success and application across computer vision. CURE represents the video as a
continuous function parameterized by a coordinate-based neural network, whose
inputs are the spatiotemporal coordinates and outputs are the corresponding RGB
values. CURE introduces a new architecture that conditions the neural network
on the input frames for imposing space-time consistency in the synthesized
video. This not only improves the final interpolation quality, but also enables
CURE to learn a prior across multiple videos. Experimental evaluations show
that CURE achieves the state-of-the-art performance on video interpolation on
several benchmark datasets.</abstract><doi>10.48550/arxiv.2203.00137</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2203.00137 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2203_00137 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition |
title | Learning Cross-Video Neural Representations for High-Quality Frame Interpolation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-05T23%3A49%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Learning%20Cross-Video%20Neural%20Representations%20for%20High-Quality%20Frame%20Interpolation&rft.au=Shangguan,%20Wentao&rft.date=2022-02-28&rft_id=info:doi/10.48550/arxiv.2203.00137&rft_dat=%3Carxiv_GOX%3E2203_00137%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |