An Internal Learning Approach to Video Inpainting
We propose a novel video inpainting algorithm that simultaneously hallucinates missing appearance and motion (optical flow) information, building upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network architectures to enforce plausible texture in static images. In ext...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Zhang, Haotian Mai, Long Xu, Ning Wang, Zhaowen Collomosse, John Jin, Hailin |
description | We propose a novel video inpainting algorithm that simultaneously
hallucinates missing appearance and motion (optical flow) information, building
upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network
architectures to enforce plausible texture in static images. In extending DIP
to video we make two important contributions. First, we show that coherent
video inpainting is possible without a priori training. We take a generative
approach to inpainting based on internal (within-video) learning without
reliance upon an external corpus of visual data to train a one-size-fits-all
model for the large space of general videos. Second, we show that such a
framework can jointly generate both appearance and flow, whilst exploiting
these complementary modalities to ensure mutual consistency. We show that
leveraging appearance statistics specific to each video achieves visually
plausible results whilst handling the challenging problem of long-term
consistency. |
doi_str_mv | 10.48550/arxiv.1909.07957 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1909_07957</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1909_07957</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-4e2610185b6642d63b939de4dacbe1b1018dc39d69284ff3d67a7db39d9fc873</originalsourceid><addsrcrecordid>eNotjssKwjAURLNxIdUPcGV-oDVp0qRZFvEFBReK23LTpBrQtMQi-vfWx2rgzDAchGaUJDzPMrKA8HSPhCqiEiJVJseIFh7vfG-DhysuLQTv_BkXXRdaqC-4b_HJGdsOmw6c74dygkYNXO92-s8IHdar43Ibl_vNblmUMQgpY25TQQnNMy0ET41gWjFlLDdQa0v1pzL1QIRKc940zAgJ0uiBqKbOJYvQ_Pf6Va664G4QXtVHvfqqszfm_D2v</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>An Internal Learning Approach to Video Inpainting</title><source>arXiv.org</source><creator>Zhang, Haotian ; Mai, Long ; Xu, Ning ; Wang, Zhaowen ; Collomosse, John ; Jin, Hailin</creator><creatorcontrib>Zhang, Haotian ; Mai, Long ; Xu, Ning ; Wang, Zhaowen ; Collomosse, John ; Jin, Hailin</creatorcontrib><description>We propose a novel video inpainting algorithm that simultaneously
hallucinates missing appearance and motion (optical flow) information, building
upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network
architectures to enforce plausible texture in static images. In extending DIP
to video we make two important contributions. First, we show that coherent
video inpainting is possible without a priori training. We take a generative
approach to inpainting based on internal (within-video) learning without
reliance upon an external corpus of visual data to train a one-size-fits-all
model for the large space of general videos. Second, we show that such a
framework can jointly generate both appearance and flow, whilst exploiting
these complementary modalities to ensure mutual consistency. We show that
leveraging appearance statistics specific to each video achieves visually
plausible results whilst handling the challenging problem of long-term
consistency.</description><identifier>DOI: 10.48550/arxiv.1909.07957</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2019-09</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1909.07957$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1909.07957$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhang, Haotian</creatorcontrib><creatorcontrib>Mai, Long</creatorcontrib><creatorcontrib>Xu, Ning</creatorcontrib><creatorcontrib>Wang, Zhaowen</creatorcontrib><creatorcontrib>Collomosse, John</creatorcontrib><creatorcontrib>Jin, Hailin</creatorcontrib><title>An Internal Learning Approach to Video Inpainting</title><description>We propose a novel video inpainting algorithm that simultaneously
hallucinates missing appearance and motion (optical flow) information, building
upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network
architectures to enforce plausible texture in static images. In extending DIP
to video we make two important contributions. First, we show that coherent
video inpainting is possible without a priori training. We take a generative
approach to inpainting based on internal (within-video) learning without
reliance upon an external corpus of visual data to train a one-size-fits-all
model for the large space of general videos. Second, we show that such a
framework can jointly generate both appearance and flow, whilst exploiting
these complementary modalities to ensure mutual consistency. We show that
leveraging appearance statistics specific to each video achieves visually
plausible results whilst handling the challenging problem of long-term
consistency.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotjssKwjAURLNxIdUPcGV-oDVp0qRZFvEFBReK23LTpBrQtMQi-vfWx2rgzDAchGaUJDzPMrKA8HSPhCqiEiJVJseIFh7vfG-DhysuLQTv_BkXXRdaqC-4b_HJGdsOmw6c74dygkYNXO92-s8IHdar43Ibl_vNblmUMQgpY25TQQnNMy0ET41gWjFlLDdQa0v1pzL1QIRKc940zAgJ0uiBqKbOJYvQ_Pf6Va664G4QXtVHvfqqszfm_D2v</recordid><startdate>20190917</startdate><enddate>20190917</enddate><creator>Zhang, Haotian</creator><creator>Mai, Long</creator><creator>Xu, Ning</creator><creator>Wang, Zhaowen</creator><creator>Collomosse, John</creator><creator>Jin, Hailin</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20190917</creationdate><title>An Internal Learning Approach to Video Inpainting</title><author>Zhang, Haotian ; Mai, Long ; Xu, Ning ; Wang, Zhaowen ; Collomosse, John ; Jin, Hailin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-4e2610185b6642d63b939de4dacbe1b1018dc39d69284ff3d67a7db39d9fc873</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Haotian</creatorcontrib><creatorcontrib>Mai, Long</creatorcontrib><creatorcontrib>Xu, Ning</creatorcontrib><creatorcontrib>Wang, Zhaowen</creatorcontrib><creatorcontrib>Collomosse, John</creatorcontrib><creatorcontrib>Jin, Hailin</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zhang, Haotian</au><au>Mai, Long</au><au>Xu, Ning</au><au>Wang, Zhaowen</au><au>Collomosse, John</au><au>Jin, Hailin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>An Internal Learning Approach to Video Inpainting</atitle><date>2019-09-17</date><risdate>2019</risdate><abstract>We propose a novel video inpainting algorithm that simultaneously
hallucinates missing appearance and motion (optical flow) information, building
upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network
architectures to enforce plausible texture in static images. In extending DIP
to video we make two important contributions. First, we show that coherent
video inpainting is possible without a priori training. We take a generative
approach to inpainting based on internal (within-video) learning without
reliance upon an external corpus of visual data to train a one-size-fits-all
model for the large space of general videos. Second, we show that such a
framework can jointly generate both appearance and flow, whilst exploiting
these complementary modalities to ensure mutual consistency. We show that
leveraging appearance statistics specific to each video achieves visually
plausible results whilst handling the challenging problem of long-term
consistency.</abstract><doi>10.48550/arxiv.1909.07957</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.1909.07957 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_1909_07957 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition |
title | An Internal Learning Approach to Video Inpainting |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-13T11%3A01%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=An%20Internal%20Learning%20Approach%20to%20Video%20Inpainting&rft.au=Zhang,%20Haotian&rft.date=2019-09-17&rft_id=info:doi/10.48550/arxiv.1909.07957&rft_dat=%3Carxiv_GOX%3E1909_07957%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |