An Internal Learning Approach to Video Inpainting

We propose a novel video inpainting algorithm that simultaneously hallucinates missing appearance and motion (optical flow) information, building upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network architectures to enforce plausible texture in static images. In ext...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Zhang, Haotian, Mai, Long, Xu, Ning, Wang, Zhaowen, Collomosse, John, Jin, Hailin
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Zhang, Haotian
Mai, Long
Xu, Ning
Wang, Zhaowen
Collomosse, John
Jin, Hailin
description We propose a novel video inpainting algorithm that simultaneously hallucinates missing appearance and motion (optical flow) information, building upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network architectures to enforce plausible texture in static images. In extending DIP to video we make two important contributions. First, we show that coherent video inpainting is possible without a priori training. We take a generative approach to inpainting based on internal (within-video) learning without reliance upon an external corpus of visual data to train a one-size-fits-all model for the large space of general videos. Second, we show that such a framework can jointly generate both appearance and flow, whilst exploiting these complementary modalities to ensure mutual consistency. We show that leveraging appearance statistics specific to each video achieves visually plausible results whilst handling the challenging problem of long-term consistency.
doi_str_mv 10.48550/arxiv.1909.07957
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1909_07957</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1909_07957</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-4e2610185b6642d63b939de4dacbe1b1018dc39d69284ff3d67a7db39d9fc873</originalsourceid><addsrcrecordid>eNotjssKwjAURLNxIdUPcGV-oDVp0qRZFvEFBReK23LTpBrQtMQi-vfWx2rgzDAchGaUJDzPMrKA8HSPhCqiEiJVJseIFh7vfG-DhysuLQTv_BkXXRdaqC-4b_HJGdsOmw6c74dygkYNXO92-s8IHdar43Ibl_vNblmUMQgpY25TQQnNMy0ET41gWjFlLDdQa0v1pzL1QIRKc940zAgJ0uiBqKbOJYvQ_Pf6Va664G4QXtVHvfqqszfm_D2v</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>An Internal Learning Approach to Video Inpainting</title><source>arXiv.org</source><creator>Zhang, Haotian ; Mai, Long ; Xu, Ning ; Wang, Zhaowen ; Collomosse, John ; Jin, Hailin</creator><creatorcontrib>Zhang, Haotian ; Mai, Long ; Xu, Ning ; Wang, Zhaowen ; Collomosse, John ; Jin, Hailin</creatorcontrib><description>We propose a novel video inpainting algorithm that simultaneously hallucinates missing appearance and motion (optical flow) information, building upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network architectures to enforce plausible texture in static images. In extending DIP to video we make two important contributions. First, we show that coherent video inpainting is possible without a priori training. We take a generative approach to inpainting based on internal (within-video) learning without reliance upon an external corpus of visual data to train a one-size-fits-all model for the large space of general videos. Second, we show that such a framework can jointly generate both appearance and flow, whilst exploiting these complementary modalities to ensure mutual consistency. We show that leveraging appearance statistics specific to each video achieves visually plausible results whilst handling the challenging problem of long-term consistency.</description><identifier>DOI: 10.48550/arxiv.1909.07957</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2019-09</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1909.07957$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1909.07957$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhang, Haotian</creatorcontrib><creatorcontrib>Mai, Long</creatorcontrib><creatorcontrib>Xu, Ning</creatorcontrib><creatorcontrib>Wang, Zhaowen</creatorcontrib><creatorcontrib>Collomosse, John</creatorcontrib><creatorcontrib>Jin, Hailin</creatorcontrib><title>An Internal Learning Approach to Video Inpainting</title><description>We propose a novel video inpainting algorithm that simultaneously hallucinates missing appearance and motion (optical flow) information, building upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network architectures to enforce plausible texture in static images. In extending DIP to video we make two important contributions. First, we show that coherent video inpainting is possible without a priori training. We take a generative approach to inpainting based on internal (within-video) learning without reliance upon an external corpus of visual data to train a one-size-fits-all model for the large space of general videos. Second, we show that such a framework can jointly generate both appearance and flow, whilst exploiting these complementary modalities to ensure mutual consistency. We show that leveraging appearance statistics specific to each video achieves visually plausible results whilst handling the challenging problem of long-term consistency.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotjssKwjAURLNxIdUPcGV-oDVp0qRZFvEFBReK23LTpBrQtMQi-vfWx2rgzDAchGaUJDzPMrKA8HSPhCqiEiJVJseIFh7vfG-DhysuLQTv_BkXXRdaqC-4b_HJGdsOmw6c74dygkYNXO92-s8IHdar43Ibl_vNblmUMQgpY25TQQnNMy0ET41gWjFlLDdQa0v1pzL1QIRKc940zAgJ0uiBqKbOJYvQ_Pf6Va664G4QXtVHvfqqszfm_D2v</recordid><startdate>20190917</startdate><enddate>20190917</enddate><creator>Zhang, Haotian</creator><creator>Mai, Long</creator><creator>Xu, Ning</creator><creator>Wang, Zhaowen</creator><creator>Collomosse, John</creator><creator>Jin, Hailin</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20190917</creationdate><title>An Internal Learning Approach to Video Inpainting</title><author>Zhang, Haotian ; Mai, Long ; Xu, Ning ; Wang, Zhaowen ; Collomosse, John ; Jin, Hailin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-4e2610185b6642d63b939de4dacbe1b1018dc39d69284ff3d67a7db39d9fc873</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Haotian</creatorcontrib><creatorcontrib>Mai, Long</creatorcontrib><creatorcontrib>Xu, Ning</creatorcontrib><creatorcontrib>Wang, Zhaowen</creatorcontrib><creatorcontrib>Collomosse, John</creatorcontrib><creatorcontrib>Jin, Hailin</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zhang, Haotian</au><au>Mai, Long</au><au>Xu, Ning</au><au>Wang, Zhaowen</au><au>Collomosse, John</au><au>Jin, Hailin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>An Internal Learning Approach to Video Inpainting</atitle><date>2019-09-17</date><risdate>2019</risdate><abstract>We propose a novel video inpainting algorithm that simultaneously hallucinates missing appearance and motion (optical flow) information, building upon the recent 'Deep Image Prior' (DIP) that exploits convolutional network architectures to enforce plausible texture in static images. In extending DIP to video we make two important contributions. First, we show that coherent video inpainting is possible without a priori training. We take a generative approach to inpainting based on internal (within-video) learning without reliance upon an external corpus of visual data to train a one-size-fits-all model for the large space of general videos. Second, we show that such a framework can jointly generate both appearance and flow, whilst exploiting these complementary modalities to ensure mutual consistency. We show that leveraging appearance statistics specific to each video achieves visually plausible results whilst handling the challenging problem of long-term consistency.</abstract><doi>10.48550/arxiv.1909.07957</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1909.07957
ispartof
issn
language eng
recordid cdi_arxiv_primary_1909_07957
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
title An Internal Learning Approach to Video Inpainting
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-13T11%3A01%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=An%20Internal%20Learning%20Approach%20to%20Video%20Inpainting&rft.au=Zhang,%20Haotian&rft.date=2019-09-17&rft_id=info:doi/10.48550/arxiv.1909.07957&rft_dat=%3Carxiv_GOX%3E1909_07957%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true