Learning representations of multivariate time series with missing data

•We design a recurrent autoencoder architecture to compress multivariate time series with missing data.•An additional regularization term aligns the learned representations with a prior kernel, which accounts for missing data.•Even with many missing data, time series belonging to different classes b...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Pattern recognition 2019-12, Vol.96, p.106973, Article 106973
Hauptverfasser: Bianchi, Filippo Maria, Livi, Lorenzo, Mikalsen, Karl Øyvind, Kampffmeyer, Michael, Jenssen, Robert
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page 106973
container_title Pattern recognition
container_volume 96
creator Bianchi, Filippo Maria
Livi, Lorenzo
Mikalsen, Karl Øyvind
Kampffmeyer, Michael
Jenssen, Robert
description •We design a recurrent autoencoder architecture to compress multivariate time series with missing data.•An additional regularization term aligns the learned representations with a prior kernel, which accounts for missing data.•Even with many missing data, time series belonging to different classes become well separated in the induced latent space.•We exploit the proposed architecture to design methods for anomaly detection and for imputing missing data.•We perform an analysis to investigate which kind of time series can be effectively encoded using recurrent layers. Learning compressed representations of multivariate time series (MTS) facilitates data analysis in the presence of noise and redundant information, and for a large number of variates and time steps. However, classical dimensionality reduction approaches are designed for vectorial data and cannot deal explicitly with missing values. In this work, we propose a novel autoencoder architecture based on recurrent neural networks to generate compressed representations of MTS. The proposed model can process inputs characterized by variable lengths and it is specifically designed to handle missing data. Our autoencoder learns fixed-length vectorial representations, whose pairwise similarities are aligned to a kernel function that operates in input space and that handles missing values. This allows to learn good representations, even in the presence of a significant amount of missing data. To show the effectiveness of the proposed approach, we evaluate the quality of the learned representations in several classification tasks, including those involving medical data, and we compare to other methods for dimensionality reduction. Successively, we design two frameworks based on the proposed architecture: one for imputing missing data and another for one-class classification. Finally, we analyze under what circumstances an autoencoder with recurrent layers can learn better compressed representations of MTS than feed-forward architectures.
doi_str_mv 10.1016/j.patcog.2019.106973
format Article
fullrecord <record><control><sourceid>elsevier_crist</sourceid><recordid>TN_cdi_cristin_nora_10037_18341</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0031320319302766</els_id><sourcerecordid>S0031320319302766</sourcerecordid><originalsourceid>FETCH-LOGICAL-c376t-4bb25388c5af3b9039189a94d55d110fccafbfab0f85fb8fa337754b4cb75cf3</originalsourceid><addsrcrecordid>eNp9kN1KAzEQhYMoWKtvILgvsDXZJM3ujSDFPyh40_swyU7qlO5uSWLFt3fL6q1XAzPnHOZ8jN0KvhBcLO93iwNkP2wXFRfNuFo2Rp6xmaiNLLVQ1TmbcS5FKSsuL9lVSjvOhRkPM_a8Rog99dsi4iFiwj5DpqFPxRCK7nOf6QiRIGORqcMiYSRMxRflj6KjlE7GFjJcs4sA-4Q3v3PONs9Pm9VruX5_eVs9rksvzTKXyrlKy7r2GoJ0DZeNqBtoVKt1KwQP3kNwARwPtQ6uDiClMVo55Z3RPsg5u5tifaSUqbf9EMGKsZyxopZKjAr1pxhSihjsIVIH8XtU2RMsu7MTLHuCZSdYo-1hsuH4_JEw2uQJe48tRfTZtgP9H_ADYs91BQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Learning representations of multivariate time series with missing data</title><source>NORA - Norwegian Open Research Archives</source><source>Elsevier ScienceDirect Journals Complete</source><creator>Bianchi, Filippo Maria ; Livi, Lorenzo ; Mikalsen, Karl Øyvind ; Kampffmeyer, Michael ; Jenssen, Robert</creator><creatorcontrib>Bianchi, Filippo Maria ; Livi, Lorenzo ; Mikalsen, Karl Øyvind ; Kampffmeyer, Michael ; Jenssen, Robert</creatorcontrib><description>•We design a recurrent autoencoder architecture to compress multivariate time series with missing data.•An additional regularization term aligns the learned representations with a prior kernel, which accounts for missing data.•Even with many missing data, time series belonging to different classes become well separated in the induced latent space.•We exploit the proposed architecture to design methods for anomaly detection and for imputing missing data.•We perform an analysis to investigate which kind of time series can be effectively encoded using recurrent layers. Learning compressed representations of multivariate time series (MTS) facilitates data analysis in the presence of noise and redundant information, and for a large number of variates and time steps. However, classical dimensionality reduction approaches are designed for vectorial data and cannot deal explicitly with missing values. In this work, we propose a novel autoencoder architecture based on recurrent neural networks to generate compressed representations of MTS. The proposed model can process inputs characterized by variable lengths and it is specifically designed to handle missing data. Our autoencoder learns fixed-length vectorial representations, whose pairwise similarities are aligned to a kernel function that operates in input space and that handles missing values. This allows to learn good representations, even in the presence of a significant amount of missing data. To show the effectiveness of the proposed approach, we evaluate the quality of the learned representations in several classification tasks, including those involving medical data, and we compare to other methods for dimensionality reduction. Successively, we design two frameworks based on the proposed architecture: one for imputing missing data and another for one-class classification. Finally, we analyze under what circumstances an autoencoder with recurrent layers can learn better compressed representations of MTS than feed-forward architectures.</description><identifier>ISSN: 0031-3203</identifier><identifier>ISSN: 1873-5142</identifier><identifier>EISSN: 1873-5142</identifier><identifier>DOI: 10.1016/j.patcog.2019.106973</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Autoencoders ; Kernel methods ; Multivariate time series ; Recurrent neural networks ; Representation learning</subject><ispartof>Pattern recognition, 2019-12, Vol.96, p.106973, Article 106973</ispartof><rights>2019</rights><rights>info:eu-repo/semantics/openAccess</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c376t-4bb25388c5af3b9039189a94d55d110fccafbfab0f85fb8fa337754b4cb75cf3</citedby><cites>FETCH-LOGICAL-c376t-4bb25388c5af3b9039189a94d55d110fccafbfab0f85fb8fa337754b4cb75cf3</cites><orcidid>0000-0002-7699-0405 ; 0000-0002-7145-3846 ; 0000-0001-6384-4743</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.patcog.2019.106973$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>230,314,780,784,885,3548,26566,27923,27924,45994</link.rule.ids></links><search><creatorcontrib>Bianchi, Filippo Maria</creatorcontrib><creatorcontrib>Livi, Lorenzo</creatorcontrib><creatorcontrib>Mikalsen, Karl Øyvind</creatorcontrib><creatorcontrib>Kampffmeyer, Michael</creatorcontrib><creatorcontrib>Jenssen, Robert</creatorcontrib><title>Learning representations of multivariate time series with missing data</title><title>Pattern recognition</title><description>•We design a recurrent autoencoder architecture to compress multivariate time series with missing data.•An additional regularization term aligns the learned representations with a prior kernel, which accounts for missing data.•Even with many missing data, time series belonging to different classes become well separated in the induced latent space.•We exploit the proposed architecture to design methods for anomaly detection and for imputing missing data.•We perform an analysis to investigate which kind of time series can be effectively encoded using recurrent layers. Learning compressed representations of multivariate time series (MTS) facilitates data analysis in the presence of noise and redundant information, and for a large number of variates and time steps. However, classical dimensionality reduction approaches are designed for vectorial data and cannot deal explicitly with missing values. In this work, we propose a novel autoencoder architecture based on recurrent neural networks to generate compressed representations of MTS. The proposed model can process inputs characterized by variable lengths and it is specifically designed to handle missing data. Our autoencoder learns fixed-length vectorial representations, whose pairwise similarities are aligned to a kernel function that operates in input space and that handles missing values. This allows to learn good representations, even in the presence of a significant amount of missing data. To show the effectiveness of the proposed approach, we evaluate the quality of the learned representations in several classification tasks, including those involving medical data, and we compare to other methods for dimensionality reduction. Successively, we design two frameworks based on the proposed architecture: one for imputing missing data and another for one-class classification. Finally, we analyze under what circumstances an autoencoder with recurrent layers can learn better compressed representations of MTS than feed-forward architectures.</description><subject>Autoencoders</subject><subject>Kernel methods</subject><subject>Multivariate time series</subject><subject>Recurrent neural networks</subject><subject>Representation learning</subject><issn>0031-3203</issn><issn>1873-5142</issn><issn>1873-5142</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>3HK</sourceid><recordid>eNp9kN1KAzEQhYMoWKtvILgvsDXZJM3ujSDFPyh40_swyU7qlO5uSWLFt3fL6q1XAzPnHOZ8jN0KvhBcLO93iwNkP2wXFRfNuFo2Rp6xmaiNLLVQ1TmbcS5FKSsuL9lVSjvOhRkPM_a8Rog99dsi4iFiwj5DpqFPxRCK7nOf6QiRIGORqcMiYSRMxRflj6KjlE7GFjJcs4sA-4Q3v3PONs9Pm9VruX5_eVs9rksvzTKXyrlKy7r2GoJ0DZeNqBtoVKt1KwQP3kNwARwPtQ6uDiClMVo55Z3RPsg5u5tifaSUqbf9EMGKsZyxopZKjAr1pxhSihjsIVIH8XtU2RMsu7MTLHuCZSdYo-1hsuH4_JEw2uQJe48tRfTZtgP9H_ADYs91BQ</recordid><startdate>20191201</startdate><enddate>20191201</enddate><creator>Bianchi, Filippo Maria</creator><creator>Livi, Lorenzo</creator><creator>Mikalsen, Karl Øyvind</creator><creator>Kampffmeyer, Michael</creator><creator>Jenssen, Robert</creator><general>Elsevier Ltd</general><general>Elsevier</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3HK</scope><orcidid>https://orcid.org/0000-0002-7699-0405</orcidid><orcidid>https://orcid.org/0000-0002-7145-3846</orcidid><orcidid>https://orcid.org/0000-0001-6384-4743</orcidid></search><sort><creationdate>20191201</creationdate><title>Learning representations of multivariate time series with missing data</title><author>Bianchi, Filippo Maria ; Livi, Lorenzo ; Mikalsen, Karl Øyvind ; Kampffmeyer, Michael ; Jenssen, Robert</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c376t-4bb25388c5af3b9039189a94d55d110fccafbfab0f85fb8fa337754b4cb75cf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Autoencoders</topic><topic>Kernel methods</topic><topic>Multivariate time series</topic><topic>Recurrent neural networks</topic><topic>Representation learning</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Bianchi, Filippo Maria</creatorcontrib><creatorcontrib>Livi, Lorenzo</creatorcontrib><creatorcontrib>Mikalsen, Karl Øyvind</creatorcontrib><creatorcontrib>Kampffmeyer, Michael</creatorcontrib><creatorcontrib>Jenssen, Robert</creatorcontrib><collection>CrossRef</collection><collection>NORA - Norwegian Open Research Archives</collection><jtitle>Pattern recognition</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Bianchi, Filippo Maria</au><au>Livi, Lorenzo</au><au>Mikalsen, Karl Øyvind</au><au>Kampffmeyer, Michael</au><au>Jenssen, Robert</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Learning representations of multivariate time series with missing data</atitle><jtitle>Pattern recognition</jtitle><date>2019-12-01</date><risdate>2019</risdate><volume>96</volume><spage>106973</spage><pages>106973-</pages><artnum>106973</artnum><issn>0031-3203</issn><issn>1873-5142</issn><eissn>1873-5142</eissn><abstract>•We design a recurrent autoencoder architecture to compress multivariate time series with missing data.•An additional regularization term aligns the learned representations with a prior kernel, which accounts for missing data.•Even with many missing data, time series belonging to different classes become well separated in the induced latent space.•We exploit the proposed architecture to design methods for anomaly detection and for imputing missing data.•We perform an analysis to investigate which kind of time series can be effectively encoded using recurrent layers. Learning compressed representations of multivariate time series (MTS) facilitates data analysis in the presence of noise and redundant information, and for a large number of variates and time steps. However, classical dimensionality reduction approaches are designed for vectorial data and cannot deal explicitly with missing values. In this work, we propose a novel autoencoder architecture based on recurrent neural networks to generate compressed representations of MTS. The proposed model can process inputs characterized by variable lengths and it is specifically designed to handle missing data. Our autoencoder learns fixed-length vectorial representations, whose pairwise similarities are aligned to a kernel function that operates in input space and that handles missing values. This allows to learn good representations, even in the presence of a significant amount of missing data. To show the effectiveness of the proposed approach, we evaluate the quality of the learned representations in several classification tasks, including those involving medical data, and we compare to other methods for dimensionality reduction. Successively, we design two frameworks based on the proposed architecture: one for imputing missing data and another for one-class classification. Finally, we analyze under what circumstances an autoencoder with recurrent layers can learn better compressed representations of MTS than feed-forward architectures.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.patcog.2019.106973</doi><orcidid>https://orcid.org/0000-0002-7699-0405</orcidid><orcidid>https://orcid.org/0000-0002-7145-3846</orcidid><orcidid>https://orcid.org/0000-0001-6384-4743</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0031-3203
ispartof Pattern recognition, 2019-12, Vol.96, p.106973, Article 106973
issn 0031-3203
1873-5142
1873-5142
language eng
recordid cdi_cristin_nora_10037_18341
source NORA - Norwegian Open Research Archives; Elsevier ScienceDirect Journals Complete
subjects Autoencoders
Kernel methods
Multivariate time series
Recurrent neural networks
Representation learning
title Learning representations of multivariate time series with missing data
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-10T21%3A40%3A10IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_crist&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Learning%20representations%20of%20multivariate%20time%20series%20with%20missing%20data&rft.jtitle=Pattern%20recognition&rft.au=Bianchi,%20Filippo%20Maria&rft.date=2019-12-01&rft.volume=96&rft.spage=106973&rft.pages=106973-&rft.artnum=106973&rft.issn=0031-3203&rft.eissn=1873-5142&rft_id=info:doi/10.1016/j.patcog.2019.106973&rft_dat=%3Celsevier_crist%3ES0031320319302766%3C/elsevier_crist%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_els_id=S0031320319302766&rfr_iscdi=true