DAM: Towards A Foundation Model for Time Series Forecasting

It is challenging to scale time series forecasting models such that they forecast accurately for multiple distinct domains and datasets, all with potentially different underlying collection procedures (e.g., sample resolution), patterns (e.g., periodicity), and prediction requirements (e.g., reconst...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Darlow, Luke, Deng, Qiwen, Hassan, Ahmed, Asenov, Martin, Singh, Rajkarn, Joosen, Artjom, Barker, Adam, Storkey, Amos
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Darlow, Luke
Deng, Qiwen
Hassan, Ahmed
Asenov, Martin
Singh, Rajkarn
Joosen, Artjom
Barker, Adam
Storkey, Amos
description It is challenging to scale time series forecasting models such that they forecast accurately for multiple distinct domains and datasets, all with potentially different underlying collection procedures (e.g., sample resolution), patterns (e.g., periodicity), and prediction requirements (e.g., reconstruction vs. forecasting). We call this general task universal forecasting. Existing methods usually assume that input data is regularly sampled, and they forecast to pre-determined horizons, resulting in failure to generalise outside of the scope of their training. We propose the DAM - a neural model that takes randomly sampled histories and outputs an adjustable basis composition as a continuous function of time for forecasting to non-fixed horizons. It involves three key components: (1) a flexible approach for using randomly sampled histories from a long-tail distribution, that enables an efficient global perspective of the underlying temporal dynamics while retaining focus on the recent history; (2) a transformer backbone that is trained on these actively sampled histories to produce, as representational output, (3) the basis coefficients of a continuous function of time. We show that a single univariate DAM, trained on 25 time series datasets, either outperformed or closely matched existing SoTA models at multivariate long-term forecasting across 18 datasets, including 8 held-out for zero-shot transfer, even though these models were trained to specialise for each dataset-horizon combination. This single DAM excels at zero-shot transfer and very-long-term forecasting, performs well at imputation, is interpretable via basis function composition and attention, can be tuned for different inference-cost requirements, is robust to missing and irregularly sampled data {by design}.
doi_str_mv 10.48550/arxiv.2407.17880
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2407_17880</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2407_17880</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2407_178803</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjEw1zM0t7Aw4GSwdnH0tVIIyS9PLEopVnBUcMsvzUtJLMnMz1PwzU9JzVFIyy9SCMnMTVUITi3KTC0GKihKTU4sLsnMS-dhYE1LzClO5YXS3Azybq4hzh66YGviC4oycxOLKuNB1sWDrTMmrAIAMeAzzQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>DAM: Towards A Foundation Model for Time Series Forecasting</title><source>arXiv.org</source><creator>Darlow, Luke ; Deng, Qiwen ; Hassan, Ahmed ; Asenov, Martin ; Singh, Rajkarn ; Joosen, Artjom ; Barker, Adam ; Storkey, Amos</creator><creatorcontrib>Darlow, Luke ; Deng, Qiwen ; Hassan, Ahmed ; Asenov, Martin ; Singh, Rajkarn ; Joosen, Artjom ; Barker, Adam ; Storkey, Amos</creatorcontrib><description>It is challenging to scale time series forecasting models such that they forecast accurately for multiple distinct domains and datasets, all with potentially different underlying collection procedures (e.g., sample resolution), patterns (e.g., periodicity), and prediction requirements (e.g., reconstruction vs. forecasting). We call this general task universal forecasting. Existing methods usually assume that input data is regularly sampled, and they forecast to pre-determined horizons, resulting in failure to generalise outside of the scope of their training. We propose the DAM - a neural model that takes randomly sampled histories and outputs an adjustable basis composition as a continuous function of time for forecasting to non-fixed horizons. It involves three key components: (1) a flexible approach for using randomly sampled histories from a long-tail distribution, that enables an efficient global perspective of the underlying temporal dynamics while retaining focus on the recent history; (2) a transformer backbone that is trained on these actively sampled histories to produce, as representational output, (3) the basis coefficients of a continuous function of time. We show that a single univariate DAM, trained on 25 time series datasets, either outperformed or closely matched existing SoTA models at multivariate long-term forecasting across 18 datasets, including 8 held-out for zero-shot transfer, even though these models were trained to specialise for each dataset-horizon combination. This single DAM excels at zero-shot transfer and very-long-term forecasting, performs well at imputation, is interpretable via basis function composition and attention, can be tuned for different inference-cost requirements, is robust to missing and irregularly sampled data {by design}.</description><identifier>DOI: 10.48550/arxiv.2407.17880</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2024-07</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2407.17880$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2407.17880$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Darlow, Luke</creatorcontrib><creatorcontrib>Deng, Qiwen</creatorcontrib><creatorcontrib>Hassan, Ahmed</creatorcontrib><creatorcontrib>Asenov, Martin</creatorcontrib><creatorcontrib>Singh, Rajkarn</creatorcontrib><creatorcontrib>Joosen, Artjom</creatorcontrib><creatorcontrib>Barker, Adam</creatorcontrib><creatorcontrib>Storkey, Amos</creatorcontrib><title>DAM: Towards A Foundation Model for Time Series Forecasting</title><description>It is challenging to scale time series forecasting models such that they forecast accurately for multiple distinct domains and datasets, all with potentially different underlying collection procedures (e.g., sample resolution), patterns (e.g., periodicity), and prediction requirements (e.g., reconstruction vs. forecasting). We call this general task universal forecasting. Existing methods usually assume that input data is regularly sampled, and they forecast to pre-determined horizons, resulting in failure to generalise outside of the scope of their training. We propose the DAM - a neural model that takes randomly sampled histories and outputs an adjustable basis composition as a continuous function of time for forecasting to non-fixed horizons. It involves three key components: (1) a flexible approach for using randomly sampled histories from a long-tail distribution, that enables an efficient global perspective of the underlying temporal dynamics while retaining focus on the recent history; (2) a transformer backbone that is trained on these actively sampled histories to produce, as representational output, (3) the basis coefficients of a continuous function of time. We show that a single univariate DAM, trained on 25 time series datasets, either outperformed or closely matched existing SoTA models at multivariate long-term forecasting across 18 datasets, including 8 held-out for zero-shot transfer, even though these models were trained to specialise for each dataset-horizon combination. This single DAM excels at zero-shot transfer and very-long-term forecasting, performs well at imputation, is interpretable via basis function composition and attention, can be tuned for different inference-cost requirements, is robust to missing and irregularly sampled data {by design}.</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjEw1zM0t7Aw4GSwdnH0tVIIyS9PLEopVnBUcMsvzUtJLMnMz1PwzU9JzVFIyy9SCMnMTVUITi3KTC0GKihKTU4sLsnMS-dhYE1LzClO5YXS3Azybq4hzh66YGviC4oycxOLKuNB1sWDrTMmrAIAMeAzzQ</recordid><startdate>20240725</startdate><enddate>20240725</enddate><creator>Darlow, Luke</creator><creator>Deng, Qiwen</creator><creator>Hassan, Ahmed</creator><creator>Asenov, Martin</creator><creator>Singh, Rajkarn</creator><creator>Joosen, Artjom</creator><creator>Barker, Adam</creator><creator>Storkey, Amos</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240725</creationdate><title>DAM: Towards A Foundation Model for Time Series Forecasting</title><author>Darlow, Luke ; Deng, Qiwen ; Hassan, Ahmed ; Asenov, Martin ; Singh, Rajkarn ; Joosen, Artjom ; Barker, Adam ; Storkey, Amos</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2407_178803</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Darlow, Luke</creatorcontrib><creatorcontrib>Deng, Qiwen</creatorcontrib><creatorcontrib>Hassan, Ahmed</creatorcontrib><creatorcontrib>Asenov, Martin</creatorcontrib><creatorcontrib>Singh, Rajkarn</creatorcontrib><creatorcontrib>Joosen, Artjom</creatorcontrib><creatorcontrib>Barker, Adam</creatorcontrib><creatorcontrib>Storkey, Amos</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Darlow, Luke</au><au>Deng, Qiwen</au><au>Hassan, Ahmed</au><au>Asenov, Martin</au><au>Singh, Rajkarn</au><au>Joosen, Artjom</au><au>Barker, Adam</au><au>Storkey, Amos</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DAM: Towards A Foundation Model for Time Series Forecasting</atitle><date>2024-07-25</date><risdate>2024</risdate><abstract>It is challenging to scale time series forecasting models such that they forecast accurately for multiple distinct domains and datasets, all with potentially different underlying collection procedures (e.g., sample resolution), patterns (e.g., periodicity), and prediction requirements (e.g., reconstruction vs. forecasting). We call this general task universal forecasting. Existing methods usually assume that input data is regularly sampled, and they forecast to pre-determined horizons, resulting in failure to generalise outside of the scope of their training. We propose the DAM - a neural model that takes randomly sampled histories and outputs an adjustable basis composition as a continuous function of time for forecasting to non-fixed horizons. It involves three key components: (1) a flexible approach for using randomly sampled histories from a long-tail distribution, that enables an efficient global perspective of the underlying temporal dynamics while retaining focus on the recent history; (2) a transformer backbone that is trained on these actively sampled histories to produce, as representational output, (3) the basis coefficients of a continuous function of time. We show that a single univariate DAM, trained on 25 time series datasets, either outperformed or closely matched existing SoTA models at multivariate long-term forecasting across 18 datasets, including 8 held-out for zero-shot transfer, even though these models were trained to specialise for each dataset-horizon combination. This single DAM excels at zero-shot transfer and very-long-term forecasting, performs well at imputation, is interpretable via basis function composition and attention, can be tuned for different inference-cost requirements, is robust to missing and irregularly sampled data {by design}.</abstract><doi>10.48550/arxiv.2407.17880</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2407.17880
ispartof
issn
language eng
recordid cdi_arxiv_primary_2407_17880
source arXiv.org
subjects Computer Science - Learning
title DAM: Towards A Foundation Model for Time Series Forecasting
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T16%3A24%3A16IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DAM:%20Towards%20A%20Foundation%20Model%20for%20Time%20Series%20Forecasting&rft.au=Darlow,%20Luke&rft.date=2024-07-25&rft_id=info:doi/10.48550/arxiv.2407.17880&rft_dat=%3Carxiv_GOX%3E2407_17880%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true