Online Residential Demand Response via Contextual Multi-Armed Bandits

Residential loads have great potential to enhance the efficiency and reliability of electricity systems via demand response (DR) programs. One major challenge in residential DR is to handle the unknown and uncertain customer behaviors. Previous works use learning techniques to predict customer DR be...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Chen, Xin, Nie, Yutong, Li, Na
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Chen, Xin
Nie, Yutong
Li, Na
description Residential loads have great potential to enhance the efficiency and reliability of electricity systems via demand response (DR) programs. One major challenge in residential DR is to handle the unknown and uncertain customer behaviors. Previous works use learning techniques to predict customer DR behaviors, while the influence of time-varying environmental factors is generally neglected, which may lead to inaccurate prediction and inefficient load adjustment. In this paper, we consider the residential DR problem where the load service entity (LSE) aims to select an optimal subset of customers to maximize the expected load reduction with a financial budget. To learn the uncertain customer behaviors under the environmental influence, we formulate the residential DR as a contextual multi-armed bandit (MAB) problem, and the online learning and selection (OLS) algorithm based on Thompson sampling is proposed to solve it. This algorithm takes the contextual information into consideration and is applicable to complicated DR settings. Numerical simulations are performed to demonstrate the learning effectiveness of the proposed algorithm.
doi_str_mv 10.48550/arxiv.2003.03627
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2003_03627</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2003_03627</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-7718571e89212013b3374648ab2b78f33b3ec5fee5e2b7aeb1975d5e05a0b2d53</originalsourceid><addsrcrecordid>eNotj8tqwzAQRbXpoqT9gK6qH7Crh8ejLFM3fUBKoGRvRtUYBLYSbCWkf18n7erCuZcLR4gHrcrKAagnGs_xVBqlbKlsbfBWrLepj4nlF08xcMqRevnCA6VwQYd9mlieIslmnzKf83GuP499jsVqHDjI53kY83QnbjrqJ77_z4XYva53zXux2b59NKtNQTVigagdoGa3NNoobb21WNWVI288us7OgL-hYwaeAbHXS4QArICUNwHsQjz-3V492sMYBxp_2otPe_Wxv1QARPk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Online Residential Demand Response via Contextual Multi-Armed Bandits</title><source>arXiv.org</source><creator>Chen, Xin ; Nie, Yutong ; Li, Na</creator><creatorcontrib>Chen, Xin ; Nie, Yutong ; Li, Na</creatorcontrib><description>Residential loads have great potential to enhance the efficiency and reliability of electricity systems via demand response (DR) programs. One major challenge in residential DR is to handle the unknown and uncertain customer behaviors. Previous works use learning techniques to predict customer DR behaviors, while the influence of time-varying environmental factors is generally neglected, which may lead to inaccurate prediction and inefficient load adjustment. In this paper, we consider the residential DR problem where the load service entity (LSE) aims to select an optimal subset of customers to maximize the expected load reduction with a financial budget. To learn the uncertain customer behaviors under the environmental influence, we formulate the residential DR as a contextual multi-armed bandit (MAB) problem, and the online learning and selection (OLS) algorithm based on Thompson sampling is proposed to solve it. This algorithm takes the contextual information into consideration and is applicable to complicated DR settings. Numerical simulations are performed to demonstrate the learning effectiveness of the proposed algorithm.</description><identifier>DOI: 10.48550/arxiv.2003.03627</identifier><language>eng</language><subject>Computer Science - Learning ; Computer Science - Systems and Control</subject><creationdate>2020-03</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2003.03627$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2003.03627$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chen, Xin</creatorcontrib><creatorcontrib>Nie, Yutong</creatorcontrib><creatorcontrib>Li, Na</creatorcontrib><title>Online Residential Demand Response via Contextual Multi-Armed Bandits</title><description>Residential loads have great potential to enhance the efficiency and reliability of electricity systems via demand response (DR) programs. One major challenge in residential DR is to handle the unknown and uncertain customer behaviors. Previous works use learning techniques to predict customer DR behaviors, while the influence of time-varying environmental factors is generally neglected, which may lead to inaccurate prediction and inefficient load adjustment. In this paper, we consider the residential DR problem where the load service entity (LSE) aims to select an optimal subset of customers to maximize the expected load reduction with a financial budget. To learn the uncertain customer behaviors under the environmental influence, we formulate the residential DR as a contextual multi-armed bandit (MAB) problem, and the online learning and selection (OLS) algorithm based on Thompson sampling is proposed to solve it. This algorithm takes the contextual information into consideration and is applicable to complicated DR settings. Numerical simulations are performed to demonstrate the learning effectiveness of the proposed algorithm.</description><subject>Computer Science - Learning</subject><subject>Computer Science - Systems and Control</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj8tqwzAQRbXpoqT9gK6qH7Crh8ejLFM3fUBKoGRvRtUYBLYSbCWkf18n7erCuZcLR4gHrcrKAagnGs_xVBqlbKlsbfBWrLepj4nlF08xcMqRevnCA6VwQYd9mlieIslmnzKf83GuP499jsVqHDjI53kY83QnbjrqJ77_z4XYva53zXux2b59NKtNQTVigagdoGa3NNoobb21WNWVI288us7OgL-hYwaeAbHXS4QArICUNwHsQjz-3V492sMYBxp_2otPe_Wxv1QARPk</recordid><startdate>20200307</startdate><enddate>20200307</enddate><creator>Chen, Xin</creator><creator>Nie, Yutong</creator><creator>Li, Na</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20200307</creationdate><title>Online Residential Demand Response via Contextual Multi-Armed Bandits</title><author>Chen, Xin ; Nie, Yutong ; Li, Na</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-7718571e89212013b3374648ab2b78f33b3ec5fee5e2b7aeb1975d5e05a0b2d53</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Learning</topic><topic>Computer Science - Systems and Control</topic><toplevel>online_resources</toplevel><creatorcontrib>Chen, Xin</creatorcontrib><creatorcontrib>Nie, Yutong</creatorcontrib><creatorcontrib>Li, Na</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chen, Xin</au><au>Nie, Yutong</au><au>Li, Na</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Online Residential Demand Response via Contextual Multi-Armed Bandits</atitle><date>2020-03-07</date><risdate>2020</risdate><abstract>Residential loads have great potential to enhance the efficiency and reliability of electricity systems via demand response (DR) programs. One major challenge in residential DR is to handle the unknown and uncertain customer behaviors. Previous works use learning techniques to predict customer DR behaviors, while the influence of time-varying environmental factors is generally neglected, which may lead to inaccurate prediction and inefficient load adjustment. In this paper, we consider the residential DR problem where the load service entity (LSE) aims to select an optimal subset of customers to maximize the expected load reduction with a financial budget. To learn the uncertain customer behaviors under the environmental influence, we formulate the residential DR as a contextual multi-armed bandit (MAB) problem, and the online learning and selection (OLS) algorithm based on Thompson sampling is proposed to solve it. This algorithm takes the contextual information into consideration and is applicable to complicated DR settings. Numerical simulations are performed to demonstrate the learning effectiveness of the proposed algorithm.</abstract><doi>10.48550/arxiv.2003.03627</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2003.03627
ispartof
issn
language eng
recordid cdi_arxiv_primary_2003_03627
source arXiv.org
subjects Computer Science - Learning
Computer Science - Systems and Control
title Online Residential Demand Response via Contextual Multi-Armed Bandits
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T00%3A50%3A52IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Online%20Residential%20Demand%20Response%20via%20Contextual%20Multi-Armed%20Bandits&rft.au=Chen,%20Xin&rft.date=2020-03-07&rft_id=info:doi/10.48550/arxiv.2003.03627&rft_dat=%3Carxiv_GOX%3E2003_03627%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true