Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses

Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Sustainability 2020-09, Vol.12 (18), p.7727
Hauptverfasser: Kurte, Kuldeep, Munk, Jeffrey, Kotevska, Olivera, Amasyali, Kadir, Smith, Robert, McKee, Evan, Du, Yan, Cui, Borui, Kuruganti, Teja, Zandi, Helia
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue 18
container_start_page 7727
container_title Sustainability
container_volume 12
creator Kurte, Kuldeep
Munk, Jeffrey
Kotevska, Olivera
Amasyali, Kadir
Smith, Robert
McKee, Evan
Du, Yan
Cui, Borui
Kuruganti, Teja
Zandi, Helia
description Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability challenges that the DRL agent could face during the deployment phase. Using online learning for such applications is not realistic due to the long learning period and likely poor comfort control during the learning process. Alternatively, DRL can be pre-trained using a building model prior to deployment. However, developing an accurate building model for every house and deploying a pre-trained DRL model for HVAC control would not be cost-effective. In this study, we focus on evaluating the ability of DRL-based HVAC control to provide cost savings when pre-trained on one building model and deployed on different house models with varying user comforts. We observed around 30% of cost reduction by pre-trained model over baseline when validated in a simulation environment and achieved up to 21% cost reduction when deployed in the real house. This finding provides experimental evidence that the pre-trained DRL has the potential to adapt to different house environments and comfort settings.
doi_str_mv 10.3390/su12187727
format Article
fullrecord <record><control><sourceid>crossref_osti_</sourceid><recordid>TN_cdi_osti_scitechconnect_1665985</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>10_3390_su12187727</sourcerecordid><originalsourceid>FETCH-LOGICAL-c294t-b5eaf46c296827bb4d1e37980260742d9057d06dcd3b85879dd370100e0af2cf3</originalsourceid><addsrcrecordid>eNpN0FFLwzAQB_AgCo65Fz9B8FGoXpK2aR9nmU4YCKK-ljS5uEjXjCQV9u3tmKD3cveHH_fwJ-SawZ0QNdzHkXFWScnlGZlxkCxjUMD5v_uSLGL8gmmEYDUrZ6Rdfat-VMkNnzRtkS6N2ifVud6lA_WWvqIbrA8adzgkukEVhiN9UBENXX8sG9r4IQXf00lNOjozQad6uvZjxHhFLqzqIy5-95y8P67emnW2eXl6bpabTPM6T1lXoLJ5OYWy4rLrcsNQyLoCXoLMuamhkAZKo43oqqKStTFCAgNAUJZrK-bk5vTXx-TaqF1CvdV-GFCnlpVlUVfFhG5PSAcfY0Db7oPbqXBoGbTHCtu_CsUPfBhjIg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses</title><source>MDPI - Multidisciplinary Digital Publishing Institute</source><source>EZB-FREE-00999 freely available EZB journals</source><creator>Kurte, Kuldeep ; Munk, Jeffrey ; Kotevska, Olivera ; Amasyali, Kadir ; Smith, Robert ; McKee, Evan ; Du, Yan ; Cui, Borui ; Kuruganti, Teja ; Zandi, Helia</creator><creatorcontrib>Kurte, Kuldeep ; Munk, Jeffrey ; Kotevska, Olivera ; Amasyali, Kadir ; Smith, Robert ; McKee, Evan ; Du, Yan ; Cui, Borui ; Kuruganti, Teja ; Zandi, Helia ; Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)</creatorcontrib><description>Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability challenges that the DRL agent could face during the deployment phase. Using online learning for such applications is not realistic due to the long learning period and likely poor comfort control during the learning process. Alternatively, DRL can be pre-trained using a building model prior to deployment. However, developing an accurate building model for every house and deploying a pre-trained DRL model for HVAC control would not be cost-effective. In this study, we focus on evaluating the ability of DRL-based HVAC control to provide cost savings when pre-trained on one building model and deployed on different house models with varying user comforts. We observed around 30% of cost reduction by pre-trained model over baseline when validated in a simulation environment and achieved up to 21% cost reduction when deployed in the real house. This finding provides experimental evidence that the pre-trained DRL has the potential to adapt to different house environments and comfort settings.</description><identifier>ISSN: 2071-1050</identifier><identifier>EISSN: 2071-1050</identifier><identifier>DOI: 10.3390/su12187727</identifier><language>eng</language><publisher>United States: MDPI</publisher><subject>adaptability ; building energy ; building simulation ; deep reinforcement learning ; demand response ; ENERGY CONSERVATION, CONSUMPTION, AND UTILIZATION ; optimal HVAC control ; smart grid</subject><ispartof>Sustainability, 2020-09, Vol.12 (18), p.7727</ispartof><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c294t-b5eaf46c296827bb4d1e37980260742d9057d06dcd3b85879dd370100e0af2cf3</citedby><cites>FETCH-LOGICAL-c294t-b5eaf46c296827bb4d1e37980260742d9057d06dcd3b85879dd370100e0af2cf3</cites><orcidid>0000-0001-8555-2573 ; 0000-0003-3966-7454 ; 0000-0001-5797-7654 ; 0000000316772243 ; 0000000185552573 ; 0000000305850103 ; 0000000339667454 ; 0000000260581025 ; 0000000337044026 ; 000000021229657X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,776,780,881,27901,27902</link.rule.ids><backlink>$$Uhttps://www.osti.gov/servlets/purl/1665985$$D View this record in Osti.gov$$Hfree_for_read</backlink></links><search><creatorcontrib>Kurte, Kuldeep</creatorcontrib><creatorcontrib>Munk, Jeffrey</creatorcontrib><creatorcontrib>Kotevska, Olivera</creatorcontrib><creatorcontrib>Amasyali, Kadir</creatorcontrib><creatorcontrib>Smith, Robert</creatorcontrib><creatorcontrib>McKee, Evan</creatorcontrib><creatorcontrib>Du, Yan</creatorcontrib><creatorcontrib>Cui, Borui</creatorcontrib><creatorcontrib>Kuruganti, Teja</creatorcontrib><creatorcontrib>Zandi, Helia</creatorcontrib><creatorcontrib>Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)</creatorcontrib><title>Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses</title><title>Sustainability</title><description>Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability challenges that the DRL agent could face during the deployment phase. Using online learning for such applications is not realistic due to the long learning period and likely poor comfort control during the learning process. Alternatively, DRL can be pre-trained using a building model prior to deployment. However, developing an accurate building model for every house and deploying a pre-trained DRL model for HVAC control would not be cost-effective. In this study, we focus on evaluating the ability of DRL-based HVAC control to provide cost savings when pre-trained on one building model and deployed on different house models with varying user comforts. We observed around 30% of cost reduction by pre-trained model over baseline when validated in a simulation environment and achieved up to 21% cost reduction when deployed in the real house. This finding provides experimental evidence that the pre-trained DRL has the potential to adapt to different house environments and comfort settings.</description><subject>adaptability</subject><subject>building energy</subject><subject>building simulation</subject><subject>deep reinforcement learning</subject><subject>demand response</subject><subject>ENERGY CONSERVATION, CONSUMPTION, AND UTILIZATION</subject><subject>optimal HVAC control</subject><subject>smart grid</subject><issn>2071-1050</issn><issn>2071-1050</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNpN0FFLwzAQB_AgCo65Fz9B8FGoXpK2aR9nmU4YCKK-ljS5uEjXjCQV9u3tmKD3cveHH_fwJ-SawZ0QNdzHkXFWScnlGZlxkCxjUMD5v_uSLGL8gmmEYDUrZ6Rdfat-VMkNnzRtkS6N2ifVud6lA_WWvqIbrA8adzgkukEVhiN9UBENXX8sG9r4IQXf00lNOjozQad6uvZjxHhFLqzqIy5-95y8P67emnW2eXl6bpabTPM6T1lXoLJ5OYWy4rLrcsNQyLoCXoLMuamhkAZKo43oqqKStTFCAgNAUJZrK-bk5vTXx-TaqF1CvdV-GFCnlpVlUVfFhG5PSAcfY0Db7oPbqXBoGbTHCtu_CsUPfBhjIg</recordid><startdate>20200918</startdate><enddate>20200918</enddate><creator>Kurte, Kuldeep</creator><creator>Munk, Jeffrey</creator><creator>Kotevska, Olivera</creator><creator>Amasyali, Kadir</creator><creator>Smith, Robert</creator><creator>McKee, Evan</creator><creator>Du, Yan</creator><creator>Cui, Borui</creator><creator>Kuruganti, Teja</creator><creator>Zandi, Helia</creator><general>MDPI</general><scope>AAYXX</scope><scope>CITATION</scope><scope>OIOZB</scope><scope>OTOTI</scope><orcidid>https://orcid.org/0000-0001-8555-2573</orcidid><orcidid>https://orcid.org/0000-0003-3966-7454</orcidid><orcidid>https://orcid.org/0000-0001-5797-7654</orcidid><orcidid>https://orcid.org/0000000316772243</orcidid><orcidid>https://orcid.org/0000000185552573</orcidid><orcidid>https://orcid.org/0000000305850103</orcidid><orcidid>https://orcid.org/0000000339667454</orcidid><orcidid>https://orcid.org/0000000260581025</orcidid><orcidid>https://orcid.org/0000000337044026</orcidid><orcidid>https://orcid.org/000000021229657X</orcidid></search><sort><creationdate>20200918</creationdate><title>Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses</title><author>Kurte, Kuldeep ; Munk, Jeffrey ; Kotevska, Olivera ; Amasyali, Kadir ; Smith, Robert ; McKee, Evan ; Du, Yan ; Cui, Borui ; Kuruganti, Teja ; Zandi, Helia</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c294t-b5eaf46c296827bb4d1e37980260742d9057d06dcd3b85879dd370100e0af2cf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>adaptability</topic><topic>building energy</topic><topic>building simulation</topic><topic>deep reinforcement learning</topic><topic>demand response</topic><topic>ENERGY CONSERVATION, CONSUMPTION, AND UTILIZATION</topic><topic>optimal HVAC control</topic><topic>smart grid</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Kurte, Kuldeep</creatorcontrib><creatorcontrib>Munk, Jeffrey</creatorcontrib><creatorcontrib>Kotevska, Olivera</creatorcontrib><creatorcontrib>Amasyali, Kadir</creatorcontrib><creatorcontrib>Smith, Robert</creatorcontrib><creatorcontrib>McKee, Evan</creatorcontrib><creatorcontrib>Du, Yan</creatorcontrib><creatorcontrib>Cui, Borui</creatorcontrib><creatorcontrib>Kuruganti, Teja</creatorcontrib><creatorcontrib>Zandi, Helia</creatorcontrib><creatorcontrib>Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)</creatorcontrib><collection>CrossRef</collection><collection>OSTI.GOV - Hybrid</collection><collection>OSTI.GOV</collection><jtitle>Sustainability</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Kurte, Kuldeep</au><au>Munk, Jeffrey</au><au>Kotevska, Olivera</au><au>Amasyali, Kadir</au><au>Smith, Robert</au><au>McKee, Evan</au><au>Du, Yan</au><au>Cui, Borui</au><au>Kuruganti, Teja</au><au>Zandi, Helia</au><aucorp>Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)</aucorp><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses</atitle><jtitle>Sustainability</jtitle><date>2020-09-18</date><risdate>2020</risdate><volume>12</volume><issue>18</issue><spage>7727</spage><pages>7727-</pages><issn>2071-1050</issn><eissn>2071-1050</eissn><abstract>Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability challenges that the DRL agent could face during the deployment phase. Using online learning for such applications is not realistic due to the long learning period and likely poor comfort control during the learning process. Alternatively, DRL can be pre-trained using a building model prior to deployment. However, developing an accurate building model for every house and deploying a pre-trained DRL model for HVAC control would not be cost-effective. In this study, we focus on evaluating the ability of DRL-based HVAC control to provide cost savings when pre-trained on one building model and deployed on different house models with varying user comforts. We observed around 30% of cost reduction by pre-trained model over baseline when validated in a simulation environment and achieved up to 21% cost reduction when deployed in the real house. This finding provides experimental evidence that the pre-trained DRL has the potential to adapt to different house environments and comfort settings.</abstract><cop>United States</cop><pub>MDPI</pub><doi>10.3390/su12187727</doi><orcidid>https://orcid.org/0000-0001-8555-2573</orcidid><orcidid>https://orcid.org/0000-0003-3966-7454</orcidid><orcidid>https://orcid.org/0000-0001-5797-7654</orcidid><orcidid>https://orcid.org/0000000316772243</orcidid><orcidid>https://orcid.org/0000000185552573</orcidid><orcidid>https://orcid.org/0000000305850103</orcidid><orcidid>https://orcid.org/0000000339667454</orcidid><orcidid>https://orcid.org/0000000260581025</orcidid><orcidid>https://orcid.org/0000000337044026</orcidid><orcidid>https://orcid.org/000000021229657X</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2071-1050
ispartof Sustainability, 2020-09, Vol.12 (18), p.7727
issn 2071-1050
2071-1050
language eng
recordid cdi_osti_scitechconnect_1665985
source MDPI - Multidisciplinary Digital Publishing Institute; EZB-FREE-00999 freely available EZB journals
subjects adaptability
building energy
building simulation
deep reinforcement learning
demand response
ENERGY CONSERVATION, CONSUMPTION, AND UTILIZATION
optimal HVAC control
smart grid
title Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T23%3A44%3A23IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-crossref_osti_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Evaluating%20the%20Adaptability%20of%20Reinforcement%20Learning%20Based%20HVAC%20Control%20for%20Residential%20Houses&rft.jtitle=Sustainability&rft.au=Kurte,%20Kuldeep&rft.aucorp=Oak%20Ridge%20National%20Lab.%20(ORNL),%20Oak%20Ridge,%20TN%20(United%20States)&rft.date=2020-09-18&rft.volume=12&rft.issue=18&rft.spage=7727&rft.pages=7727-&rft.issn=2071-1050&rft.eissn=2071-1050&rft_id=info:doi/10.3390/su12187727&rft_dat=%3Ccrossref_osti_%3E10_3390_su12187727%3C/crossref_osti_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true