Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses
Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability...
Gespeichert in:
Veröffentlicht in: | Sustainability 2020-09, Vol.12 (18), p.7727 |
---|---|
Hauptverfasser: | , , , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | 18 |
container_start_page | 7727 |
container_title | Sustainability |
container_volume | 12 |
creator | Kurte, Kuldeep Munk, Jeffrey Kotevska, Olivera Amasyali, Kadir Smith, Robert McKee, Evan Du, Yan Cui, Borui Kuruganti, Teja Zandi, Helia |
description | Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability challenges that the DRL agent could face during the deployment phase. Using online learning for such applications is not realistic due to the long learning period and likely poor comfort control during the learning process. Alternatively, DRL can be pre-trained using a building model prior to deployment. However, developing an accurate building model for every house and deploying a pre-trained DRL model for HVAC control would not be cost-effective. In this study, we focus on evaluating the ability of DRL-based HVAC control to provide cost savings when pre-trained on one building model and deployed on different house models with varying user comforts. We observed around 30% of cost reduction by pre-trained model over baseline when validated in a simulation environment and achieved up to 21% cost reduction when deployed in the real house. This finding provides experimental evidence that the pre-trained DRL has the potential to adapt to different house environments and comfort settings. |
doi_str_mv | 10.3390/su12187727 |
format | Article |
fullrecord | <record><control><sourceid>crossref_osti_</sourceid><recordid>TN_cdi_osti_scitechconnect_1665985</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>10_3390_su12187727</sourcerecordid><originalsourceid>FETCH-LOGICAL-c294t-b5eaf46c296827bb4d1e37980260742d9057d06dcd3b85879dd370100e0af2cf3</originalsourceid><addsrcrecordid>eNpN0FFLwzAQB_AgCo65Fz9B8FGoXpK2aR9nmU4YCKK-ljS5uEjXjCQV9u3tmKD3cveHH_fwJ-SawZ0QNdzHkXFWScnlGZlxkCxjUMD5v_uSLGL8gmmEYDUrZ6Rdfat-VMkNnzRtkS6N2ifVud6lA_WWvqIbrA8adzgkukEVhiN9UBENXX8sG9r4IQXf00lNOjozQad6uvZjxHhFLqzqIy5-95y8P67emnW2eXl6bpabTPM6T1lXoLJ5OYWy4rLrcsNQyLoCXoLMuamhkAZKo43oqqKStTFCAgNAUJZrK-bk5vTXx-TaqF1CvdV-GFCnlpVlUVfFhG5PSAcfY0Db7oPbqXBoGbTHCtu_CsUPfBhjIg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses</title><source>MDPI - Multidisciplinary Digital Publishing Institute</source><source>EZB-FREE-00999 freely available EZB journals</source><creator>Kurte, Kuldeep ; Munk, Jeffrey ; Kotevska, Olivera ; Amasyali, Kadir ; Smith, Robert ; McKee, Evan ; Du, Yan ; Cui, Borui ; Kuruganti, Teja ; Zandi, Helia</creator><creatorcontrib>Kurte, Kuldeep ; Munk, Jeffrey ; Kotevska, Olivera ; Amasyali, Kadir ; Smith, Robert ; McKee, Evan ; Du, Yan ; Cui, Borui ; Kuruganti, Teja ; Zandi, Helia ; Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)</creatorcontrib><description>Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability challenges that the DRL agent could face during the deployment phase. Using online learning for such applications is not realistic due to the long learning period and likely poor comfort control during the learning process. Alternatively, DRL can be pre-trained using a building model prior to deployment. However, developing an accurate building model for every house and deploying a pre-trained DRL model for HVAC control would not be cost-effective. In this study, we focus on evaluating the ability of DRL-based HVAC control to provide cost savings when pre-trained on one building model and deployed on different house models with varying user comforts. We observed around 30% of cost reduction by pre-trained model over baseline when validated in a simulation environment and achieved up to 21% cost reduction when deployed in the real house. This finding provides experimental evidence that the pre-trained DRL has the potential to adapt to different house environments and comfort settings.</description><identifier>ISSN: 2071-1050</identifier><identifier>EISSN: 2071-1050</identifier><identifier>DOI: 10.3390/su12187727</identifier><language>eng</language><publisher>United States: MDPI</publisher><subject>adaptability ; building energy ; building simulation ; deep reinforcement learning ; demand response ; ENERGY CONSERVATION, CONSUMPTION, AND UTILIZATION ; optimal HVAC control ; smart grid</subject><ispartof>Sustainability, 2020-09, Vol.12 (18), p.7727</ispartof><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c294t-b5eaf46c296827bb4d1e37980260742d9057d06dcd3b85879dd370100e0af2cf3</citedby><cites>FETCH-LOGICAL-c294t-b5eaf46c296827bb4d1e37980260742d9057d06dcd3b85879dd370100e0af2cf3</cites><orcidid>0000-0001-8555-2573 ; 0000-0003-3966-7454 ; 0000-0001-5797-7654 ; 0000000316772243 ; 0000000185552573 ; 0000000305850103 ; 0000000339667454 ; 0000000260581025 ; 0000000337044026 ; 000000021229657X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,776,780,881,27901,27902</link.rule.ids><backlink>$$Uhttps://www.osti.gov/servlets/purl/1665985$$D View this record in Osti.gov$$Hfree_for_read</backlink></links><search><creatorcontrib>Kurte, Kuldeep</creatorcontrib><creatorcontrib>Munk, Jeffrey</creatorcontrib><creatorcontrib>Kotevska, Olivera</creatorcontrib><creatorcontrib>Amasyali, Kadir</creatorcontrib><creatorcontrib>Smith, Robert</creatorcontrib><creatorcontrib>McKee, Evan</creatorcontrib><creatorcontrib>Du, Yan</creatorcontrib><creatorcontrib>Cui, Borui</creatorcontrib><creatorcontrib>Kuruganti, Teja</creatorcontrib><creatorcontrib>Zandi, Helia</creatorcontrib><creatorcontrib>Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)</creatorcontrib><title>Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses</title><title>Sustainability</title><description>Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability challenges that the DRL agent could face during the deployment phase. Using online learning for such applications is not realistic due to the long learning period and likely poor comfort control during the learning process. Alternatively, DRL can be pre-trained using a building model prior to deployment. However, developing an accurate building model for every house and deploying a pre-trained DRL model for HVAC control would not be cost-effective. In this study, we focus on evaluating the ability of DRL-based HVAC control to provide cost savings when pre-trained on one building model and deployed on different house models with varying user comforts. We observed around 30% of cost reduction by pre-trained model over baseline when validated in a simulation environment and achieved up to 21% cost reduction when deployed in the real house. This finding provides experimental evidence that the pre-trained DRL has the potential to adapt to different house environments and comfort settings.</description><subject>adaptability</subject><subject>building energy</subject><subject>building simulation</subject><subject>deep reinforcement learning</subject><subject>demand response</subject><subject>ENERGY CONSERVATION, CONSUMPTION, AND UTILIZATION</subject><subject>optimal HVAC control</subject><subject>smart grid</subject><issn>2071-1050</issn><issn>2071-1050</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNpN0FFLwzAQB_AgCo65Fz9B8FGoXpK2aR9nmU4YCKK-ljS5uEjXjCQV9u3tmKD3cveHH_fwJ-SawZ0QNdzHkXFWScnlGZlxkCxjUMD5v_uSLGL8gmmEYDUrZ6Rdfat-VMkNnzRtkS6N2ifVud6lA_WWvqIbrA8adzgkukEVhiN9UBENXX8sG9r4IQXf00lNOjozQad6uvZjxHhFLqzqIy5-95y8P67emnW2eXl6bpabTPM6T1lXoLJ5OYWy4rLrcsNQyLoCXoLMuamhkAZKo43oqqKStTFCAgNAUJZrK-bk5vTXx-TaqF1CvdV-GFCnlpVlUVfFhG5PSAcfY0Db7oPbqXBoGbTHCtu_CsUPfBhjIg</recordid><startdate>20200918</startdate><enddate>20200918</enddate><creator>Kurte, Kuldeep</creator><creator>Munk, Jeffrey</creator><creator>Kotevska, Olivera</creator><creator>Amasyali, Kadir</creator><creator>Smith, Robert</creator><creator>McKee, Evan</creator><creator>Du, Yan</creator><creator>Cui, Borui</creator><creator>Kuruganti, Teja</creator><creator>Zandi, Helia</creator><general>MDPI</general><scope>AAYXX</scope><scope>CITATION</scope><scope>OIOZB</scope><scope>OTOTI</scope><orcidid>https://orcid.org/0000-0001-8555-2573</orcidid><orcidid>https://orcid.org/0000-0003-3966-7454</orcidid><orcidid>https://orcid.org/0000-0001-5797-7654</orcidid><orcidid>https://orcid.org/0000000316772243</orcidid><orcidid>https://orcid.org/0000000185552573</orcidid><orcidid>https://orcid.org/0000000305850103</orcidid><orcidid>https://orcid.org/0000000339667454</orcidid><orcidid>https://orcid.org/0000000260581025</orcidid><orcidid>https://orcid.org/0000000337044026</orcidid><orcidid>https://orcid.org/000000021229657X</orcidid></search><sort><creationdate>20200918</creationdate><title>Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses</title><author>Kurte, Kuldeep ; Munk, Jeffrey ; Kotevska, Olivera ; Amasyali, Kadir ; Smith, Robert ; McKee, Evan ; Du, Yan ; Cui, Borui ; Kuruganti, Teja ; Zandi, Helia</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c294t-b5eaf46c296827bb4d1e37980260742d9057d06dcd3b85879dd370100e0af2cf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>adaptability</topic><topic>building energy</topic><topic>building simulation</topic><topic>deep reinforcement learning</topic><topic>demand response</topic><topic>ENERGY CONSERVATION, CONSUMPTION, AND UTILIZATION</topic><topic>optimal HVAC control</topic><topic>smart grid</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Kurte, Kuldeep</creatorcontrib><creatorcontrib>Munk, Jeffrey</creatorcontrib><creatorcontrib>Kotevska, Olivera</creatorcontrib><creatorcontrib>Amasyali, Kadir</creatorcontrib><creatorcontrib>Smith, Robert</creatorcontrib><creatorcontrib>McKee, Evan</creatorcontrib><creatorcontrib>Du, Yan</creatorcontrib><creatorcontrib>Cui, Borui</creatorcontrib><creatorcontrib>Kuruganti, Teja</creatorcontrib><creatorcontrib>Zandi, Helia</creatorcontrib><creatorcontrib>Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)</creatorcontrib><collection>CrossRef</collection><collection>OSTI.GOV - Hybrid</collection><collection>OSTI.GOV</collection><jtitle>Sustainability</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Kurte, Kuldeep</au><au>Munk, Jeffrey</au><au>Kotevska, Olivera</au><au>Amasyali, Kadir</au><au>Smith, Robert</au><au>McKee, Evan</au><au>Du, Yan</au><au>Cui, Borui</au><au>Kuruganti, Teja</au><au>Zandi, Helia</au><aucorp>Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)</aucorp><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses</atitle><jtitle>Sustainability</jtitle><date>2020-09-18</date><risdate>2020</risdate><volume>12</volume><issue>18</issue><spage>7727</spage><pages>7727-</pages><issn>2071-1050</issn><eissn>2071-1050</eissn><abstract>Intelligent Heating, Ventilation, and Air Conditioning (HVAC) control using deep reinforcement learning (DRL) has recently gained a lot of attention due to its ability to optimally control the complex behavior of the HVAC system. However, more exploration is needed on understanding the adaptability challenges that the DRL agent could face during the deployment phase. Using online learning for such applications is not realistic due to the long learning period and likely poor comfort control during the learning process. Alternatively, DRL can be pre-trained using a building model prior to deployment. However, developing an accurate building model for every house and deploying a pre-trained DRL model for HVAC control would not be cost-effective. In this study, we focus on evaluating the ability of DRL-based HVAC control to provide cost savings when pre-trained on one building model and deployed on different house models with varying user comforts. We observed around 30% of cost reduction by pre-trained model over baseline when validated in a simulation environment and achieved up to 21% cost reduction when deployed in the real house. This finding provides experimental evidence that the pre-trained DRL has the potential to adapt to different house environments and comfort settings.</abstract><cop>United States</cop><pub>MDPI</pub><doi>10.3390/su12187727</doi><orcidid>https://orcid.org/0000-0001-8555-2573</orcidid><orcidid>https://orcid.org/0000-0003-3966-7454</orcidid><orcidid>https://orcid.org/0000-0001-5797-7654</orcidid><orcidid>https://orcid.org/0000000316772243</orcidid><orcidid>https://orcid.org/0000000185552573</orcidid><orcidid>https://orcid.org/0000000305850103</orcidid><orcidid>https://orcid.org/0000000339667454</orcidid><orcidid>https://orcid.org/0000000260581025</orcidid><orcidid>https://orcid.org/0000000337044026</orcidid><orcidid>https://orcid.org/000000021229657X</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2071-1050 |
ispartof | Sustainability, 2020-09, Vol.12 (18), p.7727 |
issn | 2071-1050 2071-1050 |
language | eng |
recordid | cdi_osti_scitechconnect_1665985 |
source | MDPI - Multidisciplinary Digital Publishing Institute; EZB-FREE-00999 freely available EZB journals |
subjects | adaptability building energy building simulation deep reinforcement learning demand response ENERGY CONSERVATION, CONSUMPTION, AND UTILIZATION optimal HVAC control smart grid |
title | Evaluating the Adaptability of Reinforcement Learning Based HVAC Control for Residential Houses |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T23%3A44%3A23IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-crossref_osti_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Evaluating%20the%20Adaptability%20of%20Reinforcement%20Learning%20Based%20HVAC%20Control%20for%20Residential%20Houses&rft.jtitle=Sustainability&rft.au=Kurte,%20Kuldeep&rft.aucorp=Oak%20Ridge%20National%20Lab.%20(ORNL),%20Oak%20Ridge,%20TN%20(United%20States)&rft.date=2020-09-18&rft.volume=12&rft.issue=18&rft.spage=7727&rft.pages=7727-&rft.issn=2071-1050&rft.eissn=2071-1050&rft_id=info:doi/10.3390/su12187727&rft_dat=%3Ccrossref_osti_%3E10_3390_su12187727%3C/crossref_osti_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |