A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite
Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods...
Gespeichert in:
Veröffentlicht in: | Wireless communications and mobile computing 2021, Vol.2021 (1) |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | 1 |
container_start_page | |
container_title | Wireless communications and mobile computing |
container_volume | 2021 |
creator | Zhang, Jian Wu, Fengge |
description | Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods in real-world systems faces many challenges, such as insufficient samples, exploration safety issues, unknown actuator delays, and noise in the raw sensor data. In this work, a mixed model with different input sizes was proposed to represent the environmental dynamics model. The predication accuracy of the environmental dynamics model and the performance of the policy trained in this paper were gradually improved. Our method reduces the impact of noisy data on the model’s accuracy and improves the sampling efficiency. The experiments showed that the agent trained with our method completed a goal-state reaching task in a real-world system under wireless circumstances whose actuators were reaction wheels, whereas the soft actor-critic method failed in the same training process. The method’s effectiveness is ensured theoretically under given conditions. |
doi_str_mv | 10.1155/2021/7331894 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2550177337</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2550177337</sourcerecordid><originalsourceid>FETCH-LOGICAL-c337t-e31d05eaf65f40f115beca991ddd6c7721d1ba672664d69f10b51c0ad4defcf33</originalsourceid><addsrcrecordid>eNp9kMtOwzAQRS0EEqWw4wMssYRQO47tZlkiXlIBidc2uPGYukrjYjug_j2uWrFkNXdx7ozmIHRKySWlnI9yktORZIyOy2IPDShnJBsLKff_sigP0VEIC0IIS_AAfUzwo_uGFj84DW12pQJo_Ay2M843sIQu4iko39nuE09itLHXgCvXRe9SB-LcaZxQ_G597FWbqqq1cY1fVIQ2JThGB0a1AU52c4jebq5fq7ts-nR7X02mWcOYjBkwqgkHZQQ3BTHpnRk0qiyp1lo0UuZU05kSMhei0KI0lMw4bYjShQbTGMaG6Gy7d-XdVw8h1gvX-y6drHPOCZXJi0zUxZZqvAvBg6lX3i6VX9eU1BuH9cZhvXOY8PMtPredVj_2f_oX4yhxTw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2550177337</pqid></control><display><type>article</type><title>A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite</title><source>EZB-FREE-00999 freely available EZB journals</source><source>Wiley Online Library (Open Access Collection)</source><source>Alma/SFX Local Collection</source><creator>Zhang, Jian ; Wu, Fengge</creator><contributor>Nagaraj, Balakrishnan ; Balakrishnan Nagaraj</contributor><creatorcontrib>Zhang, Jian ; Wu, Fengge ; Nagaraj, Balakrishnan ; Balakrishnan Nagaraj</creatorcontrib><description>Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods in real-world systems faces many challenges, such as insufficient samples, exploration safety issues, unknown actuator delays, and noise in the raw sensor data. In this work, a mixed model with different input sizes was proposed to represent the environmental dynamics model. The predication accuracy of the environmental dynamics model and the performance of the policy trained in this paper were gradually improved. Our method reduces the impact of noisy data on the model’s accuracy and improves the sampling efficiency. The experiments showed that the agent trained with our method completed a goal-state reaching task in a real-world system under wireless circumstances whose actuators were reaction wheels, whereas the soft actor-critic method failed in the same training process. The method’s effectiveness is ensured theoretically under given conditions.</description><identifier>ISSN: 1530-8669</identifier><identifier>EISSN: 1530-8677</identifier><identifier>DOI: 10.1155/2021/7331894</identifier><language>eng</language><publisher>Oxford: Hindawi</publisher><subject>Actuators ; Control methods ; Control theory ; Learning ; Methods ; Model accuracy ; Neural networks ; Reaction wheels ; Satellite attitude control ; Satellite observation ; Satellites ; Sensors ; Simulation ; Virtual reality</subject><ispartof>Wireless communications and mobile computing, 2021, Vol.2021 (1)</ispartof><rights>Copyright © 2021 Jian Zhang and Fengge Wu.</rights><rights>Copyright © 2021 Jian Zhang and Fengge Wu. This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c337t-e31d05eaf65f40f115beca991ddd6c7721d1ba672664d69f10b51c0ad4defcf33</citedby><cites>FETCH-LOGICAL-c337t-e31d05eaf65f40f115beca991ddd6c7721d1ba672664d69f10b51c0ad4defcf33</cites><orcidid>0000-0002-1958-8321 ; 0000-0003-1375-4947</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,4024,27923,27924,27925</link.rule.ids></links><search><contributor>Nagaraj, Balakrishnan</contributor><contributor>Balakrishnan Nagaraj</contributor><creatorcontrib>Zhang, Jian</creatorcontrib><creatorcontrib>Wu, Fengge</creatorcontrib><title>A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite</title><title>Wireless communications and mobile computing</title><description>Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods in real-world systems faces many challenges, such as insufficient samples, exploration safety issues, unknown actuator delays, and noise in the raw sensor data. In this work, a mixed model with different input sizes was proposed to represent the environmental dynamics model. The predication accuracy of the environmental dynamics model and the performance of the policy trained in this paper were gradually improved. Our method reduces the impact of noisy data on the model’s accuracy and improves the sampling efficiency. The experiments showed that the agent trained with our method completed a goal-state reaching task in a real-world system under wireless circumstances whose actuators were reaction wheels, whereas the soft actor-critic method failed in the same training process. The method’s effectiveness is ensured theoretically under given conditions.</description><subject>Actuators</subject><subject>Control methods</subject><subject>Control theory</subject><subject>Learning</subject><subject>Methods</subject><subject>Model accuracy</subject><subject>Neural networks</subject><subject>Reaction wheels</subject><subject>Satellite attitude control</subject><subject>Satellite observation</subject><subject>Satellites</subject><subject>Sensors</subject><subject>Simulation</subject><subject>Virtual reality</subject><issn>1530-8669</issn><issn>1530-8677</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>RHX</sourceid><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><recordid>eNp9kMtOwzAQRS0EEqWw4wMssYRQO47tZlkiXlIBidc2uPGYukrjYjug_j2uWrFkNXdx7ozmIHRKySWlnI9yktORZIyOy2IPDShnJBsLKff_sigP0VEIC0IIS_AAfUzwo_uGFj84DW12pQJo_Ay2M843sIQu4iko39nuE09itLHXgCvXRe9SB-LcaZxQ_G597FWbqqq1cY1fVIQ2JThGB0a1AU52c4jebq5fq7ts-nR7X02mWcOYjBkwqgkHZQQ3BTHpnRk0qiyp1lo0UuZU05kSMhei0KI0lMw4bYjShQbTGMaG6Gy7d-XdVw8h1gvX-y6drHPOCZXJi0zUxZZqvAvBg6lX3i6VX9eU1BuH9cZhvXOY8PMtPredVj_2f_oX4yhxTw</recordid><startdate>2021</startdate><enddate>2021</enddate><creator>Zhang, Jian</creator><creator>Wu, Fengge</creator><general>Hindawi</general><general>Hindawi Limited</general><scope>RHU</scope><scope>RHW</scope><scope>RHX</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7XB</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0N</scope><scope>P5Z</scope><scope>P62</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-1958-8321</orcidid><orcidid>https://orcid.org/0000-0003-1375-4947</orcidid></search><sort><creationdate>2021</creationdate><title>A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite</title><author>Zhang, Jian ; Wu, Fengge</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c337t-e31d05eaf65f40f115beca991ddd6c7721d1ba672664d69f10b51c0ad4defcf33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Actuators</topic><topic>Control methods</topic><topic>Control theory</topic><topic>Learning</topic><topic>Methods</topic><topic>Model accuracy</topic><topic>Neural networks</topic><topic>Reaction wheels</topic><topic>Satellite attitude control</topic><topic>Satellite observation</topic><topic>Satellites</topic><topic>Sensors</topic><topic>Simulation</topic><topic>Virtual reality</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Jian</creatorcontrib><creatorcontrib>Wu, Fengge</creatorcontrib><collection>Hindawi Publishing Complete</collection><collection>Hindawi Publishing Subscription Journals</collection><collection>Hindawi Publishing Open Access Journals</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Computing Database</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><jtitle>Wireless communications and mobile computing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhang, Jian</au><au>Wu, Fengge</au><au>Nagaraj, Balakrishnan</au><au>Balakrishnan Nagaraj</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite</atitle><jtitle>Wireless communications and mobile computing</jtitle><date>2021</date><risdate>2021</risdate><volume>2021</volume><issue>1</issue><issn>1530-8669</issn><eissn>1530-8677</eissn><abstract>Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods in real-world systems faces many challenges, such as insufficient samples, exploration safety issues, unknown actuator delays, and noise in the raw sensor data. In this work, a mixed model with different input sizes was proposed to represent the environmental dynamics model. The predication accuracy of the environmental dynamics model and the performance of the policy trained in this paper were gradually improved. Our method reduces the impact of noisy data on the model’s accuracy and improves the sampling efficiency. The experiments showed that the agent trained with our method completed a goal-state reaching task in a real-world system under wireless circumstances whose actuators were reaction wheels, whereas the soft actor-critic method failed in the same training process. The method’s effectiveness is ensured theoretically under given conditions.</abstract><cop>Oxford</cop><pub>Hindawi</pub><doi>10.1155/2021/7331894</doi><orcidid>https://orcid.org/0000-0002-1958-8321</orcidid><orcidid>https://orcid.org/0000-0003-1375-4947</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1530-8669 |
ispartof | Wireless communications and mobile computing, 2021, Vol.2021 (1) |
issn | 1530-8669 1530-8677 |
language | eng |
recordid | cdi_proquest_journals_2550177337 |
source | EZB-FREE-00999 freely available EZB journals; Wiley Online Library (Open Access Collection); Alma/SFX Local Collection |
subjects | Actuators Control methods Control theory Learning Methods Model accuracy Neural networks Reaction wheels Satellite attitude control Satellite observation Satellites Sensors Simulation Virtual reality |
title | A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T19%3A29%3A52IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Novel%20Model-Based%20Reinforcement%20Learning%20Attitude%20Control%20Method%20for%20Virtual%20Reality%20Satellite&rft.jtitle=Wireless%20communications%20and%20mobile%20computing&rft.au=Zhang,%20Jian&rft.date=2021&rft.volume=2021&rft.issue=1&rft.issn=1530-8669&rft.eissn=1530-8677&rft_id=info:doi/10.1155/2021/7331894&rft_dat=%3Cproquest_cross%3E2550177337%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2550177337&rft_id=info:pmid/&rfr_iscdi=true |