A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite

Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Wireless communications and mobile computing 2021, Vol.2021 (1)
Hauptverfasser: Zhang, Jian, Wu, Fengge
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue 1
container_start_page
container_title Wireless communications and mobile computing
container_volume 2021
creator Zhang, Jian
Wu, Fengge
description Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods in real-world systems faces many challenges, such as insufficient samples, exploration safety issues, unknown actuator delays, and noise in the raw sensor data. In this work, a mixed model with different input sizes was proposed to represent the environmental dynamics model. The predication accuracy of the environmental dynamics model and the performance of the policy trained in this paper were gradually improved. Our method reduces the impact of noisy data on the model’s accuracy and improves the sampling efficiency. The experiments showed that the agent trained with our method completed a goal-state reaching task in a real-world system under wireless circumstances whose actuators were reaction wheels, whereas the soft actor-critic method failed in the same training process. The method’s effectiveness is ensured theoretically under given conditions.
doi_str_mv 10.1155/2021/7331894
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2550177337</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2550177337</sourcerecordid><originalsourceid>FETCH-LOGICAL-c337t-e31d05eaf65f40f115beca991ddd6c7721d1ba672664d69f10b51c0ad4defcf33</originalsourceid><addsrcrecordid>eNp9kMtOwzAQRS0EEqWw4wMssYRQO47tZlkiXlIBidc2uPGYukrjYjug_j2uWrFkNXdx7ozmIHRKySWlnI9yktORZIyOy2IPDShnJBsLKff_sigP0VEIC0IIS_AAfUzwo_uGFj84DW12pQJo_Ay2M843sIQu4iko39nuE09itLHXgCvXRe9SB-LcaZxQ_G597FWbqqq1cY1fVIQ2JThGB0a1AU52c4jebq5fq7ts-nR7X02mWcOYjBkwqgkHZQQ3BTHpnRk0qiyp1lo0UuZU05kSMhei0KI0lMw4bYjShQbTGMaG6Gy7d-XdVw8h1gvX-y6drHPOCZXJi0zUxZZqvAvBg6lX3i6VX9eU1BuH9cZhvXOY8PMtPredVj_2f_oX4yhxTw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2550177337</pqid></control><display><type>article</type><title>A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite</title><source>EZB-FREE-00999 freely available EZB journals</source><source>Wiley Online Library (Open Access Collection)</source><source>Alma/SFX Local Collection</source><creator>Zhang, Jian ; Wu, Fengge</creator><contributor>Nagaraj, Balakrishnan ; Balakrishnan Nagaraj</contributor><creatorcontrib>Zhang, Jian ; Wu, Fengge ; Nagaraj, Balakrishnan ; Balakrishnan Nagaraj</creatorcontrib><description>Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods in real-world systems faces many challenges, such as insufficient samples, exploration safety issues, unknown actuator delays, and noise in the raw sensor data. In this work, a mixed model with different input sizes was proposed to represent the environmental dynamics model. The predication accuracy of the environmental dynamics model and the performance of the policy trained in this paper were gradually improved. Our method reduces the impact of noisy data on the model’s accuracy and improves the sampling efficiency. The experiments showed that the agent trained with our method completed a goal-state reaching task in a real-world system under wireless circumstances whose actuators were reaction wheels, whereas the soft actor-critic method failed in the same training process. The method’s effectiveness is ensured theoretically under given conditions.</description><identifier>ISSN: 1530-8669</identifier><identifier>EISSN: 1530-8677</identifier><identifier>DOI: 10.1155/2021/7331894</identifier><language>eng</language><publisher>Oxford: Hindawi</publisher><subject>Actuators ; Control methods ; Control theory ; Learning ; Methods ; Model accuracy ; Neural networks ; Reaction wheels ; Satellite attitude control ; Satellite observation ; Satellites ; Sensors ; Simulation ; Virtual reality</subject><ispartof>Wireless communications and mobile computing, 2021, Vol.2021 (1)</ispartof><rights>Copyright © 2021 Jian Zhang and Fengge Wu.</rights><rights>Copyright © 2021 Jian Zhang and Fengge Wu. This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c337t-e31d05eaf65f40f115beca991ddd6c7721d1ba672664d69f10b51c0ad4defcf33</citedby><cites>FETCH-LOGICAL-c337t-e31d05eaf65f40f115beca991ddd6c7721d1ba672664d69f10b51c0ad4defcf33</cites><orcidid>0000-0002-1958-8321 ; 0000-0003-1375-4947</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,4024,27923,27924,27925</link.rule.ids></links><search><contributor>Nagaraj, Balakrishnan</contributor><contributor>Balakrishnan Nagaraj</contributor><creatorcontrib>Zhang, Jian</creatorcontrib><creatorcontrib>Wu, Fengge</creatorcontrib><title>A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite</title><title>Wireless communications and mobile computing</title><description>Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods in real-world systems faces many challenges, such as insufficient samples, exploration safety issues, unknown actuator delays, and noise in the raw sensor data. In this work, a mixed model with different input sizes was proposed to represent the environmental dynamics model. The predication accuracy of the environmental dynamics model and the performance of the policy trained in this paper were gradually improved. Our method reduces the impact of noisy data on the model’s accuracy and improves the sampling efficiency. The experiments showed that the agent trained with our method completed a goal-state reaching task in a real-world system under wireless circumstances whose actuators were reaction wheels, whereas the soft actor-critic method failed in the same training process. The method’s effectiveness is ensured theoretically under given conditions.</description><subject>Actuators</subject><subject>Control methods</subject><subject>Control theory</subject><subject>Learning</subject><subject>Methods</subject><subject>Model accuracy</subject><subject>Neural networks</subject><subject>Reaction wheels</subject><subject>Satellite attitude control</subject><subject>Satellite observation</subject><subject>Satellites</subject><subject>Sensors</subject><subject>Simulation</subject><subject>Virtual reality</subject><issn>1530-8669</issn><issn>1530-8677</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>RHX</sourceid><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><recordid>eNp9kMtOwzAQRS0EEqWw4wMssYRQO47tZlkiXlIBidc2uPGYukrjYjug_j2uWrFkNXdx7ozmIHRKySWlnI9yktORZIyOy2IPDShnJBsLKff_sigP0VEIC0IIS_AAfUzwo_uGFj84DW12pQJo_Ay2M843sIQu4iko39nuE09itLHXgCvXRe9SB-LcaZxQ_G597FWbqqq1cY1fVIQ2JThGB0a1AU52c4jebq5fq7ts-nR7X02mWcOYjBkwqgkHZQQ3BTHpnRk0qiyp1lo0UuZU05kSMhei0KI0lMw4bYjShQbTGMaG6Gy7d-XdVw8h1gvX-y6drHPOCZXJi0zUxZZqvAvBg6lX3i6VX9eU1BuH9cZhvXOY8PMtPredVj_2f_oX4yhxTw</recordid><startdate>2021</startdate><enddate>2021</enddate><creator>Zhang, Jian</creator><creator>Wu, Fengge</creator><general>Hindawi</general><general>Hindawi Limited</general><scope>RHU</scope><scope>RHW</scope><scope>RHX</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7XB</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0N</scope><scope>P5Z</scope><scope>P62</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-1958-8321</orcidid><orcidid>https://orcid.org/0000-0003-1375-4947</orcidid></search><sort><creationdate>2021</creationdate><title>A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite</title><author>Zhang, Jian ; Wu, Fengge</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c337t-e31d05eaf65f40f115beca991ddd6c7721d1ba672664d69f10b51c0ad4defcf33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Actuators</topic><topic>Control methods</topic><topic>Control theory</topic><topic>Learning</topic><topic>Methods</topic><topic>Model accuracy</topic><topic>Neural networks</topic><topic>Reaction wheels</topic><topic>Satellite attitude control</topic><topic>Satellite observation</topic><topic>Satellites</topic><topic>Sensors</topic><topic>Simulation</topic><topic>Virtual reality</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Jian</creatorcontrib><creatorcontrib>Wu, Fengge</creatorcontrib><collection>Hindawi Publishing Complete</collection><collection>Hindawi Publishing Subscription Journals</collection><collection>Hindawi Publishing Open Access Journals</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Computing Database</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><jtitle>Wireless communications and mobile computing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhang, Jian</au><au>Wu, Fengge</au><au>Nagaraj, Balakrishnan</au><au>Balakrishnan Nagaraj</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite</atitle><jtitle>Wireless communications and mobile computing</jtitle><date>2021</date><risdate>2021</risdate><volume>2021</volume><issue>1</issue><issn>1530-8669</issn><eissn>1530-8677</eissn><abstract>Observing the universe with virtual reality satellite is an amazing experience. An intelligent method of attitude control is the core object of research to achieve this goal. Attitude control is essentially one of the goal-state reaching tasks under constraints. Using reinforcement learning methods in real-world systems faces many challenges, such as insufficient samples, exploration safety issues, unknown actuator delays, and noise in the raw sensor data. In this work, a mixed model with different input sizes was proposed to represent the environmental dynamics model. The predication accuracy of the environmental dynamics model and the performance of the policy trained in this paper were gradually improved. Our method reduces the impact of noisy data on the model’s accuracy and improves the sampling efficiency. The experiments showed that the agent trained with our method completed a goal-state reaching task in a real-world system under wireless circumstances whose actuators were reaction wheels, whereas the soft actor-critic method failed in the same training process. The method’s effectiveness is ensured theoretically under given conditions.</abstract><cop>Oxford</cop><pub>Hindawi</pub><doi>10.1155/2021/7331894</doi><orcidid>https://orcid.org/0000-0002-1958-8321</orcidid><orcidid>https://orcid.org/0000-0003-1375-4947</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1530-8669
ispartof Wireless communications and mobile computing, 2021, Vol.2021 (1)
issn 1530-8669
1530-8677
language eng
recordid cdi_proquest_journals_2550177337
source EZB-FREE-00999 freely available EZB journals; Wiley Online Library (Open Access Collection); Alma/SFX Local Collection
subjects Actuators
Control methods
Control theory
Learning
Methods
Model accuracy
Neural networks
Reaction wheels
Satellite attitude control
Satellite observation
Satellites
Sensors
Simulation
Virtual reality
title A Novel Model-Based Reinforcement Learning Attitude Control Method for Virtual Reality Satellite
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T19%3A29%3A52IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Novel%20Model-Based%20Reinforcement%20Learning%20Attitude%20Control%20Method%20for%20Virtual%20Reality%20Satellite&rft.jtitle=Wireless%20communications%20and%20mobile%20computing&rft.au=Zhang,%20Jian&rft.date=2021&rft.volume=2021&rft.issue=1&rft.issn=1530-8669&rft.eissn=1530-8677&rft_id=info:doi/10.1155/2021/7331894&rft_dat=%3Cproquest_cross%3E2550177337%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2550177337&rft_id=info:pmid/&rfr_iscdi=true