Unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and game theory

The invention discloses an unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and a game theory, and the method achieves the nonlinear modeling of a high-dimensional and continuous state and a strategy space of an unmanned aerial vehicle through dee...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: JI MENGDA, WANG LIYING, XU GENJIU, LI ZESHENG, DUAN ZEKUN
Format: Patent
Sprache:chi ; eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator JI MENGDA
WANG LIYING
XU GENJIU
LI ZESHENG
DUAN ZEKUN
description The invention discloses an unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and a game theory, and the method achieves the nonlinear modeling of a high-dimensional and continuous state and a strategy space of an unmanned aerial vehicle through deep reinforcement learning, shows high adaptability, and can respond to a complex and dynamic environment in real time. In the training construction process of the decision network, the air combat situation is divided according to the sight angles of the unmanned aerial vehicles of the two air combat parties, and meanwhile, the unmanned aerial vehicle dual-network confrontation training is designed, so that the enemy unmanned aerial vehicles make decisions by using the previously trained model, and the decision performance of the unmanned aerial vehicles is further improved. According to the method, the air combat process of the unmanned aerial vehicles of the two parties is regarded as a two-person zero sum Markov g
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN117930880A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN117930880A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN117930880A3</originalsourceid><addsrcrecordid>eNqNyzEOgkAUBFAaC6Pe4XsAEgiFWBqisbLSmnyXQX5k_5Ld1YTbK8QDWE0y82aZvG5qWRUNMbxwT290YnoQiyfj7J0jNTASxGlq-Sn6IIvYue9BGwpjiLAzFJ02D9HWeQMLjdSD_VxP9sEWFDs4P66TRct9wOaXq2R7Ol6rc4rB1QgDGyhiXV3yfLcvsrLMDsU_5gP0TUWG</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and game theory</title><source>esp@cenet</source><creator>JI MENGDA ; WANG LIYING ; XU GENJIU ; LI ZESHENG ; DUAN ZEKUN</creator><creatorcontrib>JI MENGDA ; WANG LIYING ; XU GENJIU ; LI ZESHENG ; DUAN ZEKUN</creatorcontrib><description>The invention discloses an unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and a game theory, and the method achieves the nonlinear modeling of a high-dimensional and continuous state and a strategy space of an unmanned aerial vehicle through deep reinforcement learning, shows high adaptability, and can respond to a complex and dynamic environment in real time. In the training construction process of the decision network, the air combat situation is divided according to the sight angles of the unmanned aerial vehicles of the two air combat parties, and meanwhile, the unmanned aerial vehicle dual-network confrontation training is designed, so that the enemy unmanned aerial vehicles make decisions by using the previously trained model, and the decision performance of the unmanned aerial vehicles is further improved. According to the method, the air combat process of the unmanned aerial vehicles of the two parties is regarded as a two-person zero sum Markov g</description><language>chi ; eng</language><subject>CONTROLLING ; PHYSICS ; REGULATING ; SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES</subject><creationdate>2024</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240426&amp;DB=EPODOC&amp;CC=CN&amp;NR=117930880A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,780,885,25564,76547</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240426&amp;DB=EPODOC&amp;CC=CN&amp;NR=117930880A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>JI MENGDA</creatorcontrib><creatorcontrib>WANG LIYING</creatorcontrib><creatorcontrib>XU GENJIU</creatorcontrib><creatorcontrib>LI ZESHENG</creatorcontrib><creatorcontrib>DUAN ZEKUN</creatorcontrib><title>Unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and game theory</title><description>The invention discloses an unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and a game theory, and the method achieves the nonlinear modeling of a high-dimensional and continuous state and a strategy space of an unmanned aerial vehicle through deep reinforcement learning, shows high adaptability, and can respond to a complex and dynamic environment in real time. In the training construction process of the decision network, the air combat situation is divided according to the sight angles of the unmanned aerial vehicles of the two air combat parties, and meanwhile, the unmanned aerial vehicle dual-network confrontation training is designed, so that the enemy unmanned aerial vehicles make decisions by using the previously trained model, and the decision performance of the unmanned aerial vehicles is further improved. According to the method, the air combat process of the unmanned aerial vehicles of the two parties is regarded as a two-person zero sum Markov g</description><subject>CONTROLLING</subject><subject>PHYSICS</subject><subject>REGULATING</subject><subject>SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2024</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNqNyzEOgkAUBFAaC6Pe4XsAEgiFWBqisbLSmnyXQX5k_5Ld1YTbK8QDWE0y82aZvG5qWRUNMbxwT290YnoQiyfj7J0jNTASxGlq-Sn6IIvYue9BGwpjiLAzFJ02D9HWeQMLjdSD_VxP9sEWFDs4P66TRct9wOaXq2R7Ol6rc4rB1QgDGyhiXV3yfLcvsrLMDsU_5gP0TUWG</recordid><startdate>20240426</startdate><enddate>20240426</enddate><creator>JI MENGDA</creator><creator>WANG LIYING</creator><creator>XU GENJIU</creator><creator>LI ZESHENG</creator><creator>DUAN ZEKUN</creator><scope>EVB</scope></search><sort><creationdate>20240426</creationdate><title>Unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and game theory</title><author>JI MENGDA ; WANG LIYING ; XU GENJIU ; LI ZESHENG ; DUAN ZEKUN</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN117930880A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2024</creationdate><topic>CONTROLLING</topic><topic>PHYSICS</topic><topic>REGULATING</topic><topic>SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES</topic><toplevel>online_resources</toplevel><creatorcontrib>JI MENGDA</creatorcontrib><creatorcontrib>WANG LIYING</creatorcontrib><creatorcontrib>XU GENJIU</creatorcontrib><creatorcontrib>LI ZESHENG</creatorcontrib><creatorcontrib>DUAN ZEKUN</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>JI MENGDA</au><au>WANG LIYING</au><au>XU GENJIU</au><au>LI ZESHENG</au><au>DUAN ZEKUN</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and game theory</title><date>2024-04-26</date><risdate>2024</risdate><abstract>The invention discloses an unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and a game theory, and the method achieves the nonlinear modeling of a high-dimensional and continuous state and a strategy space of an unmanned aerial vehicle through deep reinforcement learning, shows high adaptability, and can respond to a complex and dynamic environment in real time. In the training construction process of the decision network, the air combat situation is divided according to the sight angles of the unmanned aerial vehicles of the two air combat parties, and meanwhile, the unmanned aerial vehicle dual-network confrontation training is designed, so that the enemy unmanned aerial vehicles make decisions by using the previously trained model, and the decision performance of the unmanned aerial vehicles is further improved. According to the method, the air combat process of the unmanned aerial vehicles of the two parties is regarded as a two-person zero sum Markov g</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language chi ; eng
recordid cdi_epo_espacenet_CN117930880A
source esp@cenet
subjects CONTROLLING
PHYSICS
REGULATING
SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
title Unmanned aerial vehicle air combat decision-making method and system combining reinforcement learning and game theory
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T06%3A55%3A03IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=JI%20MENGDA&rft.date=2024-04-26&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN117930880A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true