Audio and video omnibearing simulation training method and system

The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremon...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: SHAO YANNING, HUANG XIAOQIANG, XIONG YING, WU HAIBO, YAN XIANRUI, SU LIWEI, ZHANG YAN, LIAO YUNTING, XIE LUBIN, WU GUANGBIN, YE ZHIPING, TAN HUOCHAO, TAO FEIDA, BAI YANLING, LIU ZHENHUA, CHEN HAIYAN
Format: Patent
Sprache:chi ; eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator SHAO YANNING
HUANG XIAOQIANG
XIONG YING
WU HAIBO
YAN XIANRUI
SU LIWEI
ZHANG YAN
LIAO YUNTING
XIE LUBIN
WU GUANGBIN
YE ZHIPING
TAN HUOCHAO
TAO FEIDA
BAI YANLING
LIU ZHENHUA
CHEN HAIYAN
description The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremony behaviors of a seat person in the video data, obtaining a gesture and an action, and carrying out the recognition of the ceremony behaviors of the seat person. Non-linear image features of gestures and actions are obtained according to a cascade classifier; wherein the image features are obtained by outputting intermediate features of the gesture and the action according to the first stage of the cascade classifier, and taking the intermediate features as at least one input of the second stage of the cascade classifier; the audio features and the image features are fused, the emotional tendency of the seat personnel is detected according to the obtained fused features, the seat personnel are evaluated according t
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN118038733A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN118038733A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN118038733A3</originalsourceid><addsrcrecordid>eNrjZHB0LE3JzFdIzEtRKMtMSc1XyM_Ny0xKTSzKzEtXKM7MLc1JLMnMz1MoKUrMzAOJ5aaWZOSngDUUVxaXpObyMLCmJeYUp_JCaW4GRTfXEGcP3dSC_PjU4oLE5NS81JJ4Zz9DQwsDYwtzY2NHY2LUAAAa1jHt</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Audio and video omnibearing simulation training method and system</title><source>esp@cenet</source><creator>SHAO YANNING ; HUANG XIAOQIANG ; XIONG YING ; WU HAIBO ; YAN XIANRUI ; SU LIWEI ; ZHANG YAN ; LIAO YUNTING ; XIE LUBIN ; WU GUANGBIN ; YE ZHIPING ; TAN HUOCHAO ; TAO FEIDA ; BAI YANLING ; LIU ZHENHUA ; CHEN HAIYAN</creator><creatorcontrib>SHAO YANNING ; HUANG XIAOQIANG ; XIONG YING ; WU HAIBO ; YAN XIANRUI ; SU LIWEI ; ZHANG YAN ; LIAO YUNTING ; XIE LUBIN ; WU GUANGBIN ; YE ZHIPING ; TAN HUOCHAO ; TAO FEIDA ; BAI YANLING ; LIU ZHENHUA ; CHEN HAIYAN</creatorcontrib><description>The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremony behaviors of a seat person in the video data, obtaining a gesture and an action, and carrying out the recognition of the ceremony behaviors of the seat person. Non-linear image features of gestures and actions are obtained according to a cascade classifier; wherein the image features are obtained by outputting intermediate features of the gesture and the action according to the first stage of the cascade classifier, and taking the intermediate features as at least one input of the second stage of the cascade classifier; the audio features and the image features are fused, the emotional tendency of the seat personnel is detected according to the obtained fused features, the seat personnel are evaluated according t</description><language>chi ; eng</language><subject>ACOUSTICS ; ADVERTISING ; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND,DEAF OR MUTE ; CALCULATING ; COMPUTING ; COUNTING ; CRYPTOGRAPHY ; DIAGRAMS ; DISPLAY ; EDUCATION ; EDUCATIONAL OR DEMONSTRATION APPLIANCES ; GLOBES ; MUSICAL INSTRUMENTS ; PHYSICS ; PLANETARIA ; SEALS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2024</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240514&amp;DB=EPODOC&amp;CC=CN&amp;NR=118038733A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,309,781,886,25566,76549</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240514&amp;DB=EPODOC&amp;CC=CN&amp;NR=118038733A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>SHAO YANNING</creatorcontrib><creatorcontrib>HUANG XIAOQIANG</creatorcontrib><creatorcontrib>XIONG YING</creatorcontrib><creatorcontrib>WU HAIBO</creatorcontrib><creatorcontrib>YAN XIANRUI</creatorcontrib><creatorcontrib>SU LIWEI</creatorcontrib><creatorcontrib>ZHANG YAN</creatorcontrib><creatorcontrib>LIAO YUNTING</creatorcontrib><creatorcontrib>XIE LUBIN</creatorcontrib><creatorcontrib>WU GUANGBIN</creatorcontrib><creatorcontrib>YE ZHIPING</creatorcontrib><creatorcontrib>TAN HUOCHAO</creatorcontrib><creatorcontrib>TAO FEIDA</creatorcontrib><creatorcontrib>BAI YANLING</creatorcontrib><creatorcontrib>LIU ZHENHUA</creatorcontrib><creatorcontrib>CHEN HAIYAN</creatorcontrib><title>Audio and video omnibearing simulation training method and system</title><description>The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremony behaviors of a seat person in the video data, obtaining a gesture and an action, and carrying out the recognition of the ceremony behaviors of the seat person. Non-linear image features of gestures and actions are obtained according to a cascade classifier; wherein the image features are obtained by outputting intermediate features of the gesture and the action according to the first stage of the cascade classifier, and taking the intermediate features as at least one input of the second stage of the cascade classifier; the audio features and the image features are fused, the emotional tendency of the seat personnel is detected according to the obtained fused features, the seat personnel are evaluated according t</description><subject>ACOUSTICS</subject><subject>ADVERTISING</subject><subject>APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND,DEAF OR MUTE</subject><subject>CALCULATING</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>CRYPTOGRAPHY</subject><subject>DIAGRAMS</subject><subject>DISPLAY</subject><subject>EDUCATION</subject><subject>EDUCATIONAL OR DEMONSTRATION APPLIANCES</subject><subject>GLOBES</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>PLANETARIA</subject><subject>SEALS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2024</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZHB0LE3JzFdIzEtRKMtMSc1XyM_Ny0xKTSzKzEtXKM7MLc1JLMnMz1MoKUrMzAOJ5aaWZOSngDUUVxaXpObyMLCmJeYUp_JCaW4GRTfXEGcP3dSC_PjU4oLE5NS81JJ4Zz9DQwsDYwtzY2NHY2LUAAAa1jHt</recordid><startdate>20240514</startdate><enddate>20240514</enddate><creator>SHAO YANNING</creator><creator>HUANG XIAOQIANG</creator><creator>XIONG YING</creator><creator>WU HAIBO</creator><creator>YAN XIANRUI</creator><creator>SU LIWEI</creator><creator>ZHANG YAN</creator><creator>LIAO YUNTING</creator><creator>XIE LUBIN</creator><creator>WU GUANGBIN</creator><creator>YE ZHIPING</creator><creator>TAN HUOCHAO</creator><creator>TAO FEIDA</creator><creator>BAI YANLING</creator><creator>LIU ZHENHUA</creator><creator>CHEN HAIYAN</creator><scope>EVB</scope></search><sort><creationdate>20240514</creationdate><title>Audio and video omnibearing simulation training method and system</title><author>SHAO YANNING ; HUANG XIAOQIANG ; XIONG YING ; WU HAIBO ; YAN XIANRUI ; SU LIWEI ; ZHANG YAN ; LIAO YUNTING ; XIE LUBIN ; WU GUANGBIN ; YE ZHIPING ; TAN HUOCHAO ; TAO FEIDA ; BAI YANLING ; LIU ZHENHUA ; CHEN HAIYAN</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN118038733A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2024</creationdate><topic>ACOUSTICS</topic><topic>ADVERTISING</topic><topic>APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND,DEAF OR MUTE</topic><topic>CALCULATING</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>CRYPTOGRAPHY</topic><topic>DIAGRAMS</topic><topic>DISPLAY</topic><topic>EDUCATION</topic><topic>EDUCATIONAL OR DEMONSTRATION APPLIANCES</topic><topic>GLOBES</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>PLANETARIA</topic><topic>SEALS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>SHAO YANNING</creatorcontrib><creatorcontrib>HUANG XIAOQIANG</creatorcontrib><creatorcontrib>XIONG YING</creatorcontrib><creatorcontrib>WU HAIBO</creatorcontrib><creatorcontrib>YAN XIANRUI</creatorcontrib><creatorcontrib>SU LIWEI</creatorcontrib><creatorcontrib>ZHANG YAN</creatorcontrib><creatorcontrib>LIAO YUNTING</creatorcontrib><creatorcontrib>XIE LUBIN</creatorcontrib><creatorcontrib>WU GUANGBIN</creatorcontrib><creatorcontrib>YE ZHIPING</creatorcontrib><creatorcontrib>TAN HUOCHAO</creatorcontrib><creatorcontrib>TAO FEIDA</creatorcontrib><creatorcontrib>BAI YANLING</creatorcontrib><creatorcontrib>LIU ZHENHUA</creatorcontrib><creatorcontrib>CHEN HAIYAN</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>SHAO YANNING</au><au>HUANG XIAOQIANG</au><au>XIONG YING</au><au>WU HAIBO</au><au>YAN XIANRUI</au><au>SU LIWEI</au><au>ZHANG YAN</au><au>LIAO YUNTING</au><au>XIE LUBIN</au><au>WU GUANGBIN</au><au>YE ZHIPING</au><au>TAN HUOCHAO</au><au>TAO FEIDA</au><au>BAI YANLING</au><au>LIU ZHENHUA</au><au>CHEN HAIYAN</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Audio and video omnibearing simulation training method and system</title><date>2024-05-14</date><risdate>2024</risdate><abstract>The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremony behaviors of a seat person in the video data, obtaining a gesture and an action, and carrying out the recognition of the ceremony behaviors of the seat person. Non-linear image features of gestures and actions are obtained according to a cascade classifier; wherein the image features are obtained by outputting intermediate features of the gesture and the action according to the first stage of the cascade classifier, and taking the intermediate features as at least one input of the second stage of the cascade classifier; the audio features and the image features are fused, the emotional tendency of the seat personnel is detected according to the obtained fused features, the seat personnel are evaluated according t</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language chi ; eng
recordid cdi_epo_espacenet_CN118038733A
source esp@cenet
subjects ACOUSTICS
ADVERTISING
APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND,DEAF OR MUTE
CALCULATING
COMPUTING
COUNTING
CRYPTOGRAPHY
DIAGRAMS
DISPLAY
EDUCATION
EDUCATIONAL OR DEMONSTRATION APPLIANCES
GLOBES
MUSICAL INSTRUMENTS
PHYSICS
PLANETARIA
SEALS
SPEECH ANALYSIS OR SYNTHESIS
SPEECH OR AUDIO CODING OR DECODING
SPEECH OR VOICE PROCESSING
SPEECH RECOGNITION
title Audio and video omnibearing simulation training method and system
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T06%3A58%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=SHAO%20YANNING&rft.date=2024-05-14&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN118038733A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true