Audio and video omnibearing simulation training method and system
The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremon...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , , , , , , , |
---|---|
Format: | Patent |
Sprache: | chi ; eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | SHAO YANNING HUANG XIAOQIANG XIONG YING WU HAIBO YAN XIANRUI SU LIWEI ZHANG YAN LIAO YUNTING XIE LUBIN WU GUANGBIN YE ZHIPING TAN HUOCHAO TAO FEIDA BAI YANLING LIU ZHENHUA CHEN HAIYAN |
description | The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremony behaviors of a seat person in the video data, obtaining a gesture and an action, and carrying out the recognition of the ceremony behaviors of the seat person. Non-linear image features of gestures and actions are obtained according to a cascade classifier; wherein the image features are obtained by outputting intermediate features of the gesture and the action according to the first stage of the cascade classifier, and taking the intermediate features as at least one input of the second stage of the cascade classifier; the audio features and the image features are fused, the emotional tendency of the seat personnel is detected according to the obtained fused features, the seat personnel are evaluated according t |
format | Patent |
fullrecord | <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN118038733A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN118038733A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN118038733A3</originalsourceid><addsrcrecordid>eNrjZHB0LE3JzFdIzEtRKMtMSc1XyM_Ny0xKTSzKzEtXKM7MLc1JLMnMz1MoKUrMzAOJ5aaWZOSngDUUVxaXpObyMLCmJeYUp_JCaW4GRTfXEGcP3dSC_PjU4oLE5NS81JJ4Zz9DQwsDYwtzY2NHY2LUAAAa1jHt</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Audio and video omnibearing simulation training method and system</title><source>esp@cenet</source><creator>SHAO YANNING ; HUANG XIAOQIANG ; XIONG YING ; WU HAIBO ; YAN XIANRUI ; SU LIWEI ; ZHANG YAN ; LIAO YUNTING ; XIE LUBIN ; WU GUANGBIN ; YE ZHIPING ; TAN HUOCHAO ; TAO FEIDA ; BAI YANLING ; LIU ZHENHUA ; CHEN HAIYAN</creator><creatorcontrib>SHAO YANNING ; HUANG XIAOQIANG ; XIONG YING ; WU HAIBO ; YAN XIANRUI ; SU LIWEI ; ZHANG YAN ; LIAO YUNTING ; XIE LUBIN ; WU GUANGBIN ; YE ZHIPING ; TAN HUOCHAO ; TAO FEIDA ; BAI YANLING ; LIU ZHENHUA ; CHEN HAIYAN</creatorcontrib><description>The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremony behaviors of a seat person in the video data, obtaining a gesture and an action, and carrying out the recognition of the ceremony behaviors of the seat person. Non-linear image features of gestures and actions are obtained according to a cascade classifier; wherein the image features are obtained by outputting intermediate features of the gesture and the action according to the first stage of the cascade classifier, and taking the intermediate features as at least one input of the second stage of the cascade classifier; the audio features and the image features are fused, the emotional tendency of the seat personnel is detected according to the obtained fused features, the seat personnel are evaluated according t</description><language>chi ; eng</language><subject>ACOUSTICS ; ADVERTISING ; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND,DEAF OR MUTE ; CALCULATING ; COMPUTING ; COUNTING ; CRYPTOGRAPHY ; DIAGRAMS ; DISPLAY ; EDUCATION ; EDUCATIONAL OR DEMONSTRATION APPLIANCES ; GLOBES ; MUSICAL INSTRUMENTS ; PHYSICS ; PLANETARIA ; SEALS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2024</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20240514&DB=EPODOC&CC=CN&NR=118038733A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,309,781,886,25566,76549</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20240514&DB=EPODOC&CC=CN&NR=118038733A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>SHAO YANNING</creatorcontrib><creatorcontrib>HUANG XIAOQIANG</creatorcontrib><creatorcontrib>XIONG YING</creatorcontrib><creatorcontrib>WU HAIBO</creatorcontrib><creatorcontrib>YAN XIANRUI</creatorcontrib><creatorcontrib>SU LIWEI</creatorcontrib><creatorcontrib>ZHANG YAN</creatorcontrib><creatorcontrib>LIAO YUNTING</creatorcontrib><creatorcontrib>XIE LUBIN</creatorcontrib><creatorcontrib>WU GUANGBIN</creatorcontrib><creatorcontrib>YE ZHIPING</creatorcontrib><creatorcontrib>TAN HUOCHAO</creatorcontrib><creatorcontrib>TAO FEIDA</creatorcontrib><creatorcontrib>BAI YANLING</creatorcontrib><creatorcontrib>LIU ZHENHUA</creatorcontrib><creatorcontrib>CHEN HAIYAN</creatorcontrib><title>Audio and video omnibearing simulation training method and system</title><description>The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremony behaviors of a seat person in the video data, obtaining a gesture and an action, and carrying out the recognition of the ceremony behaviors of the seat person. Non-linear image features of gestures and actions are obtained according to a cascade classifier; wherein the image features are obtained by outputting intermediate features of the gesture and the action according to the first stage of the cascade classifier, and taking the intermediate features as at least one input of the second stage of the cascade classifier; the audio features and the image features are fused, the emotional tendency of the seat personnel is detected according to the obtained fused features, the seat personnel are evaluated according t</description><subject>ACOUSTICS</subject><subject>ADVERTISING</subject><subject>APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND,DEAF OR MUTE</subject><subject>CALCULATING</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>CRYPTOGRAPHY</subject><subject>DIAGRAMS</subject><subject>DISPLAY</subject><subject>EDUCATION</subject><subject>EDUCATIONAL OR DEMONSTRATION APPLIANCES</subject><subject>GLOBES</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>PLANETARIA</subject><subject>SEALS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2024</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZHB0LE3JzFdIzEtRKMtMSc1XyM_Ny0xKTSzKzEtXKM7MLc1JLMnMz1MoKUrMzAOJ5aaWZOSngDUUVxaXpObyMLCmJeYUp_JCaW4GRTfXEGcP3dSC_PjU4oLE5NS81JJ4Zz9DQwsDYwtzY2NHY2LUAAAa1jHt</recordid><startdate>20240514</startdate><enddate>20240514</enddate><creator>SHAO YANNING</creator><creator>HUANG XIAOQIANG</creator><creator>XIONG YING</creator><creator>WU HAIBO</creator><creator>YAN XIANRUI</creator><creator>SU LIWEI</creator><creator>ZHANG YAN</creator><creator>LIAO YUNTING</creator><creator>XIE LUBIN</creator><creator>WU GUANGBIN</creator><creator>YE ZHIPING</creator><creator>TAN HUOCHAO</creator><creator>TAO FEIDA</creator><creator>BAI YANLING</creator><creator>LIU ZHENHUA</creator><creator>CHEN HAIYAN</creator><scope>EVB</scope></search><sort><creationdate>20240514</creationdate><title>Audio and video omnibearing simulation training method and system</title><author>SHAO YANNING ; HUANG XIAOQIANG ; XIONG YING ; WU HAIBO ; YAN XIANRUI ; SU LIWEI ; ZHANG YAN ; LIAO YUNTING ; XIE LUBIN ; WU GUANGBIN ; YE ZHIPING ; TAN HUOCHAO ; TAO FEIDA ; BAI YANLING ; LIU ZHENHUA ; CHEN HAIYAN</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN118038733A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2024</creationdate><topic>ACOUSTICS</topic><topic>ADVERTISING</topic><topic>APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND,DEAF OR MUTE</topic><topic>CALCULATING</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>CRYPTOGRAPHY</topic><topic>DIAGRAMS</topic><topic>DISPLAY</topic><topic>EDUCATION</topic><topic>EDUCATIONAL OR DEMONSTRATION APPLIANCES</topic><topic>GLOBES</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>PLANETARIA</topic><topic>SEALS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>SHAO YANNING</creatorcontrib><creatorcontrib>HUANG XIAOQIANG</creatorcontrib><creatorcontrib>XIONG YING</creatorcontrib><creatorcontrib>WU HAIBO</creatorcontrib><creatorcontrib>YAN XIANRUI</creatorcontrib><creatorcontrib>SU LIWEI</creatorcontrib><creatorcontrib>ZHANG YAN</creatorcontrib><creatorcontrib>LIAO YUNTING</creatorcontrib><creatorcontrib>XIE LUBIN</creatorcontrib><creatorcontrib>WU GUANGBIN</creatorcontrib><creatorcontrib>YE ZHIPING</creatorcontrib><creatorcontrib>TAN HUOCHAO</creatorcontrib><creatorcontrib>TAO FEIDA</creatorcontrib><creatorcontrib>BAI YANLING</creatorcontrib><creatorcontrib>LIU ZHENHUA</creatorcontrib><creatorcontrib>CHEN HAIYAN</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>SHAO YANNING</au><au>HUANG XIAOQIANG</au><au>XIONG YING</au><au>WU HAIBO</au><au>YAN XIANRUI</au><au>SU LIWEI</au><au>ZHANG YAN</au><au>LIAO YUNTING</au><au>XIE LUBIN</au><au>WU GUANGBIN</au><au>YE ZHIPING</au><au>TAN HUOCHAO</au><au>TAO FEIDA</au><au>BAI YANLING</au><au>LIU ZHENHUA</au><au>CHEN HAIYAN</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Audio and video omnibearing simulation training method and system</title><date>2024-05-14</date><risdate>2024</risdate><abstract>The invention discloses an audio and video omnibearing simulation training method and system, and the method comprises the steps: collecting audio data and video data for executing simulation operation, obtaining the audio data, marking the audio features, carrying out the recognition of the ceremony behaviors of a seat person in the video data, obtaining a gesture and an action, and carrying out the recognition of the ceremony behaviors of the seat person. Non-linear image features of gestures and actions are obtained according to a cascade classifier; wherein the image features are obtained by outputting intermediate features of the gesture and the action according to the first stage of the cascade classifier, and taking the intermediate features as at least one input of the second stage of the cascade classifier; the audio features and the image features are fused, the emotional tendency of the seat personnel is detected according to the obtained fused features, the seat personnel are evaluated according t</abstract><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | |
ispartof | |
issn | |
language | chi ; eng |
recordid | cdi_epo_espacenet_CN118038733A |
source | esp@cenet |
subjects | ACOUSTICS ADVERTISING APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND,DEAF OR MUTE CALCULATING COMPUTING COUNTING CRYPTOGRAPHY DIAGRAMS DISPLAY EDUCATION EDUCATIONAL OR DEMONSTRATION APPLIANCES GLOBES MUSICAL INSTRUMENTS PHYSICS PLANETARIA SEALS SPEECH ANALYSIS OR SYNTHESIS SPEECH OR AUDIO CODING OR DECODING SPEECH OR VOICE PROCESSING SPEECH RECOGNITION |
title | Audio and video omnibearing simulation training method and system |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T06%3A58%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=SHAO%20YANNING&rft.date=2024-05-14&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN118038733A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |