Video recommendation method based on multi-modal video content and multi-task learning

The invention discloses a video recommendation method based on multi-modal video content and multi-task learning. The method comprises the following steps: extracting visual, audio and text features of a short video through a pre-trained model; fusing the multi-modal features of the video by adoptin...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: SHI JINGLUN, LIANG KEHONG, LIN YANGCHENG, FU QIANSHUAN, DENG LI
Format: Patent
Sprache:chi ; eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator SHI JINGLUN
LIANG KEHONG
LIN YANGCHENG
FU QIANSHUAN
DENG LI
description The invention discloses a video recommendation method based on multi-modal video content and multi-task learning. The method comprises the following steps: extracting visual, audio and text features of a short video through a pre-trained model; fusing the multi-modal features of the video by adopting an attention mechanism method; learning feature representation of the social relationship of the user by adopting a deep walk method; proposing a deep neural network model based on an attention mechanism to learn multi-domain feature representation; embedding the features generated based on the above steps into a sharing layer as a multi-task model, and generating prediction results through a multi-layer perceptron. According to the method, the attention mechanism is combined with the user features to fuse the video multi-modal features, so that the whole recommendation is richer and more personalized; meanwhile, because of multi-domain features and with consideration of the importance ofinteraction features in r
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN111246256A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN111246256A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN111246256A3</originalsourceid><addsrcrecordid>eNqNizEOwjAMRbMwIOAO5gAdUqA7qkBMTKhrZWIDEYldNYbzI1APwPT19N6fu66LxAojB82ZhdCiCmS2hxJcsTDBl1_JYpWVMMH7dwgqxmKAQpM1LE9IjKNEuS_d7Iap8GrahVsfD5f2VPGgPZcBAwtb35699_W2qXfNfvNP8wEtOTmj</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Video recommendation method based on multi-modal video content and multi-task learning</title><source>esp@cenet</source><creator>SHI JINGLUN ; LIANG KEHONG ; LIN YANGCHENG ; FU QIANSHUAN ; DENG LI</creator><creatorcontrib>SHI JINGLUN ; LIANG KEHONG ; LIN YANGCHENG ; FU QIANSHUAN ; DENG LI</creatorcontrib><description>The invention discloses a video recommendation method based on multi-modal video content and multi-task learning. The method comprises the following steps: extracting visual, audio and text features of a short video through a pre-trained model; fusing the multi-modal features of the video by adopting an attention mechanism method; learning feature representation of the social relationship of the user by adopting a deep walk method; proposing a deep neural network model based on an attention mechanism to learn multi-domain feature representation; embedding the features generated based on the above steps into a sharing layer as a multi-task model, and generating prediction results through a multi-layer perceptron. According to the method, the attention mechanism is combined with the user features to fuse the video multi-modal features, so that the whole recommendation is richer and more personalized; meanwhile, because of multi-domain features and with consideration of the importance ofinteraction features in r</description><language>chi ; eng</language><subject>CALCULATING ; COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS ; COMPUTING ; COUNTING ; ELECTRIC COMMUNICATION TECHNIQUE ; ELECTRIC DIGITAL DATA PROCESSING ; ELECTRICITY ; PHYSICS ; PICTORIAL COMMUNICATION, e.g. TELEVISION</subject><creationdate>2020</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20200605&amp;DB=EPODOC&amp;CC=CN&amp;NR=111246256A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,780,885,25564,76547</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20200605&amp;DB=EPODOC&amp;CC=CN&amp;NR=111246256A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>SHI JINGLUN</creatorcontrib><creatorcontrib>LIANG KEHONG</creatorcontrib><creatorcontrib>LIN YANGCHENG</creatorcontrib><creatorcontrib>FU QIANSHUAN</creatorcontrib><creatorcontrib>DENG LI</creatorcontrib><title>Video recommendation method based on multi-modal video content and multi-task learning</title><description>The invention discloses a video recommendation method based on multi-modal video content and multi-task learning. The method comprises the following steps: extracting visual, audio and text features of a short video through a pre-trained model; fusing the multi-modal features of the video by adopting an attention mechanism method; learning feature representation of the social relationship of the user by adopting a deep walk method; proposing a deep neural network model based on an attention mechanism to learn multi-domain feature representation; embedding the features generated based on the above steps into a sharing layer as a multi-task model, and generating prediction results through a multi-layer perceptron. According to the method, the attention mechanism is combined with the user features to fuse the video multi-modal features, so that the whole recommendation is richer and more personalized; meanwhile, because of multi-domain features and with consideration of the importance ofinteraction features in r</description><subject>CALCULATING</subject><subject>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>ELECTRIC COMMUNICATION TECHNIQUE</subject><subject>ELECTRIC DIGITAL DATA PROCESSING</subject><subject>ELECTRICITY</subject><subject>PHYSICS</subject><subject>PICTORIAL COMMUNICATION, e.g. TELEVISION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2020</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNqNizEOwjAMRbMwIOAO5gAdUqA7qkBMTKhrZWIDEYldNYbzI1APwPT19N6fu66LxAojB82ZhdCiCmS2hxJcsTDBl1_JYpWVMMH7dwgqxmKAQpM1LE9IjKNEuS_d7Iap8GrahVsfD5f2VPGgPZcBAwtb35699_W2qXfNfvNP8wEtOTmj</recordid><startdate>20200605</startdate><enddate>20200605</enddate><creator>SHI JINGLUN</creator><creator>LIANG KEHONG</creator><creator>LIN YANGCHENG</creator><creator>FU QIANSHUAN</creator><creator>DENG LI</creator><scope>EVB</scope></search><sort><creationdate>20200605</creationdate><title>Video recommendation method based on multi-modal video content and multi-task learning</title><author>SHI JINGLUN ; LIANG KEHONG ; LIN YANGCHENG ; FU QIANSHUAN ; DENG LI</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN111246256A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2020</creationdate><topic>CALCULATING</topic><topic>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>ELECTRIC COMMUNICATION TECHNIQUE</topic><topic>ELECTRIC DIGITAL DATA PROCESSING</topic><topic>ELECTRICITY</topic><topic>PHYSICS</topic><topic>PICTORIAL COMMUNICATION, e.g. TELEVISION</topic><toplevel>online_resources</toplevel><creatorcontrib>SHI JINGLUN</creatorcontrib><creatorcontrib>LIANG KEHONG</creatorcontrib><creatorcontrib>LIN YANGCHENG</creatorcontrib><creatorcontrib>FU QIANSHUAN</creatorcontrib><creatorcontrib>DENG LI</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>SHI JINGLUN</au><au>LIANG KEHONG</au><au>LIN YANGCHENG</au><au>FU QIANSHUAN</au><au>DENG LI</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Video recommendation method based on multi-modal video content and multi-task learning</title><date>2020-06-05</date><risdate>2020</risdate><abstract>The invention discloses a video recommendation method based on multi-modal video content and multi-task learning. The method comprises the following steps: extracting visual, audio and text features of a short video through a pre-trained model; fusing the multi-modal features of the video by adopting an attention mechanism method; learning feature representation of the social relationship of the user by adopting a deep walk method; proposing a deep neural network model based on an attention mechanism to learn multi-domain feature representation; embedding the features generated based on the above steps into a sharing layer as a multi-task model, and generating prediction results through a multi-layer perceptron. According to the method, the attention mechanism is combined with the user features to fuse the video multi-modal features, so that the whole recommendation is richer and more personalized; meanwhile, because of multi-domain features and with consideration of the importance ofinteraction features in r</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language chi ; eng
recordid cdi_epo_espacenet_CN111246256A
source esp@cenet
subjects CALCULATING
COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
COMPUTING
COUNTING
ELECTRIC COMMUNICATION TECHNIQUE
ELECTRIC DIGITAL DATA PROCESSING
ELECTRICITY
PHYSICS
PICTORIAL COMMUNICATION, e.g. TELEVISION
title Video recommendation method based on multi-modal video content and multi-task learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-30T21%3A20%3A50IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=SHI%20JINGLUN&rft.date=2020-06-05&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN111246256A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true