A knowledge distilled attention-based latent information extraction network for sequential user behavior

When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a signi...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Multimedia tools and applications 2023, Vol.82 (1), p.1017-1043
Hauptverfasser: Huang, Ruo, McIntyre, Shelby, Song, Meina, E, Haihong, Ou, Zhonghong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1043
container_issue 1
container_start_page 1017
container_title Multimedia tools and applications
container_volume 82
creator Huang, Ruo
McIntyre, Shelby
Song, Meina
E, Haihong
Ou, Zhonghong
description When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a significant increase in computation time. Therefore, these models cannot meet the requirement of fast response needed in application scenarios such as online advertising. To deal with these issues, we propose a Knowledge Distilled Attention-based Latent Information Extraction Network for Sequential user behavior (KD-ALIENS). In this model structure, user and item attributes and history are utilized to model the latent information from high-order feature interactions in conjunction with user sequential historical behavior. With regard to the issues of long-distance dependency and noise, we have adopted the self-attention mechanism to learn the sequential patterns between items in a user-item interaction history. With regard to the issue of a complex model architecture which cannot meet the requirement of fast response, the use of model compression and acceleration is realized by: (a) use of a knowledge-distilled teacher and student module, wherein the complex teacher module extracts a user’s general preference from high-order feature interactions and sequential patterns of long history sequences; and (b) a sampling method to sample both the relatively long-term and short-term item histories. Experimental studies on two real-world datasets demonstrate considerable improvements for click-through rate (CTR) prediction accuracy relative to strong baseline models and also show the effectiveness of the student-model compression and acceleration for speed.
doi_str_mv 10.1007/s11042-022-12513-y
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2758755421</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2758755421</sourcerecordid><originalsourceid>FETCH-LOGICAL-c200t-89a652cfb5dd96843267b99521be0d0cc3f4019b87e04fdc2a0575d86349d3c33</originalsourceid><addsrcrecordid>eNp9UMtOwzAQtBBIlMIPcLLE2bC24zg5VhUvqRIXOFuO7bRp07jYKaV_j9MgceO0M9qZ2dUgdEvhngLIh0gpZIwAY4QyQTk5nqEJFZITKRk9T5gXQKQAeomuYlwD0FywbIJWM7zp_KF1dumwbWLftAlj3feu6xvfkUrHxFs9cNx0tQ9bPSyw--6DNifYuf7gwwanJY7ucz9YdYv30QVcuZX-any4Rhe1bqO7-Z1T9PH0-D5_IYu359f5bEEMA-hJUer0mKkrYW2ZFxlnuazKUjBaObBgDK8zoGVVSAdZbQ3TIKSwRc6z0nLD-RTdjbm74NMnsVdrvw9dOqmYFIUUImM0qdioMsHHGFytdqHZ6nBUFNTQqBobValRdWpUHZOJj6aYxN3Shb_of1w_vul7bg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2758755421</pqid></control><display><type>article</type><title>A knowledge distilled attention-based latent information extraction network for sequential user behavior</title><source>SpringerNature Journals</source><creator>Huang, Ruo ; McIntyre, Shelby ; Song, Meina ; E, Haihong ; Ou, Zhonghong</creator><creatorcontrib>Huang, Ruo ; McIntyre, Shelby ; Song, Meina ; E, Haihong ; Ou, Zhonghong</creatorcontrib><description>When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a significant increase in computation time. Therefore, these models cannot meet the requirement of fast response needed in application scenarios such as online advertising. To deal with these issues, we propose a Knowledge Distilled Attention-based Latent Information Extraction Network for Sequential user behavior (KD-ALIENS). In this model structure, user and item attributes and history are utilized to model the latent information from high-order feature interactions in conjunction with user sequential historical behavior. With regard to the issues of long-distance dependency and noise, we have adopted the self-attention mechanism to learn the sequential patterns between items in a user-item interaction history. With regard to the issue of a complex model architecture which cannot meet the requirement of fast response, the use of model compression and acceleration is realized by: (a) use of a knowledge-distilled teacher and student module, wherein the complex teacher module extracts a user’s general preference from high-order feature interactions and sequential patterns of long history sequences; and (b) a sampling method to sample both the relatively long-term and short-term item histories. Experimental studies on two real-world datasets demonstrate considerable improvements for click-through rate (CTR) prediction accuracy relative to strong baseline models and also show the effectiveness of the student-model compression and acceleration for speed.</description><identifier>ISSN: 1380-7501</identifier><identifier>EISSN: 1573-7721</identifier><identifier>DOI: 10.1007/s11042-022-12513-y</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Acceleration ; Complexity ; Compressive strength ; Computer Communication Networks ; Computer Science ; Data Structures and Information Theory ; Feature extraction ; Information retrieval ; Modules ; Multimedia Information Systems ; Recommender systems ; Sampling methods ; Sequences ; Special Purpose and Application-Based Systems ; Teachers ; User behavior</subject><ispartof>Multimedia tools and applications, 2023, Vol.82 (1), p.1017-1043</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022</rights><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c200t-89a652cfb5dd96843267b99521be0d0cc3f4019b87e04fdc2a0575d86349d3c33</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11042-022-12513-y$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11042-022-12513-y$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>315,781,785,27928,27929,41492,42561,51323</link.rule.ids></links><search><creatorcontrib>Huang, Ruo</creatorcontrib><creatorcontrib>McIntyre, Shelby</creatorcontrib><creatorcontrib>Song, Meina</creatorcontrib><creatorcontrib>E, Haihong</creatorcontrib><creatorcontrib>Ou, Zhonghong</creatorcontrib><title>A knowledge distilled attention-based latent information extraction network for sequential user behavior</title><title>Multimedia tools and applications</title><addtitle>Multimed Tools Appl</addtitle><description>When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a significant increase in computation time. Therefore, these models cannot meet the requirement of fast response needed in application scenarios such as online advertising. To deal with these issues, we propose a Knowledge Distilled Attention-based Latent Information Extraction Network for Sequential user behavior (KD-ALIENS). In this model structure, user and item attributes and history are utilized to model the latent information from high-order feature interactions in conjunction with user sequential historical behavior. With regard to the issues of long-distance dependency and noise, we have adopted the self-attention mechanism to learn the sequential patterns between items in a user-item interaction history. With regard to the issue of a complex model architecture which cannot meet the requirement of fast response, the use of model compression and acceleration is realized by: (a) use of a knowledge-distilled teacher and student module, wherein the complex teacher module extracts a user’s general preference from high-order feature interactions and sequential patterns of long history sequences; and (b) a sampling method to sample both the relatively long-term and short-term item histories. Experimental studies on two real-world datasets demonstrate considerable improvements for click-through rate (CTR) prediction accuracy relative to strong baseline models and also show the effectiveness of the student-model compression and acceleration for speed.</description><subject>Acceleration</subject><subject>Complexity</subject><subject>Compressive strength</subject><subject>Computer Communication Networks</subject><subject>Computer Science</subject><subject>Data Structures and Information Theory</subject><subject>Feature extraction</subject><subject>Information retrieval</subject><subject>Modules</subject><subject>Multimedia Information Systems</subject><subject>Recommender systems</subject><subject>Sampling methods</subject><subject>Sequences</subject><subject>Special Purpose and Application-Based Systems</subject><subject>Teachers</subject><subject>User behavior</subject><issn>1380-7501</issn><issn>1573-7721</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>8G5</sourceid><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><sourceid>GUQSH</sourceid><sourceid>M2O</sourceid><recordid>eNp9UMtOwzAQtBBIlMIPcLLE2bC24zg5VhUvqRIXOFuO7bRp07jYKaV_j9MgceO0M9qZ2dUgdEvhngLIh0gpZIwAY4QyQTk5nqEJFZITKRk9T5gXQKQAeomuYlwD0FywbIJWM7zp_KF1dumwbWLftAlj3feu6xvfkUrHxFs9cNx0tQ9bPSyw--6DNifYuf7gwwanJY7ucz9YdYv30QVcuZX-any4Rhe1bqO7-Z1T9PH0-D5_IYu359f5bEEMA-hJUer0mKkrYW2ZFxlnuazKUjBaObBgDK8zoGVVSAdZbQ3TIKSwRc6z0nLD-RTdjbm74NMnsVdrvw9dOqmYFIUUImM0qdioMsHHGFytdqHZ6nBUFNTQqBobValRdWpUHZOJj6aYxN3Shb_of1w_vul7bg</recordid><startdate>2023</startdate><enddate>2023</enddate><creator>Huang, Ruo</creator><creator>McIntyre, Shelby</creator><creator>Song, Meina</creator><creator>E, Haihong</creator><creator>Ou, Zhonghong</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>8G5</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M2O</scope><scope>MBDVC</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope></search><sort><creationdate>2023</creationdate><title>A knowledge distilled attention-based latent information extraction network for sequential user behavior</title><author>Huang, Ruo ; McIntyre, Shelby ; Song, Meina ; E, Haihong ; Ou, Zhonghong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c200t-89a652cfb5dd96843267b99521be0d0cc3f4019b87e04fdc2a0575d86349d3c33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Acceleration</topic><topic>Complexity</topic><topic>Compressive strength</topic><topic>Computer Communication Networks</topic><topic>Computer Science</topic><topic>Data Structures and Information Theory</topic><topic>Feature extraction</topic><topic>Information retrieval</topic><topic>Modules</topic><topic>Multimedia Information Systems</topic><topic>Recommender systems</topic><topic>Sampling methods</topic><topic>Sequences</topic><topic>Special Purpose and Application-Based Systems</topic><topic>Teachers</topic><topic>User behavior</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Huang, Ruo</creatorcontrib><creatorcontrib>McIntyre, Shelby</creatorcontrib><creatorcontrib>Song, Meina</creatorcontrib><creatorcontrib>E, Haihong</creatorcontrib><creatorcontrib>Ou, Zhonghong</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>Access via ABI/INFORM (ProQuest)</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Research Library (Alumni Edition)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Research Library</collection><collection>Research Library (Corporate)</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest One Business</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><jtitle>Multimedia tools and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Huang, Ruo</au><au>McIntyre, Shelby</au><au>Song, Meina</au><au>E, Haihong</au><au>Ou, Zhonghong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A knowledge distilled attention-based latent information extraction network for sequential user behavior</atitle><jtitle>Multimedia tools and applications</jtitle><stitle>Multimed Tools Appl</stitle><date>2023</date><risdate>2023</risdate><volume>82</volume><issue>1</issue><spage>1017</spage><epage>1043</epage><pages>1017-1043</pages><issn>1380-7501</issn><eissn>1573-7721</eissn><abstract>When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a significant increase in computation time. Therefore, these models cannot meet the requirement of fast response needed in application scenarios such as online advertising. To deal with these issues, we propose a Knowledge Distilled Attention-based Latent Information Extraction Network for Sequential user behavior (KD-ALIENS). In this model structure, user and item attributes and history are utilized to model the latent information from high-order feature interactions in conjunction with user sequential historical behavior. With regard to the issues of long-distance dependency and noise, we have adopted the self-attention mechanism to learn the sequential patterns between items in a user-item interaction history. With regard to the issue of a complex model architecture which cannot meet the requirement of fast response, the use of model compression and acceleration is realized by: (a) use of a knowledge-distilled teacher and student module, wherein the complex teacher module extracts a user’s general preference from high-order feature interactions and sequential patterns of long history sequences; and (b) a sampling method to sample both the relatively long-term and short-term item histories. Experimental studies on two real-world datasets demonstrate considerable improvements for click-through rate (CTR) prediction accuracy relative to strong baseline models and also show the effectiveness of the student-model compression and acceleration for speed.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11042-022-12513-y</doi><tpages>27</tpages></addata></record>
fulltext fulltext
identifier ISSN: 1380-7501
ispartof Multimedia tools and applications, 2023, Vol.82 (1), p.1017-1043
issn 1380-7501
1573-7721
language eng
recordid cdi_proquest_journals_2758755421
source SpringerNature Journals
subjects Acceleration
Complexity
Compressive strength
Computer Communication Networks
Computer Science
Data Structures and Information Theory
Feature extraction
Information retrieval
Modules
Multimedia Information Systems
Recommender systems
Sampling methods
Sequences
Special Purpose and Application-Based Systems
Teachers
User behavior
title A knowledge distilled attention-based latent information extraction network for sequential user behavior
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-17T02%3A34%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20knowledge%20distilled%20attention-based%20latent%20information%20extraction%20network%20for%20sequential%20user%20behavior&rft.jtitle=Multimedia%20tools%20and%20applications&rft.au=Huang,%20Ruo&rft.date=2023&rft.volume=82&rft.issue=1&rft.spage=1017&rft.epage=1043&rft.pages=1017-1043&rft.issn=1380-7501&rft.eissn=1573-7721&rft_id=info:doi/10.1007/s11042-022-12513-y&rft_dat=%3Cproquest_cross%3E2758755421%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2758755421&rft_id=info:pmid/&rfr_iscdi=true