A knowledge distilled attention-based latent information extraction network for sequential user behavior
When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a signi...
Gespeichert in:
Veröffentlicht in: | Multimedia tools and applications 2023, Vol.82 (1), p.1017-1043 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1043 |
---|---|
container_issue | 1 |
container_start_page | 1017 |
container_title | Multimedia tools and applications |
container_volume | 82 |
creator | Huang, Ruo McIntyre, Shelby Song, Meina E, Haihong Ou, Zhonghong |
description | When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a significant increase in computation time. Therefore, these models cannot meet the requirement of fast response needed in application scenarios such as online advertising. To deal with these issues, we propose a Knowledge Distilled Attention-based Latent Information Extraction Network for Sequential user behavior (KD-ALIENS). In this model structure, user and item attributes and history are utilized to model the latent information from high-order feature interactions in conjunction with user sequential historical behavior. With regard to the issues of long-distance dependency and noise, we have adopted the self-attention mechanism to learn the sequential patterns between items in a user-item interaction history. With regard to the issue of a complex model architecture which cannot meet the requirement of fast response, the use of model compression and acceleration is realized by: (a) use of a knowledge-distilled teacher and student module, wherein the complex teacher module extracts a user’s general preference from high-order feature interactions and sequential patterns of long history sequences; and (b) a sampling method to sample both the relatively long-term and short-term item histories. Experimental studies on two real-world datasets demonstrate considerable improvements for click-through rate (CTR) prediction accuracy relative to strong baseline models and also show the effectiveness of the student-model compression and acceleration for speed. |
doi_str_mv | 10.1007/s11042-022-12513-y |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2758755421</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2758755421</sourcerecordid><originalsourceid>FETCH-LOGICAL-c200t-89a652cfb5dd96843267b99521be0d0cc3f4019b87e04fdc2a0575d86349d3c33</originalsourceid><addsrcrecordid>eNp9UMtOwzAQtBBIlMIPcLLE2bC24zg5VhUvqRIXOFuO7bRp07jYKaV_j9MgceO0M9qZ2dUgdEvhngLIh0gpZIwAY4QyQTk5nqEJFZITKRk9T5gXQKQAeomuYlwD0FywbIJWM7zp_KF1dumwbWLftAlj3feu6xvfkUrHxFs9cNx0tQ9bPSyw--6DNifYuf7gwwanJY7ucz9YdYv30QVcuZX-any4Rhe1bqO7-Z1T9PH0-D5_IYu359f5bEEMA-hJUer0mKkrYW2ZFxlnuazKUjBaObBgDK8zoGVVSAdZbQ3TIKSwRc6z0nLD-RTdjbm74NMnsVdrvw9dOqmYFIUUImM0qdioMsHHGFytdqHZ6nBUFNTQqBobValRdWpUHZOJj6aYxN3Shb_of1w_vul7bg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2758755421</pqid></control><display><type>article</type><title>A knowledge distilled attention-based latent information extraction network for sequential user behavior</title><source>SpringerNature Journals</source><creator>Huang, Ruo ; McIntyre, Shelby ; Song, Meina ; E, Haihong ; Ou, Zhonghong</creator><creatorcontrib>Huang, Ruo ; McIntyre, Shelby ; Song, Meina ; E, Haihong ; Ou, Zhonghong</creatorcontrib><description>When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a significant increase in computation time. Therefore, these models cannot meet the requirement of fast response needed in application scenarios such as online advertising. To deal with these issues, we propose a Knowledge Distilled Attention-based Latent Information Extraction Network for Sequential user behavior (KD-ALIENS). In this model structure, user and item attributes and history are utilized to model the latent information from high-order feature interactions in conjunction with user sequential historical behavior. With regard to the issues of long-distance dependency and noise, we have adopted the self-attention mechanism to learn the sequential patterns between items in a user-item interaction history. With regard to the issue of a complex model architecture which cannot meet the requirement of fast response, the use of model compression and acceleration is realized by: (a) use of a knowledge-distilled teacher and student module, wherein the complex teacher module extracts a user’s general preference from high-order feature interactions and sequential patterns of long history sequences; and (b) a sampling method to sample both the relatively long-term and short-term item histories. Experimental studies on two real-world datasets demonstrate considerable improvements for click-through rate (CTR) prediction accuracy relative to strong baseline models and also show the effectiveness of the student-model compression and acceleration for speed.</description><identifier>ISSN: 1380-7501</identifier><identifier>EISSN: 1573-7721</identifier><identifier>DOI: 10.1007/s11042-022-12513-y</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Acceleration ; Complexity ; Compressive strength ; Computer Communication Networks ; Computer Science ; Data Structures and Information Theory ; Feature extraction ; Information retrieval ; Modules ; Multimedia Information Systems ; Recommender systems ; Sampling methods ; Sequences ; Special Purpose and Application-Based Systems ; Teachers ; User behavior</subject><ispartof>Multimedia tools and applications, 2023, Vol.82 (1), p.1017-1043</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022</rights><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c200t-89a652cfb5dd96843267b99521be0d0cc3f4019b87e04fdc2a0575d86349d3c33</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11042-022-12513-y$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11042-022-12513-y$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>315,781,785,27928,27929,41492,42561,51323</link.rule.ids></links><search><creatorcontrib>Huang, Ruo</creatorcontrib><creatorcontrib>McIntyre, Shelby</creatorcontrib><creatorcontrib>Song, Meina</creatorcontrib><creatorcontrib>E, Haihong</creatorcontrib><creatorcontrib>Ou, Zhonghong</creatorcontrib><title>A knowledge distilled attention-based latent information extraction network for sequential user behavior</title><title>Multimedia tools and applications</title><addtitle>Multimed Tools Appl</addtitle><description>When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a significant increase in computation time. Therefore, these models cannot meet the requirement of fast response needed in application scenarios such as online advertising. To deal with these issues, we propose a Knowledge Distilled Attention-based Latent Information Extraction Network for Sequential user behavior (KD-ALIENS). In this model structure, user and item attributes and history are utilized to model the latent information from high-order feature interactions in conjunction with user sequential historical behavior. With regard to the issues of long-distance dependency and noise, we have adopted the self-attention mechanism to learn the sequential patterns between items in a user-item interaction history. With regard to the issue of a complex model architecture which cannot meet the requirement of fast response, the use of model compression and acceleration is realized by: (a) use of a knowledge-distilled teacher and student module, wherein the complex teacher module extracts a user’s general preference from high-order feature interactions and sequential patterns of long history sequences; and (b) a sampling method to sample both the relatively long-term and short-term item histories. Experimental studies on two real-world datasets demonstrate considerable improvements for click-through rate (CTR) prediction accuracy relative to strong baseline models and also show the effectiveness of the student-model compression and acceleration for speed.</description><subject>Acceleration</subject><subject>Complexity</subject><subject>Compressive strength</subject><subject>Computer Communication Networks</subject><subject>Computer Science</subject><subject>Data Structures and Information Theory</subject><subject>Feature extraction</subject><subject>Information retrieval</subject><subject>Modules</subject><subject>Multimedia Information Systems</subject><subject>Recommender systems</subject><subject>Sampling methods</subject><subject>Sequences</subject><subject>Special Purpose and Application-Based Systems</subject><subject>Teachers</subject><subject>User behavior</subject><issn>1380-7501</issn><issn>1573-7721</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>8G5</sourceid><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><sourceid>GUQSH</sourceid><sourceid>M2O</sourceid><recordid>eNp9UMtOwzAQtBBIlMIPcLLE2bC24zg5VhUvqRIXOFuO7bRp07jYKaV_j9MgceO0M9qZ2dUgdEvhngLIh0gpZIwAY4QyQTk5nqEJFZITKRk9T5gXQKQAeomuYlwD0FywbIJWM7zp_KF1dumwbWLftAlj3feu6xvfkUrHxFs9cNx0tQ9bPSyw--6DNifYuf7gwwanJY7ucz9YdYv30QVcuZX-any4Rhe1bqO7-Z1T9PH0-D5_IYu359f5bEEMA-hJUer0mKkrYW2ZFxlnuazKUjBaObBgDK8zoGVVSAdZbQ3TIKSwRc6z0nLD-RTdjbm74NMnsVdrvw9dOqmYFIUUImM0qdioMsHHGFytdqHZ6nBUFNTQqBobValRdWpUHZOJj6aYxN3Shb_of1w_vul7bg</recordid><startdate>2023</startdate><enddate>2023</enddate><creator>Huang, Ruo</creator><creator>McIntyre, Shelby</creator><creator>Song, Meina</creator><creator>E, Haihong</creator><creator>Ou, Zhonghong</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>8G5</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M2O</scope><scope>MBDVC</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope></search><sort><creationdate>2023</creationdate><title>A knowledge distilled attention-based latent information extraction network for sequential user behavior</title><author>Huang, Ruo ; McIntyre, Shelby ; Song, Meina ; E, Haihong ; Ou, Zhonghong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c200t-89a652cfb5dd96843267b99521be0d0cc3f4019b87e04fdc2a0575d86349d3c33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Acceleration</topic><topic>Complexity</topic><topic>Compressive strength</topic><topic>Computer Communication Networks</topic><topic>Computer Science</topic><topic>Data Structures and Information Theory</topic><topic>Feature extraction</topic><topic>Information retrieval</topic><topic>Modules</topic><topic>Multimedia Information Systems</topic><topic>Recommender systems</topic><topic>Sampling methods</topic><topic>Sequences</topic><topic>Special Purpose and Application-Based Systems</topic><topic>Teachers</topic><topic>User behavior</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Huang, Ruo</creatorcontrib><creatorcontrib>McIntyre, Shelby</creatorcontrib><creatorcontrib>Song, Meina</creatorcontrib><creatorcontrib>E, Haihong</creatorcontrib><creatorcontrib>Ou, Zhonghong</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>Access via ABI/INFORM (ProQuest)</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Research Library (Alumni Edition)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Research Library</collection><collection>Research Library (Corporate)</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>ProQuest One Business</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><jtitle>Multimedia tools and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Huang, Ruo</au><au>McIntyre, Shelby</au><au>Song, Meina</au><au>E, Haihong</au><au>Ou, Zhonghong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A knowledge distilled attention-based latent information extraction network for sequential user behavior</atitle><jtitle>Multimedia tools and applications</jtitle><stitle>Multimed Tools Appl</stitle><date>2023</date><risdate>2023</risdate><volume>82</volume><issue>1</issue><spage>1017</spage><epage>1043</epage><pages>1017-1043</pages><issn>1380-7501</issn><eissn>1573-7721</eissn><abstract>When modeling user-item interaction sequences to extract sequential patterns, current recommender systems face the dual issues of a) long-distance dependencies in conjunction with b) high levels of noise. In addition, with the complexity of current recommendation model architectures there is a significant increase in computation time. Therefore, these models cannot meet the requirement of fast response needed in application scenarios such as online advertising. To deal with these issues, we propose a Knowledge Distilled Attention-based Latent Information Extraction Network for Sequential user behavior (KD-ALIENS). In this model structure, user and item attributes and history are utilized to model the latent information from high-order feature interactions in conjunction with user sequential historical behavior. With regard to the issues of long-distance dependency and noise, we have adopted the self-attention mechanism to learn the sequential patterns between items in a user-item interaction history. With regard to the issue of a complex model architecture which cannot meet the requirement of fast response, the use of model compression and acceleration is realized by: (a) use of a knowledge-distilled teacher and student module, wherein the complex teacher module extracts a user’s general preference from high-order feature interactions and sequential patterns of long history sequences; and (b) a sampling method to sample both the relatively long-term and short-term item histories. Experimental studies on two real-world datasets demonstrate considerable improvements for click-through rate (CTR) prediction accuracy relative to strong baseline models and also show the effectiveness of the student-model compression and acceleration for speed.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11042-022-12513-y</doi><tpages>27</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1380-7501 |
ispartof | Multimedia tools and applications, 2023, Vol.82 (1), p.1017-1043 |
issn | 1380-7501 1573-7721 |
language | eng |
recordid | cdi_proquest_journals_2758755421 |
source | SpringerNature Journals |
subjects | Acceleration Complexity Compressive strength Computer Communication Networks Computer Science Data Structures and Information Theory Feature extraction Information retrieval Modules Multimedia Information Systems Recommender systems Sampling methods Sequences Special Purpose and Application-Based Systems Teachers User behavior |
title | A knowledge distilled attention-based latent information extraction network for sequential user behavior |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-17T02%3A34%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20knowledge%20distilled%20attention-based%20latent%20information%20extraction%20network%20for%20sequential%20user%20behavior&rft.jtitle=Multimedia%20tools%20and%20applications&rft.au=Huang,%20Ruo&rft.date=2023&rft.volume=82&rft.issue=1&rft.spage=1017&rft.epage=1043&rft.pages=1017-1043&rft.issn=1380-7501&rft.eissn=1573-7721&rft_id=info:doi/10.1007/s11042-022-12513-y&rft_dat=%3Cproquest_cross%3E2758755421%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2758755421&rft_id=info:pmid/&rfr_iscdi=true |