MODEL DISTILLATION METHOD AND RELATED DEVICE

This disclosure relates to the field of artificial intelligence, and provides model distillation methods and apparatuses. In an implementation, a method including: obtaining first input data and second input data from a second computing node, wherein the first input data is output data of the third...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Jiang, Xin, Shang, Lifeng, Bai, Haoli, HOU, Lu, Qian, Li
Format: Patent
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Jiang, Xin
Shang, Lifeng
Bai, Haoli
HOU, Lu
Qian, Li
description This disclosure relates to the field of artificial intelligence, and provides model distillation methods and apparatuses. In an implementation, a method including: obtaining first input data and second input data from a second computing node, wherein the first input data is output data of the third sub-model, and the second input data is output data processed by the fourth sub-model, processing the first input data by using the first sub-model, to obtain a first intermediate output, processing the second input data by using the second sub-model, to obtain a second intermediate output, wherein the first intermediate output and the second intermediate output are used to determine a first gradient, and distilling the first sub-model based on the first gradient, to obtain an updated first sub-model.
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_US2024185086A1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>US2024185086A1</sourcerecordid><originalsourceid>FETCH-epo_espacenet_US2024185086A13</originalsourceid><addsrcrecordid>eNrjZNDx9Xdx9VFw8QwO8fTxcQzx9PdT8HUN8fB3UXD0c1EIcgWKuboouLiGeTq78jCwpiXmFKfyQmluBmU31xBnD93Ugvz41OKCxOTUvNSS-NBgIwMjE0MLUwMLM0dDY-JUAQAbwyX1</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>MODEL DISTILLATION METHOD AND RELATED DEVICE</title><source>esp@cenet</source><creator>Jiang, Xin ; Shang, Lifeng ; Bai, Haoli ; HOU, Lu ; Qian, Li</creator><creatorcontrib>Jiang, Xin ; Shang, Lifeng ; Bai, Haoli ; HOU, Lu ; Qian, Li</creatorcontrib><description>This disclosure relates to the field of artificial intelligence, and provides model distillation methods and apparatuses. In an implementation, a method including: obtaining first input data and second input data from a second computing node, wherein the first input data is output data of the third sub-model, and the second input data is output data processed by the fourth sub-model, processing the first input data by using the first sub-model, to obtain a first intermediate output, processing the second input data by using the second sub-model, to obtain a second intermediate output, wherein the first intermediate output and the second intermediate output are used to determine a first gradient, and distilling the first sub-model based on the first gradient, to obtain an updated first sub-model.</description><language>eng</language><subject>CALCULATING ; COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS ; COMPUTING ; COUNTING ; PHYSICS</subject><creationdate>2024</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240606&amp;DB=EPODOC&amp;CC=US&amp;NR=2024185086A1$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,776,881,25542,76289</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240606&amp;DB=EPODOC&amp;CC=US&amp;NR=2024185086A1$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>Jiang, Xin</creatorcontrib><creatorcontrib>Shang, Lifeng</creatorcontrib><creatorcontrib>Bai, Haoli</creatorcontrib><creatorcontrib>HOU, Lu</creatorcontrib><creatorcontrib>Qian, Li</creatorcontrib><title>MODEL DISTILLATION METHOD AND RELATED DEVICE</title><description>This disclosure relates to the field of artificial intelligence, and provides model distillation methods and apparatuses. In an implementation, a method including: obtaining first input data and second input data from a second computing node, wherein the first input data is output data of the third sub-model, and the second input data is output data processed by the fourth sub-model, processing the first input data by using the first sub-model, to obtain a first intermediate output, processing the second input data by using the second sub-model, to obtain a second intermediate output, wherein the first intermediate output and the second intermediate output are used to determine a first gradient, and distilling the first sub-model based on the first gradient, to obtain an updated first sub-model.</description><subject>CALCULATING</subject><subject>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>PHYSICS</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2024</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZNDx9Xdx9VFw8QwO8fTxcQzx9PdT8HUN8fB3UXD0c1EIcgWKuboouLiGeTq78jCwpiXmFKfyQmluBmU31xBnD93Ugvz41OKCxOTUvNSS-NBgIwMjE0MLUwMLM0dDY-JUAQAbwyX1</recordid><startdate>20240606</startdate><enddate>20240606</enddate><creator>Jiang, Xin</creator><creator>Shang, Lifeng</creator><creator>Bai, Haoli</creator><creator>HOU, Lu</creator><creator>Qian, Li</creator><scope>EVB</scope></search><sort><creationdate>20240606</creationdate><title>MODEL DISTILLATION METHOD AND RELATED DEVICE</title><author>Jiang, Xin ; Shang, Lifeng ; Bai, Haoli ; HOU, Lu ; Qian, Li</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_US2024185086A13</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>eng</language><creationdate>2024</creationdate><topic>CALCULATING</topic><topic>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>PHYSICS</topic><toplevel>online_resources</toplevel><creatorcontrib>Jiang, Xin</creatorcontrib><creatorcontrib>Shang, Lifeng</creatorcontrib><creatorcontrib>Bai, Haoli</creatorcontrib><creatorcontrib>HOU, Lu</creatorcontrib><creatorcontrib>Qian, Li</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Jiang, Xin</au><au>Shang, Lifeng</au><au>Bai, Haoli</au><au>HOU, Lu</au><au>Qian, Li</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>MODEL DISTILLATION METHOD AND RELATED DEVICE</title><date>2024-06-06</date><risdate>2024</risdate><abstract>This disclosure relates to the field of artificial intelligence, and provides model distillation methods and apparatuses. In an implementation, a method including: obtaining first input data and second input data from a second computing node, wherein the first input data is output data of the third sub-model, and the second input data is output data processed by the fourth sub-model, processing the first input data by using the first sub-model, to obtain a first intermediate output, processing the second input data by using the second sub-model, to obtain a second intermediate output, wherein the first intermediate output and the second intermediate output are used to determine a first gradient, and distilling the first sub-model based on the first gradient, to obtain an updated first sub-model.</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language eng
recordid cdi_epo_espacenet_US2024185086A1
source esp@cenet
subjects CALCULATING
COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
COMPUTING
COUNTING
PHYSICS
title MODEL DISTILLATION METHOD AND RELATED DEVICE
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T12%3A45%3A27IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=Jiang,%20Xin&rft.date=2024-06-06&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EUS2024185086A1%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true