METHOD AND APPARATUS FOR TRAINING NEURAL NETWORK, AND DEVICE AND STORAGE MEDIUM

Embodiments of the present invention provide a method and apparatus for training a neural network, and a device and a storage medium. The method comprises: at a first working node among a plurality of working nodes, acquiring a first set of global gradients for a first set of network layers in a neu...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: CHEN, Yangrui, XIE, Cong, GU, Juncheng, LIN, Haibin, PENG, Yanghua, ZHU, Yibo
Format: Patent
Sprache:chi ; eng ; fre
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator CHEN, Yangrui
XIE, Cong
GU, Juncheng
LIN, Haibin
PENG, Yanghua
ZHU, Yibo
description Embodiments of the present invention provide a method and apparatus for training a neural network, and a device and a storage medium. The method comprises: at a first working node among a plurality of working nodes, acquiring a first set of global gradients for a first set of network layers in a neural network, the first set of global gradients being obtained from aggregating local gradients determined by the plurality of working nodes for the first set of network layers in the current training step, and the plurality of working nodes being configured to jointly train the neural network; acquiring a second set of global gradients for a second set of network layers in the neural network, the second set of network layers being different from the first set of network layers, and the second set of global gradients being obtained from aggregating local gradients determined by the plurality of working nodes for the second set of network layers in a training step previous to the current training step; and updating p
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_WO2024104232A1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>WO2024104232A1</sourcerecordid><originalsourceid>FETCH-epo_espacenet_WO2024104232A13</originalsourceid><addsrcrecordid>eNrjZPD3dQ3x8HdRcPQD4oAAxyDHkNBgBTf_IIWQIEdPP08_dwU_19AgRx8gFRLuH-StA1bq4hrm6ewKZgaH-Ac5ursq-Lq6eIb68jCwpiXmFKfyQmluBmU31xBnD93Ugvz41OKCxOTUvNSS-HB_IwMjE0MDEyNjI0dDY-JUAQDSIi8_</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>METHOD AND APPARATUS FOR TRAINING NEURAL NETWORK, AND DEVICE AND STORAGE MEDIUM</title><source>esp@cenet</source><creator>CHEN, Yangrui ; XIE, Cong ; GU, Juncheng ; LIN, Haibin ; PENG, Yanghua ; ZHU, Yibo</creator><creatorcontrib>CHEN, Yangrui ; XIE, Cong ; GU, Juncheng ; LIN, Haibin ; PENG, Yanghua ; ZHU, Yibo</creatorcontrib><description>Embodiments of the present invention provide a method and apparatus for training a neural network, and a device and a storage medium. The method comprises: at a first working node among a plurality of working nodes, acquiring a first set of global gradients for a first set of network layers in a neural network, the first set of global gradients being obtained from aggregating local gradients determined by the plurality of working nodes for the first set of network layers in the current training step, and the plurality of working nodes being configured to jointly train the neural network; acquiring a second set of global gradients for a second set of network layers in the neural network, the second set of network layers being different from the first set of network layers, and the second set of global gradients being obtained from aggregating local gradients determined by the plurality of working nodes for the second set of network layers in a training step previous to the current training step; and updating p</description><language>chi ; eng ; fre</language><subject>CALCULATING ; COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS ; COMPUTING ; COUNTING ; PHYSICS</subject><creationdate>2024</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240523&amp;DB=EPODOC&amp;CC=WO&amp;NR=2024104232A1$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,776,881,25542,76290</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240523&amp;DB=EPODOC&amp;CC=WO&amp;NR=2024104232A1$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>CHEN, Yangrui</creatorcontrib><creatorcontrib>XIE, Cong</creatorcontrib><creatorcontrib>GU, Juncheng</creatorcontrib><creatorcontrib>LIN, Haibin</creatorcontrib><creatorcontrib>PENG, Yanghua</creatorcontrib><creatorcontrib>ZHU, Yibo</creatorcontrib><title>METHOD AND APPARATUS FOR TRAINING NEURAL NETWORK, AND DEVICE AND STORAGE MEDIUM</title><description>Embodiments of the present invention provide a method and apparatus for training a neural network, and a device and a storage medium. The method comprises: at a first working node among a plurality of working nodes, acquiring a first set of global gradients for a first set of network layers in a neural network, the first set of global gradients being obtained from aggregating local gradients determined by the plurality of working nodes for the first set of network layers in the current training step, and the plurality of working nodes being configured to jointly train the neural network; acquiring a second set of global gradients for a second set of network layers in the neural network, the second set of network layers being different from the first set of network layers, and the second set of global gradients being obtained from aggregating local gradients determined by the plurality of working nodes for the second set of network layers in a training step previous to the current training step; and updating p</description><subject>CALCULATING</subject><subject>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>PHYSICS</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2024</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZPD3dQ3x8HdRcPQD4oAAxyDHkNBgBTf_IIWQIEdPP08_dwU_19AgRx8gFRLuH-StA1bq4hrm6ewKZgaH-Ac5ursq-Lq6eIb68jCwpiXmFKfyQmluBmU31xBnD93Ugvz41OKCxOTUvNSS-HB_IwMjE0MDEyNjI0dDY-JUAQDSIi8_</recordid><startdate>20240523</startdate><enddate>20240523</enddate><creator>CHEN, Yangrui</creator><creator>XIE, Cong</creator><creator>GU, Juncheng</creator><creator>LIN, Haibin</creator><creator>PENG, Yanghua</creator><creator>ZHU, Yibo</creator><scope>EVB</scope></search><sort><creationdate>20240523</creationdate><title>METHOD AND APPARATUS FOR TRAINING NEURAL NETWORK, AND DEVICE AND STORAGE MEDIUM</title><author>CHEN, Yangrui ; XIE, Cong ; GU, Juncheng ; LIN, Haibin ; PENG, Yanghua ; ZHU, Yibo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_WO2024104232A13</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng ; fre</language><creationdate>2024</creationdate><topic>CALCULATING</topic><topic>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>PHYSICS</topic><toplevel>online_resources</toplevel><creatorcontrib>CHEN, Yangrui</creatorcontrib><creatorcontrib>XIE, Cong</creatorcontrib><creatorcontrib>GU, Juncheng</creatorcontrib><creatorcontrib>LIN, Haibin</creatorcontrib><creatorcontrib>PENG, Yanghua</creatorcontrib><creatorcontrib>ZHU, Yibo</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>CHEN, Yangrui</au><au>XIE, Cong</au><au>GU, Juncheng</au><au>LIN, Haibin</au><au>PENG, Yanghua</au><au>ZHU, Yibo</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>METHOD AND APPARATUS FOR TRAINING NEURAL NETWORK, AND DEVICE AND STORAGE MEDIUM</title><date>2024-05-23</date><risdate>2024</risdate><abstract>Embodiments of the present invention provide a method and apparatus for training a neural network, and a device and a storage medium. The method comprises: at a first working node among a plurality of working nodes, acquiring a first set of global gradients for a first set of network layers in a neural network, the first set of global gradients being obtained from aggregating local gradients determined by the plurality of working nodes for the first set of network layers in the current training step, and the plurality of working nodes being configured to jointly train the neural network; acquiring a second set of global gradients for a second set of network layers in the neural network, the second set of network layers being different from the first set of network layers, and the second set of global gradients being obtained from aggregating local gradients determined by the plurality of working nodes for the second set of network layers in a training step previous to the current training step; and updating p</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language chi ; eng ; fre
recordid cdi_epo_espacenet_WO2024104232A1
source esp@cenet
subjects CALCULATING
COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
COMPUTING
COUNTING
PHYSICS
title METHOD AND APPARATUS FOR TRAINING NEURAL NETWORK, AND DEVICE AND STORAGE MEDIUM
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-02T18%3A28%3A13IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=CHEN,%20Yangrui&rft.date=2024-05-23&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EWO2024104232A1%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true