Acceleration method of convolutional neural network parallelization training

The invention provides an acceleration method of convolutional neural network parallelized training and a mixed-batch idea. The method is applied to a complete machine system composed of a CPU and anFPGA, and mainly solves the problem that under a large-scale convolutional neural network structure,...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: HONG QIFEI, SHI AOKAI, RUAN AIWU
Format: Patent
Sprache:chi ; eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator HONG QIFEI
SHI AOKAI
RUAN AIWU
description The invention provides an acceleration method of convolutional neural network parallelized training and a mixed-batch idea. The method is applied to a complete machine system composed of a CPU and anFPGA, and mainly solves the problem that under a large-scale convolutional neural network structure, when the FPGA is used for parallelization training of one batch sample, storage space is insufficient, and the method can be applied to image recognition and target detection in the field of computer vision. The above method includes the following steps that 1, in the data preprocessing stage, thesamples of a original training library are randomly rearranged; 2, in the feedforward calculation stage, data is written in shared memory in the form of the batch, based on the parallel processing ofeach layer of the convolutional neural network achieved through an OpenCL language, data of one sample in the batch of the previous layer is randomly read in a first full-connection layer in whole internet, and the output of th
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN108090565A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN108090565A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN108090565A3</originalsourceid><addsrcrecordid>eNrjZPBxTE5OzUktSizJzM9TyE0tychPUchPU0jOzyvLzykFiSbmKOSllhaBqZLy_KJshYJEIC8nNSezCqKtpCgxMy8zL52HgTUtMac4lRdKczMourmGOHvophbkx6cWFyQmpwKNiHf2MzSwMLA0MDUzdTQmRg0AJAQ2_A</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Acceleration method of convolutional neural network parallelization training</title><source>esp@cenet</source><creator>HONG QIFEI ; SHI AOKAI ; RUAN AIWU</creator><creatorcontrib>HONG QIFEI ; SHI AOKAI ; RUAN AIWU</creatorcontrib><description>The invention provides an acceleration method of convolutional neural network parallelized training and a mixed-batch idea. The method is applied to a complete machine system composed of a CPU and anFPGA, and mainly solves the problem that under a large-scale convolutional neural network structure, when the FPGA is used for parallelization training of one batch sample, storage space is insufficient, and the method can be applied to image recognition and target detection in the field of computer vision. The above method includes the following steps that 1, in the data preprocessing stage, thesamples of a original training library are randomly rearranged; 2, in the feedforward calculation stage, data is written in shared memory in the form of the batch, based on the parallel processing ofeach layer of the convolutional neural network achieved through an OpenCL language, data of one sample in the batch of the previous layer is randomly read in a first full-connection layer in whole internet, and the output of th</description><language>chi ; eng</language><subject>CALCULATING ; COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS ; COMPUTING ; COUNTING ; PHYSICS</subject><creationdate>2018</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20180529&amp;DB=EPODOC&amp;CC=CN&amp;NR=108090565A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,780,885,25564,76547</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20180529&amp;DB=EPODOC&amp;CC=CN&amp;NR=108090565A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>HONG QIFEI</creatorcontrib><creatorcontrib>SHI AOKAI</creatorcontrib><creatorcontrib>RUAN AIWU</creatorcontrib><title>Acceleration method of convolutional neural network parallelization training</title><description>The invention provides an acceleration method of convolutional neural network parallelized training and a mixed-batch idea. The method is applied to a complete machine system composed of a CPU and anFPGA, and mainly solves the problem that under a large-scale convolutional neural network structure, when the FPGA is used for parallelization training of one batch sample, storage space is insufficient, and the method can be applied to image recognition and target detection in the field of computer vision. The above method includes the following steps that 1, in the data preprocessing stage, thesamples of a original training library are randomly rearranged; 2, in the feedforward calculation stage, data is written in shared memory in the form of the batch, based on the parallel processing ofeach layer of the convolutional neural network achieved through an OpenCL language, data of one sample in the batch of the previous layer is randomly read in a first full-connection layer in whole internet, and the output of th</description><subject>CALCULATING</subject><subject>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>PHYSICS</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2018</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZPBxTE5OzUktSizJzM9TyE0tychPUchPU0jOzyvLzykFiSbmKOSllhaBqZLy_KJshYJEIC8nNSezCqKtpCgxMy8zL52HgTUtMac4lRdKczMourmGOHvophbkx6cWFyQmpwKNiHf2MzSwMLA0MDUzdTQmRg0AJAQ2_A</recordid><startdate>20180529</startdate><enddate>20180529</enddate><creator>HONG QIFEI</creator><creator>SHI AOKAI</creator><creator>RUAN AIWU</creator><scope>EVB</scope></search><sort><creationdate>20180529</creationdate><title>Acceleration method of convolutional neural network parallelization training</title><author>HONG QIFEI ; SHI AOKAI ; RUAN AIWU</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN108090565A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2018</creationdate><topic>CALCULATING</topic><topic>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>PHYSICS</topic><toplevel>online_resources</toplevel><creatorcontrib>HONG QIFEI</creatorcontrib><creatorcontrib>SHI AOKAI</creatorcontrib><creatorcontrib>RUAN AIWU</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>HONG QIFEI</au><au>SHI AOKAI</au><au>RUAN AIWU</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Acceleration method of convolutional neural network parallelization training</title><date>2018-05-29</date><risdate>2018</risdate><abstract>The invention provides an acceleration method of convolutional neural network parallelized training and a mixed-batch idea. The method is applied to a complete machine system composed of a CPU and anFPGA, and mainly solves the problem that under a large-scale convolutional neural network structure, when the FPGA is used for parallelization training of one batch sample, storage space is insufficient, and the method can be applied to image recognition and target detection in the field of computer vision. The above method includes the following steps that 1, in the data preprocessing stage, thesamples of a original training library are randomly rearranged; 2, in the feedforward calculation stage, data is written in shared memory in the form of the batch, based on the parallel processing ofeach layer of the convolutional neural network achieved through an OpenCL language, data of one sample in the batch of the previous layer is randomly read in a first full-connection layer in whole internet, and the output of th</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language chi ; eng
recordid cdi_epo_espacenet_CN108090565A
source esp@cenet
subjects CALCULATING
COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
COMPUTING
COUNTING
PHYSICS
title Acceleration method of convolutional neural network parallelization training
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T10%3A11%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=HONG%20QIFEI&rft.date=2018-05-29&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN108090565A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true