EMPIRICAL ANALYSIS OF IEEE754, FIXED-POINT AND POSIT IN LOW PRECISION MACHINE LEARNING

Deep neural networks have changed the current algorithms' results in applications such as object classification, image segmentation or natural language processing. To increase their accuracy, they became more complex and more costly in terms of storage, computation time and energy consumption....

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Scientific Bulletin. Series C, Electrical Engineering and Computer Science Electrical Engineering and Computer Science, 2023-01 (3), p.13
Hauptverfasser: Ciocırlan, Ştefan-Dan, Neacsu, Teodor-Andrei, Rughinis, Razvan-Victor
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue 3
container_start_page 13
container_title Scientific Bulletin. Series C, Electrical Engineering and Computer Science
container_volume
creator Ciocırlan, Ştefan-Dan
Neacsu, Teodor-Andrei
Rughinis, Razvan-Victor
description Deep neural networks have changed the current algorithms' results in applications such as object classification, image segmentation or natural language processing. To increase their accuracy, they became more complex and more costly in terms of storage, computation time and energy consumption. This paper attacks the problem of storage and presents the advantages of using different number representations as fixed-point and posit numbers for deep neural network inference. The deep neural networks were trained using the proposed framework Low Precision Machine Learning (LPML) with 32-bit IEEE754. The storage was first optimized by the usage of knowledge distillation and then by modifying layer by layer the number representation together with the precision. The first significant results were made by modifying the number representation of the network but keeping the same precision per layer. For a 2-layer network (2LayerNet) using 16-bit Posit, the accuracy is 93.45%, close to 93.47%, the accuracy for using 32-bit IEEE754. Using the 8-bit Posit decreases the accuracy by 1.29%, but at the same time, it reduces the storage space by 75%. The usage of fixed point representation showed a small tolerance in the number of bits used for the fractional part. Using a 4-4 bit fixed point (4 bits for the integer part and 4 bits for the fractional part) reduces the storage used by 75% but decreases accuracy as low as 67.21%. When at least 8 bits are used for the fractional part, the results are similar to the 32-bit IEEE754. To increase accuracy before reducing precision, knowledge distillation was used. A ResNet18 network gained an 0.87% increase in accuracy by using a ResNet34 as a professor. By changing the number representation system and precision per layer, the storage was reduced by 43.47%, and the accuracy decreased by 0.26%. In conclusion, with the usage of knowledge distillation and change of number representation and precision per layer, the Resnet18 network had 66.75% smaller storage space than the ResNet34 professor network by losing only 1.38% in accuracy.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2870077890</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2870077890</sourcerecordid><originalsourceid>FETCH-LOGICAL-p183t-6e0298e7cfb82a48c6fe6d52f209ba8495957746537d91226d45d1cf4048ebc3</originalsourceid><addsrcrecordid>eNotjclqwzAUAHVooSHNPwh6rUF-2o_GkZMHjmRs0-UUvMjQUpo0Tv6_KelpLsPMHVkAGJVwKdgDWc3zR8-4AgVWygV5cbsKa8yzkmY-K98bbGgoKDrntBTPtMA3t06qgL69CmtahQZbip6W4ZVWtcuxweDpLsu36B0tXVZ79JtHcj91X3Nc_XNJ2sK1-TYpw-ZvlhxTw8-JigysiXqYegOdMIOaoholTMBs3xlhpZVaCyW5Hm0KoEYhx3SYBBMm9gNfkqdb9ng6_FzifN5_Hi6n7-txD0YzprWxjP8CUOdDeQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2870077890</pqid></control><display><type>article</type><title>EMPIRICAL ANALYSIS OF IEEE754, FIXED-POINT AND POSIT IN LOW PRECISION MACHINE LEARNING</title><source>EZB-FREE-00999 freely available EZB journals</source><creator>Ciocırlan, Ştefan-Dan ; Neacsu, Teodor-Andrei ; Rughinis, Razvan-Victor</creator><creatorcontrib>Ciocırlan, Ştefan-Dan ; Neacsu, Teodor-Andrei ; Rughinis, Razvan-Victor</creatorcontrib><description>Deep neural networks have changed the current algorithms' results in applications such as object classification, image segmentation or natural language processing. To increase their accuracy, they became more complex and more costly in terms of storage, computation time and energy consumption. This paper attacks the problem of storage and presents the advantages of using different number representations as fixed-point and posit numbers for deep neural network inference. The deep neural networks were trained using the proposed framework Low Precision Machine Learning (LPML) with 32-bit IEEE754. The storage was first optimized by the usage of knowledge distillation and then by modifying layer by layer the number representation together with the precision. The first significant results were made by modifying the number representation of the network but keeping the same precision per layer. For a 2-layer network (2LayerNet) using 16-bit Posit, the accuracy is 93.45%, close to 93.47%, the accuracy for using 32-bit IEEE754. Using the 8-bit Posit decreases the accuracy by 1.29%, but at the same time, it reduces the storage space by 75%. The usage of fixed point representation showed a small tolerance in the number of bits used for the fractional part. Using a 4-4 bit fixed point (4 bits for the integer part and 4 bits for the fractional part) reduces the storage used by 75% but decreases accuracy as low as 67.21%. When at least 8 bits are used for the fractional part, the results are similar to the 32-bit IEEE754. To increase accuracy before reducing precision, knowledge distillation was used. A ResNet18 network gained an 0.87% increase in accuracy by using a ResNet34 as a professor. By changing the number representation system and precision per layer, the storage was reduced by 43.47%, and the accuracy decreased by 0.26%. In conclusion, with the usage of knowledge distillation and change of number representation and precision per layer, the Resnet18 network had 66.75% smaller storage space than the ResNet34 professor network by losing only 1.38% in accuracy.</description><identifier>ISSN: 2286-3540</identifier><language>eng</language><publisher>Bucharest: University Polytechnica of Bucharest</publisher><subject>Accuracy ; Algorithms ; Artificial neural networks ; Distillation ; Empirical analysis ; Energy consumption ; Image classification ; Image segmentation ; Machine learning ; Natural language processing ; Neural networks ; Representations</subject><ispartof>Scientific Bulletin. Series C, Electrical Engineering and Computer Science, 2023-01 (3), p.13</ispartof><rights>Copyright University Polytechnica of Bucharest 2023</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780</link.rule.ids></links><search><creatorcontrib>Ciocırlan, Ştefan-Dan</creatorcontrib><creatorcontrib>Neacsu, Teodor-Andrei</creatorcontrib><creatorcontrib>Rughinis, Razvan-Victor</creatorcontrib><title>EMPIRICAL ANALYSIS OF IEEE754, FIXED-POINT AND POSIT IN LOW PRECISION MACHINE LEARNING</title><title>Scientific Bulletin. Series C, Electrical Engineering and Computer Science</title><description>Deep neural networks have changed the current algorithms' results in applications such as object classification, image segmentation or natural language processing. To increase their accuracy, they became more complex and more costly in terms of storage, computation time and energy consumption. This paper attacks the problem of storage and presents the advantages of using different number representations as fixed-point and posit numbers for deep neural network inference. The deep neural networks were trained using the proposed framework Low Precision Machine Learning (LPML) with 32-bit IEEE754. The storage was first optimized by the usage of knowledge distillation and then by modifying layer by layer the number representation together with the precision. The first significant results were made by modifying the number representation of the network but keeping the same precision per layer. For a 2-layer network (2LayerNet) using 16-bit Posit, the accuracy is 93.45%, close to 93.47%, the accuracy for using 32-bit IEEE754. Using the 8-bit Posit decreases the accuracy by 1.29%, but at the same time, it reduces the storage space by 75%. The usage of fixed point representation showed a small tolerance in the number of bits used for the fractional part. Using a 4-4 bit fixed point (4 bits for the integer part and 4 bits for the fractional part) reduces the storage used by 75% but decreases accuracy as low as 67.21%. When at least 8 bits are used for the fractional part, the results are similar to the 32-bit IEEE754. To increase accuracy before reducing precision, knowledge distillation was used. A ResNet18 network gained an 0.87% increase in accuracy by using a ResNet34 as a professor. By changing the number representation system and precision per layer, the storage was reduced by 43.47%, and the accuracy decreased by 0.26%. In conclusion, with the usage of knowledge distillation and change of number representation and precision per layer, the Resnet18 network had 66.75% smaller storage space than the ResNet34 professor network by losing only 1.38% in accuracy.</description><subject>Accuracy</subject><subject>Algorithms</subject><subject>Artificial neural networks</subject><subject>Distillation</subject><subject>Empirical analysis</subject><subject>Energy consumption</subject><subject>Image classification</subject><subject>Image segmentation</subject><subject>Machine learning</subject><subject>Natural language processing</subject><subject>Neural networks</subject><subject>Representations</subject><issn>2286-3540</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNotjclqwzAUAHVooSHNPwh6rUF-2o_GkZMHjmRs0-UUvMjQUpo0Tv6_KelpLsPMHVkAGJVwKdgDWc3zR8-4AgVWygV5cbsKa8yzkmY-K98bbGgoKDrntBTPtMA3t06qgL69CmtahQZbip6W4ZVWtcuxweDpLsu36B0tXVZ79JtHcj91X3Nc_XNJ2sK1-TYpw-ZvlhxTw8-JigysiXqYegOdMIOaoholTMBs3xlhpZVaCyW5Hm0KoEYhx3SYBBMm9gNfkqdb9ng6_FzifN5_Hi6n7-txD0YzprWxjP8CUOdDeQ</recordid><startdate>20230101</startdate><enddate>20230101</enddate><creator>Ciocırlan, Ştefan-Dan</creator><creator>Neacsu, Teodor-Andrei</creator><creator>Rughinis, Razvan-Victor</creator><general>University Polytechnica of Bucharest</general><scope>7SP</scope><scope>8FD</scope><scope>L7M</scope></search><sort><creationdate>20230101</creationdate><title>EMPIRICAL ANALYSIS OF IEEE754, FIXED-POINT AND POSIT IN LOW PRECISION MACHINE LEARNING</title><author>Ciocırlan, Ştefan-Dan ; Neacsu, Teodor-Andrei ; Rughinis, Razvan-Victor</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-p183t-6e0298e7cfb82a48c6fe6d52f209ba8495957746537d91226d45d1cf4048ebc3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Accuracy</topic><topic>Algorithms</topic><topic>Artificial neural networks</topic><topic>Distillation</topic><topic>Empirical analysis</topic><topic>Energy consumption</topic><topic>Image classification</topic><topic>Image segmentation</topic><topic>Machine learning</topic><topic>Natural language processing</topic><topic>Neural networks</topic><topic>Representations</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Ciocırlan, Ştefan-Dan</creatorcontrib><creatorcontrib>Neacsu, Teodor-Andrei</creatorcontrib><creatorcontrib>Rughinis, Razvan-Victor</creatorcontrib><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>Advanced Technologies Database with Aerospace</collection><jtitle>Scientific Bulletin. Series C, Electrical Engineering and Computer Science</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ciocırlan, Ştefan-Dan</au><au>Neacsu, Teodor-Andrei</au><au>Rughinis, Razvan-Victor</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>EMPIRICAL ANALYSIS OF IEEE754, FIXED-POINT AND POSIT IN LOW PRECISION MACHINE LEARNING</atitle><jtitle>Scientific Bulletin. Series C, Electrical Engineering and Computer Science</jtitle><date>2023-01-01</date><risdate>2023</risdate><issue>3</issue><spage>13</spage><pages>13-</pages><issn>2286-3540</issn><abstract>Deep neural networks have changed the current algorithms' results in applications such as object classification, image segmentation or natural language processing. To increase their accuracy, they became more complex and more costly in terms of storage, computation time and energy consumption. This paper attacks the problem of storage and presents the advantages of using different number representations as fixed-point and posit numbers for deep neural network inference. The deep neural networks were trained using the proposed framework Low Precision Machine Learning (LPML) with 32-bit IEEE754. The storage was first optimized by the usage of knowledge distillation and then by modifying layer by layer the number representation together with the precision. The first significant results were made by modifying the number representation of the network but keeping the same precision per layer. For a 2-layer network (2LayerNet) using 16-bit Posit, the accuracy is 93.45%, close to 93.47%, the accuracy for using 32-bit IEEE754. Using the 8-bit Posit decreases the accuracy by 1.29%, but at the same time, it reduces the storage space by 75%. The usage of fixed point representation showed a small tolerance in the number of bits used for the fractional part. Using a 4-4 bit fixed point (4 bits for the integer part and 4 bits for the fractional part) reduces the storage used by 75% but decreases accuracy as low as 67.21%. When at least 8 bits are used for the fractional part, the results are similar to the 32-bit IEEE754. To increase accuracy before reducing precision, knowledge distillation was used. A ResNet18 network gained an 0.87% increase in accuracy by using a ResNet34 as a professor. By changing the number representation system and precision per layer, the storage was reduced by 43.47%, and the accuracy decreased by 0.26%. In conclusion, with the usage of knowledge distillation and change of number representation and precision per layer, the Resnet18 network had 66.75% smaller storage space than the ResNet34 professor network by losing only 1.38% in accuracy.</abstract><cop>Bucharest</cop><pub>University Polytechnica of Bucharest</pub></addata></record>
fulltext fulltext
identifier ISSN: 2286-3540
ispartof Scientific Bulletin. Series C, Electrical Engineering and Computer Science, 2023-01 (3), p.13
issn 2286-3540
language eng
recordid cdi_proquest_journals_2870077890
source EZB-FREE-00999 freely available EZB journals
subjects Accuracy
Algorithms
Artificial neural networks
Distillation
Empirical analysis
Energy consumption
Image classification
Image segmentation
Machine learning
Natural language processing
Neural networks
Representations
title EMPIRICAL ANALYSIS OF IEEE754, FIXED-POINT AND POSIT IN LOW PRECISION MACHINE LEARNING
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-01T02%3A27%3A14IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=EMPIRICAL%20ANALYSIS%20OF%20IEEE754,%20FIXED-POINT%20AND%20POSIT%20IN%20LOW%20PRECISION%20MACHINE%20LEARNING&rft.jtitle=Scientific%20Bulletin.%20Series%20C,%20Electrical%20Engineering%20and%20Computer%20Science&rft.au=Cioc%C4%B1rlan,%20%C5%9Etefan-Dan&rft.date=2023-01-01&rft.issue=3&rft.spage=13&rft.pages=13-&rft.issn=2286-3540&rft_id=info:doi/&rft_dat=%3Cproquest%3E2870077890%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2870077890&rft_id=info:pmid/&rfr_iscdi=true