Rethinking FUN: Frequency-Domain Utilization Networks

The search for efficient neural network architectures has gained much focus in recent years, where modern architectures focus not only on accuracy but also on inference time and model size. Here, we present FUN, a family of novel Frequency-domain Utilization Networks. These networks utilize the inhe...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Goldberg, Kfir, Shapiro, Stav, Richardson, Elad, Avidan, Shai
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Goldberg, Kfir
Shapiro, Stav
Richardson, Elad
Avidan, Shai
description The search for efficient neural network architectures has gained much focus in recent years, where modern architectures focus not only on accuracy but also on inference time and model size. Here, we present FUN, a family of novel Frequency-domain Utilization Networks. These networks utilize the inherent efficiency of the frequency-domain by working directly in that domain, represented with the Discrete Cosine Transform. Using modern techniques and building blocks such as compound-scaling and inverted-residual layers we generate a set of such networks allowing one to balance between size, latency and accuracy while outperforming competing RGB-based models. Extensive evaluations verifies that our networks present strong alternatives to previous approaches. Moreover, we show that working in frequency domain allows for dynamic compression of the input at inference time without any explicit change to the architecture.
doi_str_mv 10.48550/arxiv.2012.03357
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2012_03357</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2012_03357</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-2acd170b7ca9a78916c9d2f0cf81ef8b5ba5b91d415fc76dbed3288c716efd963</originalsourceid><addsrcrecordid>eNotzstOAjEYQOFuWBDwAVg5LzBDL_TmjqCDJAQTA-vJ35s2QEfLCMLTE9DV2Z18CI0IriaKczyG_BuPFcWEVpgxLvuIv_vuM6ZtTB9FvVk9FXX23z8-2XP53O4hpmLTxV28QBfbVKx8d2rz9jBEvQC7g3_47wCt65f17LVcvs0Xs-myBCFlScE6IrGRFjRIpYmw2tGAbVDEB2W4AW40cRPCg5XCGe8YVcpKInxwWrABevzb3t3NV457yOfm5m_ufnYF2UJAmA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Rethinking FUN: Frequency-Domain Utilization Networks</title><source>arXiv.org</source><creator>Goldberg, Kfir ; Shapiro, Stav ; Richardson, Elad ; Avidan, Shai</creator><creatorcontrib>Goldberg, Kfir ; Shapiro, Stav ; Richardson, Elad ; Avidan, Shai</creatorcontrib><description>The search for efficient neural network architectures has gained much focus in recent years, where modern architectures focus not only on accuracy but also on inference time and model size. Here, we present FUN, a family of novel Frequency-domain Utilization Networks. These networks utilize the inherent efficiency of the frequency-domain by working directly in that domain, represented with the Discrete Cosine Transform. Using modern techniques and building blocks such as compound-scaling and inverted-residual layers we generate a set of such networks allowing one to balance between size, latency and accuracy while outperforming competing RGB-based models. Extensive evaluations verifies that our networks present strong alternatives to previous approaches. Moreover, we show that working in frequency domain allows for dynamic compression of the input at inference time without any explicit change to the architecture.</description><identifier>DOI: 10.48550/arxiv.2012.03357</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2020-12</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2012.03357$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2012.03357$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Goldberg, Kfir</creatorcontrib><creatorcontrib>Shapiro, Stav</creatorcontrib><creatorcontrib>Richardson, Elad</creatorcontrib><creatorcontrib>Avidan, Shai</creatorcontrib><title>Rethinking FUN: Frequency-Domain Utilization Networks</title><description>The search for efficient neural network architectures has gained much focus in recent years, where modern architectures focus not only on accuracy but also on inference time and model size. Here, we present FUN, a family of novel Frequency-domain Utilization Networks. These networks utilize the inherent efficiency of the frequency-domain by working directly in that domain, represented with the Discrete Cosine Transform. Using modern techniques and building blocks such as compound-scaling and inverted-residual layers we generate a set of such networks allowing one to balance between size, latency and accuracy while outperforming competing RGB-based models. Extensive evaluations verifies that our networks present strong alternatives to previous approaches. Moreover, we show that working in frequency domain allows for dynamic compression of the input at inference time without any explicit change to the architecture.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzstOAjEYQOFuWBDwAVg5LzBDL_TmjqCDJAQTA-vJ35s2QEfLCMLTE9DV2Z18CI0IriaKczyG_BuPFcWEVpgxLvuIv_vuM6ZtTB9FvVk9FXX23z8-2XP53O4hpmLTxV28QBfbVKx8d2rz9jBEvQC7g3_47wCt65f17LVcvs0Xs-myBCFlScE6IrGRFjRIpYmw2tGAbVDEB2W4AW40cRPCg5XCGe8YVcpKInxwWrABevzb3t3NV457yOfm5m_ufnYF2UJAmA</recordid><startdate>20201206</startdate><enddate>20201206</enddate><creator>Goldberg, Kfir</creator><creator>Shapiro, Stav</creator><creator>Richardson, Elad</creator><creator>Avidan, Shai</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20201206</creationdate><title>Rethinking FUN: Frequency-Domain Utilization Networks</title><author>Goldberg, Kfir ; Shapiro, Stav ; Richardson, Elad ; Avidan, Shai</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-2acd170b7ca9a78916c9d2f0cf81ef8b5ba5b91d415fc76dbed3288c716efd963</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Goldberg, Kfir</creatorcontrib><creatorcontrib>Shapiro, Stav</creatorcontrib><creatorcontrib>Richardson, Elad</creatorcontrib><creatorcontrib>Avidan, Shai</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Goldberg, Kfir</au><au>Shapiro, Stav</au><au>Richardson, Elad</au><au>Avidan, Shai</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Rethinking FUN: Frequency-Domain Utilization Networks</atitle><date>2020-12-06</date><risdate>2020</risdate><abstract>The search for efficient neural network architectures has gained much focus in recent years, where modern architectures focus not only on accuracy but also on inference time and model size. Here, we present FUN, a family of novel Frequency-domain Utilization Networks. These networks utilize the inherent efficiency of the frequency-domain by working directly in that domain, represented with the Discrete Cosine Transform. Using modern techniques and building blocks such as compound-scaling and inverted-residual layers we generate a set of such networks allowing one to balance between size, latency and accuracy while outperforming competing RGB-based models. Extensive evaluations verifies that our networks present strong alternatives to previous approaches. Moreover, we show that working in frequency domain allows for dynamic compression of the input at inference time without any explicit change to the architecture.</abstract><doi>10.48550/arxiv.2012.03357</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2012.03357
ispartof
issn
language eng
recordid cdi_arxiv_primary_2012_03357
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
title Rethinking FUN: Frequency-Domain Utilization Networks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-02T01%3A17%3A08IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Rethinking%20FUN:%20Frequency-Domain%20Utilization%20Networks&rft.au=Goldberg,%20Kfir&rft.date=2020-12-06&rft_id=info:doi/10.48550/arxiv.2012.03357&rft_dat=%3Carxiv_GOX%3E2012_03357%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true