Rethinking FUN: Frequency-Domain Utilization Networks
The search for efficient neural network architectures has gained much focus in recent years, where modern architectures focus not only on accuracy but also on inference time and model size. Here, we present FUN, a family of novel Frequency-domain Utilization Networks. These networks utilize the inhe...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Goldberg, Kfir Shapiro, Stav Richardson, Elad Avidan, Shai |
description | The search for efficient neural network architectures has gained much focus
in recent years, where modern architectures focus not only on accuracy but also
on inference time and model size. Here, we present FUN, a family of novel
Frequency-domain Utilization Networks. These networks utilize the inherent
efficiency of the frequency-domain by working directly in that domain,
represented with the Discrete Cosine Transform. Using modern techniques and
building blocks such as compound-scaling and inverted-residual layers we
generate a set of such networks allowing one to balance between size, latency
and accuracy while outperforming competing RGB-based models. Extensive
evaluations verifies that our networks present strong alternatives to previous
approaches. Moreover, we show that working in frequency domain allows for
dynamic compression of the input at inference time without any explicit change
to the architecture. |
doi_str_mv | 10.48550/arxiv.2012.03357 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2012_03357</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2012_03357</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-2acd170b7ca9a78916c9d2f0cf81ef8b5ba5b91d415fc76dbed3288c716efd963</originalsourceid><addsrcrecordid>eNotzstOAjEYQOFuWBDwAVg5LzBDL_TmjqCDJAQTA-vJ35s2QEfLCMLTE9DV2Z18CI0IriaKczyG_BuPFcWEVpgxLvuIv_vuM6ZtTB9FvVk9FXX23z8-2XP53O4hpmLTxV28QBfbVKx8d2rz9jBEvQC7g3_47wCt65f17LVcvs0Xs-myBCFlScE6IrGRFjRIpYmw2tGAbVDEB2W4AW40cRPCg5XCGe8YVcpKInxwWrABevzb3t3NV457yOfm5m_ufnYF2UJAmA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Rethinking FUN: Frequency-Domain Utilization Networks</title><source>arXiv.org</source><creator>Goldberg, Kfir ; Shapiro, Stav ; Richardson, Elad ; Avidan, Shai</creator><creatorcontrib>Goldberg, Kfir ; Shapiro, Stav ; Richardson, Elad ; Avidan, Shai</creatorcontrib><description>The search for efficient neural network architectures has gained much focus
in recent years, where modern architectures focus not only on accuracy but also
on inference time and model size. Here, we present FUN, a family of novel
Frequency-domain Utilization Networks. These networks utilize the inherent
efficiency of the frequency-domain by working directly in that domain,
represented with the Discrete Cosine Transform. Using modern techniques and
building blocks such as compound-scaling and inverted-residual layers we
generate a set of such networks allowing one to balance between size, latency
and accuracy while outperforming competing RGB-based models. Extensive
evaluations verifies that our networks present strong alternatives to previous
approaches. Moreover, we show that working in frequency domain allows for
dynamic compression of the input at inference time without any explicit change
to the architecture.</description><identifier>DOI: 10.48550/arxiv.2012.03357</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2020-12</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2012.03357$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2012.03357$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Goldberg, Kfir</creatorcontrib><creatorcontrib>Shapiro, Stav</creatorcontrib><creatorcontrib>Richardson, Elad</creatorcontrib><creatorcontrib>Avidan, Shai</creatorcontrib><title>Rethinking FUN: Frequency-Domain Utilization Networks</title><description>The search for efficient neural network architectures has gained much focus
in recent years, where modern architectures focus not only on accuracy but also
on inference time and model size. Here, we present FUN, a family of novel
Frequency-domain Utilization Networks. These networks utilize the inherent
efficiency of the frequency-domain by working directly in that domain,
represented with the Discrete Cosine Transform. Using modern techniques and
building blocks such as compound-scaling and inverted-residual layers we
generate a set of such networks allowing one to balance between size, latency
and accuracy while outperforming competing RGB-based models. Extensive
evaluations verifies that our networks present strong alternatives to previous
approaches. Moreover, we show that working in frequency domain allows for
dynamic compression of the input at inference time without any explicit change
to the architecture.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzstOAjEYQOFuWBDwAVg5LzBDL_TmjqCDJAQTA-vJ35s2QEfLCMLTE9DV2Z18CI0IriaKczyG_BuPFcWEVpgxLvuIv_vuM6ZtTB9FvVk9FXX23z8-2XP53O4hpmLTxV28QBfbVKx8d2rz9jBEvQC7g3_47wCt65f17LVcvs0Xs-myBCFlScE6IrGRFjRIpYmw2tGAbVDEB2W4AW40cRPCg5XCGe8YVcpKInxwWrABevzb3t3NV457yOfm5m_ufnYF2UJAmA</recordid><startdate>20201206</startdate><enddate>20201206</enddate><creator>Goldberg, Kfir</creator><creator>Shapiro, Stav</creator><creator>Richardson, Elad</creator><creator>Avidan, Shai</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20201206</creationdate><title>Rethinking FUN: Frequency-Domain Utilization Networks</title><author>Goldberg, Kfir ; Shapiro, Stav ; Richardson, Elad ; Avidan, Shai</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-2acd170b7ca9a78916c9d2f0cf81ef8b5ba5b91d415fc76dbed3288c716efd963</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Goldberg, Kfir</creatorcontrib><creatorcontrib>Shapiro, Stav</creatorcontrib><creatorcontrib>Richardson, Elad</creatorcontrib><creatorcontrib>Avidan, Shai</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Goldberg, Kfir</au><au>Shapiro, Stav</au><au>Richardson, Elad</au><au>Avidan, Shai</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Rethinking FUN: Frequency-Domain Utilization Networks</atitle><date>2020-12-06</date><risdate>2020</risdate><abstract>The search for efficient neural network architectures has gained much focus
in recent years, where modern architectures focus not only on accuracy but also
on inference time and model size. Here, we present FUN, a family of novel
Frequency-domain Utilization Networks. These networks utilize the inherent
efficiency of the frequency-domain by working directly in that domain,
represented with the Discrete Cosine Transform. Using modern techniques and
building blocks such as compound-scaling and inverted-residual layers we
generate a set of such networks allowing one to balance between size, latency
and accuracy while outperforming competing RGB-based models. Extensive
evaluations verifies that our networks present strong alternatives to previous
approaches. Moreover, we show that working in frequency domain allows for
dynamic compression of the input at inference time without any explicit change
to the architecture.</abstract><doi>10.48550/arxiv.2012.03357</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2012.03357 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2012_03357 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition |
title | Rethinking FUN: Frequency-Domain Utilization Networks |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-02T01%3A17%3A08IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Rethinking%20FUN:%20Frequency-Domain%20Utilization%20Networks&rft.au=Goldberg,%20Kfir&rft.date=2020-12-06&rft_id=info:doi/10.48550/arxiv.2012.03357&rft_dat=%3Carxiv_GOX%3E2012_03357%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |