Deep Blind Image Quality Assessment Powered by Online Hard Example Mining

Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example min...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on multimedia 2023-01, Vol.25, p.1-11
Hauptverfasser: Wang, Zhihua, Jiang, Qiuping, Zhao, Shanshan, Feng, Wensen, Lin, Weisi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 11
container_issue
container_start_page 1
container_title IEEE transactions on multimedia
container_volume 25
creator Wang, Zhihua
Jiang, Qiuping
Zhao, Shanshan
Feng, Wensen
Lin, Weisi
description Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example mining (OHEM) is a promising way to alleviate this issue. Inspired by the recent finding that network pruning disproportionately hampers the model's memorization of a tractable subset, atypical, low-quality, long-tailed samples, that are hard-to-memorize during training and easily "forgotten" during pruning, we propose an effective "plug-and-play" OHEM pipeline, especially for generalizable deep BIQA. Specifically, we train two parallel weight-sharing branches simultaneously, where one is full model and other is a "self-competitor" generated from the full model online by network pruning. Then, we leverage the prediction disagreement between the full model and its pruned variant (i.e., the self-competitor) to expose easily "forgettable" samples, which are therefore regarded as the hard ones. We then enforce the prediction consistency between the full model and its pruned variant to implicitly put more focus on these hard samples, which benefits the full model to recover forgettable information introduced by pruning. Extensive experiments across multiple datasets and BIQA models demonstrate that the proposed OHEM can improve the model performance and generalizability as measured by correlation numbers and group maximum differentiation (gMAD) competition. Our code are available at: https://github.com/wangzhihua520/IQA_with_OHEM
doi_str_mv 10.1109/TMM.2023.3257564
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_ieee_primary_10070789</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10070789</ieee_id><sourcerecordid>2904709180</sourcerecordid><originalsourceid>FETCH-LOGICAL-c292t-d6b609ff0c008b778f1c2ee7703ec80ac60bdc740209d9b2e1432ccb999e75673</originalsourceid><addsrcrecordid>eNpNkD1PwzAQhi0EEqWwMzBYYk45Ox-Ox1IKjdSqIJXZcpxLlSpf2Klo_z2u2oHpbnje93QPIY8MJoyBfNmsVhMOPJyEPBZxEl2REZMRCwCEuPZ7zCGQnMEtuXNuB8CiGMSIZG-IPX2tq7agWaO3SL_2uq6GI506h8412A70s_tFiwXNj3TdehTpQtuCzg-66Wukq6qt2u09uSl17fDhMsfk-32-mS2C5fojm02XgeGSD0GR5AnIsgQDkOZCpCUzHFEICNGkoE0CeWFEBBxkIXOOLAq5MbmUEv1fIhyT53Nvb7ufPbpB7bq9bf1JxSVEAiRLwVNwpoztnLNYqt5WjbZHxUCdhCkvTJ2EqYswH3k6RypE_IeDAJHK8A9cNWWD</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2904709180</pqid></control><display><type>article</type><title>Deep Blind Image Quality Assessment Powered by Online Hard Example Mining</title><source>IEEE Electronic Library (IEL)</source><creator>Wang, Zhihua ; Jiang, Qiuping ; Zhao, Shanshan ; Feng, Wensen ; Lin, Weisi</creator><creatorcontrib>Wang, Zhihua ; Jiang, Qiuping ; Zhao, Shanshan ; Feng, Wensen ; Lin, Weisi</creatorcontrib><description>Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example mining (OHEM) is a promising way to alleviate this issue. Inspired by the recent finding that network pruning disproportionately hampers the model's memorization of a tractable subset, atypical, low-quality, long-tailed samples, that are hard-to-memorize during training and easily "forgotten" during pruning, we propose an effective "plug-and-play" OHEM pipeline, especially for generalizable deep BIQA. Specifically, we train two parallel weight-sharing branches simultaneously, where one is full model and other is a "self-competitor" generated from the full model online by network pruning. Then, we leverage the prediction disagreement between the full model and its pruned variant (i.e., the self-competitor) to expose easily "forgettable" samples, which are therefore regarded as the hard ones. We then enforce the prediction consistency between the full model and its pruned variant to implicitly put more focus on these hard samples, which benefits the full model to recover forgettable information introduced by pruning. Extensive experiments across multiple datasets and BIQA models demonstrate that the proposed OHEM can improve the model performance and generalizability as measured by correlation numbers and group maximum differentiation (gMAD) competition. Our code are available at: https://github.com/wangzhihua520/IQA_with_OHEM</description><identifier>ISSN: 1520-9210</identifier><identifier>EISSN: 1941-0077</identifier><identifier>DOI: 10.1109/TMM.2023.3257564</identifier><identifier>CODEN: ITMUF8</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Artificial neural networks ; Blind image quality assessment ; co-evolution ; Computational modeling ; Data models ; Datasets ; hard example mining ; Image quality ; network pruning ; Pipelines ; Predictive models ; Pruning ; Quality assessment ; Training ; Transformers</subject><ispartof>IEEE transactions on multimedia, 2023-01, Vol.25, p.1-11</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c292t-d6b609ff0c008b778f1c2ee7703ec80ac60bdc740209d9b2e1432ccb999e75673</citedby><cites>FETCH-LOGICAL-c292t-d6b609ff0c008b778f1c2ee7703ec80ac60bdc740209d9b2e1432ccb999e75673</cites><orcidid>0000-0001-9866-1947 ; 0000-0002-6025-9343 ; 0000-0002-4398-536X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10070789$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,796,27924,27925,54758</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10070789$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Wang, Zhihua</creatorcontrib><creatorcontrib>Jiang, Qiuping</creatorcontrib><creatorcontrib>Zhao, Shanshan</creatorcontrib><creatorcontrib>Feng, Wensen</creatorcontrib><creatorcontrib>Lin, Weisi</creatorcontrib><title>Deep Blind Image Quality Assessment Powered by Online Hard Example Mining</title><title>IEEE transactions on multimedia</title><addtitle>TMM</addtitle><description>Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example mining (OHEM) is a promising way to alleviate this issue. Inspired by the recent finding that network pruning disproportionately hampers the model's memorization of a tractable subset, atypical, low-quality, long-tailed samples, that are hard-to-memorize during training and easily "forgotten" during pruning, we propose an effective "plug-and-play" OHEM pipeline, especially for generalizable deep BIQA. Specifically, we train two parallel weight-sharing branches simultaneously, where one is full model and other is a "self-competitor" generated from the full model online by network pruning. Then, we leverage the prediction disagreement between the full model and its pruned variant (i.e., the self-competitor) to expose easily "forgettable" samples, which are therefore regarded as the hard ones. We then enforce the prediction consistency between the full model and its pruned variant to implicitly put more focus on these hard samples, which benefits the full model to recover forgettable information introduced by pruning. Extensive experiments across multiple datasets and BIQA models demonstrate that the proposed OHEM can improve the model performance and generalizability as measured by correlation numbers and group maximum differentiation (gMAD) competition. Our code are available at: https://github.com/wangzhihua520/IQA_with_OHEM</description><subject>Artificial neural networks</subject><subject>Blind image quality assessment</subject><subject>co-evolution</subject><subject>Computational modeling</subject><subject>Data models</subject><subject>Datasets</subject><subject>hard example mining</subject><subject>Image quality</subject><subject>network pruning</subject><subject>Pipelines</subject><subject>Predictive models</subject><subject>Pruning</subject><subject>Quality assessment</subject><subject>Training</subject><subject>Transformers</subject><issn>1520-9210</issn><issn>1941-0077</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpNkD1PwzAQhi0EEqWwMzBYYk45Ox-Ox1IKjdSqIJXZcpxLlSpf2Klo_z2u2oHpbnje93QPIY8MJoyBfNmsVhMOPJyEPBZxEl2REZMRCwCEuPZ7zCGQnMEtuXNuB8CiGMSIZG-IPX2tq7agWaO3SL_2uq6GI506h8412A70s_tFiwXNj3TdehTpQtuCzg-66Wukq6qt2u09uSl17fDhMsfk-32-mS2C5fojm02XgeGSD0GR5AnIsgQDkOZCpCUzHFEICNGkoE0CeWFEBBxkIXOOLAq5MbmUEv1fIhyT53Nvb7ufPbpB7bq9bf1JxSVEAiRLwVNwpoztnLNYqt5WjbZHxUCdhCkvTJ2EqYswH3k6RypE_IeDAJHK8A9cNWWD</recordid><startdate>20230101</startdate><enddate>20230101</enddate><creator>Wang, Zhihua</creator><creator>Jiang, Qiuping</creator><creator>Zhao, Shanshan</creator><creator>Feng, Wensen</creator><creator>Lin, Weisi</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0001-9866-1947</orcidid><orcidid>https://orcid.org/0000-0002-6025-9343</orcidid><orcidid>https://orcid.org/0000-0002-4398-536X</orcidid></search><sort><creationdate>20230101</creationdate><title>Deep Blind Image Quality Assessment Powered by Online Hard Example Mining</title><author>Wang, Zhihua ; Jiang, Qiuping ; Zhao, Shanshan ; Feng, Wensen ; Lin, Weisi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c292t-d6b609ff0c008b778f1c2ee7703ec80ac60bdc740209d9b2e1432ccb999e75673</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Artificial neural networks</topic><topic>Blind image quality assessment</topic><topic>co-evolution</topic><topic>Computational modeling</topic><topic>Data models</topic><topic>Datasets</topic><topic>hard example mining</topic><topic>Image quality</topic><topic>network pruning</topic><topic>Pipelines</topic><topic>Predictive models</topic><topic>Pruning</topic><topic>Quality assessment</topic><topic>Training</topic><topic>Transformers</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wang, Zhihua</creatorcontrib><creatorcontrib>Jiang, Qiuping</creatorcontrib><creatorcontrib>Zhao, Shanshan</creatorcontrib><creatorcontrib>Feng, Wensen</creatorcontrib><creatorcontrib>Lin, Weisi</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on multimedia</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Wang, Zhihua</au><au>Jiang, Qiuping</au><au>Zhao, Shanshan</au><au>Feng, Wensen</au><au>Lin, Weisi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep Blind Image Quality Assessment Powered by Online Hard Example Mining</atitle><jtitle>IEEE transactions on multimedia</jtitle><stitle>TMM</stitle><date>2023-01-01</date><risdate>2023</risdate><volume>25</volume><spage>1</spage><epage>11</epage><pages>1-11</pages><issn>1520-9210</issn><eissn>1941-0077</eissn><coden>ITMUF8</coden><abstract>Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example mining (OHEM) is a promising way to alleviate this issue. Inspired by the recent finding that network pruning disproportionately hampers the model's memorization of a tractable subset, atypical, low-quality, long-tailed samples, that are hard-to-memorize during training and easily "forgotten" during pruning, we propose an effective "plug-and-play" OHEM pipeline, especially for generalizable deep BIQA. Specifically, we train two parallel weight-sharing branches simultaneously, where one is full model and other is a "self-competitor" generated from the full model online by network pruning. Then, we leverage the prediction disagreement between the full model and its pruned variant (i.e., the self-competitor) to expose easily "forgettable" samples, which are therefore regarded as the hard ones. We then enforce the prediction consistency between the full model and its pruned variant to implicitly put more focus on these hard samples, which benefits the full model to recover forgettable information introduced by pruning. Extensive experiments across multiple datasets and BIQA models demonstrate that the proposed OHEM can improve the model performance and generalizability as measured by correlation numbers and group maximum differentiation (gMAD) competition. Our code are available at: https://github.com/wangzhihua520/IQA_with_OHEM</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/TMM.2023.3257564</doi><tpages>11</tpages><orcidid>https://orcid.org/0000-0001-9866-1947</orcidid><orcidid>https://orcid.org/0000-0002-6025-9343</orcidid><orcidid>https://orcid.org/0000-0002-4398-536X</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 1520-9210
ispartof IEEE transactions on multimedia, 2023-01, Vol.25, p.1-11
issn 1520-9210
1941-0077
language eng
recordid cdi_ieee_primary_10070789
source IEEE Electronic Library (IEL)
subjects Artificial neural networks
Blind image quality assessment
co-evolution
Computational modeling
Data models
Datasets
hard example mining
Image quality
network pruning
Pipelines
Predictive models
Pruning
Quality assessment
Training
Transformers
title Deep Blind Image Quality Assessment Powered by Online Hard Example Mining
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-05T05%3A59%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20Blind%20Image%20Quality%20Assessment%20Powered%20by%20Online%20Hard%20Example%20Mining&rft.jtitle=IEEE%20transactions%20on%20multimedia&rft.au=Wang,%20Zhihua&rft.date=2023-01-01&rft.volume=25&rft.spage=1&rft.epage=11&rft.pages=1-11&rft.issn=1520-9210&rft.eissn=1941-0077&rft.coden=ITMUF8&rft_id=info:doi/10.1109/TMM.2023.3257564&rft_dat=%3Cproquest_RIE%3E2904709180%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2904709180&rft_id=info:pmid/&rft_ieee_id=10070789&rfr_iscdi=true