Deep Blind Image Quality Assessment Powered by Online Hard Example Mining
Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example min...
Gespeichert in:
Veröffentlicht in: | IEEE transactions on multimedia 2023-01, Vol.25, p.1-11 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 11 |
---|---|
container_issue | |
container_start_page | 1 |
container_title | IEEE transactions on multimedia |
container_volume | 25 |
creator | Wang, Zhihua Jiang, Qiuping Zhao, Shanshan Feng, Wensen Lin, Weisi |
description | Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example mining (OHEM) is a promising way to alleviate this issue. Inspired by the recent finding that network pruning disproportionately hampers the model's memorization of a tractable subset, atypical, low-quality, long-tailed samples, that are hard-to-memorize during training and easily "forgotten" during pruning, we propose an effective "plug-and-play" OHEM pipeline, especially for generalizable deep BIQA. Specifically, we train two parallel weight-sharing branches simultaneously, where one is full model and other is a "self-competitor" generated from the full model online by network pruning. Then, we leverage the prediction disagreement between the full model and its pruned variant (i.e., the self-competitor) to expose easily "forgettable" samples, which are therefore regarded as the hard ones. We then enforce the prediction consistency between the full model and its pruned variant to implicitly put more focus on these hard samples, which benefits the full model to recover forgettable information introduced by pruning. Extensive experiments across multiple datasets and BIQA models demonstrate that the proposed OHEM can improve the model performance and generalizability as measured by correlation numbers and group maximum differentiation (gMAD) competition. Our code are available at: https://github.com/wangzhihua520/IQA_with_OHEM |
doi_str_mv | 10.1109/TMM.2023.3257564 |
format | Article |
fullrecord | <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_ieee_primary_10070789</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10070789</ieee_id><sourcerecordid>2904709180</sourcerecordid><originalsourceid>FETCH-LOGICAL-c292t-d6b609ff0c008b778f1c2ee7703ec80ac60bdc740209d9b2e1432ccb999e75673</originalsourceid><addsrcrecordid>eNpNkD1PwzAQhi0EEqWwMzBYYk45Ox-Ox1IKjdSqIJXZcpxLlSpf2Klo_z2u2oHpbnje93QPIY8MJoyBfNmsVhMOPJyEPBZxEl2REZMRCwCEuPZ7zCGQnMEtuXNuB8CiGMSIZG-IPX2tq7agWaO3SL_2uq6GI506h8412A70s_tFiwXNj3TdehTpQtuCzg-66Wukq6qt2u09uSl17fDhMsfk-32-mS2C5fojm02XgeGSD0GR5AnIsgQDkOZCpCUzHFEICNGkoE0CeWFEBBxkIXOOLAq5MbmUEv1fIhyT53Nvb7ufPbpB7bq9bf1JxSVEAiRLwVNwpoztnLNYqt5WjbZHxUCdhCkvTJ2EqYswH3k6RypE_IeDAJHK8A9cNWWD</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2904709180</pqid></control><display><type>article</type><title>Deep Blind Image Quality Assessment Powered by Online Hard Example Mining</title><source>IEEE Electronic Library (IEL)</source><creator>Wang, Zhihua ; Jiang, Qiuping ; Zhao, Shanshan ; Feng, Wensen ; Lin, Weisi</creator><creatorcontrib>Wang, Zhihua ; Jiang, Qiuping ; Zhao, Shanshan ; Feng, Wensen ; Lin, Weisi</creatorcontrib><description>Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example mining (OHEM) is a promising way to alleviate this issue. Inspired by the recent finding that network pruning disproportionately hampers the model's memorization of a tractable subset, atypical, low-quality, long-tailed samples, that are hard-to-memorize during training and easily "forgotten" during pruning, we propose an effective "plug-and-play" OHEM pipeline, especially for generalizable deep BIQA. Specifically, we train two parallel weight-sharing branches simultaneously, where one is full model and other is a "self-competitor" generated from the full model online by network pruning. Then, we leverage the prediction disagreement between the full model and its pruned variant (i.e., the self-competitor) to expose easily "forgettable" samples, which are therefore regarded as the hard ones. We then enforce the prediction consistency between the full model and its pruned variant to implicitly put more focus on these hard samples, which benefits the full model to recover forgettable information introduced by pruning. Extensive experiments across multiple datasets and BIQA models demonstrate that the proposed OHEM can improve the model performance and generalizability as measured by correlation numbers and group maximum differentiation (gMAD) competition. Our code are available at: https://github.com/wangzhihua520/IQA_with_OHEM</description><identifier>ISSN: 1520-9210</identifier><identifier>EISSN: 1941-0077</identifier><identifier>DOI: 10.1109/TMM.2023.3257564</identifier><identifier>CODEN: ITMUF8</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Artificial neural networks ; Blind image quality assessment ; co-evolution ; Computational modeling ; Data models ; Datasets ; hard example mining ; Image quality ; network pruning ; Pipelines ; Predictive models ; Pruning ; Quality assessment ; Training ; Transformers</subject><ispartof>IEEE transactions on multimedia, 2023-01, Vol.25, p.1-11</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c292t-d6b609ff0c008b778f1c2ee7703ec80ac60bdc740209d9b2e1432ccb999e75673</citedby><cites>FETCH-LOGICAL-c292t-d6b609ff0c008b778f1c2ee7703ec80ac60bdc740209d9b2e1432ccb999e75673</cites><orcidid>0000-0001-9866-1947 ; 0000-0002-6025-9343 ; 0000-0002-4398-536X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10070789$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,796,27924,27925,54758</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10070789$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Wang, Zhihua</creatorcontrib><creatorcontrib>Jiang, Qiuping</creatorcontrib><creatorcontrib>Zhao, Shanshan</creatorcontrib><creatorcontrib>Feng, Wensen</creatorcontrib><creatorcontrib>Lin, Weisi</creatorcontrib><title>Deep Blind Image Quality Assessment Powered by Online Hard Example Mining</title><title>IEEE transactions on multimedia</title><addtitle>TMM</addtitle><description>Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example mining (OHEM) is a promising way to alleviate this issue. Inspired by the recent finding that network pruning disproportionately hampers the model's memorization of a tractable subset, atypical, low-quality, long-tailed samples, that are hard-to-memorize during training and easily "forgotten" during pruning, we propose an effective "plug-and-play" OHEM pipeline, especially for generalizable deep BIQA. Specifically, we train two parallel weight-sharing branches simultaneously, where one is full model and other is a "self-competitor" generated from the full model online by network pruning. Then, we leverage the prediction disagreement between the full model and its pruned variant (i.e., the self-competitor) to expose easily "forgettable" samples, which are therefore regarded as the hard ones. We then enforce the prediction consistency between the full model and its pruned variant to implicitly put more focus on these hard samples, which benefits the full model to recover forgettable information introduced by pruning. Extensive experiments across multiple datasets and BIQA models demonstrate that the proposed OHEM can improve the model performance and generalizability as measured by correlation numbers and group maximum differentiation (gMAD) competition. Our code are available at: https://github.com/wangzhihua520/IQA_with_OHEM</description><subject>Artificial neural networks</subject><subject>Blind image quality assessment</subject><subject>co-evolution</subject><subject>Computational modeling</subject><subject>Data models</subject><subject>Datasets</subject><subject>hard example mining</subject><subject>Image quality</subject><subject>network pruning</subject><subject>Pipelines</subject><subject>Predictive models</subject><subject>Pruning</subject><subject>Quality assessment</subject><subject>Training</subject><subject>Transformers</subject><issn>1520-9210</issn><issn>1941-0077</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpNkD1PwzAQhi0EEqWwMzBYYk45Ox-Ox1IKjdSqIJXZcpxLlSpf2Klo_z2u2oHpbnje93QPIY8MJoyBfNmsVhMOPJyEPBZxEl2REZMRCwCEuPZ7zCGQnMEtuXNuB8CiGMSIZG-IPX2tq7agWaO3SL_2uq6GI506h8412A70s_tFiwXNj3TdehTpQtuCzg-66Wukq6qt2u09uSl17fDhMsfk-32-mS2C5fojm02XgeGSD0GR5AnIsgQDkOZCpCUzHFEICNGkoE0CeWFEBBxkIXOOLAq5MbmUEv1fIhyT53Nvb7ufPbpB7bq9bf1JxSVEAiRLwVNwpoztnLNYqt5WjbZHxUCdhCkvTJ2EqYswH3k6RypE_IeDAJHK8A9cNWWD</recordid><startdate>20230101</startdate><enddate>20230101</enddate><creator>Wang, Zhihua</creator><creator>Jiang, Qiuping</creator><creator>Zhao, Shanshan</creator><creator>Feng, Wensen</creator><creator>Lin, Weisi</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0001-9866-1947</orcidid><orcidid>https://orcid.org/0000-0002-6025-9343</orcidid><orcidid>https://orcid.org/0000-0002-4398-536X</orcidid></search><sort><creationdate>20230101</creationdate><title>Deep Blind Image Quality Assessment Powered by Online Hard Example Mining</title><author>Wang, Zhihua ; Jiang, Qiuping ; Zhao, Shanshan ; Feng, Wensen ; Lin, Weisi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c292t-d6b609ff0c008b778f1c2ee7703ec80ac60bdc740209d9b2e1432ccb999e75673</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Artificial neural networks</topic><topic>Blind image quality assessment</topic><topic>co-evolution</topic><topic>Computational modeling</topic><topic>Data models</topic><topic>Datasets</topic><topic>hard example mining</topic><topic>Image quality</topic><topic>network pruning</topic><topic>Pipelines</topic><topic>Predictive models</topic><topic>Pruning</topic><topic>Quality assessment</topic><topic>Training</topic><topic>Transformers</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wang, Zhihua</creatorcontrib><creatorcontrib>Jiang, Qiuping</creatorcontrib><creatorcontrib>Zhao, Shanshan</creatorcontrib><creatorcontrib>Feng, Wensen</creatorcontrib><creatorcontrib>Lin, Weisi</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on multimedia</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Wang, Zhihua</au><au>Jiang, Qiuping</au><au>Zhao, Shanshan</au><au>Feng, Wensen</au><au>Lin, Weisi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep Blind Image Quality Assessment Powered by Online Hard Example Mining</atitle><jtitle>IEEE transactions on multimedia</jtitle><stitle>TMM</stitle><date>2023-01-01</date><risdate>2023</risdate><volume>25</volume><spage>1</spage><epage>11</epage><pages>1-11</pages><issn>1520-9210</issn><eissn>1941-0077</eissn><coden>ITMUF8</coden><abstract>Recently, blind image quality assessment (BIQA) models based on deep neural networks (DNNs) have achieved impressive performance on existing datasets. However, due to the intrinsic imbalance property of the training set, not all distortions or images are handled equally well. Online hard example mining (OHEM) is a promising way to alleviate this issue. Inspired by the recent finding that network pruning disproportionately hampers the model's memorization of a tractable subset, atypical, low-quality, long-tailed samples, that are hard-to-memorize during training and easily "forgotten" during pruning, we propose an effective "plug-and-play" OHEM pipeline, especially for generalizable deep BIQA. Specifically, we train two parallel weight-sharing branches simultaneously, where one is full model and other is a "self-competitor" generated from the full model online by network pruning. Then, we leverage the prediction disagreement between the full model and its pruned variant (i.e., the self-competitor) to expose easily "forgettable" samples, which are therefore regarded as the hard ones. We then enforce the prediction consistency between the full model and its pruned variant to implicitly put more focus on these hard samples, which benefits the full model to recover forgettable information introduced by pruning. Extensive experiments across multiple datasets and BIQA models demonstrate that the proposed OHEM can improve the model performance and generalizability as measured by correlation numbers and group maximum differentiation (gMAD) competition. Our code are available at: https://github.com/wangzhihua520/IQA_with_OHEM</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/TMM.2023.3257564</doi><tpages>11</tpages><orcidid>https://orcid.org/0000-0001-9866-1947</orcidid><orcidid>https://orcid.org/0000-0002-6025-9343</orcidid><orcidid>https://orcid.org/0000-0002-4398-536X</orcidid></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISSN: 1520-9210 |
ispartof | IEEE transactions on multimedia, 2023-01, Vol.25, p.1-11 |
issn | 1520-9210 1941-0077 |
language | eng |
recordid | cdi_ieee_primary_10070789 |
source | IEEE Electronic Library (IEL) |
subjects | Artificial neural networks Blind image quality assessment co-evolution Computational modeling Data models Datasets hard example mining Image quality network pruning Pipelines Predictive models Pruning Quality assessment Training Transformers |
title | Deep Blind Image Quality Assessment Powered by Online Hard Example Mining |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-05T05%3A59%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20Blind%20Image%20Quality%20Assessment%20Powered%20by%20Online%20Hard%20Example%20Mining&rft.jtitle=IEEE%20transactions%20on%20multimedia&rft.au=Wang,%20Zhihua&rft.date=2023-01-01&rft.volume=25&rft.spage=1&rft.epage=11&rft.pages=1-11&rft.issn=1520-9210&rft.eissn=1941-0077&rft.coden=ITMUF8&rft_id=info:doi/10.1109/TMM.2023.3257564&rft_dat=%3Cproquest_RIE%3E2904709180%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2904709180&rft_id=info:pmid/&rft_ieee_id=10070789&rfr_iscdi=true |