Bridging Precision and Confidence: A Train-Time Loss for Calibrating Object Detection
Deep neural networks (DNNs) have enabled astounding progress in several vision-based problems. Despite showing high predictive accuracy, recently, several works have revealed that they tend to provide overconfident predictions and thus are poorly calibrated. The majority of the works addressing the...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Munir, Muhammad Akhtar Khan, Muhammad Haris Khan, Salman Khan, Fahad Shahbaz |
description | Deep neural networks (DNNs) have enabled astounding progress in several
vision-based problems. Despite showing high predictive accuracy, recently,
several works have revealed that they tend to provide overconfident predictions
and thus are poorly calibrated. The majority of the works addressing the
miscalibration of DNNs fall under the scope of classification and consider only
in-domain predictions. However, there is little to no progress in studying the
calibration of DNN-based object detection models, which are central to many
vision-based safety-critical applications. In this paper, inspired by the
train-time calibration methods, we propose a novel auxiliary loss formulation
that explicitly aims to align the class confidence of bounding boxes with the
accurateness of predictions (i.e. precision). Since the original formulation of
our loss depends on the counts of true positives and false positives in a
minibatch, we develop a differentiable proxy of our loss that can be used
during training with other application-specific loss functions. We perform
extensive experiments on challenging in-domain and out-domain scenarios with
six benchmark datasets including MS-COCO, Cityscapes, Sim10k, and BDD100k. Our
results reveal that our train-time loss surpasses strong calibration baselines
in reducing calibration error for both in and out-domain scenarios. Our source
code and pre-trained models are available at
https://github.com/akhtarvision/bpc_calibration |
doi_str_mv | 10.48550/arxiv.2303.14404 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2303_14404</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2303_14404</sourcerecordid><originalsourceid>FETCH-LOGICAL-a674-c1f709f6bfc3ac95d35b276a50401d8a396ee0830201ec4aca0b177393bab4b3</originalsourceid><addsrcrecordid>eNotj0tOwzAUAL1hgQoHYIUvkPAc23HCroSvFKlIhHX0_Kseah3kRAhuT1tYzWpGGsauBJSq0RpuMH_TV1lJkKVQCtQ5e7_L5LeUtvw1B0czTYlj8rybUiQfkgu3fM2HjJSKgfaB99M88zhl3uGObMbl6G7sR3ALvw_LAYfEBTuLuJvD5T9X7O3xYeiei37z9NKt-wJrowonooE21jY6ia7VXmpbmRo1KBC-QdnWIUAjoQIRnEKHYIUxspUWrbJyxa7_qqet8TPTHvPPeNwbT3vyFw9pSko</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Bridging Precision and Confidence: A Train-Time Loss for Calibrating Object Detection</title><source>arXiv.org</source><creator>Munir, Muhammad Akhtar ; Khan, Muhammad Haris ; Khan, Salman ; Khan, Fahad Shahbaz</creator><creatorcontrib>Munir, Muhammad Akhtar ; Khan, Muhammad Haris ; Khan, Salman ; Khan, Fahad Shahbaz</creatorcontrib><description>Deep neural networks (DNNs) have enabled astounding progress in several
vision-based problems. Despite showing high predictive accuracy, recently,
several works have revealed that they tend to provide overconfident predictions
and thus are poorly calibrated. The majority of the works addressing the
miscalibration of DNNs fall under the scope of classification and consider only
in-domain predictions. However, there is little to no progress in studying the
calibration of DNN-based object detection models, which are central to many
vision-based safety-critical applications. In this paper, inspired by the
train-time calibration methods, we propose a novel auxiliary loss formulation
that explicitly aims to align the class confidence of bounding boxes with the
accurateness of predictions (i.e. precision). Since the original formulation of
our loss depends on the counts of true positives and false positives in a
minibatch, we develop a differentiable proxy of our loss that can be used
during training with other application-specific loss functions. We perform
extensive experiments on challenging in-domain and out-domain scenarios with
six benchmark datasets including MS-COCO, Cityscapes, Sim10k, and BDD100k. Our
results reveal that our train-time loss surpasses strong calibration baselines
in reducing calibration error for both in and out-domain scenarios. Our source
code and pre-trained models are available at
https://github.com/akhtarvision/bpc_calibration</description><identifier>DOI: 10.48550/arxiv.2303.14404</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2023-03</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2303.14404$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2303.14404$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Munir, Muhammad Akhtar</creatorcontrib><creatorcontrib>Khan, Muhammad Haris</creatorcontrib><creatorcontrib>Khan, Salman</creatorcontrib><creatorcontrib>Khan, Fahad Shahbaz</creatorcontrib><title>Bridging Precision and Confidence: A Train-Time Loss for Calibrating Object Detection</title><description>Deep neural networks (DNNs) have enabled astounding progress in several
vision-based problems. Despite showing high predictive accuracy, recently,
several works have revealed that they tend to provide overconfident predictions
and thus are poorly calibrated. The majority of the works addressing the
miscalibration of DNNs fall under the scope of classification and consider only
in-domain predictions. However, there is little to no progress in studying the
calibration of DNN-based object detection models, which are central to many
vision-based safety-critical applications. In this paper, inspired by the
train-time calibration methods, we propose a novel auxiliary loss formulation
that explicitly aims to align the class confidence of bounding boxes with the
accurateness of predictions (i.e. precision). Since the original formulation of
our loss depends on the counts of true positives and false positives in a
minibatch, we develop a differentiable proxy of our loss that can be used
during training with other application-specific loss functions. We perform
extensive experiments on challenging in-domain and out-domain scenarios with
six benchmark datasets including MS-COCO, Cityscapes, Sim10k, and BDD100k. Our
results reveal that our train-time loss surpasses strong calibration baselines
in reducing calibration error for both in and out-domain scenarios. Our source
code and pre-trained models are available at
https://github.com/akhtarvision/bpc_calibration</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj0tOwzAUAL1hgQoHYIUvkPAc23HCroSvFKlIhHX0_Kseah3kRAhuT1tYzWpGGsauBJSq0RpuMH_TV1lJkKVQCtQ5e7_L5LeUtvw1B0czTYlj8rybUiQfkgu3fM2HjJSKgfaB99M88zhl3uGObMbl6G7sR3ALvw_LAYfEBTuLuJvD5T9X7O3xYeiei37z9NKt-wJrowonooE21jY6ia7VXmpbmRo1KBC-QdnWIUAjoQIRnEKHYIUxspUWrbJyxa7_qqet8TPTHvPPeNwbT3vyFw9pSko</recordid><startdate>20230325</startdate><enddate>20230325</enddate><creator>Munir, Muhammad Akhtar</creator><creator>Khan, Muhammad Haris</creator><creator>Khan, Salman</creator><creator>Khan, Fahad Shahbaz</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230325</creationdate><title>Bridging Precision and Confidence: A Train-Time Loss for Calibrating Object Detection</title><author>Munir, Muhammad Akhtar ; Khan, Muhammad Haris ; Khan, Salman ; Khan, Fahad Shahbaz</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a674-c1f709f6bfc3ac95d35b276a50401d8a396ee0830201ec4aca0b177393bab4b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Munir, Muhammad Akhtar</creatorcontrib><creatorcontrib>Khan, Muhammad Haris</creatorcontrib><creatorcontrib>Khan, Salman</creatorcontrib><creatorcontrib>Khan, Fahad Shahbaz</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Munir, Muhammad Akhtar</au><au>Khan, Muhammad Haris</au><au>Khan, Salman</au><au>Khan, Fahad Shahbaz</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Bridging Precision and Confidence: A Train-Time Loss for Calibrating Object Detection</atitle><date>2023-03-25</date><risdate>2023</risdate><abstract>Deep neural networks (DNNs) have enabled astounding progress in several
vision-based problems. Despite showing high predictive accuracy, recently,
several works have revealed that they tend to provide overconfident predictions
and thus are poorly calibrated. The majority of the works addressing the
miscalibration of DNNs fall under the scope of classification and consider only
in-domain predictions. However, there is little to no progress in studying the
calibration of DNN-based object detection models, which are central to many
vision-based safety-critical applications. In this paper, inspired by the
train-time calibration methods, we propose a novel auxiliary loss formulation
that explicitly aims to align the class confidence of bounding boxes with the
accurateness of predictions (i.e. precision). Since the original formulation of
our loss depends on the counts of true positives and false positives in a
minibatch, we develop a differentiable proxy of our loss that can be used
during training with other application-specific loss functions. We perform
extensive experiments on challenging in-domain and out-domain scenarios with
six benchmark datasets including MS-COCO, Cityscapes, Sim10k, and BDD100k. Our
results reveal that our train-time loss surpasses strong calibration baselines
in reducing calibration error for both in and out-domain scenarios. Our source
code and pre-trained models are available at
https://github.com/akhtarvision/bpc_calibration</abstract><doi>10.48550/arxiv.2303.14404</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2303.14404 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2303_14404 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition |
title | Bridging Precision and Confidence: A Train-Time Loss for Calibrating Object Detection |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-30T02%3A25%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Bridging%20Precision%20and%20Confidence:%20A%20Train-Time%20Loss%20for%20Calibrating%20Object%20Detection&rft.au=Munir,%20Muhammad%20Akhtar&rft.date=2023-03-25&rft_id=info:doi/10.48550/arxiv.2303.14404&rft_dat=%3Carxiv_GOX%3E2303_14404%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |