Deep Neural Network Uncertainty Quantification for LArTPC Reconstruction
We evaluate uncertainty quantification (UQ) methods for deep learning applied to liquid argon time projection chamber (LArTPC) physics analysis tasks. As deep learning applications enter widespread usage among physics data analysis, neural networks with reliable estimates of prediction uncertainty a...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | We evaluate uncertainty quantification (UQ) methods for deep learning applied
to liquid argon time projection chamber (LArTPC) physics analysis tasks. As
deep learning applications enter widespread usage among physics data analysis,
neural networks with reliable estimates of prediction uncertainty and robust
performance against overconfidence and out-of-distribution (OOD) samples are
critical for their full deployment in analyzing experimental data. While
numerous UQ methods have been tested on simple datasets, performance
evaluations for more complex tasks and datasets are scarce. We assess the
application of selected deep learning UQ methods on the task of particle
classification using the PiLArNet [1] monte carlo 3D LArTPC point cloud
dataset. We observe that UQ methods not only allow for better rejection of
prediction mistakes and OOD detection, but also generally achieve higher
overall accuracy across different task settings. We assess the precision of
uncertainty quantification using different evaluation metrics, such as
distributional separation of prediction entropy across correctly and
incorrectly identified samples, receiver operating characteristic curves
(ROCs), and expected calibration error from observed empirical accuracy. We
conclude that ensembling methods can obtain well calibrated classification
probabilities and generally perform better than other existing methods in deep
learning UQ literature. |
---|---|
DOI: | 10.48550/arxiv.2302.03787 |