TY - GEN
T1 - Fixing Overconfidence in Dynamic Neural Networks
AU - Meronen, Lassi
AU - Trapp, Martin
AU - Pilzer, Andrea
AU - Yang, Le
AU - Solin, Arno
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/1/3
Y1 - 2024/1/3
N2 - Dynamic neural networks are a recent technique that promises a remedy for the increasing size of modern deep learning models by dynamically adapting their computational cost to the difficulty of the inputs. In this way, the model can adjust to a limited computational budget. However, the poor quality of uncertainty estimates in deep learning models makes it difficult to distinguish between hard and easy samples. To address this challenge, we present a computationally efficient approach for post-hoc uncertainty quantification in dynamic neural networks. We show that adequately quantifying and accounting for both aleatoric and epistemic uncertainty through a probabilistic treatment of the last layers improves the predictive performance and aids decision-making when determining the computational budget. In the experiments, we show improvements on CIFAR100, ImageNet, and Caltech-256 in terms of accuracy, capturing uncertainty, and calibration error.
AB - Dynamic neural networks are a recent technique that promises a remedy for the increasing size of modern deep learning models by dynamically adapting their computational cost to the difficulty of the inputs. In this way, the model can adjust to a limited computational budget. However, the poor quality of uncertainty estimates in deep learning models makes it difficult to distinguish between hard and easy samples. To address this challenge, we present a computationally efficient approach for post-hoc uncertainty quantification in dynamic neural networks. We show that adequately quantifying and accounting for both aleatoric and epistemic uncertainty through a probabilistic treatment of the last layers improves the predictive performance and aids decision-making when determining the computational budget. In the experiments, we show improvements on CIFAR100, ImageNet, and Caltech-256 in terms of accuracy, capturing uncertainty, and calibration error.
KW - Algorithms
KW - and algorithms
KW - formulations
KW - Machine learning architectures
UR - http://www.scopus.com/inward/record.url?scp=85191991682&partnerID=8YFLogxK
U2 - 10.1109/WACV57701.2024.00266
DO - 10.1109/WACV57701.2024.00266
M3 - Conference article in proceedings
AN - SCOPUS:85191991682
T3 - IEEE Winter Conference on Applications of Computer Vision
SP - 2668
EP - 2678
BT - Proceedings - 2024 IEEE Winter Conference on Applications of Computer Vision, WACV 2024
PB - IEEE
T2 - IEEE Winter Conference on Applications of Computer Vision
Y2 - 4 January 2024 through 8 January 2024
ER -