Bayesian neural networks (BNN) can estimate the uncertainty in predictions, as opposed to non-Bayesian neural networks (NNs). However, BNNs have been far less widely used than non-Bayesian NNs in practice since they need iterative NN executions to predict a result for one data, and it gives rise to prohibitive computational cost. This computational burden is a critical problem when processing data streams with low-latency. To address this problem, we propose a novel model VQ-BNN, which approximates BNN inference for data streams. In order to reduce the computational burden, VQ-BNN inference predicts NN only once and compensates the result with previously memorized predictions. To be specific, VQ-BNN inference for data streams is given by temporal exponential smoothing of recent predictions. The computational cost of this model is almost the same as that of non-Bayesian NNs. Experiments including semantic segmentation on real-world data show that this model performs significantly faster than BNNs while estimating predictive results comparable to or superior to the results of BNNs.
|Title of host publication||35th AAAI Conference on Artificial Intelligence, AAAI 2021|
|Publisher||Association for the Advancement of Artificial Intelligence|
|Number of pages||9|
|Publication status||Published - 2021|
|Event||35th AAAI Conference on Artificial Intelligence, AAAI 2021 - Virtual, Online|
Duration: 2021 Feb 2 → 2021 Feb 9
|Name||35th AAAI Conference on Artificial Intelligence, AAAI 2021|
|Conference||35th AAAI Conference on Artificial Intelligence, AAAI 2021|
|Period||21/2/2 → 21/2/9|
Bibliographical noteFunding Information:
This work was supported by Samsung Research Funding Center of Samsung Electronics under Project Number SRFC-IT1801-10.
© 2021, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.
All Science Journal Classification (ASJC) codes
- Artificial Intelligence