Akademska digitalna zbirka SLovenije - logo
E-resources
Full text
Peer reviewed
  • Reparameterized attention f...
    Wu, Yiming; Li, Ruixiang; Yu, Yunlong; Li, Xi

    Pattern recognition letters, December 2022, 2022-12-00, Volume: 164
    Journal Article

    •Modeling the uncertainty of attention modules.•Improving the generalization ability of attention models.•Mitigating the degradation issue that appears in the reparameterized attention.•Improving the image classification performance of different attention models on different datasets consistently. The attention mechanism has been widely explored for neural networks as it could effectively model the interdependencies among channels, spatial positions, and frames. A neural network with attention modules has uncertainties in its parameters, but training the models deterministically hardly captures the uncertainties. Modeling the parameters’ uncertainty of the attention module could facilitate flexibly capturing the representative patterns, thus promoting the generalization of the models. In this work, we propose a novel reparameterized attention strategy by modeling the uncertainty of the parameters in the attention module and performing uncertainty-aware optimization. Instead of learning deterministic parameters for the attention modules, our strategy learns variational posterior distributions. The experimental results show that our strategy could consistently improve different models’ accuracy and reduce the generalization gap without extra computation.