Documents
Presentation Slides
Information Preserving Dimensionality Reduction for Mutual Information Analysis of Deep Learning
- Citation Author(s):
- Submitted by:
- Taro Tezuka
- Last updated:
- 6 March 2022 - 9:44am
- Document Type:
- Presentation Slides
- Document Year:
- 2022
- Event:
- Presenters:
- Shizuma Namekawa
- Categories:
- Log in to post comments
Mutual information has been actively investigated as a tool for analyzing neural networks' behavior, most notably the information bottleneck theory. However, estimating mutual information is a notoriously tricky task, especially for high-dimensional stochastic variables. Recently, mutual information neural estimation (MINE) was proposed as a non-parametric method to estimate mutual information for continuous variables without discretization. Unfortunately, MINE also produces significant errors for high-dimensional variables. Analyzing the activity of neural networks requires a dimensionality reduction mechanism, with the resulting low-dimensional representations retaining as much information as possible. We investigated different dimensionality reduction methods to determine their capabilities in estimating mutual information regarding the activity of trained neural networks. We combined MIME with principal component analysis (PCA), a convolutional neural network (CNN), and global average pooling (GAP). The experiments showed that introducing dimensionality reduction provides more stable results than the baseline method. In terms of stability, PCA-MINE and GAP-MINE performed better than CNN-MINE. They also require much less computation time than CNN. Another advantage of GAP-MINE is that it requires no hyperparameter optimization. However, the results suggested that GAP-MINE may underestimate mutual information.