TY - GEN
T1 - Enforcing Sparsity on Latent Space for Robust and Explainable Representations
AU - Li, Hanao
AU - Han, Tian
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/1/3
Y1 - 2024/1/3
N2 - Recently, dense latent variable models have shown promising results, but their distributed and potentially redundant codes make them less interpretable and less robust to noise. On the other hand, sparse representations are more parsimonious, providing better explainability and noise robustness, but it is difficult to enforce sparsity due to the complexity and computational cost involved. In this paper, we propose a novel unsupervised learning approach to enforce sparsity on the latent space for the generator model, utilizing a gradually sparsified spike and slab distribution as our prior. Our model is composed of a top-down generator network that maps the latent variable to the observations. We use maximum likelihood sampling to infer latent variables in the generator's posterior direction, and spike and slab regularization in the inference stage can induce sparsity by pushing non-informative latent dimensions toward zero. Our experiments show that the learned sparse latent representations preserve the majority of the information, and our model can learn disentangled semantics, increase the explainability of the latent codes, and enhance the robustness of the classification and denoising tasks.
AB - Recently, dense latent variable models have shown promising results, but their distributed and potentially redundant codes make them less interpretable and less robust to noise. On the other hand, sparse representations are more parsimonious, providing better explainability and noise robustness, but it is difficult to enforce sparsity due to the complexity and computational cost involved. In this paper, we propose a novel unsupervised learning approach to enforce sparsity on the latent space for the generator model, utilizing a gradually sparsified spike and slab distribution as our prior. Our model is composed of a top-down generator network that maps the latent variable to the observations. We use maximum likelihood sampling to infer latent variables in the generator's posterior direction, and spike and slab regularization in the inference stage can induce sparsity by pushing non-informative latent dimensions toward zero. Our experiments show that the learned sparse latent representations preserve the majority of the information, and our model can learn disentangled semantics, increase the explainability of the latent codes, and enhance the robustness of the classification and denoising tasks.
KW - 3D
KW - accountable
KW - Algorithms
KW - Algorithms
KW - Algorithms
KW - and algorithms
KW - etc.
KW - ethical computer vision
KW - Explainable
KW - fair
KW - formulations
KW - Generative models for image
KW - Machine learning architectures
KW - privacy-preserving
KW - video
UR - http://www.scopus.com/inward/record.url?scp=85191964073&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85191964073&partnerID=8YFLogxK
U2 - 10.1109/WACV57701.2024.00520
DO - 10.1109/WACV57701.2024.00520
M3 - Conference contribution
AN - SCOPUS:85191964073
T3 - Proceedings - 2024 IEEE Winter Conference on Applications of Computer Vision, WACV 2024
SP - 5270
EP - 5279
BT - Proceedings - 2024 IEEE Winter Conference on Applications of Computer Vision, WACV 2024
T2 - 2024 IEEE Winter Conference on Applications of Computer Vision, WACV 2024
Y2 - 4 January 2024 through 8 January 2024
ER -