Back to Search Start Over

Improving Dictionary Learning with Gated Sparse Autoencoders

Authors :
Rajamanoharan, Senthooran
Conmy, Arthur
Smith, Lewis
Lieberum, Tom
Varma, Vikrant
Kramár, János
Shah, Rohin
Nanda, Neel
Publication Year :
2024

Abstract

Recent work has found that sparse autoencoders (SAEs) are an effective technique for unsupervised discovery of interpretable features in language models' (LMs) activations, by finding sparse, linear reconstructions of LM activations. We introduce the Gated Sparse Autoencoder (Gated SAE), which achieves a Pareto improvement over training with prevailing methods. In SAEs, the L1 penalty used to encourage sparsity introduces many undesirable biases, such as shrinkage -- systematic underestimation of feature activations. The key insight of Gated SAEs is to separate the functionality of (a) determining which directions to use and (b) estimating the magnitudes of those directions: this enables us to apply the L1 penalty only to the former, limiting the scope of undesirable side effects. Through training SAEs on LMs of up to 7B parameters we find that, in typical hyper-parameter ranges, Gated SAEs solve shrinkage, are similarly interpretable, and require half as many firing features to achieve comparable reconstruction fidelity.<br />Comment: 15 main text pages, 22 appendix pages

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2404.16014
Document Type :
Working Paper