File size: 1,364 Bytes
c3c81ad 211a82e c3c81ad 211a82e c3c81ad 0ff1877 211a82e 0ff1877 211a82e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 |
---
language: en
library_name: mlsae
license: mit
tags:
- arxiv:2409.04185
- model_hub_mixin
- pytorch_model_hub_mixin
---
# Model Card for tim-lawson/sae-pythia-410m-deduped-x64-k32-tfm-layers-23
A Multi-Layer Sparse Autoencoder (MLSAE) trained on the residual stream activation
vectors from [EleutherAI/pythia-410m-deduped](https://huggingface.co/EleutherAI/pythia-410m-deduped) with an
expansion factor of R = 64 and sparsity k = 32, over 1 billion
tokens from [monology/pile-uncopyrighted](https://huggingface.co/datasets/monology/pile-uncopyrighted).
This model is a PyTorch Lightning MLSAETransformer module, which includes the underlying
transformer.
### Model Sources
- **Repository:** <https://github.com/tim-lawson/mlsae>
- **Paper:** <https://arxiv.org/abs/2409.04185>
- **Weights & Biases:** <https://wandb.ai/timlawson-/mlsae>
## Citation
**BibTeX:**
```bibtex
@misc{lawson_residual_2024,
title = {Residual {{Stream Analysis}} with {{Multi-Layer SAEs}}},
author = {Lawson, Tim and Farnik, Lucy and Houghton, Conor and Aitchison, Laurence},
year = {2024},
month = oct,
number = {arXiv:2409.04185},
eprint = {2409.04185},
primaryclass = {cs},
publisher = {arXiv},
doi = {10.48550/arXiv.2409.04185},
urldate = {2024-10-08},
archiveprefix = {arXiv}
}
``` |