SAELens
ArthurConmyGDM's picture
Update README.md
1828c4a verified
---
license: cc-by-4.0
library_name: saelens
---
# 1. Gemma Scope
Gemma Scope is a comprehensive, open suite of sparse autoencoders for Gemma 2 9B and 2B. Sparse Autoencoders are a "microscope" of sorts that can help us break down a model’s internal activations into the underlying concepts, just as biologists use microscopes to study the individual cells of plants and animals.
See our [landing page](https://huggingface.co/google/gemma-scope) for details on the whole suite. This is a specific set of SAEs:
# 2. What Is `gemma-scope-9b-it-res`?
- `gemma-scope-`: See 1.
- `9b-it-`: These SAEs were trained on Gemma v2 9B instruction-tuned model.
- `res`: These SAEs were trained on the model's residual stream.
# 3. Why aren't there more IT SAEs?
To summarise our [technical report, Section 4.5](https://storage.googleapis.com/gemma-scope/gemma-scope-report.pdf), we find the same results as [Kissane et al., 2024](https://www.alignmentforum.org/posts/fmwk6qxrpW8d4jvbd/saes-usually-transfer-between-base-and-chat-models), that SAEs trained on Gemma 2 9B base transfer very well to the IT model, and these IT SAEs only work marginally better. Therefore in many cases we expect it is sufficient to use our PT SAEs for the equivalent IT model, e.g. using the [Gemma 2 9B PT SAEs](https://huggingface.co/google/gemma-scope-9b-pt-res) to interpret Gemma 2 9B IT.
# 4. Point of Contact
Point of contact: Arthur Conmy
Contact by email:
```python
''.join(list('moc.elgoog@ymnoc')[::-1])
```
HuggingFace account:
https://huggingface.co/ArthurConmyGDM