BatchTopK SAE trained on gemma-2-2B-it on a mix of lmsys and fineweb data

This model has been pushed to the Hub using the PytorchModelHubMixin integration:

  • Library: [More Information Needed]
  • Docs: [More Information Needed]
Downloads last month
2
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including science-of-finetuning/SAE-chat-gemma-2-2b-L13-k100-lr1e-04-local-shuffling