{ "model_type": "sparse-autoencoder", "base_model": "google/t5gemma-2-270m-270m", "pipeline_tag": "text2text-generation", "architecture": "T5 Encoder-Decoder", "colab_link": "https://colab.research.google.com/github/mindchain/t5gemma2-sae-all-layers/blob/main/t5gemma2-sae-all-layers.ipynb", "tasks": [ "mechanistic-interpretability", "activation-steering", "feature-visualization", "model-interventions" ], "languages": ["en"], "tags": [ "sae", "sparse-autoencoder", "t5gemma", "t5gemma2", "google-t5gemma", "mechanistic-interpretability", "activation-steering", "steering", "feature-visualization", "neuronpedia", "gemma-scope", "sae-lens", "transformer-lens", "llm-interpretability", "explainable-ai", "xai", "model-steering", "feature-engineering", "representation-learning", "dictionary-learning", "layer-interpretation", "hidden-state-analysis", "nlp", "natural-language-processing", "text-to-text", "language-model", "llm", "large-language-model", "encoder-decoder", "transformer" ], "datasets": [], "metrics": { "mse": 0.105, "cosine_similarity": 0.80, "l0_sparsity": 0.33, "final_loss": 0.0014 }, "training": { "epochs": 5, "batch_size": 2, "learning_rate": 0.0001, "optimizer": "AdamW", "l1_coefficient": 0.01, "d_sae": 4096, "d_in": 640 } }