mindchain's picture
Upload card.json with huggingface_hub
7ea5199 verified
{
"model_type": "sparse-autoencoder",
"base_model": "google/t5gemma-2-270m-270m",
"pipeline_tag": "text2text-generation",
"architecture": "T5 Encoder-Decoder",
"colab_link": "https://colab.research.google.com/github/mindchain/t5gemma2-sae-all-layers/blob/main/t5gemma2-sae-all-layers.ipynb",
"tasks": [
"mechanistic-interpretability",
"activation-steering",
"feature-visualization",
"model-interventions"
],
"languages": ["en"],
"tags": [
"sae",
"sparse-autoencoder",
"t5gemma",
"t5gemma2",
"google-t5gemma",
"mechanistic-interpretability",
"activation-steering",
"steering",
"feature-visualization",
"neuronpedia",
"gemma-scope",
"sae-lens",
"transformer-lens",
"llm-interpretability",
"explainable-ai",
"xai",
"model-steering",
"feature-engineering",
"representation-learning",
"dictionary-learning",
"layer-interpretation",
"hidden-state-analysis",
"nlp",
"natural-language-processing",
"text-to-text",
"language-model",
"llm",
"large-language-model",
"encoder-decoder",
"transformer"
],
"datasets": [],
"metrics": {
"mse": 0.105,
"cosine_similarity": 0.80,
"l0_sparsity": 0.33,
"final_loss": 0.0014
},
"training": {
"epochs": 5,
"batch_size": 2,
"learning_rate": 0.0001,
"optimizer": "AdamW",
"l1_coefficient": 0.01,
"d_sae": 4096,
"d_in": 640
}
}