vae-lyra / config.json
AbstractPhil's picture
Config update at step 5625
60adb7c verified
{
"modality_dims": {
"clip": 768,
"t5": 768
},
"latent_dim": 768,
"seq_len": 77,
"encoder_layers": 3,
"decoder_layers": 3,
"hidden_dim": 1024,
"dropout": 0.1,
"fusion_strategy": "cantor",
"fusion_heads": 8,
"fusion_dropout": 0.1,
"beta_kl": 0.1,
"beta_reconstruction": 1.0,
"beta_cross_modal": 0.05,
"recon_type": "mse",
"use_kl_annealing": true,
"kl_anneal_epochs": 10,
"kl_start_beta": 0.0,
"batch_size": 16,
"num_epochs": 50,
"learning_rate": 0.0001,
"weight_decay": 1e-05,
"gradient_clip": 1.0,
"use_scheduler": true,
"scheduler_type": "cosine",
"num_samples": 10000,
"synthetic_ratio": 0.15,
"checkpoint_dir": "./checkpoints_lyra",
"save_every": 1000,
"keep_last_n": 10,
"hf_repo": "AbstractPhil/vae-lyra",
"push_to_hub": true,
"push_every": 2000,
"auto_load_from_hub": true,
"use_wandb": false,
"wandb_project": "vae-lyra",
"wandb_entity": null,
"log_every": 50,
"device": "cuda",
"mixed_precision": true,
"seed": 42,
"num_workers": 0
}