experiment: tokenizer_checkpoint: "titok_ll32_vae_c16.bin" output_dir: "titok_ll32_vae_c16" model: vq_model: quantize_mode: "vae" token_size: 16 # vit arch vit_enc_model_size: "large" vit_dec_model_size: "large" vit_enc_patch_size: 16 vit_dec_patch_size: 16 num_latent_tokens: 32 finetune_decoder: False is_legacy: False dataset: preprocessing: crop_size: 256