{ "architectures": [ "GPTVForCausalLM" ], "attn_dropout": 0.0, "auto_map": { "AutoConfig": "model_cfg.ModelConfig", "AutoModelForCausalLM": "model.GPTVForCausalLM" }, "bos_token_id": 259, "dropout": 0.01, "dtype": "float32", "embed_dim": 256, "eos_token_id": 260, "ffn_mult": 4.0, "hidden_dim": 256, "hidden_squeeze_ratio": 0.5, "max_length": 4096, "model_type": "custom-gpt", "num_heads": 4, "num_hidden_layers": 22, "num_kv_groups": 2, "num_layers": 22, "pad_token_id": 0, "patch_size": 8, "tie_word_embeddings": false, "transformers_version": "4.57.1", "use_bias": true, "vocab_size": 512 }