Update config.json
Browse files- config.json +5 -2
config.json
CHANGED
|
@@ -1,8 +1,11 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "./llava-v1.5-13b",
|
| 3 |
"architectures": [
|
| 4 |
"LlavaQwen2ForCausalLM"
|
| 5 |
],
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"bos_token_id": 151643,
|
| 8 |
"eos_token_id": 151645,
|
|
@@ -42,4 +45,4 @@
|
|
| 42 |
"use_mm_proj": true,
|
| 43 |
"use_sliding_window": false,
|
| 44 |
"vocab_size": 151936
|
| 45 |
-
}
|
|
|
|
| 1 |
{
|
|
|
|
| 2 |
"architectures": [
|
| 3 |
"LlavaQwen2ForCausalLM"
|
| 4 |
],
|
| 5 |
+
"auto_map": {
|
| 6 |
+
"AutoConfig": "llava_qwen.LlavaConfig",
|
| 7 |
+
"AutoModelForCausalLM": "llava_qwen.LlavaQwen2ForCausalLM"
|
| 8 |
+
},
|
| 9 |
"attention_dropout": 0.0,
|
| 10 |
"bos_token_id": 151643,
|
| 11 |
"eos_token_id": 151645,
|
|
|
|
| 45 |
"use_mm_proj": true,
|
| 46 |
"use_sliding_window": false,
|
| 47 |
"vocab_size": 151936
|
| 48 |
+
}
|