| { | |
| "attn_implementation": "flash_attention_2", | |
| "bos_token_id": 128000, | |
| "do_sample": true, | |
| "eos_token_id": [ | |
| 128001, | |
| 128008, | |
| 128009 | |
| ], | |
| "num_assistant_tokens": 5, | |
| "num_assistant_tokens_schedule": "heuristic", | |
| "return_legacy_cache": true, | |
| "temperature": 0.6, | |
| "top_p": 0.9, | |
| "transformers_version": "4.45.1" | |
| } | |