Update README.md
Browse files
README.md
CHANGED
|
@@ -57,7 +57,7 @@ For more details, including benchmark evaluation, hardware requirements, and inf
|
|
| 57 |
```python
|
| 58 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 59 |
|
| 60 |
-
model_name = "Kwaipilot/KAT-Dev
|
| 61 |
|
| 62 |
# load the tokenizer and the model
|
| 63 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
@@ -94,7 +94,7 @@ print("content:", content)
|
|
| 94 |
## Claude Code
|
| 95 |
### vllm server
|
| 96 |
```
|
| 97 |
-
MODEL_PATH="Kwaipilot/KAT-Dev
|
| 98 |
|
| 99 |
vllm serve $MODEL_PATH \
|
| 100 |
--enable-prefix-caching \
|
|
|
|
| 57 |
```python
|
| 58 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 59 |
|
| 60 |
+
model_name = "Kwaipilot/KAT-Dev"
|
| 61 |
|
| 62 |
# load the tokenizer and the model
|
| 63 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
|
|
| 94 |
## Claude Code
|
| 95 |
### vllm server
|
| 96 |
```
|
| 97 |
+
MODEL_PATH="Kwaipilot/KAT-Dev"
|
| 98 |
|
| 99 |
vllm serve $MODEL_PATH \
|
| 100 |
--enable-prefix-caching \
|