Update README.md
Browse files
README.md
CHANGED
|
@@ -174,16 +174,12 @@ messages = messages_ja
|
|
| 174 |
for i in range(len(messages)-1):
|
| 175 |
|
| 176 |
inputs = [messages[0], messages[i+1]]
|
| 177 |
-
#prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
| 178 |
prompt = pipe.tokenizer.apply_chat_template(inputs, tokenize=False, add_generation_prompt=True)
|
| 179 |
-
#print(f"prompt:{prompt}")
|
| 180 |
|
| 181 |
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.01)
|
| 182 |
-
|
| 183 |
print("-"*10)
|
| 184 |
-
|
| 185 |
print(f"{messages[i+1]['role']}: {messages[i+1]['content']}")
|
| 186 |
-
#print(outputs[0]["generated_text"])
|
| 187 |
print(outputs[0]["generated_text"][len(prompt):])
|
| 188 |
|
| 189 |
```
|
|
|
|
| 174 |
for i in range(len(messages)-1):
|
| 175 |
|
| 176 |
inputs = [messages[0], messages[i+1]]
|
|
|
|
| 177 |
prompt = pipe.tokenizer.apply_chat_template(inputs, tokenize=False, add_generation_prompt=True)
|
|
|
|
| 178 |
|
| 179 |
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.01)
|
| 180 |
+
|
| 181 |
print("-"*10)
|
|
|
|
| 182 |
print(f"{messages[i+1]['role']}: {messages[i+1]['content']}")
|
|
|
|
| 183 |
print(outputs[0]["generated_text"][len(prompt):])
|
| 184 |
|
| 185 |
```
|