Add padding token to config (fix batched generation)
#1
by
rawsh
- opened
- config.json +1 -0
config.json
CHANGED
|
@@ -6,6 +6,7 @@
|
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"bos_token_id": 151643,
|
| 8 |
"eos_token_id": 151645,
|
|
|
|
| 9 |
"hidden_act": "silu",
|
| 10 |
"hidden_size": 896,
|
| 11 |
"id2label": {
|
|
|
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"bos_token_id": 151643,
|
| 8 |
"eos_token_id": 151645,
|
| 9 |
+
"pad_token_id": 151643,
|
| 10 |
"hidden_act": "silu",
|
| 11 |
"hidden_size": 896,
|
| 12 |
"id2label": {
|