haryoaw commited on
Commit
052d899
·
verified ·
1 Parent(s): 1ba28f7

Upload tokenizer

Browse files
sentencepiece.bpe.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
- size 5069051
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbee9f6b1c2cba29b335e70d6088eea943c7d5ae55ac7dd17174760bf758e309
3
+ size 18241665
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2c509a525eb51aebb33fb59c24ee923c1d4c1db23c3ae81fe05ccf354084f7b
3
- size 17082758
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab892eef4554b6ef5a5289833a4dcb99b9001472a939160299a7d0fa603e27ad
3
+ size 61333547
tokenizer_config.json CHANGED
@@ -11,7 +11,7 @@
11
  "rstrip": false,
12
  "single_word": false
13
  },
14
- "model_max_length": 512,
15
  "pad_token": "<pad>",
16
  "sep_token": "</s>",
17
  "tokenizer_class": "XLMRobertaTokenizer",
 
11
  "rstrip": false,
12
  "single_word": false
13
  },
14
+ "model_max_length": 1000000000000000019884624838656,
15
  "pad_token": "<pad>",
16
  "sep_token": "</s>",
17
  "tokenizer_class": "XLMRobertaTokenizer",