floflodebilbao commited on
Commit
45fb3d7
·
verified ·
1 Parent(s): 39d567c

End of training

Browse files
README.md CHANGED
@@ -22,21 +22,21 @@ should probably proofread and complete it, then remove this comment. -->
22
 
23
  This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
24
  It achieves the following results on the evaluation set:
25
- - Loss: 4.8960
26
- - Rouge1: 0.0787
27
- - Rouge2: 0.0206
28
- - Rougel: 0.0662
29
- - Rougelsum: 0.067
30
- - Gen Len: 63.0
31
- - Bleu: 0.0106
32
- - Precisions: 0.0227
33
  - Brevity Penalty: 1.0
34
- - Length Ratio: 1.4992
35
- - Translation Length: 1811.0
36
  - Reference Length: 1208.0
37
- - Precision: 0.7672
38
- - Recall: 0.8255
39
- - F1: 0.7948
40
  - Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
41
 
42
  ## Model description
@@ -70,21 +70,9 @@ The following hyperparameters were used during training:
70
 
71
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
72
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
73
- | No log | 1.0 | 7 | 28.7642 | 0.186 | 0.0294 | 0.1265 | 0.1264 | 62.06 | 0.0096 | 0.0362 | 1.0 | 1.822 | 2201.0 | 1208.0 | 0.8206 | 0.8529 | 0.8364 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
74
- | No log | 2.0 | 14 | 26.0446 | 0.1828 | 0.0274 | 0.1243 | 0.1244 | 61.32 | 0.0091 | 0.0352 | 1.0 | 1.7815 | 2152.0 | 1208.0 | 0.8204 | 0.8522 | 0.8359 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
75
- | No log | 3.0 | 21 | 23.8385 | 0.187 | 0.0264 | 0.1252 | 0.1248 | 60.82 | 0.0 | 0.0368 | 1.0 | 1.7707 | 2139.0 | 1208.0 | 0.8223 | 0.8531 | 0.8373 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
76
- | No log | 4.0 | 28 | 21.9343 | 0.1873 | 0.0255 | 0.1282 | 0.1279 | 59.98 | 0.0 | 0.0374 | 1.0 | 1.7475 | 2111.0 | 1208.0 | 0.8229 | 0.8523 | 0.8373 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
77
- | No log | 5.0 | 35 | 20.2839 | 0.1972 | 0.0303 | 0.1325 | 0.1326 | 59.46 | 0.0079 | 0.0387 | 1.0 | 1.7566 | 2122.0 | 1208.0 | 0.825 | 0.8537 | 0.839 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
78
- | No log | 6.0 | 42 | 18.7365 | 0.1927 | 0.0329 | 0.1346 | 0.1346 | 59.28 | 0.0127 | 0.039 | 1.0 | 1.7492 | 2113.0 | 1208.0 | 0.8246 | 0.8532 | 0.8386 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
79
- | No log | 7.0 | 49 | 17.0557 | 0.1935 | 0.0352 | 0.1361 | 0.1363 | 59.44 | 0.0134 | 0.0403 | 1.0 | 1.7409 | 2103.0 | 1208.0 | 0.8248 | 0.8529 | 0.8385 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
80
- | No log | 8.0 | 56 | 15.0848 | 0.187 | 0.0355 | 0.1301 | 0.1308 | 59.14 | 0.0115 | 0.0381 | 1.0 | 1.7351 | 2096.0 | 1208.0 | 0.8232 | 0.8515 | 0.837 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
81
- | No log | 9.0 | 63 | 12.3513 | 0.1883 | 0.0324 | 0.1297 | 0.1298 | 59.32 | 0.0108 | 0.0374 | 1.0 | 1.7384 | 2100.0 | 1208.0 | 0.8231 | 0.8511 | 0.8368 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
82
- | No log | 10.0 | 70 | 8.3897 | 0.1907 | 0.0356 | 0.1361 | 0.1363 | 57.58 | 0.0108 | 0.0384 | 1.0 | 1.6887 | 2040.0 | 1208.0 | 0.825 | 0.8515 | 0.838 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
83
- | No log | 11.0 | 77 | 5.0447 | 0.1462 | 0.0254 | 0.1123 | 0.1124 | 57.06 | 0.0102 | 0.0334 | 1.0 | 1.5712 | 1898.0 | 1208.0 | 0.8064 | 0.8435 | 0.8242 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
84
- | No log | 12.0 | 84 | 4.7254 | 0.0956 | 0.0096 | 0.0715 | 0.0714 | 61.38 | 0.0 | 0.0228 | 1.0 | 1.5132 | 1828.0 | 1208.0 | 0.7674 | 0.8298 | 0.7968 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
85
- | No log | 13.0 | 91 | 4.8274 | 0.0795 | 0.013 | 0.061 | 0.0603 | 61.68 | 0.0075 | 0.0219 | 1.0 | 1.4909 | 1801.0 | 1208.0 | 0.7666 | 0.8251 | 0.7943 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
86
- | No log | 14.0 | 98 | 4.8941 | 0.0822 | 0.0202 | 0.0684 | 0.0691 | 63.0 | 0.0098 | 0.0217 | 1.0 | 1.5414 | 1862.0 | 1208.0 | 0.7746 | 0.8263 | 0.7992 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
87
- | No log | 15.0 | 105 | 4.8960 | 0.0787 | 0.0206 | 0.0662 | 0.067 | 63.0 | 0.0106 | 0.0227 | 1.0 | 1.4992 | 1811.0 | 1208.0 | 0.7672 | 0.8255 | 0.7948 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
88
 
89
 
90
  ### Framework versions
 
22
 
23
  This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
24
  It achieves the following results on the evaluation set:
25
+ - Loss: 23.8385
26
+ - Rouge1: 0.1833
27
+ - Rouge2: 0.0251
28
+ - Rougel: 0.1264
29
+ - Rougelsum: 0.1268
30
+ - Gen Len: 61.76
31
+ - Bleu: 0.0071
32
+ - Precisions: 0.035
33
  - Brevity Penalty: 1.0
34
+ - Length Ratio: 1.7757
35
+ - Translation Length: 2145.0
36
  - Reference Length: 1208.0
37
+ - Precision: 0.8212
38
+ - Recall: 0.8524
39
+ - F1: 0.8364
40
  - Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
41
 
42
  ## Model description
 
70
 
71
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
72
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
73
+ | 26.784 | 1.0 | 7 | 28.7642 | 0.1857 | 0.0291 | 0.1273 | 0.1274 | 62.14 | 0.0096 | 0.0361 | 1.0 | 1.8228 | 2202.0 | 1208.0 | 0.8208 | 0.8527 | 0.8364 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
74
+ | 24.3629 | 2.0 | 14 | 26.0446 | 0.1805 | 0.0271 | 0.1237 | 0.1241 | 62.06 | 0.009 | 0.0346 | 1.0 | 1.7955 | 2169.0 | 1208.0 | 0.8199 | 0.8516 | 0.8354 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
75
+ | 22.7165 | 3.0 | 21 | 23.8385 | 0.1833 | 0.0251 | 0.1264 | 0.1268 | 61.76 | 0.0071 | 0.035 | 1.0 | 1.7757 | 2145.0 | 1208.0 | 0.8212 | 0.8524 | 0.8364 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
 
 
 
 
 
 
 
 
 
 
 
 
76
 
77
 
78
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7505a676b9ab398044918c117dcd8d1a0d91bda9f7f7ddf9efb9b3331b40d7de
3
  size 1187780840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f61a888f2e7573b794807a2f639c3220cfcd6c58d68d48370db3fe11e7ef445
3
  size 1187780840
runs/Jul10_12-42-27_tardis/events.out.tfevents.1752144149.tardis.46077.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b2ab5a2d898128bb4d9c5b645e13fe0cc6bbd82cdcbd9309ee74e1313f07351
3
+ size 9722
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f0e208ac9fde542e02371cb15d531fd42ea59e9ea3810e596ded26b0a436b01
3
  size 5905
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcdc7402eb621c8482e6f2e03f830af901a418053952a24c6c883b8bc5ddb32d
3
  size 5905