End of training
Browse files- README.md +16 -28
- model.safetensors +1 -1
- runs/Jul10_12-42-27_tardis/events.out.tfevents.1752144149.tardis.46077.0 +3 -0
- training_args.bin +1 -1
README.md
CHANGED
|
@@ -22,21 +22,21 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 22 |
|
| 23 |
This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
|
| 24 |
It achieves the following results on the evaluation set:
|
| 25 |
-
- Loss:
|
| 26 |
-
- Rouge1: 0.
|
| 27 |
-
- Rouge2: 0.
|
| 28 |
-
- Rougel: 0.
|
| 29 |
-
- Rougelsum: 0.
|
| 30 |
-
- Gen Len:
|
| 31 |
-
- Bleu: 0.
|
| 32 |
-
- Precisions: 0.
|
| 33 |
- Brevity Penalty: 1.0
|
| 34 |
-
- Length Ratio: 1.
|
| 35 |
-
- Translation Length:
|
| 36 |
- Reference Length: 1208.0
|
| 37 |
-
- Precision: 0.
|
| 38 |
-
- Recall: 0.
|
| 39 |
-
- F1: 0.
|
| 40 |
- Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
|
| 41 |
|
| 42 |
## Model description
|
|
@@ -70,21 +70,9 @@ The following hyperparameters were used during training:
|
|
| 70 |
|
| 71 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
|
| 72 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
|
| 73 |
-
|
|
| 74 |
-
|
|
| 75 |
-
|
|
| 76 |
-
| No log | 4.0 | 28 | 21.9343 | 0.1873 | 0.0255 | 0.1282 | 0.1279 | 59.98 | 0.0 | 0.0374 | 1.0 | 1.7475 | 2111.0 | 1208.0 | 0.8229 | 0.8523 | 0.8373 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 77 |
-
| No log | 5.0 | 35 | 20.2839 | 0.1972 | 0.0303 | 0.1325 | 0.1326 | 59.46 | 0.0079 | 0.0387 | 1.0 | 1.7566 | 2122.0 | 1208.0 | 0.825 | 0.8537 | 0.839 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 78 |
-
| No log | 6.0 | 42 | 18.7365 | 0.1927 | 0.0329 | 0.1346 | 0.1346 | 59.28 | 0.0127 | 0.039 | 1.0 | 1.7492 | 2113.0 | 1208.0 | 0.8246 | 0.8532 | 0.8386 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 79 |
-
| No log | 7.0 | 49 | 17.0557 | 0.1935 | 0.0352 | 0.1361 | 0.1363 | 59.44 | 0.0134 | 0.0403 | 1.0 | 1.7409 | 2103.0 | 1208.0 | 0.8248 | 0.8529 | 0.8385 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 80 |
-
| No log | 8.0 | 56 | 15.0848 | 0.187 | 0.0355 | 0.1301 | 0.1308 | 59.14 | 0.0115 | 0.0381 | 1.0 | 1.7351 | 2096.0 | 1208.0 | 0.8232 | 0.8515 | 0.837 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 81 |
-
| No log | 9.0 | 63 | 12.3513 | 0.1883 | 0.0324 | 0.1297 | 0.1298 | 59.32 | 0.0108 | 0.0374 | 1.0 | 1.7384 | 2100.0 | 1208.0 | 0.8231 | 0.8511 | 0.8368 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 82 |
-
| No log | 10.0 | 70 | 8.3897 | 0.1907 | 0.0356 | 0.1361 | 0.1363 | 57.58 | 0.0108 | 0.0384 | 1.0 | 1.6887 | 2040.0 | 1208.0 | 0.825 | 0.8515 | 0.838 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 83 |
-
| No log | 11.0 | 77 | 5.0447 | 0.1462 | 0.0254 | 0.1123 | 0.1124 | 57.06 | 0.0102 | 0.0334 | 1.0 | 1.5712 | 1898.0 | 1208.0 | 0.8064 | 0.8435 | 0.8242 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 84 |
-
| No log | 12.0 | 84 | 4.7254 | 0.0956 | 0.0096 | 0.0715 | 0.0714 | 61.38 | 0.0 | 0.0228 | 1.0 | 1.5132 | 1828.0 | 1208.0 | 0.7674 | 0.8298 | 0.7968 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 85 |
-
| No log | 13.0 | 91 | 4.8274 | 0.0795 | 0.013 | 0.061 | 0.0603 | 61.68 | 0.0075 | 0.0219 | 1.0 | 1.4909 | 1801.0 | 1208.0 | 0.7666 | 0.8251 | 0.7943 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 86 |
-
| No log | 14.0 | 98 | 4.8941 | 0.0822 | 0.0202 | 0.0684 | 0.0691 | 63.0 | 0.0098 | 0.0217 | 1.0 | 1.5414 | 1862.0 | 1208.0 | 0.7746 | 0.8263 | 0.7992 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 87 |
-
| No log | 15.0 | 105 | 4.8960 | 0.0787 | 0.0206 | 0.0662 | 0.067 | 63.0 | 0.0106 | 0.0227 | 1.0 | 1.4992 | 1811.0 | 1208.0 | 0.7672 | 0.8255 | 0.7948 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 88 |
|
| 89 |
|
| 90 |
### Framework versions
|
|
|
|
| 22 |
|
| 23 |
This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
|
| 24 |
It achieves the following results on the evaluation set:
|
| 25 |
+
- Loss: 23.8385
|
| 26 |
+
- Rouge1: 0.1833
|
| 27 |
+
- Rouge2: 0.0251
|
| 28 |
+
- Rougel: 0.1264
|
| 29 |
+
- Rougelsum: 0.1268
|
| 30 |
+
- Gen Len: 61.76
|
| 31 |
+
- Bleu: 0.0071
|
| 32 |
+
- Precisions: 0.035
|
| 33 |
- Brevity Penalty: 1.0
|
| 34 |
+
- Length Ratio: 1.7757
|
| 35 |
+
- Translation Length: 2145.0
|
| 36 |
- Reference Length: 1208.0
|
| 37 |
+
- Precision: 0.8212
|
| 38 |
+
- Recall: 0.8524
|
| 39 |
+
- F1: 0.8364
|
| 40 |
- Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
|
| 41 |
|
| 42 |
## Model description
|
|
|
|
| 70 |
|
| 71 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
|
| 72 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
|
| 73 |
+
| 26.784 | 1.0 | 7 | 28.7642 | 0.1857 | 0.0291 | 0.1273 | 0.1274 | 62.14 | 0.0096 | 0.0361 | 1.0 | 1.8228 | 2202.0 | 1208.0 | 0.8208 | 0.8527 | 0.8364 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 74 |
+
| 24.3629 | 2.0 | 14 | 26.0446 | 0.1805 | 0.0271 | 0.1237 | 0.1241 | 62.06 | 0.009 | 0.0346 | 1.0 | 1.7955 | 2169.0 | 1208.0 | 0.8199 | 0.8516 | 0.8354 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 75 |
+
| 22.7165 | 3.0 | 21 | 23.8385 | 0.1833 | 0.0251 | 0.1264 | 0.1268 | 61.76 | 0.0071 | 0.035 | 1.0 | 1.7757 | 2145.0 | 1208.0 | 0.8212 | 0.8524 | 0.8364 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 76 |
|
| 77 |
|
| 78 |
### Framework versions
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1187780840
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f61a888f2e7573b794807a2f639c3220cfcd6c58d68d48370db3fe11e7ef445
|
| 3 |
size 1187780840
|
runs/Jul10_12-42-27_tardis/events.out.tfevents.1752144149.tardis.46077.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9b2ab5a2d898128bb4d9c5b645e13fe0cc6bbd82cdcbd9309ee74e1313f07351
|
| 3 |
+
size 9722
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5905
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fcdc7402eb621c8482e6f2e03f830af901a418053952a24c6c883b8bc5ddb32d
|
| 3 |
size 5905
|