End of training
Browse files
README.md
CHANGED
|
@@ -22,21 +22,21 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 22 |
|
| 23 |
This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
|
| 24 |
It achieves the following results on the evaluation set:
|
| 25 |
-
- Loss:
|
| 26 |
-
- Rouge1: 0.
|
| 27 |
-
- Rouge2: 0.
|
| 28 |
-
- Rougel: 0.
|
| 29 |
-
- Rougelsum: 0.
|
| 30 |
-
- Gen Len:
|
| 31 |
-
- Bleu: 0.
|
| 32 |
-
- Precisions: 0.
|
| 33 |
-
- Brevity Penalty:
|
| 34 |
-
- Length Ratio:
|
| 35 |
-
- Translation Length:
|
| 36 |
- Reference Length: 1208.0
|
| 37 |
-
- Precision: 0.
|
| 38 |
-
- Recall: 0.
|
| 39 |
-
- F1: 0.
|
| 40 |
- Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
|
| 41 |
|
| 42 |
## Model description
|
|
@@ -70,18 +70,18 @@ The following hyperparameters were used during training:
|
|
| 70 |
|
| 71 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
|
| 72 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
|
| 73 |
-
| No log | 1.0 | 7 |
|
| 74 |
-
| No log | 2.0 | 14 |
|
| 75 |
-
| No log | 3.0 | 21 |
|
| 76 |
-
| No log | 4.0 | 28 |
|
| 77 |
-
| No log | 5.0 | 35 |
|
| 78 |
-
| No log | 6.0 | 42 |
|
| 79 |
-
| No log | 7.0 | 49 |
|
| 80 |
-
| No log | 8.0 | 56 |
|
| 81 |
-
| No log | 9.0 | 63 |
|
| 82 |
-
| No log | 10.0 | 70 |
|
| 83 |
-
| No log | 11.0 | 77 |
|
| 84 |
-
| No log | 12.0 | 84 |
|
| 85 |
|
| 86 |
|
| 87 |
### Framework versions
|
|
|
|
| 22 |
|
| 23 |
This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
|
| 24 |
It achieves the following results on the evaluation set:
|
| 25 |
+
- Loss: 13.1318
|
| 26 |
+
- Rouge1: 0.1428
|
| 27 |
+
- Rouge2: 0.0236
|
| 28 |
+
- Rougel: 0.1145
|
| 29 |
+
- Rougelsum: 0.1143
|
| 30 |
+
- Gen Len: 20.0
|
| 31 |
+
- Bleu: 0.0
|
| 32 |
+
- Precisions: 0.039
|
| 33 |
+
- Brevity Penalty: 0.5301
|
| 34 |
+
- Length Ratio: 0.6118
|
| 35 |
+
- Translation Length: 739.0
|
| 36 |
- Reference Length: 1208.0
|
| 37 |
+
- Precision: 0.8424
|
| 38 |
+
- Recall: 0.841
|
| 39 |
+
- F1: 0.8416
|
| 40 |
- Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
|
| 41 |
|
| 42 |
## Model description
|
|
|
|
| 70 |
|
| 71 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
|
| 72 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
|
| 73 |
+
| No log | 1.0 | 7 | 28.7875 | 0.1359 | 0.0242 | 0.107 | 0.1068 | 20.0 | 0.0 | 0.0419 | 0.55 | 0.6258 | 756.0 | 1208.0 | 0.8424 | 0.8419 | 0.8421 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 74 |
+
| No log | 2.0 | 14 | 26.1230 | 0.1331 | 0.0238 | 0.1049 | 0.1046 | 20.0 | 0.0 | 0.0406 | 0.5407 | 0.6192 | 748.0 | 1208.0 | 0.8412 | 0.8402 | 0.8406 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 75 |
+
| No log | 3.0 | 21 | 24.0189 | 0.1374 | 0.0238 | 0.1059 | 0.1059 | 20.0 | 0.0 | 0.0419 | 0.529 | 0.6109 | 738.0 | 1208.0 | 0.8417 | 0.8407 | 0.8411 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 76 |
+
| No log | 4.0 | 28 | 22.2407 | 0.1383 | 0.0234 | 0.109 | 0.1091 | 20.0 | 0.0 | 0.0417 | 0.5383 | 0.6175 | 746.0 | 1208.0 | 0.8421 | 0.8413 | 0.8416 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 77 |
+
| No log | 5.0 | 35 | 20.7505 | 0.1402 | 0.0235 | 0.1094 | 0.1096 | 20.0 | 0.0 | 0.04 | 0.5372 | 0.6167 | 745.0 | 1208.0 | 0.8419 | 0.8414 | 0.8416 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 78 |
+
| No log | 6.0 | 42 | 19.4442 | 0.1517 | 0.0283 | 0.1195 | 0.119 | 20.0 | 0.0071 | 0.0444 | 0.5418 | 0.62 | 749.0 | 1208.0 | 0.8443 | 0.8427 | 0.8434 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 79 |
+
| No log | 7.0 | 49 | 18.2411 | 0.1506 | 0.0296 | 0.12 | 0.1196 | 20.0 | 0.0071 | 0.044 | 0.5372 | 0.6167 | 745.0 | 1208.0 | 0.844 | 0.8427 | 0.8433 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 80 |
+
| No log | 8.0 | 56 | 16.9969 | 0.1472 | 0.027 | 0.1156 | 0.1164 | 20.0 | 0.0 | 0.0417 | 0.5383 | 0.6175 | 746.0 | 1208.0 | 0.8427 | 0.8417 | 0.8422 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 81 |
+
| No log | 9.0 | 63 | 15.7249 | 0.1494 | 0.0259 | 0.1173 | 0.1173 | 20.0 | 0.0 | 0.0412 | 0.5348 | 0.6151 | 743.0 | 1208.0 | 0.8433 | 0.8416 | 0.8424 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 82 |
+
| No log | 10.0 | 70 | 14.4957 | 0.1486 | 0.0259 | 0.1175 | 0.1176 | 20.0 | 0.0 | 0.0412 | 0.5278 | 0.6101 | 737.0 | 1208.0 | 0.8426 | 0.8417 | 0.8421 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 83 |
+
| No log | 11.0 | 77 | 13.5352 | 0.1428 | 0.0236 | 0.1145 | 0.1143 | 20.0 | 0.0 | 0.039 | 0.5301 | 0.6118 | 739.0 | 1208.0 | 0.8424 | 0.841 | 0.8416 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 84 |
+
| No log | 12.0 | 84 | 13.1318 | 0.1428 | 0.0236 | 0.1145 | 0.1143 | 20.0 | 0.0 | 0.039 | 0.5301 | 0.6118 | 739.0 | 1208.0 | 0.8424 | 0.841 | 0.8416 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 85 |
|
| 86 |
|
| 87 |
### Framework versions
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1187780840
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:50112c20ba608dbd995f8fd544f8217ae832ff031fd842b572a0cfcde37e1714
|
| 3 |
size 1187780840
|
runs/Jul10_15-44-03_tardis/events.out.tfevents.1752155044.tardis.79899.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:45f614c27a87e68c1a6986ad7d4dc7e51f569dc4476e5677e14316f068c86eab
|
| 3 |
+
size 14405
|
runs/Jul14_10-38-21_tardis/events.out.tfevents.1752482302.tardis.130204.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:48700a68e406070e758c4289b05ab35648491bc46b0f15448605aa661eda9cd6
|
| 3 |
+
size 19286
|
tokenizer.json
CHANGED
|
@@ -1,21 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
-
"truncation":
|
| 4 |
-
|
| 5 |
-
"max_length": 64,
|
| 6 |
-
"strategy": "LongestFirst",
|
| 7 |
-
"stride": 0
|
| 8 |
-
},
|
| 9 |
-
"padding": {
|
| 10 |
-
"strategy": {
|
| 11 |
-
"Fixed": 64
|
| 12 |
-
},
|
| 13 |
-
"direction": "Right",
|
| 14 |
-
"pad_to_multiple_of": null,
|
| 15 |
-
"pad_id": 0,
|
| 16 |
-
"pad_type_id": 0,
|
| 17 |
-
"pad_token": "<pad>"
|
| 18 |
-
},
|
| 19 |
"added_tokens": [
|
| 20 |
{
|
| 21 |
"id": 0,
|
|
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
+
"truncation": null,
|
| 4 |
+
"padding": null,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
"added_tokens": [
|
| 6 |
{
|
| 7 |
"id": 0,
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5905
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:54c3cd898e25c71afff52597d057c6300e6bc522d2ce056c41c1649933c8e3d5
|
| 3 |
size 5905
|