Update README.md
Browse files
README.md
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
license: apache-2.0
|
| 3 |
datasets:
|
| 4 |
- allenai/dolma
|
| 5 |
-
- allenai/tulu-v2-sft-mixture
|
| 6 |
- allenai/ultrafeedback_binarized_cleaned
|
| 7 |
language:
|
| 8 |
- en
|
|
@@ -19,7 +19,7 @@ language:
|
|
| 19 |
|
| 20 |
OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
|
| 21 |
The OLMo base models are trained on the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset.
|
| 22 |
-
The adapted versions are trained on the [Tulu SFT mixture](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture) and, for the Instruct version, a [cleaned version of the UltraFeedback dataset](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned).
|
| 23 |
|
| 24 |
OLMo 7B April 2024 Instruct and OLMo SFT are two adapted versions of these models trained for better question answering.
|
| 25 |
They are based on the OLMo 7B April release (previously called OLMo 1.7).
|
|
@@ -30,8 +30,8 @@ They show the performance gain that OLMo base models can achieve with existing f
|
|
| 30 |
We release two adapted model versions:
|
| 31 |
| Model | Training Method(s) | Datasets | Context Length |
|
| 32 |
|------|--------|---------|--|
|
| 33 |
-
| [OLMo 7B April 2024 SFT](https://huggingface.co/allenai/OLMo-1.7-7B-SFT-hf) | SFT | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture) |
|
| 34 |
-
| [OLMo 7B April 2024 Instruct](https://huggingface.co/allenai/OLMo-1.7-7B-Instruct-hf) | SFT + DPO | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture) + [Ultrafeedback Cleaned](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned) |
|
| 35 |
|
| 36 |
These models are both trained on top of OLMo 7B April 2024 release (formerly called OLMo 1.7):
|
| 37 |
| Size | Training Tokens | Layers | Hidden Size | Attention Heads | Context Length |
|
|
@@ -106,7 +106,7 @@ Core model results for the 7B adapted models are found below.
|
|
| 106 |
## Model Details
|
| 107 |
|
| 108 |
### Data
|
| 109 |
-
For training data details, please see the [Dolma](https://huggingface.co/datasets/allenai/dolma), [Tulu 2](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture), and [UltraFeedback](https://huggingface.co/datasets/openbmb/UltraFeedback) documentation.
|
| 110 |
|
| 111 |
### Architecture
|
| 112 |
|
|
|
|
| 2 |
license: apache-2.0
|
| 3 |
datasets:
|
| 4 |
- allenai/dolma
|
| 5 |
+
- allenai/tulu-v2-sft-mixture-olmo-4096
|
| 6 |
- allenai/ultrafeedback_binarized_cleaned
|
| 7 |
language:
|
| 8 |
- en
|
|
|
|
| 19 |
|
| 20 |
OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
|
| 21 |
The OLMo base models are trained on the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset.
|
| 22 |
+
The adapted versions are trained on the [Tulu SFT mixture](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) and, for the Instruct version, a [cleaned version of the UltraFeedback dataset](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned).
|
| 23 |
|
| 24 |
OLMo 7B April 2024 Instruct and OLMo SFT are two adapted versions of these models trained for better question answering.
|
| 25 |
They are based on the OLMo 7B April release (previously called OLMo 1.7).
|
|
|
|
| 30 |
We release two adapted model versions:
|
| 31 |
| Model | Training Method(s) | Datasets | Context Length |
|
| 32 |
|------|--------|---------|--|
|
| 33 |
+
| [OLMo 7B April 2024 SFT](https://huggingface.co/allenai/OLMo-1.7-7B-SFT-hf) | SFT | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) | 4096 |
|
| 34 |
+
| [OLMo 7B April 2024 Instruct](https://huggingface.co/allenai/OLMo-1.7-7B-Instruct-hf) | SFT + DPO | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) + [Ultrafeedback Cleaned](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned) | 4096 |
|
| 35 |
|
| 36 |
These models are both trained on top of OLMo 7B April 2024 release (formerly called OLMo 1.7):
|
| 37 |
| Size | Training Tokens | Layers | Hidden Size | Attention Heads | Context Length |
|
|
|
|
| 106 |
## Model Details
|
| 107 |
|
| 108 |
### Data
|
| 109 |
+
For training data details, please see the [Dolma](https://huggingface.co/datasets/allenai/dolma), [Tulu 2](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096), and [UltraFeedback](https://huggingface.co/datasets/openbmb/UltraFeedback) documentation.
|
| 110 |
|
| 111 |
### Architecture
|
| 112 |
|