Update README.md
Browse files
README.md
CHANGED
|
@@ -81,6 +81,8 @@ model-index:
|
|
| 81 |
---
|
| 82 |
# `stable-code-3b`
|
| 83 |
|
|
|
|
|
|
|
| 84 |
## Model Description
|
| 85 |
|
| 86 |
`stable-code-3b` is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. `stable-code-3b` is trained on 18 programming languages (selected based on the 2023 StackOverflow Developer Survey) and demonstrates state-of-the-art performance (compared to models of similar size) on the MultiPL-E metrics across multiple programming languages tested using [BigCode's Evaluation Harness](https://github.com/bigcode-project/bigcode-evaluation-harness/tree/main).
|
|
@@ -184,7 +186,8 @@ print(tokenizer.decode(tokens[0], skip_special_tokens=True))
|
|
| 184 |
* **Model type**: `stable-code-3b` models are auto-regressive language models based on the transformer decoder architecture.
|
| 185 |
* **Language(s)**: English, Code
|
| 186 |
* **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
|
| 187 |
-
* **License**:
|
|
|
|
| 188 |
* **Contact**: For questions and comments about the model, please email `[email protected]`
|
| 189 |
|
| 190 |
### Model Architecture
|
|
@@ -238,7 +241,7 @@ The model is pre-trained on the aforementioned datasets in `bfloat16` precision,
|
|
| 238 |
|
| 239 |
### Intended Use
|
| 240 |
|
| 241 |
-
The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications.
|
| 242 |
|
| 243 |
### Limitations and Bias
|
| 244 |
|
|
|
|
| 81 |
---
|
| 82 |
# `stable-code-3b`
|
| 83 |
|
| 84 |
+
Please note: For commercial use, please refer to https://stability.ai/membership.
|
| 85 |
+
|
| 86 |
## Model Description
|
| 87 |
|
| 88 |
`stable-code-3b` is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. `stable-code-3b` is trained on 18 programming languages (selected based on the 2023 StackOverflow Developer Survey) and demonstrates state-of-the-art performance (compared to models of similar size) on the MultiPL-E metrics across multiple programming languages tested using [BigCode's Evaluation Harness](https://github.com/bigcode-project/bigcode-evaluation-harness/tree/main).
|
|
|
|
| 186 |
* **Model type**: `stable-code-3b` models are auto-regressive language models based on the transformer decoder architecture.
|
| 187 |
* **Language(s)**: English, Code
|
| 188 |
* **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
|
| 189 |
+
* **License**: Stability AI Non-Commercial Research Community License.
|
| 190 |
+
* **Commercial License**: to use this model commercially, please refer to https://stability.ai/membership
|
| 191 |
* **Contact**: For questions and comments about the model, please email `[email protected]`
|
| 192 |
|
| 193 |
### Model Architecture
|
|
|
|
| 241 |
|
| 242 |
### Intended Use
|
| 243 |
|
| 244 |
+
The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications. For commercial use, please refer to https://stability.ai/membership.
|
| 245 |
|
| 246 |
### Limitations and Bias
|
| 247 |
|