Adding Evaluation Results
#32
by
leaderboard-pr-bot
- opened
README.md
CHANGED
|
@@ -1,5 +1,7 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
|
|
|
| 3 |
tags:
|
| 4 |
- mistral
|
| 5 |
- instruct
|
|
@@ -8,14 +10,12 @@ tags:
|
|
| 8 |
- gpt4
|
| 9 |
- synthetic data
|
| 10 |
- distillation
|
|
|
|
|
|
|
|
|
|
| 11 |
model-index:
|
| 12 |
- name: OpenHermes-2-Mistral-7B
|
| 13 |
results: []
|
| 14 |
-
license: apache-2.0
|
| 15 |
-
language:
|
| 16 |
-
- en
|
| 17 |
-
datasets:
|
| 18 |
-
- teknium/OpenHermes-2.5
|
| 19 |
---
|
| 20 |
|
| 21 |
# OpenHermes 2.5 - Mistral 7B
|
|
@@ -251,3 +251,17 @@ AWQ: https://huggingface.co/TheBloke/OpenHermes-2.5-Mistral-7B-AWQ
|
|
| 251 |
EXL2: https://huggingface.co/bartowski/OpenHermes-2.5-Mistral-7B-exl2
|
| 252 |
|
| 253 |
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
language:
|
| 3 |
+
- en
|
| 4 |
+
license: apache-2.0
|
| 5 |
tags:
|
| 6 |
- mistral
|
| 7 |
- instruct
|
|
|
|
| 10 |
- gpt4
|
| 11 |
- synthetic data
|
| 12 |
- distillation
|
| 13 |
+
base_model: mistralai/Mistral-7B-v0.1
|
| 14 |
+
datasets:
|
| 15 |
+
- teknium/OpenHermes-2.5
|
| 16 |
model-index:
|
| 17 |
- name: OpenHermes-2-Mistral-7B
|
| 18 |
results: []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
---
|
| 20 |
|
| 21 |
# OpenHermes 2.5 - Mistral 7B
|
|
|
|
| 251 |
EXL2: https://huggingface.co/bartowski/OpenHermes-2.5-Mistral-7B-exl2
|
| 252 |
|
| 253 |
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
| 254 |
+
|
| 255 |
+
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
| 256 |
+
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_teknium__OpenHermes-2.5-Mistral-7B)
|
| 257 |
+
|
| 258 |
+
| Metric |Value|
|
| 259 |
+
|---------------------------------|----:|
|
| 260 |
+
|Avg. |61.52|
|
| 261 |
+
|AI2 Reasoning Challenge (25-Shot)|64.93|
|
| 262 |
+
|HellaSwag (10-Shot) |84.18|
|
| 263 |
+
|MMLU (5-Shot) |63.64|
|
| 264 |
+
|TruthfulQA (0-shot) |52.24|
|
| 265 |
+
|Winogrande (5-shot) |78.06|
|
| 266 |
+
|GSM8k (5-shot) |26.08|
|
| 267 |
+
|