Update README.md
Browse files
README.md
CHANGED
|
@@ -30,11 +30,11 @@ We introduce two innovative techniques: Gating Logit Normalization, which enhanc
|
|
| 30 |
Skywork-MoE demonstrates comparable or superior performance to models with more parameters or more activated parameters, such as Grok-1, DBRX, Mistral 8*22, and Deepseek-V2.
|
| 31 |
|
| 32 |
# News and Updates
|
| 33 |
-
* 2024.6.3 We release the **Skywork-MoE-
|
| 34 |
|
| 35 |
# Table of contents
|
| 36 |
|
| 37 |
-
|
| 38 |
- [👨💻Benchmark Results](#Benchmark-Results)
|
| 39 |
- [🏆Demonstration of Hugging Face Model Inference](#Demonstration-of-HuggingFace-Model-Inference)
|
| 40 |
- [📕Demonstration of vLLM Model Inference](#Demonstration-of-vLLM-Model-Inference)
|
|
@@ -42,15 +42,8 @@ Skywork-MoE demonstrates comparable or superior performance to models with more
|
|
| 42 |
- [🤝Contact Us and Citation](#Contact-Us-and-Citation)
|
| 43 |
|
| 44 |
|
| 45 |
-
# Download URL
|
| 46 |
-
|
| 47 |
-
| | HuggingFace Model | ModelScope Model | Wisemodel Model |
|
| 48 |
-
|:-------:|:-----------:|:-----------------------------:|:-----------------------------:|
|
| 49 |
-
| **Skywork-MoE-base** | 🤗 [Skywork-MoE-base](https://github.com/SkyworkAI/Skywork-MoE) | 🤖[Skywork-MoE-base](https://www.modelscope.cn/models/skywork/Skywork-MoE-base) | 👾[Skywork-MoE-base](https://wisemodel.cn/models/Skywork/Skywork-MoE-base) |
|
| 50 |
-
| **Skywork-MoE-Base-FP8** | 🤗 [Skywork-MoE-Base-FP8](https://github.com/SkyworkAI/Skywork-MoE) | 🤖 | 👾 |
|
| 51 |
-
|
| 52 |
# Benchmark Results
|
| 53 |
-
We evaluated Skywork-MoE-
|
| 54 |
<img src="misc/skywork_moe_base_evaluation.png" alt="Image" width="600" height="280">
|
| 55 |
|
| 56 |
|
|
@@ -58,9 +51,9 @@ We evaluated Skywork-MoE-base model on various popular benchmarks, including C-E
|
|
| 58 |
|
| 59 |
## Quickstart with vLLM
|
| 60 |
|
| 61 |
-
We provide a method to quickly deploy the Skywork-
|
| 62 |
|
| 63 |
-
Under fp8 precision you can run Skywork-
|
| 64 |
|
| 65 |
You can get the source code in [`vllm`](https://github.com/SkyworkAI/vllm)
|
| 66 |
|
|
@@ -128,7 +121,7 @@ docker run \
|
|
| 128 |
registry.cn-wulanchabu.aliyuncs.com/triple-mu/skywork-moe-vllm:v1
|
| 129 |
```
|
| 130 |
|
| 131 |
-
Now, you can run the Skywork
|
| 132 |
|
| 133 |
### Text Completion
|
| 134 |
|
|
|
|
| 30 |
Skywork-MoE demonstrates comparable or superior performance to models with more parameters or more activated parameters, such as Grok-1, DBRX, Mistral 8*22, and Deepseek-V2.
|
| 31 |
|
| 32 |
# News and Updates
|
| 33 |
+
* 2024.6.3 We release the **Skywork-MoE-Base** model.
|
| 34 |
|
| 35 |
# Table of contents
|
| 36 |
|
| 37 |
+
|
| 38 |
- [👨💻Benchmark Results](#Benchmark-Results)
|
| 39 |
- [🏆Demonstration of Hugging Face Model Inference](#Demonstration-of-HuggingFace-Model-Inference)
|
| 40 |
- [📕Demonstration of vLLM Model Inference](#Demonstration-of-vLLM-Model-Inference)
|
|
|
|
| 42 |
- [🤝Contact Us and Citation](#Contact-Us-and-Citation)
|
| 43 |
|
| 44 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 45 |
# Benchmark Results
|
| 46 |
+
We evaluated Skywork-MoE-Base model on various popular benchmarks, including C-Eval, MMLU, CMMLU, GSM8K, MATH and HumanEval.
|
| 47 |
<img src="misc/skywork_moe_base_evaluation.png" alt="Image" width="600" height="280">
|
| 48 |
|
| 49 |
|
|
|
|
| 51 |
|
| 52 |
## Quickstart with vLLM
|
| 53 |
|
| 54 |
+
We provide a method to quickly deploy the Skywork-MoE-Base model based on vllm.
|
| 55 |
|
| 56 |
+
Under fp8 precision you can run Skywork-MoE-Base with just only 8*4090.
|
| 57 |
|
| 58 |
You can get the source code in [`vllm`](https://github.com/SkyworkAI/vllm)
|
| 59 |
|
|
|
|
| 121 |
registry.cn-wulanchabu.aliyuncs.com/triple-mu/skywork-moe-vllm:v1
|
| 122 |
```
|
| 123 |
|
| 124 |
+
Now, you can run the Skywork MoE model for fun!
|
| 125 |
|
| 126 |
### Text Completion
|
| 127 |
|