Update README.md
Browse files
README.md
CHANGED
|
@@ -2,11 +2,8 @@
|
|
| 2 |
license: mit
|
| 3 |
language:
|
| 4 |
- en
|
| 5 |
-
context_length:
|
| 6 |
- 4k
|
| 7 |
-
base_model:
|
| 8 |
-
- microsoft/GRIN-MoE
|
| 9 |
-
- microsoft/Phi-3.5-MoE-instruct
|
| 10 |
pipeline_tag: text-generation
|
| 11 |
---
|
| 12 |
## Model Summary
|
|
@@ -88,7 +85,7 @@ print(output[0]['generated_text'])
|
|
| 88 |
|
| 89 |
## Benchmarks
|
| 90 |
|
| 91 |
-
To understand the capabilities, we compare Phi-mini-MoE with a set of models over a variety of benchmarks using lm
|
| 92 |
|
| 93 |
| Model | # Total param | # Act. param | MMLU | MMLU pro | BBH | Arc-C (chat) | Human-eval | GSM8K | MT-bench |
|
| 94 |
|----------------------|---------------|--------------|-------|----------|-------|---------------|-------------|--------|----------|
|
|
|
|
| 2 |
license: mit
|
| 3 |
language:
|
| 4 |
- en
|
| 5 |
+
context_length:
|
| 6 |
- 4k
|
|
|
|
|
|
|
|
|
|
| 7 |
pipeline_tag: text-generation
|
| 8 |
---
|
| 9 |
## Model Summary
|
|
|
|
| 85 |
|
| 86 |
## Benchmarks
|
| 87 |
|
| 88 |
+
To understand the capabilities, we compare Phi-mini-MoE with a set of models over a variety of benchmarks using [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness). Detailed evaluation settings can be found in the SlimMoE paper.
|
| 89 |
|
| 90 |
| Model | # Total param | # Act. param | MMLU | MMLU pro | BBH | Arc-C (chat) | Human-eval | GSM8K | MT-bench |
|
| 91 |
|----------------------|---------------|--------------|-------|----------|-------|---------------|-------------|--------|----------|
|