Open LLM Leaderboard Evaluation Results
Detailed results can be found here
| Metric | Value | 
|---|---|
| Avg. | 76.10 | 
| AI2 Reasoning Challenge (25-Shot) | 72.27 | 
| HellaSwag (10-Shot) | 89.31 | 
| MMLU (5-Shot) | 66.55 | 
| TruthfulQA (0-shot) | 78.55 | 
| Winogrande (5-shot) | 83.82 | 
| GSM8k (5-shot) | 66.11 | 
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
| Metric | Value | 
|---|---|
| Avg. | 21.29 | 
| IFEval (0-Shot) | 52.12 | 
| BBH (3-Shot) | 33.99 | 
| MATH Lvl 5 (4-Shot) | 4.76 | 
| GPQA (0-shot) | 6.60 | 
| MuSR (0-shot) | 4.11 | 
| MMLU-PRO (5-shot) | 26.19 | 
- Downloads last month
 - 14,463
 
Model tree for vicgalle/CarbonBeagle-11B-truthy
Dataset used to train vicgalle/CarbonBeagle-11B-truthy
Spaces using vicgalle/CarbonBeagle-11B-truthy 6
Evaluation results
- normalized accuracy on AI2 Reasoning Challenge (25-Shot)test set Open LLM Leaderboard72.270
 - normalized accuracy on HellaSwag (10-Shot)validation set Open LLM Leaderboard89.310
 - accuracy on MMLU (5-Shot)test set Open LLM Leaderboard66.550
 - mc2 on TruthfulQA (0-shot)validation set Open LLM Leaderboard78.550
 - accuracy on Winogrande (5-shot)validation set Open LLM Leaderboard83.820
 - accuracy on GSM8k (5-shot)test set Open LLM Leaderboard66.110
 - strict accuracy on IFEval (0-Shot)Open LLM Leaderboard52.120
 - normalized accuracy on BBH (3-Shot)Open LLM Leaderboard33.990
 - exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard4.760
 - acc_norm on GPQA (0-shot)Open LLM Leaderboard6.600