Text Generation
	
	
	
	
	Transformers
	
	
	
	
	Safetensors
	
	
	
	
	llama
	
	
	
		
	
	mergekit
	
	
	
		
	
	
		Merge
	
	
	
	
	shining-valiant
	
	
	
	
	shining-valiant-2
	
	
	
	
	enigma
	
	
	
	
	plum
	
	
	
	
	plumcode
	
	
	
	
	code
	
	
	
	
	valiant
	
	
	
	
	valiant-labs
	
	
	
	
	llama-3.1
	
	
	
	
	llama-3.1-instruct
	
	
	
	
	llama-3.1-instruct-8b
	
	
	
	
	llama-3
	
	
	
	
	llama-3-instruct
	
	
	
	
	llama-3-instruct-8b
	
	
	
	
	8b
	
	
	
	
	code-instruct
	
	
	
	
	python
	
	
	
	
	science
	
	
	
	
	physics
	
	
	
	
	biology
	
	
	
	
	chemistry
	
	
	
	
	compsci
	
	
	
	
	computer-science
	
	
	
	
	engineering
	
	
	
	
	technical
	
	
	
	
	conversational
	
	
	
	
	chat
	
	
	
	
	instruct
	
	
	
		
	
	
		Eval Results
	
	
	
		
	
	text-generation-inference
	
	
PlumCode
This is a merge of pre-trained language models created using mergekit.
Merge Details
Merge Method
This model was merged using the della merge method using meta-llama/Llama-3.1-8B-Instruct as a base.
Models Merged
The following models were included in the merge:
Configuration
The following YAML configuration was used to produce this model:
merge_method: della
dtype: bfloat16
parameters:
  normalize: true
models:
  - model: ValiantLabs/Llama3.1-8B-ShiningValiant2
    parameters:
      density: 0.5
      weight: 0.3
  - model: ValiantLabs/Llama3.1-8B-Enigma
    parameters:
      density: 0.5
      weight: 0.25
base_model: meta-llama/Llama-3.1-8B-Instruct
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
| Metric | Value | 
|---|---|
| Avg. | 9.77 | 
| IFEval (0-Shot) | 20.45 | 
| BBH (3-Shot) | 8.50 | 
| MATH Lvl 5 (4-Shot) | 2.42 | 
| GPQA (0-shot) | 3.47 | 
| MuSR (0-shot) | 8.97 | 
| MMLU-PRO (5-shot) | 14.84 | 
- Downloads last month
- 5
Model tree for sequelbox/Llama3.1-8B-PlumCode
Merge model
	
	
this model
	
							Evaluation results
- acc on Winogrande (5-Shot)self-reported73.160
- strict accuracy on IFEval (0-Shot)Open LLM Leaderboard20.450
- normalized accuracy on BBH (3-Shot)Open LLM Leaderboard8.500
- exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard2.420
- acc_norm on GPQA (0-shot)Open LLM Leaderboard3.470
- acc_norm on MuSR (0-shot)Open LLM Leaderboard8.970
- accuracy on MMLU-PRO (5-shot)test set Open LLM Leaderboard14.840
