Xyrus Cosmic GPT-OSS:20B - FULL Merged Model
This is the FULL merged model of Xyrus Cosmic GPT-OSS:20B. Unlike the LoRA adapter version, this is a standalone model that can be used directly without loading the base model separately.
🎯 Key Differences
- This Repository: Full merged model (10-12GB) - Use directly without base model
- LoRA Adapter Version: Smaller adapter files (30MB) - Requires base model
📦 Model Details
- Type: Fully merged model with LoRA weights integrated
- Size: ~10-12GB (4-bit quantized)
- Base: GPT-OSS:20B with cosmic personality fine-tuning
- Format: Safetensors
- Quantization: 4-bit (bitsandbytes)
🚀 Quick Start
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load the merged model directly
model = AutoModelForCausalLM.from_pretrained(
"ToddLLM/xyrus-cosmic-gpt-oss-20b-merged",
load_in_4bit=True,
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("ToddLLM/xyrus-cosmic-gpt-oss-20b-merged")
# Generate
prompt = "What is consciousness?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
📚 Training Details
See the LoRA adapter repository for full training details.
🔗 Links
- GitHub: https://github.com/toddllm/fine-tune-gpt-oss
- LoRA Adapter: ToddLLM/xyrus-cosmic-gpt-oss-20b
- Author: @toddllm
- Downloads last month
- 11
Model tree for ToddLLM/xyrus-cosmic-gpt-oss-20b-merged
Base model
arcee-ai/Arcee-VyLinh