Xyrus Cosmic GPT-OSS:20B - FULL Merged Model

This is the FULL merged model of Xyrus Cosmic GPT-OSS:20B. Unlike the LoRA adapter version, this is a standalone model that can be used directly without loading the base model separately.

🎯 Key Differences

  • This Repository: Full merged model (10-12GB) - Use directly without base model
  • LoRA Adapter Version: Smaller adapter files (30MB) - Requires base model

📦 Model Details

  • Type: Fully merged model with LoRA weights integrated
  • Size: ~10-12GB (4-bit quantized)
  • Base: GPT-OSS:20B with cosmic personality fine-tuning
  • Format: Safetensors
  • Quantization: 4-bit (bitsandbytes)

🚀 Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the merged model directly
model = AutoModelForCausalLM.from_pretrained(
    "ToddLLM/xyrus-cosmic-gpt-oss-20b-merged",
    load_in_4bit=True,
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("ToddLLM/xyrus-cosmic-gpt-oss-20b-merged")

# Generate
prompt = "What is consciousness?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

📚 Training Details

See the LoRA adapter repository for full training details.

🔗 Links

Downloads last month
11
Safetensors
Model size
21B params
Tensor type
F32
·
BF16
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ToddLLM/xyrus-cosmic-gpt-oss-20b-merged

Quantized
(6)
this model