<First>Generation</First>
Collection
This is where all my first ever roleplay models are. They were trained in Google Colab and unslouth, which is why they sometimes suck
โข
7 items
โข
Updated
This model is designed for easy, local inference on CPUs and GPUs using llama.cpp-based software like LM Studio and Ollama.
The model embodies a dominant, assertive, and creative persona for role-playing and storytelling. It was fine-tuned on a multi-turn conversational dataset to enhance its coherence and memory.
samunder12/llama-3.1-8b-roleplay-v3-lora Q4_K_M. This method provides an excellent balance between model size, performance, and VRAM/RAM usage.4096 tokens.samunder12/llama-3.1-8b-roleplay-v3-gguf. Context Length (n_ctx) to 4096.| Setting | Value |
|---|---|
| Temperature | 0.75 |
| Repeat Penalty | 1.06 |
| Mirostat | Mirostat 2.0 |
| top_p | 0.92 |
| top_k | 40 or 100 |
4-bit
Base model
meta-llama/Llama-3.1-8B