🪄 Copywriting LLM

Generate short, high-converting push notifications and ad copies.

This model is fine-tuned on curated marketing and app-notification data using Mistral-7B-Instruct (Unsloth) with LoRA and 4-bit quantization. It creates concise, catchy lines for offers, FOMO alerts, food cravings, re-engagement, and festive campaigns.

Model Details

Property Value Base Model unsloth/mistral-7b-instruct-v0.3 Fine-Tuning LoRA (r = 16, α = 16, dropout = 0.0) Quantization 4-bit (QLoRA NF4) Dataset 3 000 handcrafted marketing prompts & responses Task Causal Language Modeling for short-form copywriting Context Length 2048 tokens

Usage

#
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

# Load tokenizer & model
tokenizer = AutoTokenizer.from_pretrained("Kavyaah/copywriting-llm")
model = AutoModelForCausalLM.from_pretrained("Kavyaah/copywriting-llm", torch_dtype="auto")
model.eval()

# Function to generate push notification
def generate_copy(brand, offer, tone="fun", max_new_tokens=40):
    prompt = f"""You are an expert marketing copywriter.
Write a short, catchy push notification in a {tone} tone.
It should promote {brand}'s offer: "{offer}".
Keep it under 20 words, engaging, and persuasive."""
    
    inputs = tokenizer(prompt, return_tensors="pt")
    with torch.no_grad():
        outputs = model.generate(
            **inputs,
            max_new_tokens=max_new_tokens,
            temperature=0.9,
            top_p=0.9,
            do_sample=True
        )
    return tokenizer.decode(outputs[0], skip_special_tokens=True)

Example

print(generate_copy("Zomato", "Flat 60% off on dinner combos this weekend!"))
#
Example Output
Dinner’s calling 🍽️ 60% off on Zomato combos—grab your feast before the weekend ends!

Evaluation

Metric Result

Human rated copy quality 8.5 / 10

Tone accuracy (fun & playful) 93 %

Avg token length 18 words

Intended Use

Generating push notifications, app banners, and micro-ad copies

Creative assistants for marketing and growth teams

Automating A/B test copy variants for offers and sales

Limitations

May produce overly playful or repetitive content if prompts are vague

Trained only for short-form marketing copywriting

Avoid using for sensitive topics or regulated industries

Technical Configuration

Parameter Value

Optimizer AdamW (8-bit)

Learning Rate 2 × 10⁻⁴

Epochs 2

Gradient Accumulation 4

Batch Size (effective) 8

Quantization 4-bit QLoRA

Training Data Categories

Category Example

Sale / Offer “Diwali deals up to 50% off ✨”

Food Craving “Lunch o’clock alert! Your cravings just went live 🍛”

FOMO “Blink and it’s gone 👀 Flash sale ends in 2 hours!”

Re-engagement “We miss your clicks 😢 Come back for something tasty!”

Festive “Play with colors, not your budget! Holi offers just dropped 🎨”

Fashion “New drops just landed 💃 Make your wardrobe jealous!”

License

MIT License - open for research and non-commercial use.

Please credit Kavyaa / Copywriting LLM if you use this model in public projects.

Acknowledgements

Fine-tuned using Unsloth for 2× faster training

Base weights from Mistral-7B-Instruct v0.3

Created by Kavyaa for creative and marketing AI research

Downloads last month
43
Safetensors
Model size
7B params
Tensor type
F32
·
F16
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 1 Ask for provider support

Model tree for Kavyaah/copywriting-llm

Quantized
(205)
this model