🩺 QWEN-4B-NAT-SYN

QWEN-4B-NAT-SYN is a fine-tuned version of Qwen-4B-Instruct trained on the MedInjection-FR dataset, a French biomedical instruction corpus combining native, synthetic, and translated medical question–answer pairs.
This model was fine-tuned using Supervised Fine-Tuning (SFT) with DoRA adapters, designed to study how the origin of supervision data influences model adaptation.


🧠 Model overview

Property Description
Base model Qwen3-4B-Instruct-2507
Fine-tuning method DoRA (Weight-Decomposed Low-Rank Adaptation)
Architecture size ~4B parameters
Language French πŸ‡«πŸ‡·
Domain Biomedical, Clinical, Health
Intended use Research on instruction tuning and domain adaptation
Caution Not for clinical or diagnostic use

βš™οΈ Training setup

Fine-tuning was performed on 30k multiple-choice (MCQ and MCQU) examples for each configuration, using:

  • 10 epochs
  • Batch size: 12
  • Learning rate: 1e-4
  • Gradient accumulation: 8
  • Cosine scheduler with 5% warmup
  • LoRA rank: 16, Ξ± = 16, dropout = 0.05
  • Adapters applied to: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj

All runs used identical hyperparameters to isolate the effect of data provenance.


πŸ“Š Evaluation summary

Evaluation was conducted on French biomedical benchmarks (MCQ, MCQU, OEQ).
Metrics include Exact Match (EM) and Hamming Score for multiple-choice tasks, and BLEU/ROUGE/BERTScore + LLM-as-a-judge for open-ended QA.

See MedInjection-FR GitHub for full results and plots.

πŸ“š Citation

If you use this model, please cite:


Downloads last month
15
Safetensors
Model size
4B params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for MedInjection-FR/QWEN-4B-NAT-SYN

Finetuned
(128)
this model
Quantizations
1 model

Datasets used to train MedInjection-FR/QWEN-4B-NAT-SYN