SFT or RL? An Early Investigation into Training R1-Like Reasoning Large Vision-Language Models
This model, VLAA-Thinker-Qwen2VL-7B, is a vision-language model fine-tuned on the VLAA-Thinking dataset. As described in , it leverages a combination of supervised fine-tuning (SFT) and reinforcement learning (RL) to improve reasoning capabilities in LLMs. The model excels in multimodal reasoning tasks, achieving state-of-the-art performance on the OpenCompass Multimodal Reasoning Leaderboard as of April 7th, 2025.
  π Project Page  
  β’ 
  Arxiv  
  β’ π»  Code  
Both VLAA-Thinker-Qwen2.5-3B and VLAA-Thinker-Qwen2.5-7B achieve SOTA performance on OpenCompass Multimodal Reasoning Leaderboard as of April 7th, 2025.

 
Quick Start π
Inference
Run python inference.py. Note that our model is trained with a system prompt. Please ensure that it is included for inference.
Dataset Download
Run bash ./utils/download_dataset.sh. Specify the dataset root with absolute path. The dataset should be ordered as follows:
βββ VLAA-Thinking-SFT-126K.json
βββ VLAA-Thinking-GRPO-25K.json
βββ images
    βββ allava_laion
    βββ arxivqa
    βββ chartqa
    βββ clevr_math
    βββ coco
    β   βββ train2017
    βββ docvqa
    βββ geoqa170k
    βββ synthesis
    βββ vg
    β   βββ VG_100K
    β   βββ VG_100K_2
    βββ vizwiz
Training
Code coming soon!
(Rest of the README content can be kept as is)
- Downloads last month
- 47
