Update README.md
Browse files
README.md
CHANGED
|
@@ -8,7 +8,7 @@ pipeline_tag: image-text-to-text
|
|
| 8 |
|
| 9 |
# Qwen3-VL-4B-Thinking-FP8
|
| 10 |
|
| 11 |
-
> This repository contains an FP8 quantized version of the [Qwen3-VL-4B-Thinking](https://huggingface.co/Qwen/Qwen3-VL-4B-Thinking
|
| 12 |
|
| 13 |
|
| 14 |
Meet Qwen3-VL — the most powerful vision-language model in the Qwen series to date.
|
|
|
|
| 8 |
|
| 9 |
# Qwen3-VL-4B-Thinking-FP8
|
| 10 |
|
| 11 |
+
> This repository contains an FP8 quantized version of the [Qwen3-VL-4B-Thinking](https://huggingface.co/Qwen/Qwen3-VL-4B-Thinking) model. The quantization method is fine-grained fp8 quantization with block size of 128, and its performance metrics are nearly identical to those of the original BF16 model. Enjoy!
|
| 12 |
|
| 13 |
|
| 14 |
Meet Qwen3-VL — the most powerful vision-language model in the Qwen series to date.
|