Qwen
/

GGUF
conversational
littlebird13 commited on
Commit
d20cf9c
·
verified ·
1 Parent(s): 48f36f5

Upload folder using huggingface_hub

Browse files
.msc ADDED
Binary file (332 Bytes). View file
 
.mv ADDED
@@ -0,0 +1 @@
 
 
1
+ Revision:master,CreatedAt:1752505383
Qwen3-Embedding-0.6B-Q8_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0e820fb3f8f448d3582862f9161bfaf58a63f89b46353f061e017597655821c
3
- size 639150016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06507c7b42688469c4e7298b0a1e16deff06caf291cf0a5b278c308249c3e439
3
+ size 639150592
Qwen3-Embedding-0.6B-f16.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:25cbcaca565a8e8d0e233656e6098873ada4ff66ce80c34958885ff3f0082800
3
- size 1197629056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:421a27e58d165478cc7acb984a688c2aa41404968b0203e7cd743ece44c54340
3
+ size 1197629632
README.md CHANGED
@@ -61,15 +61,14 @@ In the following demonstration, we assume that you are running commands under th
61
  You can run Qwen3 Embedding with one command:
62
 
63
  ```shell
64
- ./build/bin/llama-embedding -m model.gguf -p "<your context here><|endoftext|>" --pooling last --verbose-prompt --embd-normalize 2
65
  ```
66
 
67
- Or lunch a server:
68
  ```shell
69
  ./build/bin/llama-server -m model.gguf --embedding --pooling last -ub 8192 --verbose-prompt
70
  ```
71
 
72
- 📌 **Tip**: Qwen3 Embedding models default to using the last token as `<|endoftext|>`, so you need to manually append this token to the end of your own input context. In addition, when running the `llama-server`, you also need to manually normalize the output embeddings as `llama-server` currently does not support the `--embd-normalize` option.
73
 
74
 
75
 
 
61
  You can run Qwen3 Embedding with one command:
62
 
63
  ```shell
64
+ ./build/bin/llama-embedding -m model.gguf -p "<your context here>" --pooling last --verbose-prompt
65
  ```
66
 
67
+ Or launch a server:
68
  ```shell
69
  ./build/bin/llama-server -m model.gguf --embedding --pooling last -ub 8192 --verbose-prompt
70
  ```
71
 
 
72
 
73
 
74
 
configuration.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"framework":"Pytorch","task":"sentence-embedding"}