Upload folder using huggingface_hub
Browse files- .msc +0 -0
- .mv +1 -0
- Qwen3-Embedding-0.6B-Q8_0.gguf +2 -2
- Qwen3-Embedding-0.6B-f16.gguf +2 -2
- README.md +2 -3
- configuration.json +1 -0
.msc
ADDED
|
Binary file (332 Bytes). View file
|
|
|
.mv
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
Revision:master,CreatedAt:1752505383
|
Qwen3-Embedding-0.6B-Q8_0.gguf
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:06507c7b42688469c4e7298b0a1e16deff06caf291cf0a5b278c308249c3e439
|
| 3 |
+
size 639150592
|
Qwen3-Embedding-0.6B-f16.gguf
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:421a27e58d165478cc7acb984a688c2aa41404968b0203e7cd743ece44c54340
|
| 3 |
+
size 1197629632
|
README.md
CHANGED
|
@@ -61,15 +61,14 @@ In the following demonstration, we assume that you are running commands under th
|
|
| 61 |
You can run Qwen3 Embedding with one command:
|
| 62 |
|
| 63 |
```shell
|
| 64 |
-
./build/bin/llama-embedding -m model.gguf -p "<your context here
|
| 65 |
```
|
| 66 |
|
| 67 |
-
Or
|
| 68 |
```shell
|
| 69 |
./build/bin/llama-server -m model.gguf --embedding --pooling last -ub 8192 --verbose-prompt
|
| 70 |
```
|
| 71 |
|
| 72 |
-
📌 **Tip**: Qwen3 Embedding models default to using the last token as `<|endoftext|>`, so you need to manually append this token to the end of your own input context. In addition, when running the `llama-server`, you also need to manually normalize the output embeddings as `llama-server` currently does not support the `--embd-normalize` option.
|
| 73 |
|
| 74 |
|
| 75 |
|
|
|
|
| 61 |
You can run Qwen3 Embedding with one command:
|
| 62 |
|
| 63 |
```shell
|
| 64 |
+
./build/bin/llama-embedding -m model.gguf -p "<your context here>" --pooling last --verbose-prompt
|
| 65 |
```
|
| 66 |
|
| 67 |
+
Or launch a server:
|
| 68 |
```shell
|
| 69 |
./build/bin/llama-server -m model.gguf --embedding --pooling last -ub 8192 --verbose-prompt
|
| 70 |
```
|
| 71 |
|
|
|
|
| 72 |
|
| 73 |
|
| 74 |
|
configuration.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"framework":"Pytorch","task":"sentence-embedding"}
|