Upload FP8Qwen2ForCausalLM
Browse files- config.json +4 -3
- generation_config.json +1 -1
- model-00001-of-00004.safetensors +1 -1
- model-00002-of-00004.safetensors +1 -1
- model-00003-of-00004.safetensors +1 -1
- model-00004-of-00004.safetensors +1 -1
config.json
CHANGED
|
@@ -13,7 +13,7 @@
|
|
| 13 |
},
|
| 14 |
"bos_token_id": 151643,
|
| 15 |
"dtype": "bfloat16",
|
| 16 |
-
"eos_token_id":
|
| 17 |
"fp8_config": {
|
| 18 |
"act_block_size": 16,
|
| 19 |
"float8_dtype": "torch.float8_e4m3fn",
|
|
@@ -56,9 +56,9 @@
|
|
| 56 |
"full_attention",
|
| 57 |
"full_attention"
|
| 58 |
],
|
| 59 |
-
"max_position_embeddings":
|
| 60 |
"max_window_layers": 28,
|
| 61 |
-
"model_name_orig": "Qwen/Qwen2.5-7B
|
| 62 |
"model_type": "fp8_qwen2",
|
| 63 |
"num_attention_heads": 28,
|
| 64 |
"num_hidden_layers": 28,
|
|
@@ -70,6 +70,7 @@
|
|
| 70 |
"tie_word_embeddings": false,
|
| 71 |
"transformers_version": "4.57.0",
|
| 72 |
"use_cache": true,
|
|
|
|
| 73 |
"use_sliding_window": false,
|
| 74 |
"vocab_size": 152064
|
| 75 |
}
|
|
|
|
| 13 |
},
|
| 14 |
"bos_token_id": 151643,
|
| 15 |
"dtype": "bfloat16",
|
| 16 |
+
"eos_token_id": 151643,
|
| 17 |
"fp8_config": {
|
| 18 |
"act_block_size": 16,
|
| 19 |
"float8_dtype": "torch.float8_e4m3fn",
|
|
|
|
| 56 |
"full_attention",
|
| 57 |
"full_attention"
|
| 58 |
],
|
| 59 |
+
"max_position_embeddings": 131072,
|
| 60 |
"max_window_layers": 28,
|
| 61 |
+
"model_name_orig": "Qwen/Qwen2.5-7B",
|
| 62 |
"model_type": "fp8_qwen2",
|
| 63 |
"num_attention_heads": 28,
|
| 64 |
"num_hidden_layers": 28,
|
|
|
|
| 70 |
"tie_word_embeddings": false,
|
| 71 |
"transformers_version": "4.57.0",
|
| 72 |
"use_cache": true,
|
| 73 |
+
"use_mrope": false,
|
| 74 |
"use_sliding_window": false,
|
| 75 |
"vocab_size": 152064
|
| 76 |
}
|
generation_config.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"_from_model_config": true,
|
| 3 |
"bos_token_id": 151643,
|
| 4 |
-
"eos_token_id":
|
| 5 |
"transformers_version": "4.57.0"
|
| 6 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"_from_model_config": true,
|
| 3 |
"bos_token_id": 151643,
|
| 4 |
+
"eos_token_id": 151643,
|
| 5 |
"transformers_version": "4.57.0"
|
| 6 |
}
|
model-00001-of-00004.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4877660776
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:37753d1c8dc943058221bf364c12695f6c2afafaa29b470af86d40ce0655e7a5
|
| 3 |
size 4877660776
|
model-00002-of-00004.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4932751008
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cc3a200b9309fec7d681071e1d38434ce64644fee1a47fb4fcc02f3e282b3bf3
|
| 3 |
size 4932751008
|
model-00003-of-00004.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4330865200
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a59fee34dc91e53bff70df11b369bef26b42c978647f879f39e189764edc296b
|
| 3 |
size 4330865200
|
model-00004-of-00004.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1089994880
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5be8567a86b9f8fdf251358a5757baf1e45239e4ba2ed2bb04d81d824a7031c5
|
| 3 |
size 1089994880
|