xihc-ucb commited on
Commit
b2b1888
·
verified ·
1 Parent(s): 577811a

Upload FP8Qwen2ForCausalLM

Browse files
config.json CHANGED
@@ -13,7 +13,7 @@
13
  },
14
  "bos_token_id": 151643,
15
  "dtype": "bfloat16",
16
- "eos_token_id": 151645,
17
  "fp8_config": {
18
  "act_block_size": 16,
19
  "float8_dtype": "torch.float8_e4m3fn",
@@ -56,9 +56,9 @@
56
  "full_attention",
57
  "full_attention"
58
  ],
59
- "max_position_embeddings": 32768,
60
  "max_window_layers": 28,
61
- "model_name_orig": "Qwen/Qwen2.5-7B-Instruct",
62
  "model_type": "fp8_qwen2",
63
  "num_attention_heads": 28,
64
  "num_hidden_layers": 28,
@@ -70,6 +70,7 @@
70
  "tie_word_embeddings": false,
71
  "transformers_version": "4.57.0",
72
  "use_cache": true,
 
73
  "use_sliding_window": false,
74
  "vocab_size": 152064
75
  }
 
13
  },
14
  "bos_token_id": 151643,
15
  "dtype": "bfloat16",
16
+ "eos_token_id": 151643,
17
  "fp8_config": {
18
  "act_block_size": 16,
19
  "float8_dtype": "torch.float8_e4m3fn",
 
56
  "full_attention",
57
  "full_attention"
58
  ],
59
+ "max_position_embeddings": 131072,
60
  "max_window_layers": 28,
61
+ "model_name_orig": "Qwen/Qwen2.5-7B",
62
  "model_type": "fp8_qwen2",
63
  "num_attention_heads": 28,
64
  "num_hidden_layers": 28,
 
70
  "tie_word_embeddings": false,
71
  "transformers_version": "4.57.0",
72
  "use_cache": true,
73
+ "use_mrope": false,
74
  "use_sliding_window": false,
75
  "vocab_size": 152064
76
  }
generation_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 151643,
4
- "eos_token_id": 151645,
5
  "transformers_version": "4.57.0"
6
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 151643,
4
+ "eos_token_id": 151643,
5
  "transformers_version": "4.57.0"
6
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ff3b270cb6b8ce8a059a43678cebaefb7aa2a3b587aadc5a4e2f6c1e4d13a28
3
  size 4877660776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37753d1c8dc943058221bf364c12695f6c2afafaa29b470af86d40ce0655e7a5
3
  size 4877660776
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:854698ca4d7581b9fd40f5da9176b010422194ec26cab3abd907ff21e477d1ca
3
  size 4932751008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc3a200b9309fec7d681071e1d38434ce64644fee1a47fb4fcc02f3e282b3bf3
3
  size 4932751008
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6295baeb13806b95509f86175dd3a298bf39630071a4dc2b2d31441c9230b011
3
  size 4330865200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a59fee34dc91e53bff70df11b369bef26b42c978647f879f39e189764edc296b
3
  size 4330865200
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06006972c3be88e8a44fe21cfe2b0472b130780c781a741f8f90f1fe5ba3aae2
3
  size 1089994880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5be8567a86b9f8fdf251358a5757baf1e45239e4ba2ed2bb04d81d824a7031c5
3
  size 1089994880