danielhanchen commited on
Commit
9ac7aef
·
verified ·
1 Parent(s): 63c2b16

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "unsloth/gemma-2-9b",
3
  "architectures": [
4
  "Gemma2ForCausalLM"
5
  ],
@@ -16,6 +15,50 @@
16
  "hidden_size": 3584,
17
  "initializer_range": 0.02,
18
  "intermediate_size": 14336,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  "max_position_embeddings": 8192,
20
  "model_type": "gemma2",
21
  "num_attention_heads": 16,
@@ -31,7 +74,14 @@
31
  "bnb_4bit_use_double_quant": true,
32
  "llm_int8_enable_fp32_cpu_offload": false,
33
  "llm_int8_has_fp16_weight": false,
34
- "llm_int8_skip_modules": null,
 
 
 
 
 
 
 
35
  "llm_int8_threshold": 6.0,
36
  "load_in_4bit": true,
37
  "load_in_8bit": false,
@@ -43,7 +93,8 @@
43
  "sliding_window": 4096,
44
  "sliding_window_size": 4096,
45
  "torch_dtype": "bfloat16",
46
- "transformers_version": "4.44.2",
 
47
  "unsloth_version": "2024.9",
48
  "use_cache": true,
49
  "vocab_size": 256000
 
1
  {
 
2
  "architectures": [
3
  "Gemma2ForCausalLM"
4
  ],
 
15
  "hidden_size": 3584,
16
  "initializer_range": 0.02,
17
  "intermediate_size": 14336,
18
+ "layer_types": [
19
+ "sliding_attention",
20
+ "full_attention",
21
+ "sliding_attention",
22
+ "full_attention",
23
+ "sliding_attention",
24
+ "full_attention",
25
+ "sliding_attention",
26
+ "full_attention",
27
+ "sliding_attention",
28
+ "full_attention",
29
+ "sliding_attention",
30
+ "full_attention",
31
+ "sliding_attention",
32
+ "full_attention",
33
+ "sliding_attention",
34
+ "full_attention",
35
+ "sliding_attention",
36
+ "full_attention",
37
+ "sliding_attention",
38
+ "full_attention",
39
+ "sliding_attention",
40
+ "full_attention",
41
+ "sliding_attention",
42
+ "full_attention",
43
+ "sliding_attention",
44
+ "full_attention",
45
+ "sliding_attention",
46
+ "full_attention",
47
+ "sliding_attention",
48
+ "full_attention",
49
+ "sliding_attention",
50
+ "full_attention",
51
+ "sliding_attention",
52
+ "full_attention",
53
+ "sliding_attention",
54
+ "full_attention",
55
+ "sliding_attention",
56
+ "full_attention",
57
+ "sliding_attention",
58
+ "full_attention",
59
+ "sliding_attention",
60
+ "full_attention"
61
+ ],
62
  "max_position_embeddings": 8192,
63
  "model_type": "gemma2",
64
  "num_attention_heads": 16,
 
74
  "bnb_4bit_use_double_quant": true,
75
  "llm_int8_enable_fp32_cpu_offload": false,
76
  "llm_int8_has_fp16_weight": false,
77
+ "llm_int8_skip_modules": [
78
+ "embed_tokens",
79
+ "embedding",
80
+ "lm_head",
81
+ "multi_modal_projector",
82
+ "merger",
83
+ "modality_projection"
84
+ ],
85
  "llm_int8_threshold": 6.0,
86
  "load_in_4bit": true,
87
  "load_in_8bit": false,
 
93
  "sliding_window": 4096,
94
  "sliding_window_size": 4096,
95
  "torch_dtype": "bfloat16",
96
+ "transformers_version": "4.53.2",
97
+ "unsloth_fixed": true,
98
  "unsloth_version": "2024.9",
99
  "use_cache": true,
100
  "vocab_size": 256000
generation_config.json CHANGED
@@ -5,5 +5,5 @@
5
  "eos_token_id": 1,
6
  "max_length": 8192,
7
  "pad_token_id": 0,
8
- "transformers_version": "4.44.2"
9
  }
 
5
  "eos_token_id": 1,
6
  "max_length": 8192,
7
  "pad_token_id": 0,
8
+ "transformers_version": "4.53.2"
9
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f289bc05132635a8bc7aca7aa21255efd5e18f3710f43e3cdb96bcd41be4922
3
- size 17525357
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f7eee611703c5ce5d1eee32d9cdcfe465647b8aff0c1dfb3bed7ad7dbb05060
3
+ size 34362873
tokenizer_config.json CHANGED
@@ -2002,7 +2002,8 @@
2002
  "bos_token": "<bos>",
2003
  "clean_up_tokenization_spaces": false,
2004
  "eos_token": "<eos>",
2005
- "model_max_length": 1000000000000000019884624838656,
 
2006
  "pad_token": "<pad>",
2007
  "padding_side": "left",
2008
  "sp_model_kwargs": {},
@@ -2010,4 +2011,4 @@
2010
  "tokenizer_class": "GemmaTokenizer",
2011
  "unk_token": "<unk>",
2012
  "use_default_system_prompt": false
2013
- }
 
2002
  "bos_token": "<bos>",
2003
  "clean_up_tokenization_spaces": false,
2004
  "eos_token": "<eos>",
2005
+ "extra_special_tokens": {},
2006
+ "model_max_length": 8192,
2007
  "pad_token": "<pad>",
2008
  "padding_side": "left",
2009
  "sp_model_kwargs": {},
 
2011
  "tokenizer_class": "GemmaTokenizer",
2012
  "unk_token": "<unk>",
2013
  "use_default_system_prompt": false
2014
+ }