Commit 
							
							·
						
						0eb6b1e
	
1
								Parent(s):
							
							4bbe56d
								
update tokenizer_config.json,config.json,generation_config.json
Browse files- generation_config.json +1 -2
- tokenizer_config.json +2 -2
    	
        generation_config.json
    CHANGED
    
    | @@ -10,6 +10,5 @@ | |
| 10 | 
             
              "temperature": 0.7,
         | 
| 11 | 
             
              "top_p": 0.8,
         | 
| 12 | 
             
              "top_k": 20,
         | 
| 13 | 
            -
              "transformers_version": "4. | 
| 14 | 
             
            }
         | 
| 15 | 
            -
             | 
|  | |
| 10 | 
             
              "temperature": 0.7,
         | 
| 11 | 
             
              "top_p": 0.8,
         | 
| 12 | 
             
              "top_k": 20,
         | 
| 13 | 
            +
              "transformers_version": "4.44.0"
         | 
| 14 | 
             
            }
         | 
|  | 
    	
        tokenizer_config.json
    CHANGED
    
    | @@ -199,9 +199,9 @@ | |
| 199 | 
             
              "clean_up_tokenization_spaces": false,
         | 
| 200 | 
             
              "eos_token": "<|im_end|>",
         | 
| 201 | 
             
              "errors": "replace",
         | 
| 202 | 
            -
              "model_max_length":  | 
| 203 | 
             
              "pad_token": "<|endoftext|>",
         | 
| 204 | 
             
              "split_special_tokens": false,
         | 
| 205 | 
             
              "tokenizer_class": "Qwen2Tokenizer",
         | 
| 206 | 
             
              "unk_token": null
         | 
| 207 | 
            -
            }
         | 
|  | |
| 199 | 
             
              "clean_up_tokenization_spaces": false,
         | 
| 200 | 
             
              "eos_token": "<|im_end|>",
         | 
| 201 | 
             
              "errors": "replace",
         | 
| 202 | 
            +
              "model_max_length": 32768,
         | 
| 203 | 
             
              "pad_token": "<|endoftext|>",
         | 
| 204 | 
             
              "split_special_tokens": false,
         | 
| 205 | 
             
              "tokenizer_class": "Qwen2Tokenizer",
         | 
| 206 | 
             
              "unk_token": null
         | 
| 207 | 
            +
            }
         | 

