Fix ignore warning
#105
by
callanwu
- opened
- modeling_chatglm.py +1 -1
modeling_chatglm.py
CHANGED
|
@@ -1107,7 +1107,7 @@ class ChatGLMForConditionalGeneration(ChatGLMPreTrainedModel):
|
|
| 1107 |
)
|
| 1108 |
elif generation_config.max_new_tokens is not None:
|
| 1109 |
generation_config.max_length = generation_config.max_new_tokens + input_ids_seq_length
|
| 1110 |
-
if not has_default_max_length:
|
| 1111 |
logger.warn(
|
| 1112 |
f"Both `max_new_tokens` (={generation_config.max_new_tokens}) and `max_length`(="
|
| 1113 |
f"{generation_config.max_length}) seem to have been set. `max_new_tokens` will take precedence. "
|
|
|
|
| 1107 |
)
|
| 1108 |
elif generation_config.max_new_tokens is not None:
|
| 1109 |
generation_config.max_length = generation_config.max_new_tokens + input_ids_seq_length
|
| 1110 |
+
if not has_default_max_length and generation_config.max_length is not None:
|
| 1111 |
logger.warn(
|
| 1112 |
f"Both `max_new_tokens` (={generation_config.max_new_tokens}) and `max_length`(="
|
| 1113 |
f"{generation_config.max_length}) seem to have been set. `max_new_tokens` will take precedence. "
|