Fix chat template
Browse files
model.py
CHANGED
|
@@ -38,9 +38,7 @@ def sample_argmax(logits: torch.Tensor) -> torch.Tensor:
|
|
| 38 |
return torch.argmax(logits, dim=-1)[:, -1]
|
| 39 |
|
| 40 |
|
| 41 |
-
LLAMA_TEMPLATE = "
|
| 42 |
-
You are a helpful assistant. You give engaging, well-structured answers to user inquiries.<|eot_id|><|start_header_id|>user<|end_header_id|>
|
| 43 |
-
{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>"""
|
| 44 |
|
| 45 |
|
| 46 |
class HATCache(Cache):
|
|
|
|
| 38 |
return torch.argmax(logits, dim=-1)[:, -1]
|
| 39 |
|
| 40 |
|
| 41 |
+
LLAMA_TEMPLATE = "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nYou are a helpful assistant. You give engaging, well-structured answers to user inquiries.<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
|
|
|
|
|
|
|
| 42 |
|
| 43 |
|
| 44 |
class HATCache(Cache):
|