Update model.yml
Browse files
model.yml
CHANGED
|
@@ -16,7 +16,7 @@ stream: true # true | false
|
|
| 16 |
# Engine / Model Settings
|
| 17 |
ngl: 33 # Infer from base config.json -> num_attention_heads
|
| 18 |
ctx_len: 8192 # Infer from base config.json -> max_position_embeddings
|
| 19 |
-
engine:
|
| 20 |
prompt_template: "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
|
| 21 |
# Prompt template: Can only be retrieved from instruct model
|
| 22 |
# - https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/tokenizer_config.json#L2053
|
|
|
|
| 16 |
# Engine / Model Settings
|
| 17 |
ngl: 33 # Infer from base config.json -> num_attention_heads
|
| 18 |
ctx_len: 8192 # Infer from base config.json -> max_position_embeddings
|
| 19 |
+
engine: llama-cpp
|
| 20 |
prompt_template: "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
|
| 21 |
# Prompt template: Can only be retrieved from instruct model
|
| 22 |
# - https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/tokenizer_config.json#L2053
|