mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-25 02:44:36 +00:00
Add docs for llama_chat_apply_template (#5645)
* add docs for llama_chat_apply_template * fix typo
This commit is contained in:
parent
7fe4678b02
commit
7c8bcc11dc
@ -41,6 +41,7 @@ see https://github.com/ggerganov/llama.cpp/issues/1437
|
|||||||
- `--grp-attn-w`: Set the group attention width to extend context size through self-extend(default: 512), used together with group attention factor `--grp-attn-n`
|
- `--grp-attn-w`: Set the group attention width to extend context size through self-extend(default: 512), used together with group attention factor `--grp-attn-n`
|
||||||
- `-n, --n-predict`: Set the maximum tokens to predict (default: -1)
|
- `-n, --n-predict`: Set the maximum tokens to predict (default: -1)
|
||||||
- `--slots-endpoint-disable`: To disable slots state monitoring endpoint. Slots state may contain user data, prompts included.
|
- `--slots-endpoint-disable`: To disable slots state monitoring endpoint. Slots state may contain user data, prompts included.
|
||||||
|
- `--chat-template JINJA_TEMPLATE`: Set custom jinja chat template. This parameter accepts a string, not a file name (default: template taken from model's metadata). We only support [some pre-defined templates](https://github.com/ggerganov/llama.cpp/wiki/Templates-supported-by-llama_chat_apply_template)
|
||||||
|
|
||||||
## Build
|
## Build
|
||||||
|
|
||||||
|
2
llama.h
2
llama.h
@ -708,7 +708,7 @@ extern "C" {
|
|||||||
|
|
||||||
/// Apply chat template. Inspired by hf apply_chat_template() on python.
|
/// Apply chat template. Inspired by hf apply_chat_template() on python.
|
||||||
/// Both "model" and "custom_template" are optional, but at least one is required. "custom_template" has higher precedence than "model"
|
/// Both "model" and "custom_template" are optional, but at least one is required. "custom_template" has higher precedence than "model"
|
||||||
/// NOTE: This function only support some known jinja templates. It is not a jinja parser.
|
/// NOTE: This function does not use a jinja parser. It only support a pre-defined list of template. See more: https://github.com/ggerganov/llama.cpp/wiki/Templates-supported-by-llama_chat_apply_template
|
||||||
/// @param tmpl A Jinja template to use for this chat. If this is nullptr, the model’s default chat template will be used instead.
|
/// @param tmpl A Jinja template to use for this chat. If this is nullptr, the model’s default chat template will be used instead.
|
||||||
/// @param chat Pointer to a list of multiple llama_chat_message
|
/// @param chat Pointer to a list of multiple llama_chat_message
|
||||||
/// @param n_msg Number of llama_chat_message in this chat
|
/// @param n_msg Number of llama_chat_message in this chat
|
||||||
|
Loading…
Reference in New Issue
Block a user