mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-24 18:34:36 +00:00
server: Propagate standby_timeout after it has been initialized
This commit is contained in:
parent
4fd58a8013
commit
7006dd784c
@ -1430,10 +1430,6 @@ struct server_context {
|
||||
// Necessary similarity of prompt for slot selection
|
||||
float slot_prompt_similarity = 0.0f;
|
||||
|
||||
server_context() {
|
||||
queue_tasks.standby_timeout = params_base.standby_timeout;
|
||||
}
|
||||
|
||||
~server_context() {
|
||||
if (ctx) {
|
||||
llama_free(ctx);
|
||||
@ -1485,6 +1481,8 @@ struct server_context {
|
||||
|
||||
n_ctx = llama_n_ctx(ctx);
|
||||
|
||||
queue_tasks.standby_timeout = params.standby_timeout;
|
||||
|
||||
add_bos_token = llama_add_bos_token(model);
|
||||
has_eos_token = !llama_add_eos_token(model);
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user