mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 03:44:35 +00:00
server : bug fix in ingest_images
n_tokens is incremented internally by llama_batch_add
This commit is contained in:
parent
a8c981b734
commit
3d5929e8ee
@ -963,7 +963,6 @@ struct llama_server_context
|
||||
{
|
||||
llama_batch_add(batch, append_tokens[i], slot.n_past, { slot.id }, true);
|
||||
slot.n_past += 1;
|
||||
batch.n_tokens += 1;
|
||||
}
|
||||
}
|
||||
return true;
|
||||
|
Loading…
Reference in New Issue
Block a user