mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 03:44:35 +00:00
server : bug fix in ingest_images
n_tokens is incremented internally by llama_batch_add
This commit is contained in:
parent
a8c981b734
commit
3d5929e8ee
@ -963,7 +963,6 @@ struct llama_server_context
|
|||||||
{
|
{
|
||||||
llama_batch_add(batch, append_tokens[i], slot.n_past, { slot.id }, true);
|
llama_batch_add(batch, append_tokens[i], slot.n_past, { slot.id }, true);
|
||||||
slot.n_past += 1;
|
slot.n_past += 1;
|
||||||
batch.n_tokens += 1;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return true;
|
return true;
|
||||||
|
Loading…
Reference in New Issue
Block a user