mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-25 10:54:36 +00:00
parent
3df784b305
commit
c2a16c0bdb
@ -62,6 +62,10 @@ struct common_speculative * common_speculative_init(
|
|||||||
}
|
}
|
||||||
|
|
||||||
void common_speculative_free(struct common_speculative * spec) {
|
void common_speculative_free(struct common_speculative * spec) {
|
||||||
|
if (spec == nullptr) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
common_sampler_free(spec->smpl);
|
common_sampler_free(spec->smpl);
|
||||||
|
|
||||||
llama_batch_free(spec->batch);
|
llama_batch_free(spec->batch);
|
||||||
|
@ -720,7 +720,7 @@ struct server_slot {
|
|||||||
int id;
|
int id;
|
||||||
int id_task = -1;
|
int id_task = -1;
|
||||||
|
|
||||||
llama_batch batch_spec;
|
llama_batch batch_spec = {};
|
||||||
|
|
||||||
llama_context * ctx_dft = nullptr;
|
llama_context * ctx_dft = nullptr;
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user