mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 03:44:35 +00:00
speculative : print draft token pieces
This commit is contained in:
parent
ebe41d49a6
commit
6c150d763e
@ -241,7 +241,7 @@ int main(int argc, char ** argv) {
|
|||||||
llama_sample_softmax(ctx_dft, &cur_p);
|
llama_sample_softmax(ctx_dft, &cur_p);
|
||||||
|
|
||||||
for (int i = 0; i < 3; ++i) {
|
for (int i = 0; i < 3; ++i) {
|
||||||
LOG(" - draft candidate %d: %d (%.3f)\n", i, cur_p.data[i].id, cur_p.data[i].p);
|
LOG(" - draft candidate %3d: %6d (%8.3f) '%s'\n", i, cur_p.data[i].id, cur_p.data[i].p, llama_token_to_piece(ctx_dft, cur_p.data[i].id).c_str());
|
||||||
}
|
}
|
||||||
|
|
||||||
// too low probability, stop drafting
|
// too low probability, stop drafting
|
||||||
|
Loading…
Reference in New Issue
Block a user