.. |
CMakeLists.txt
|
Work on the BPE tokenizer (#3252)
|
2023-10-03 09:16:26 +02:00 |
test-c.c
|
tests : add a C compliance test (#2848)
|
2023-08-30 09:20:26 +03:00 |
test-double-float.cpp
|
gguf : support big endian platform (#3552)
|
2023-10-20 14:19:40 +03:00 |
test-grad0.cpp
|
sync : ggml (conv 1d + 2d updates, UB fixes) (#3468)
|
2023-10-04 15:29:58 +03:00 |
test-grammar-parser.cpp
|
gguf : new file format with flexible meta data (beta) (#2398)
|
2023-08-21 23:07:43 +03:00 |
test-llama-grammar.cpp
|
gguf : new file format with flexible meta data (beta) (#2398)
|
2023-08-21 23:07:43 +03:00 |
test-opt.cpp
|
sync : ggml (conv 1d + 2d updates, UB fixes) (#3468)
|
2023-10-04 15:29:58 +03:00 |
test-quantize-fns.cpp
|
check C++ code with -Wmissing-declarations (#3184)
|
2023-09-15 15:38:27 -04:00 |
test-quantize-perf.cpp
|
sync : ggml (conv 1d + 2d updates, UB fixes) (#3468)
|
2023-10-04 15:29:58 +03:00 |
test-rope.cpp
|
llama : custom attention mask + parallel decoding + no context swaps (#3228)
|
2023-09-28 19:04:36 +03:00 |
test-sampling.cpp
|
llama : combine repetition, frequency and presence penalties in 1 call
|
2023-10-20 17:05:46 +03:00 |
test-tokenizer-0-falcon.cpp
|
Minor improvements in GPT2 tokenizer (#3567)
|
2023-10-10 18:59:52 +02:00 |
test-tokenizer-0-falcon.py
|
Minor improvements in GPT2 tokenizer (#3567)
|
2023-10-10 18:59:52 +02:00 |
test-tokenizer-0-llama.cpp
|
Minor improvements in GPT2 tokenizer (#3567)
|
2023-10-10 18:59:52 +02:00 |
test-tokenizer-0-llama.py
|
Minor improvements in GPT2 tokenizer (#3567)
|
2023-10-10 18:59:52 +02:00 |
test-tokenizer-1-bpe.cpp
|
Work on the BPE tokenizer (#3252)
|
2023-10-03 09:16:26 +02:00 |
test-tokenizer-1-llama.cpp
|
Work on the BPE tokenizer (#3252)
|
2023-10-03 09:16:26 +02:00 |