mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-24 02:14:35 +00:00
ec893798b7
* tests : verify that RoPE is "additive" * llama : replace ggml_diag_mask_inf with ggml_add (custom -inf mask) * ggml : ggml_rope now takes a vector with positions instead of n_past * metal : add rope_f16 kernel + optimize cpy kernels * llama : unified KV cache + batch inference API * llama : add new llama_decode() API that works with llama_batch * llama : add cell_max heuristic for more efficient kv_cache * llama : extend llama_kv_cache API * llama : more robust cell_max heuristic + wip shift * metal : disable concurrency optimization * llama : add llama_kv_cache_shift_seq + no more context swaps * llama : apply K-cache roping for Falcon and Baichuan * speculative : fix KV cache management * parallel : example for serving multiple users in parallel * parallel : disable hot-plug to avoid cache fragmentation * fixes : speculative KV cache + llama worst-case graph * llama : extend batch API to select which logits to output * llama : fix worst case graph build * ggml-cuda : update rope implementation for parallel decoding (#3254) * ggml-cuda : update rope implementation for parallel decoding * better solution for p0 computation * fix rope * simpler rope implementation --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * make : add parallel to build + fix static functions in llama.cpp * simple : fix token counting * parallel : various improvements * llama : fix cell_max logic + rename functions * parallel : try smaller batches when the KV cache is fragmented * parallel : fix sequence termination criteria * llama : silence errors KV cache errors * parallel : remove new line from prompt * parallel : process system prompt once + configurable paramters + llama API * parallel : remove question with short answers * parallel : count cache misses * parallel : print misses on each request * parallel : minor * llama : fix n_kv to never become 0 * parallel : rename hot-plug to continuous-batching * llama : improve llama_batch API + simplify parallel example * simple : add parallel decoding support * simple : improve comments + free batch * ggml-cuda : add rope f16, restore performance with parallel decoding (#3272) * ggml-cuda : add rope f16, restore performance * offload KQ_mask with all models * fix rope shift --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * llama : disable MPI for now ggml-ci * train : make KQ_pos memory buffer permanent via dummy scale op * ggml : revert change to ggml_cpy, add ggml_cont_Nd instead (#3275) ggml-ci * parallel : fix bug (extra BOS) + smaller token_prev array * parallel : fix cases where the input prompts can overflow the batch * parallel : add disabled experimental batch chunking in powers of two * llama : llama.h formatting + comments * simple : add README.md * llama : fix kv cache heuristic when context is less than 32 * parallel : fix crash when `-n -1` * llama : simplify returns if/else branches * metal : use mm kernels for batch size > 2 * examples : utilize new llama_get_logits_ith() * examples : add example for batched decoding * examples : do not eval prompt 2 times (close #3348) * server : clear the KV cache beyond n_past before llama_decode * server : avoid context swaps by shifting the KV cache --------- Co-authored-by: slaren <slarengh@gmail.com>
91 lines
1002 B
Plaintext
91 lines
1002 B
Plaintext
*.o
|
|
*.a
|
|
*.so
|
|
*.gguf
|
|
*.bin
|
|
*.exe
|
|
*.dll
|
|
*.log
|
|
*.gcov
|
|
*.gcno
|
|
*.gcda
|
|
*.dot
|
|
.DS_Store
|
|
.build/
|
|
.cache/
|
|
.direnv/
|
|
.envrc
|
|
.swiftpm
|
|
.venv
|
|
.clang-tidy
|
|
.vs/
|
|
.vscode/
|
|
|
|
lcov-report/
|
|
gcovr-report/
|
|
|
|
build*/
|
|
out/
|
|
tmp/
|
|
|
|
models/*
|
|
models-mnt
|
|
|
|
/Pipfile
|
|
/baby-llama
|
|
/beam-search
|
|
/benchmark-matmult
|
|
/convert-llama2c-to-ggml
|
|
/embd-input-test
|
|
/embedding
|
|
/gguf
|
|
/gguf-llama-simple
|
|
/libllama.so
|
|
/llama-bench
|
|
/main
|
|
/metal
|
|
/perplexity
|
|
/quantize
|
|
/quantize-stats
|
|
/result
|
|
/save-load-state
|
|
/server
|
|
/simple
|
|
/batched
|
|
/speculative
|
|
/parallel
|
|
/train-text-from-scratch
|
|
/vdot
|
|
build-info.h
|
|
arm_neon.h
|
|
compile_commands.json
|
|
CMakeSettings.json
|
|
|
|
__pycache__
|
|
dist
|
|
|
|
zig-out/
|
|
zig-cache/
|
|
|
|
ppl-*.txt
|
|
qnt-*.txt
|
|
perf-*.txt
|
|
|
|
examples/jeopardy/results.txt
|
|
|
|
poetry.lock
|
|
poetry.toml
|
|
|
|
# Test binaries
|
|
tests/test-grammar-parser
|
|
tests/test-llama-grammar
|
|
tests/test-double-float
|
|
tests/test-grad0
|
|
tests/test-opt
|
|
tests/test-quantize-fns
|
|
tests/test-quantize-perf
|
|
tests/test-sampling
|
|
tests/test-tokenizer-0-llama
|
|
tests/test-tokenizer-0-falcon
|
|
tests/test-tokenizer-1
|