mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-25 02:44:36 +00:00
62cfc54f77
Command that calculates some statistics over the errors introduced by quantization, like mean square error, max error and some percentile errors for layer weights. Should be useful for testing quantization improvements. Exposes some internal state from ggml and llama for testing
40 lines
351 B
Plaintext
40 lines
351 B
Plaintext
*.o
|
|
*.a
|
|
.cache/
|
|
.vs/
|
|
.vscode/
|
|
.DS_Store
|
|
|
|
.build/
|
|
build/
|
|
build-em/
|
|
build-debug/
|
|
build-release/
|
|
build-static/
|
|
build-no-accel/
|
|
build-sanitize-addr/
|
|
build-sanitize-thread/
|
|
|
|
models/*
|
|
|
|
/main
|
|
/quantize
|
|
/quantize-stats
|
|
/result
|
|
/perplexity
|
|
/embedding
|
|
/Pipfile
|
|
|
|
arm_neon.h
|
|
compile_commands.json
|
|
|
|
.envrc
|
|
.direnv/
|
|
|
|
.venv
|
|
__pycache__
|
|
.swiftpm
|
|
|
|
zig-out/
|
|
zig-cache/
|