llama.cpp/examples/server/tests/features
Olivier Chafik 8843a98c2b
Improve usability of --model-url & related flags (#6930)
* args: default --model to models/ + filename from --model-url or --hf-file (or else legacy models/7B/ggml-model-f16.gguf)

* args: main & server now call gpt_params_handle_model_default

* args: define DEFAULT_MODEL_PATH + update cli docs

* curl: check url of previous download (.json metadata w/ url, etag & lastModified)

* args: fix update to quantize-stats.cpp

* curl: support legacy .etag / .lastModified companion files

* curl: rm legacy .etag file support

* curl: reuse regex across headers callback calls

* curl: unique_ptr to manage lifecycle of curl & outfile

* curl: nit: no need for multiline regex flag

* curl: update failed test (model file collision) + gitignore *.gguf.json
2024-04-30 00:52:50 +01:00
..
steps Server: fix seed for multiple slots (#6835) 2024-04-24 11:08:36 +02:00
embeddings.feature Improve usability of --model-url & related flags (#6930) 2024-04-30 00:52:50 +01:00
environment.py server tests : more pythonic process management; fix bare except: (#6146) 2024-03-20 06:33:49 +01:00
issues.feature server: tests: passkey challenge / self-extend with context shift demo (#5832) 2024-03-02 22:00:14 +01:00
parallel.feature common: llama_load_model_from_url split support (#6192) 2024-03-23 18:07:00 +01:00
passkey.feature server: tests: passkey challenge / self-extend with context shift demo (#5832) 2024-03-02 22:00:14 +01:00
results.feature Server: fix seed for multiple slots (#6835) 2024-04-24 11:08:36 +02:00
security.feature json-schema-to-grammar improvements (+ added to server) (#5978) 2024-03-21 11:50:43 +00:00
server.feature common: llama_load_model_from_url split support (#6192) 2024-03-23 18:07:00 +01:00
slotsave.feature llama : save and restore kv cache for single seq id (#6341) 2024-04-08 15:43:30 +03:00
wrong_usages.feature server: tests: passkey challenge / self-extend with context shift demo (#5832) 2024-03-02 22:00:14 +01:00