mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-28 12:24:35 +00:00
b782e5c7d4
* server : add split model test * add test speculative * add invalid cases
49 lines
1.2 KiB
Python
49 lines
1.2 KiB
Python
import pytest
|
|
from utils import *
|
|
|
|
server = ServerPreset.tinyllama2()
|
|
|
|
|
|
@pytest.fixture(scope="module", autouse=True)
|
|
def create_server():
|
|
global server
|
|
server = ServerPreset.tinyllama2()
|
|
|
|
|
|
def test_server_start_simple():
|
|
global server
|
|
server.start()
|
|
res = server.make_request("GET", "/health")
|
|
assert res.status_code == 200
|
|
|
|
|
|
def test_server_props():
|
|
global server
|
|
server.start()
|
|
res = server.make_request("GET", "/props")
|
|
assert res.status_code == 200
|
|
assert res.body["total_slots"] == server.n_slots
|
|
|
|
|
|
def test_server_models():
|
|
global server
|
|
server.start()
|
|
res = server.make_request("GET", "/models")
|
|
assert res.status_code == 200
|
|
assert len(res.body["data"]) == 1
|
|
assert res.body["data"][0]["id"] == server.model_alias
|
|
|
|
def test_load_split_model():
|
|
global server
|
|
server.model_hf_repo = "ggml-org/models"
|
|
server.model_hf_file = "tinyllamas/split/stories15M-q8_0-00001-of-00003.gguf"
|
|
server.model_alias = "tinyllama-split"
|
|
server.start()
|
|
res = server.make_request("POST", "/completion", data={
|
|
"n_predict": 16,
|
|
"prompt": "Hello",
|
|
"temperature": 0.0,
|
|
})
|
|
assert res.status_code == 200
|
|
assert match_regex("(little|girl)+", res.body["content"])
|