mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-28 04:14:35 +00:00
6262d13e0b
Some checks are pending
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/full-cuda.Dockerfile platforms:linux/amd64 tag:full-cuda]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/full.Dockerfile platforms:linux/amd64,linux/arm64 tag:full]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-cli-cuda.Dockerfile platforms:linux/amd64 tag:light-cuda]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-cli-intel.Dockerfile platforms:linux/amd64 tag:light-intel]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-cli.Dockerfile platforms:linux/amd64,linux/arm64 tag:light]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-server-cuda.Dockerfile platforms:linux/amd64 tag:server-cuda]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-server-intel.Dockerfile platforms:linux/amd64 tag:server-intel]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-server.Dockerfile platforms:linux/amd64,linux/arm64 tag:server]) (push) Waiting to run
Nix CI / nix-eval (macos-latest) (push) Waiting to run
Nix CI / nix-eval (ubuntu-latest) (push) Waiting to run
Nix CI / nix-build (macos-latest) (push) Waiting to run
Nix CI / nix-build (ubuntu-latest) (push) Waiting to run
flake8 Lint / Lint (push) Waiting to run
Python Type-Check / pyright type-check (push) Waiting to run
https://github.com/ggerganov/llama.cpp/pull/9418
132 lines
5.1 KiB
C++
132 lines
5.1 KiB
C++
#include "arg.h"
|
|
#include "common.h"
|
|
|
|
#include <string>
|
|
#include <vector>
|
|
#include <sstream>
|
|
#include <unordered_set>
|
|
|
|
#undef NDEBUG
|
|
#include <cassert>
|
|
|
|
int main(void) {
|
|
gpt_params params;
|
|
|
|
printf("test-arg-parser: make sure there is no duplicated arguments in any examples\n\n");
|
|
for (int ex = 0; ex < LLAMA_EXAMPLE_COUNT; ex++) {
|
|
try {
|
|
auto ctx_arg = gpt_params_parser_init(params, (enum llama_example)ex);
|
|
std::unordered_set<std::string> seen_args;
|
|
std::unordered_set<std::string> seen_env_vars;
|
|
for (const auto & opt : ctx_arg.options) {
|
|
// check for args duplications
|
|
for (const auto & arg : opt.args) {
|
|
if (seen_args.find(arg) == seen_args.end()) {
|
|
seen_args.insert(arg);
|
|
} else {
|
|
fprintf(stderr, "test-arg-parser: found different handlers for the same argument: %s", arg);
|
|
exit(1);
|
|
}
|
|
}
|
|
// check for env var duplications
|
|
if (opt.env) {
|
|
if (seen_env_vars.find(opt.env) == seen_env_vars.end()) {
|
|
seen_env_vars.insert(opt.env);
|
|
} else {
|
|
fprintf(stderr, "test-arg-parser: found different handlers for the same env var: %s", opt.env);
|
|
exit(1);
|
|
}
|
|
}
|
|
}
|
|
} catch (std::exception & e) {
|
|
printf("%s\n", e.what());
|
|
assert(false);
|
|
}
|
|
}
|
|
|
|
auto list_str_to_char = [](std::vector<std::string> & argv) -> std::vector<char *> {
|
|
std::vector<char *> res;
|
|
for (auto & arg : argv) {
|
|
res.push_back(const_cast<char *>(arg.data()));
|
|
}
|
|
return res;
|
|
};
|
|
|
|
std::vector<std::string> argv;
|
|
|
|
printf("test-arg-parser: test invalid usage\n\n");
|
|
|
|
// missing value
|
|
argv = {"binary_name", "-m"};
|
|
assert(false == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
|
|
// wrong value (int)
|
|
argv = {"binary_name", "-ngl", "hello"};
|
|
assert(false == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
|
|
// wrong value (enum)
|
|
argv = {"binary_name", "-sm", "hello"};
|
|
assert(false == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
|
|
// non-existence arg in specific example (--draft cannot be used outside llama-speculative)
|
|
argv = {"binary_name", "--draft", "123"};
|
|
assert(false == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_SERVER));
|
|
|
|
|
|
printf("test-arg-parser: test valid usage\n\n");
|
|
|
|
argv = {"binary_name", "-m", "model_file.gguf"};
|
|
assert(true == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
assert(params.model == "model_file.gguf");
|
|
|
|
argv = {"binary_name", "-t", "1234"};
|
|
assert(true == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
assert(params.cpuparams.n_threads == 1234);
|
|
|
|
argv = {"binary_name", "--verbose"};
|
|
assert(true == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
assert(params.verbosity > 1);
|
|
|
|
argv = {"binary_name", "-m", "abc.gguf", "--predict", "6789", "--batch-size", "9090"};
|
|
assert(true == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
assert(params.model == "abc.gguf");
|
|
assert(params.n_predict == 6789);
|
|
assert(params.n_batch == 9090);
|
|
|
|
// --draft cannot be used outside llama-speculative
|
|
argv = {"binary_name", "--draft", "123"};
|
|
assert(true == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_SPECULATIVE));
|
|
assert(params.n_draft == 123);
|
|
|
|
// skip this part on windows, because setenv is not supported
|
|
#ifdef _WIN32
|
|
printf("test-arg-parser: skip on windows build\n");
|
|
#else
|
|
printf("test-arg-parser: test environment variables (valid + invalid usages)\n\n");
|
|
|
|
setenv("LLAMA_ARG_THREADS", "blah", true);
|
|
argv = {"binary_name"};
|
|
assert(false == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
|
|
setenv("LLAMA_ARG_MODEL", "blah.gguf", true);
|
|
setenv("LLAMA_ARG_THREADS", "1010", true);
|
|
argv = {"binary_name"};
|
|
assert(true == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
assert(params.model == "blah.gguf");
|
|
assert(params.cpuparams.n_threads == 1010);
|
|
|
|
|
|
printf("test-arg-parser: test environment variables being overwritten\n\n");
|
|
|
|
setenv("LLAMA_ARG_MODEL", "blah.gguf", true);
|
|
setenv("LLAMA_ARG_THREADS", "1010", true);
|
|
argv = {"binary_name", "-m", "overwritten.gguf"};
|
|
assert(true == gpt_params_parse(argv.size(), list_str_to_char(argv).data(), params, LLAMA_EXAMPLE_COMMON));
|
|
assert(params.model == "overwritten.gguf");
|
|
assert(params.cpuparams.n_threads == 1010);
|
|
#endif // _WIN32
|
|
|
|
|
|
printf("test-arg-parser: all tests OK\n\n");
|
|
}
|