mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-11-11 21:39:52 +00:00
6262d13e0b
Some checks are pending
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/full-cuda.Dockerfile platforms:linux/amd64 tag:full-cuda]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/full.Dockerfile platforms:linux/amd64,linux/arm64 tag:full]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-cli-cuda.Dockerfile platforms:linux/amd64 tag:light-cuda]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-cli-intel.Dockerfile platforms:linux/amd64 tag:light-intel]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-cli.Dockerfile platforms:linux/amd64,linux/arm64 tag:light]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-server-cuda.Dockerfile platforms:linux/amd64 tag:server-cuda]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-server-intel.Dockerfile platforms:linux/amd64 tag:server-intel]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-server.Dockerfile platforms:linux/amd64,linux/arm64 tag:server]) (push) Waiting to run
Nix CI / nix-eval (macos-latest) (push) Waiting to run
Nix CI / nix-eval (ubuntu-latest) (push) Waiting to run
Nix CI / nix-build (macos-latest) (push) Waiting to run
Nix CI / nix-build (ubuntu-latest) (push) Waiting to run
flake8 Lint / Lint (push) Waiting to run
Python Type-Check / pyright type-check (push) Waiting to run
https://github.com/ggerganov/llama.cpp/pull/9418
40 lines
1.0 KiB
C++
40 lines
1.0 KiB
C++
#include "log.h"
|
|
|
|
#include <cstdlib>
|
|
#include <thread>
|
|
|
|
int main() {
|
|
const int n_thread = 8;
|
|
|
|
std::thread threads[n_thread];
|
|
for (int i = 0; i < n_thread; i++) {
|
|
threads[i] = std::thread([i]() {
|
|
const int n_msg = 1000;
|
|
|
|
for (int j = 0; j < n_msg; j++) {
|
|
const int log_type = std::rand() % 4;
|
|
|
|
switch (log_type) {
|
|
case 0: LOG_INF("Thread %d: %d\n", i, j); break;
|
|
case 1: LOG_WRN("Thread %d: %d\n", i, j); break;
|
|
case 2: LOG_ERR("Thread %d: %d\n", i, j); break;
|
|
case 3: LOG_DBG("Thread %d: %d\n", i, j); break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
if (rand () % 10 < 5) {
|
|
gpt_log_set_timestamps(gpt_log_main(), rand() % 2);
|
|
gpt_log_set_prefix (gpt_log_main(), rand() % 2);
|
|
}
|
|
}
|
|
});
|
|
}
|
|
|
|
for (int i = 0; i < n_thread; i++) {
|
|
threads[i].join();
|
|
}
|
|
|
|
return 0;
|
|
}
|