mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-11-13 14:29:52 +00:00
ggml : add GGML_DEFAULT_N_THREADS
This commit is contained in:
parent
d990e3fffc
commit
a3a2a0eda8
4
ggml.c
4
ggml.c
@ -9363,7 +9363,7 @@ struct ggml_cgraph ggml_build_forward(struct ggml_tensor * tensor) {
|
|||||||
struct ggml_cgraph result = {
|
struct ggml_cgraph result = {
|
||||||
/*.n_nodes =*/ 0,
|
/*.n_nodes =*/ 0,
|
||||||
/*.n_leafs =*/ 0,
|
/*.n_leafs =*/ 0,
|
||||||
/*.n_threads =*/ 0,
|
/*.n_threads =*/ GGML_DEFAULT_N_THREADS,
|
||||||
/*.work_size =*/ 0,
|
/*.work_size =*/ 0,
|
||||||
/*.work =*/ NULL,
|
/*.work =*/ NULL,
|
||||||
/*.nodes =*/ { NULL },
|
/*.nodes =*/ { NULL },
|
||||||
@ -9984,7 +9984,7 @@ void ggml_graph_print(const struct ggml_cgraph * cgraph) {
|
|||||||
GGML_PRINT("=== GRAPH ===\n");
|
GGML_PRINT("=== GRAPH ===\n");
|
||||||
|
|
||||||
GGML_PRINT_DEBUG("n_threads = %d\n", cgraph->n_threads);
|
GGML_PRINT_DEBUG("n_threads = %d\n", cgraph->n_threads);
|
||||||
GGML_PRINT_DEBUG("total work size = %zu bytes\n",cgraph->work_size);
|
GGML_PRINT_DEBUG("total work size = %zu bytes\n", cgraph->work_size);
|
||||||
|
|
||||||
GGML_PRINT("n_nodes = %d\n", cgraph->n_nodes);
|
GGML_PRINT("n_nodes = %d\n", cgraph->n_nodes);
|
||||||
for (int i = 0; i < cgraph->n_nodes; i++) {
|
for (int i = 0; i < cgraph->n_nodes; i++) {
|
||||||
|
1
ggml.h
1
ggml.h
@ -182,6 +182,7 @@ extern "C" {
|
|||||||
#define GGML_MAX_PARAMS 16
|
#define GGML_MAX_PARAMS 16
|
||||||
#define GGML_MAX_CONTEXTS 64
|
#define GGML_MAX_CONTEXTS 64
|
||||||
#define GGML_MAX_OPT 4
|
#define GGML_MAX_OPT 4
|
||||||
|
#define GGML_DEFAULT_N_THREADS 4
|
||||||
|
|
||||||
#ifdef __ARM_NEON
|
#ifdef __ARM_NEON
|
||||||
// we use the built-in 16-bit float type
|
// we use the built-in 16-bit float type
|
||||||
|
Loading…
Reference in New Issue
Block a user