mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-11-15 07:19:53 +00:00
26 lines
498 B
C
26 lines
498 B
C
|
#pragma once
|
||
|
|
||
|
#include "ggml.h"
|
||
|
#include "ggml-backend.h"
|
||
|
|
||
|
|
||
|
#ifdef __cplusplus
|
||
|
extern "C" {
|
||
|
#endif
|
||
|
|
||
|
// buffer_type API
|
||
|
GGML_API ggml_backend_buffer_type_t ggml_backend_amx_buffer_type(void);
|
||
|
|
||
|
GGML_API bool ggml_backend_is_amx(ggml_backend_t backend);
|
||
|
|
||
|
// backend API
|
||
|
GGML_API ggml_backend_t ggml_backend_amx_init(void);
|
||
|
|
||
|
GGML_API void ggml_backend_amx_set_n_threads(ggml_backend_t backend_amx, int n_threads);
|
||
|
|
||
|
GGML_API ggml_backend_reg_t ggml_backend_amx_reg(void);
|
||
|
|
||
|
#ifdef __cplusplus
|
||
|
}
|
||
|
#endif
|