mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-26 11:24:35 +00:00
ggml : skip nops
This commit is contained in:
parent
2471d56a2e
commit
ed9fde7a1e
4
ggml.c
4
ggml.c
@ -16602,6 +16602,10 @@ static void ggml_compute_forward_cross_entropy_loss_back(
|
|||||||
static void ggml_compute_forward(struct ggml_compute_params * params, struct ggml_tensor * tensor) {
|
static void ggml_compute_forward(struct ggml_compute_params * params, struct ggml_tensor * tensor) {
|
||||||
GGML_ASSERT(params);
|
GGML_ASSERT(params);
|
||||||
|
|
||||||
|
if (tensor->op == GGML_OP_NONE) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef GGML_USE_CUBLAS
|
#ifdef GGML_USE_CUBLAS
|
||||||
bool skip_cpu = ggml_cuda_compute_forward(params, tensor);
|
bool skip_cpu = ggml_cuda_compute_forward(params, tensor);
|
||||||
if (skip_cpu) {
|
if (skip_cpu) {
|
||||||
|
Loading…
Reference in New Issue
Block a user