mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 03:44:35 +00:00
ggml : skip nops in compute_forward
This commit is contained in:
parent
6a30bf3e51
commit
8d8d54f834
4
ggml.c
4
ggml.c
@ -16602,6 +16602,10 @@ static void ggml_compute_forward_cross_entropy_loss_back(
|
||||
static void ggml_compute_forward(struct ggml_compute_params * params, struct ggml_tensor * tensor) {
|
||||
GGML_ASSERT(params);
|
||||
|
||||
if (tensor->op == GGML_OP_NONE) {
|
||||
return;
|
||||
}
|
||||
|
||||
#ifdef GGML_USE_CUBLAS
|
||||
bool skip_cpu = ggml_cuda_compute_forward(params, tensor);
|
||||
if (skip_cpu) {
|
||||
|
Loading…
Reference in New Issue
Block a user