llama : fix op mul check with command-r-plus (#10476)

This commit is contained in:
Diego Devesa 2024-11-24 16:10:26 +01:00 committed by GitHub
parent 9336db462c
commit dc39012cba
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -7181,12 +7181,12 @@ static bool weight_buft_supported(const llama_hparams & hparams, ggml_tensor * w
} break; } break;
case GGML_OP_ADD: case GGML_OP_ADD:
{ {
ggml_tensor * a = ggml_new_tensor_2d(ctx, GGML_TYPE_F32, w->ne[0], 512); ggml_tensor * a = ggml_new_tensor_4d(ctx, GGML_TYPE_F32, w->ne[0], w->ne[1], w->ne[2], w->ne[3]);
op_tensor = ggml_add(ctx, a, w); op_tensor = ggml_add(ctx, a, w);
} break; } break;
case GGML_OP_MUL: case GGML_OP_MUL:
{ {
ggml_tensor * a = ggml_new_tensor_2d(ctx, GGML_TYPE_F32, w->ne[0], 512); ggml_tensor * a = ggml_new_tensor_4d(ctx, GGML_TYPE_F32, w->ne[0], w->ne[1], w->ne[2], w->ne[3]);
op_tensor = ggml_mul(ctx, a, w); op_tensor = ggml_mul(ctx, a, w);
} break; } break;
case GGML_OP_DIV: case GGML_OP_DIV: