mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-26 03:14:35 +00:00
llama : fix op mul check with command-r-plus (#10476)
This commit is contained in:
parent
9336db462c
commit
dc39012cba
@ -7181,12 +7181,12 @@ static bool weight_buft_supported(const llama_hparams & hparams, ggml_tensor * w
|
|||||||
} break;
|
} break;
|
||||||
case GGML_OP_ADD:
|
case GGML_OP_ADD:
|
||||||
{
|
{
|
||||||
ggml_tensor * a = ggml_new_tensor_2d(ctx, GGML_TYPE_F32, w->ne[0], 512);
|
ggml_tensor * a = ggml_new_tensor_4d(ctx, GGML_TYPE_F32, w->ne[0], w->ne[1], w->ne[2], w->ne[3]);
|
||||||
op_tensor = ggml_add(ctx, a, w);
|
op_tensor = ggml_add(ctx, a, w);
|
||||||
} break;
|
} break;
|
||||||
case GGML_OP_MUL:
|
case GGML_OP_MUL:
|
||||||
{
|
{
|
||||||
ggml_tensor * a = ggml_new_tensor_2d(ctx, GGML_TYPE_F32, w->ne[0], 512);
|
ggml_tensor * a = ggml_new_tensor_4d(ctx, GGML_TYPE_F32, w->ne[0], w->ne[1], w->ne[2], w->ne[3]);
|
||||||
op_tensor = ggml_mul(ctx, a, w);
|
op_tensor = ggml_mul(ctx, a, w);
|
||||||
} break;
|
} break;
|
||||||
case GGML_OP_DIV:
|
case GGML_OP_DIV:
|
||||||
|
Loading…
Reference in New Issue
Block a user