mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-02 14:54:35 +00:00
kompute : remove Q6_K from list of supported quant types
This commit is contained in:
parent
24a4a5956a
commit
3d850db767
@ -6498,8 +6498,7 @@ struct llama_context * llama_new_context_with_model(
|
|||||||
&& (model->ftype == LLAMA_FTYPE_ALL_F32
|
&& (model->ftype == LLAMA_FTYPE_ALL_F32
|
||||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_F16
|
|| model->ftype == LLAMA_FTYPE_MOSTLY_F16
|
||||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_0
|
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_0
|
||||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_1
|
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_1)) {
|
||||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q6_K)) {
|
|
||||||
// this allocates all Vulkan resources and memory buffers
|
// this allocates all Vulkan resources and memory buffers
|
||||||
ctx->ctx_kompute = ggml_vk_init();
|
ctx->ctx_kompute = ggml_vk_init();
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user