mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-30 21:34:36 +00:00
kompute : remove Q6_K from list of supported quant types
This commit is contained in:
parent
24a4a5956a
commit
3d850db767
@ -6498,8 +6498,7 @@ struct llama_context * llama_new_context_with_model(
|
||||
&& (model->ftype == LLAMA_FTYPE_ALL_F32
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_F16
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_0
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_1
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q6_K)) {
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_1)) {
|
||||
// this allocates all Vulkan resources and memory buffers
|
||||
ctx->ctx_kompute = ggml_vk_init();
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user