mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-02 14:54:35 +00:00
Delete TODO now that we have q8_0.
This commit is contained in:
parent
8564f79036
commit
09d83f0401
@ -6510,7 +6510,6 @@ struct llama_context * llama_new_context_with_model(
|
||||
#undef LLAMA_METAL_CHECK_BUF
|
||||
}
|
||||
#elif defined(GGML_USE_KOMPUTE)
|
||||
// TODO(cebtenzzre): we need to check the type of each tensor because Q8_0 is not currently supported
|
||||
if (ggml_vk_has_device() && params.n_gpu_layers > 0
|
||||
&& (model->arch == LLM_ARCH_LLAMA || model->arch == LLM_ARCH_FALCON)
|
||||
&& (model->ftype == LLAMA_FTYPE_ALL_F32
|
||||
|
Loading…
Reference in New Issue
Block a user