mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-08 09:41:45 +00:00
ggml: Add POOL2D OP for GPU acceleration to the Vulkan backend in the MobileVLM model. (#9763)
close-issue.yml #801:Scheduled
llama : switch KQ multiplication to F32 precision by default (#10015)
nix-ci-aarch64.yml #777:Scheduled
llama : switch KQ multiplication to F32 precision by default (#10015)
close-issue.yml #776:Scheduled