llama.cpp/ggml/src/ggml-cpu
2024-12-17 18:35:49 +02:00
..
amx remove CMAKE_WINDOWS_EXPORT_ALL_SYMBOLS (#10797) 2024-12-12 19:02:49 +01:00
cmake ggml : build backends as libraries (#10256) 2024-11-14 18:04:35 +01:00
llamafile ggml : move AMX to the CPU backend (#10570) 2024-11-29 21:54:58 +01:00
CMakeLists.txt cmake : fix "amd64" processor string (whisper/2638) 2024-12-17 18:35:49 +02:00
cpu-feats-x86.cpp ggml : add predefined list of CPU backend variants to build (#10626) 2024-12-04 14:45:40 +01:00
ggml-cpu-aarch64.cpp ggml : disable iq4_nl interleave size 8 (#10709) 2024-12-07 18:38:15 +02:00
ggml-cpu-aarch64.h ggml : refactor online repacking (#10446) 2024-12-07 14:37:50 +02:00
ggml-cpu-hbm.cpp ggml : refactor online repacking (#10446) 2024-12-07 14:37:50 +02:00
ggml-cpu-hbm.h ggml : refactor online repacking (#10446) 2024-12-07 14:37:50 +02:00
ggml-cpu-impl.h ggml : move AMX to the CPU backend (#10570) 2024-11-29 21:54:58 +01:00
ggml-cpu-quants.c ggml : fix I8MM Q4_1 scaling factor conversion (#10562) 2024-11-29 16:25:39 +02:00
ggml-cpu-quants.h ggml : build backends as libraries (#10256) 2024-11-14 18:04:35 +01:00
ggml-cpu-traits.cpp ggml : refactor online repacking (#10446) 2024-12-07 14:37:50 +02:00
ggml-cpu-traits.h ggml : refactor online repacking (#10446) 2024-12-07 14:37:50 +02:00
ggml-cpu.c llama : add Qwen2VL support + multimodal RoPE (#10361) 2024-12-14 14:43:46 +02:00
ggml-cpu.cpp ggml : update ggml_backend_cpu_device_supports_op (#10867) 2024-12-17 18:35:42 +02:00