llama.cpp/ggml/src
Sigbjørn Skjæret b72c20b85c
Fix conversion of unnormalized BF16->BF16 weights (#7843)
* add truncate_bf16

* truncate intermediate fp32 if converting bf16 to bf16

* fix masking in __compute_fp32_to_bf16

* np.int16 no longer used

* missing cast and additional numpy 2.x fix

* ggml-impl : do not flush bf16 subnormals to zero

* ggml : add reference fp32 to bf16 conversion

The fast version is no longer equivalent for all platforms
because of the handling of subnormal values.

* gguf-py : remove flush to zero for bf16 subnormals

* gguf-py : remove float32 truncation to bf16

Rounding achieves the same thing in the cases where this was used.

* missed prototype update in merge

* merge cleanup

---------

Co-authored-by: Francis Couture-Harpin <git@compilade.net>
2024-08-02 15:11:39 -04:00
..
ggml-cann cann: Fix ggml_cann_im2col for 1D im2col (#8819) 2024-08-02 16:50:53 +08:00
ggml-cuda cuda : fix dmmv cols requirement to 2*GGML_CUDA_DMMV_X (#8800) 2024-08-01 15:26:22 +02:00
ggml-sycl [SYCL] Fixing wrong VDR iq4nl value (#8812) 2024-08-02 08:55:17 +08:00
kompute@4565194ed7 llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
kompute-shaders llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
llamafile ggml : move sgemm sources to llamafile subfolder (#8394) 2024-07-10 15:23:29 +03:00
vulkan-shaders chore : Fix vulkan related compiler warnings, add help text, improve CLI options (#8477) 2024-07-28 09:52:42 +02:00
CMakeLists.txt cann: update cmake (#8765) 2024-07-30 12:37:35 +02:00
ggml-aarch64.c ggml : fix build on Windows with Snapdragon X (#8531) 2024-07-25 19:01:00 +03:00
ggml-aarch64.h ggml : minor naming changes (#8433) 2024-07-12 10:46:02 +03:00
ggml-alloc.c ggml : reduce hash table reset cost (#8698) 2024-07-27 04:41:55 +02:00
ggml-backend-impl.h llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-backend.c ggml : reduce hash table reset cost (#8698) 2024-07-27 04:41:55 +02:00
ggml-blas.cpp ggml : reduce hash table reset cost (#8698) 2024-07-27 04:41:55 +02:00
ggml-cann.cpp cann: Fix Multi-NPU execution error (#8710) 2024-07-27 16:36:44 +08:00
ggml-common.h feat: Support Moore Threads GPU (#8383) 2024-07-28 01:41:25 +02:00
ggml-cuda.cu ggml-cuda: Adding support for unified memory (#8035) 2024-08-01 23:28:28 +02:00
ggml-impl.h Fix conversion of unnormalized BF16->BF16 weights (#7843) 2024-08-02 15:11:39 -04:00
ggml-kompute.cpp ggml : reduce hash table reset cost (#8698) 2024-07-27 04:41:55 +02:00
ggml-metal.m ggml : reduce hash table reset cost (#8698) 2024-07-27 04:41:55 +02:00
ggml-metal.metal ggml : fix quant dot product with odd number of blocks (#8549) 2024-07-19 17:17:27 +02:00
ggml-quants.c ggml: bugfix: fix the inactive elements is agnostic for risc-v vector (#8748) 2024-07-29 18:38:34 +02:00
ggml-quants.h ggml : minor naming changes (#8433) 2024-07-12 10:46:02 +03:00
ggml-rpc.cpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-sycl.cpp [SYCL] Add TIMESTEP_EMBEDDING OP (#8707) 2024-07-30 14:56:51 +08:00
ggml-vulkan.cpp vulkan : initialize vk_buffer_struct members to VK_NULL_HANDLE (ggml/893) 2024-07-27 17:43:44 +03:00
ggml.c Fix conversion of unnormalized BF16->BF16 weights (#7843) 2024-08-02 15:11:39 -04:00