mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 22:59:24 +01:00
b72c20b85c
* add truncate_bf16 * truncate intermediate fp32 if converting bf16 to bf16 * fix masking in __compute_fp32_to_bf16 * np.int16 no longer used * missing cast and additional numpy 2.x fix * ggml-impl : do not flush bf16 subnormals to zero * ggml : add reference fp32 to bf16 conversion The fast version is no longer equivalent for all platforms because of the handling of subnormal values. * gguf-py : remove flush to zero for bf16 subnormals * gguf-py : remove float32 truncation to bf16 Rounding achieves the same thing in the cases where this was used. * missed prototype update in merge * merge cleanup --------- Co-authored-by: Francis Couture-Harpin <git@compilade.net> |
||
---|---|---|
.. | ||
ggml-cann | ||
ggml-cuda | ||
ggml-sycl | ||
kompute@4565194ed7 | ||
kompute-shaders | ||
llamafile | ||
vulkan-shaders | ||
CMakeLists.txt | ||
ggml-aarch64.c | ||
ggml-aarch64.h | ||
ggml-alloc.c | ||
ggml-backend-impl.h | ||
ggml-backend.c | ||
ggml-blas.cpp | ||
ggml-cann.cpp | ||
ggml-common.h | ||
ggml-cuda.cu | ||
ggml-impl.h | ||
ggml-kompute.cpp | ||
ggml-metal.m | ||
ggml-metal.metal | ||
ggml-quants.c | ||
ggml-quants.h | ||
ggml-rpc.cpp | ||
ggml-sycl.cpp | ||
ggml-vulkan.cpp | ||
ggml.c |