llama.cpp/ggml
Rémy O fc1b0d0936
vulkan: initial support for IQ1_S and IQ1_M quantizations (#11528)
* vulkan: initial support for IQ1_S and IQ1_M quantizations

* vulkan: define MMV kernels for IQ1 quantizations

* devops: increase timeout of Vulkan tests again

* vulkan: simplify ifdef for init_iq_shmem
2025-02-15 09:01:40 +01:00
..
cmake cmake: add ggml find package (#11369) 2025-01-26 12:07:48 -04:00
include cleanup: fix compile warnings associated with gnu_printf (#11811) 2025-02-12 10:06:53 -04:00
src vulkan: initial support for IQ1_S and IQ1_M quantizations (#11528) 2025-02-15 09:01:40 +01:00
.gitignore vulkan : cmake integration (#8119) 2024-07-13 18:12:39 +02:00
CMakeLists.txt cmake: Add ability to pass in GGML_BUILD_NUMBER (ggml/1096) 2025-02-04 12:59:15 +02:00