llama.cpp/ggml
2025-05-11 16:09:33 +02:00
..
cmake scripts : update sync + fix cmake merge 2025-03-27 10:09:29 +02:00
include Add --no-op-offload to improve -ot pp perf in MoE models like llama4 400B (#13386) 2025-05-11 14:18:39 +02:00
src CUDA: fix crash with partial offloading of MoE (#13439) 2025-05-11 16:09:33 +02:00
.gitignore vulkan : cmake integration (#8119) 2024-07-13 18:12:39 +02:00
CMakeLists.txt whisper: remove MSVC warnings pragmas (whisper/3090) 2025-05-07 17:28:36 +03:00