llama.cpp/ggml
2025-06-10 13:01:33 +01:00
..
cmake cmake: Factor out CPU architecture detection (#13883) 2025-05-29 12:50:25 +02:00
include ggml : remove ggml_graph_import and ggml_graph_export declarations (ggml/1247) 2025-06-01 13:43:57 +03:00
src Vulkan: Don't default to CPU device (like llvmpipe), even if no other device is available, to allow fallback to CPU backend (#14099) 2025-06-10 13:01:33 +01:00
.gitignore vulkan : cmake integration (#8119) 2024-07-13 18:12:39 +02:00
CMakeLists.txt ggml-cpu : split arch-specific implementations (#13892) 2025-06-09 16:47:13 +02:00