llama.cpp/tests
Jeff Bolz a813badbbd
vulkan: im2col and matmul optimizations for stable diffusion (#10942)
* tests: Add im2col perf tests

* vulkan: optimize im2col, more elements per thread

* vulkan: increase small tile size for NV_coopmat2

* vulkan: change im2col to 512 elements per workgroup
2024-12-29 10:16:34 +01:00
..
.gitignore
CMakeLists.txt tests: add tests for GGUF (#10830) 2024-12-17 19:09:35 +01:00
get-model.cpp
get-model.h
run-json-schema-to-grammar.mjs
test-arg-parser.cpp
test-autorelease.cpp
test-backend-ops.cpp vulkan: im2col and matmul optimizations for stable diffusion (#10942) 2024-12-29 10:16:34 +01:00
test-barrier.cpp
test-c.c
test-chat-template.cpp llama : support InfiniAI Megrez 3b (#10893) 2024-12-23 01:35:44 +01:00
test-double-float.cpp
test-gguf.cpp tests: disable GGUF test for bad value size (#10886) 2024-12-19 08:53:58 +01:00
test-grammar-integration.cpp llama : minor grammar refactor (#10897) 2024-12-19 17:42:13 +02:00
test-grammar-parser.cpp
test-json-schema-to-grammar.cpp
test-llama-grammar.cpp llama : minor grammar refactor (#10897) 2024-12-19 17:42:13 +02:00
test-log.cpp
test-lora-conversion-inference.sh Fix HF repo commit to clone lora test models (#10649) 2024-12-04 10:45:48 +01:00
test-model-load-cancel.cpp
test-opt.cpp
test-quantize-fns.cpp
test-quantize-perf.cpp
test-rope.cpp llama : add Qwen2VL support + multimodal RoPE (#10361) 2024-12-14 14:43:46 +02:00
test-sampling.cpp sampling : refactor + optimize penalties sampler (#10803) 2024-12-16 12:31:14 +02:00
test-tokenizer-0.cpp
test-tokenizer-0.py
test-tokenizer-0.sh
test-tokenizer-1-bpe.cpp
test-tokenizer-1-spm.cpp
test-tokenizer-random.py