llama.cpp/tools
Đinh Trọng Huy e0e3aa231d
llama : add support for BertForSequenceClassification reranker (#13858)
* convert: add support for BertForSequenceClassification

* add support for reranking using BertForSequenceClassification

* merge checks of eos and sep

* fix lint

---------

Co-authored-by: dinhhuy <huy.dinh@brains-tech.co.jp>
2025-05-28 19:01:58 +02:00
..
batched-bench batched-bench : fix pp batch contents (#13492) 2025-05-13 18:01:53 +03:00
cvector-generator llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
export-lora llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
gguf-split llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
imatrix imatrix : Add --parse-special for enabling parsing of special tokens in imatrix calculation (#13389) 2025-05-09 11:53:58 +02:00
llama-bench kv-cache : add SWA support (#13194) 2025-05-20 08:05:46 +03:00
main llama : do not crash if there is no CPU backend (#13395) 2025-05-09 13:02:07 +02:00
mtmd mtmd : support Qwen 2.5 Omni (input audio+vision, no audio output) (#13784) 2025-05-27 14:06:10 +02:00
perplexity context : remove logits_all flag (#13284) 2025-05-08 14:26:50 +03:00
quantize quantize : improve tensor-type pattern matching (#13033) 2025-05-13 19:12:31 +02:00
rpc rpc : Fix build on OpenBSD (#13541) 2025-05-25 15:35:53 +03:00
run kv-cache : simplify the interface (#13660) 2025-05-21 15:11:13 +03:00
server llama : add support for BertForSequenceClassification reranker (#13858) 2025-05-28 19:01:58 +02:00
tokenize llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
tts tts : fix n_ubatch + make WavTokenizer cache-less (#13713) 2025-05-22 22:21:07 +03:00
CMakeLists.txt mtmd : rename llava directory to mtmd (#13311) 2025-05-05 16:02:55 +02:00