llama.cpp/gguf-py/gguf
Xuan-Son Nguyen 5215b91e93
clip : fix confused naming ffn_up and ffn_down (#13290)
* clip :  fix confused naming ffn_up and ffn_down

* rm ffn_i/o/g naming

* rename n_embd, n_ff

* small fix

* no check n_ff
2025-05-05 12:54:44 +02:00
..
scripts gguf-py : GGUF Editor GUI - Python + Qt6 (#12930) 2025-04-18 20:30:41 +02:00
__init__.py
constants.py llama-model : support Qwen2 embedding models and pooling_mode_lasttoken (#13245) 2025-05-02 11:42:30 -04:00
gguf.py
gguf_reader.py
gguf_writer.py convert : converting mmproj for Qwen2/2.5VL from convert_hf_to_gguf (#13209) 2025-05-02 17:17:15 +02:00
lazy.py gguf-py : support lazy tensor splitting (#12809) 2025-04-08 09:03:07 +02:00
metadata.py convert : fix Norway problem when parsing YAML (#12114) 2025-02-28 17:44:46 +01:00
py.typed
quants.py
tensor_mapping.py clip : fix confused naming ffn_up and ffn_down (#13290) 2025-05-05 12:54:44 +02:00
utility.py convert : ability to lazy-load safetensors remotely without downloading to disk (#12820) 2025-04-10 17:24:44 +02:00
vocab.py convert : Support chat_template.json (#12460) 2025-03-19 08:58:13 +01:00