llama.cpp/gguf-py/gguf
Xuan-Son Nguyen 64eda5deb9
convert : ability to lazy-load safetensors remotely without downloading to disk (#12820)
* gguf util : add SafetensorRemote

* fix style

* convert: add --remote option

* convert : allow using lazy remote tensors

It's a bit slow for now since everything is blocking and single-threaded.

* correct metadata.name

* small style fix

* support HF_TOKEN

* convert : use writeable buffer for remote lazy tensors

* convert : fix flake8 lint regarding lamdba assigment

* multithreaded download

* multithread: print debug

* fix style

* Revert "multithreaded download"

This reverts commit 42fc895ace385edc972ad819c76c704aeea61791.

* bring back _get_request_headers

---------

Co-authored-by: Francis Couture-Harpin <git@compilade.net>
2025-04-10 17:24:44 +02:00
..
scripts Refactor gguf scripts to improve metadata handling (#11909) 2025-02-26 08:04:48 -05:00
__init__.py convert-*.py: GGUF Naming Convention Refactor and Metadata Override Refactor (#7499) 2024-07-18 20:40:15 +10:00
constants.py llama : Support Qwen3 and Qwen3MoE (#12828) 2025-04-09 11:47:36 +02:00
gguf.py gguf-py: Refactor and allow reading/modifying existing GGUF files (#3981) 2023-11-11 08:04:50 +03:00
gguf_reader.py Refactor gguf scripts to improve metadata handling (#11909) 2025-02-26 08:04:48 -05:00
gguf_writer.py llama : Support llama 4 text-only (#12791) 2025-04-07 23:06:44 +02:00
lazy.py gguf-py : support lazy tensor splitting (#12809) 2025-04-08 09:03:07 +02:00
metadata.py convert : fix Norway problem when parsing YAML (#12114) 2025-02-28 17:44:46 +01:00
py.typed convert : various script cleanups/fixes + merges and special token handling (#2842) 2023-08-30 11:25:50 +03:00
quants.py ggml-quants : ternary packing for TriLMs and BitNet b1.58 (#8151) 2024-09-05 21:48:47 -04:00
tensor_mapping.py llama : support BailingMoE (Ling) (#12634) 2025-03-30 22:21:03 +02:00
utility.py convert : ability to lazy-load safetensors remotely without downloading to disk (#12820) 2025-04-10 17:24:44 +02:00
vocab.py convert : Support chat_template.json (#12460) 2025-03-19 08:58:13 +01:00