llama.cpp/examples/batched.swift
Georgi Gerganov 745aa5319b
llama : deprecate llama_kv_self_ API (#14030)
* llama : deprecate llama_kv_self_ API

ggml-ci

* llama : allow llama_memory_(nullptr)

ggml-ci

* memory : add flag for optional data clear in llama_memory_clear

ggml-ci
2025-06-06 14:11:15 +03:00
..
Sources llama : deprecate llama_kv_self_ API (#14030) 2025-06-06 14:11:15 +03:00
.gitignore examples : add batched.swift + improve CI for swift (#3562) 2023-10-11 06:14:05 -05:00
Makefile build: rename main → llama-cli, server → llama-server, llava-cli → llama-llava-cli, etc... (#7809) 2024-06-13 00:41:52 +01:00
Package.swift build: rename main → llama-cli, server → llama-server, llava-cli → llama-llava-cli, etc... (#7809) 2024-06-13 00:41:52 +01:00
README.md build: rename main → llama-cli, server → llama-server, llava-cli → llama-llava-cli, etc... (#7809) 2024-06-13 00:41:52 +01:00

This is a swift clone of examples/batched.

$ make $ ./llama-batched-swift MODEL_PATH [PROMPT] [PARALLEL]