Files
llama.cpp/common
Radoslav Gerganov bcf7546160 server : add arg for disabling prompt caching (#18776)
* server : add arg for disabling prompt caching

Disabling prompt caching is useful for clients who are restricted to
sending only OpenAI-compat requests and want deterministic
responses.

* address review comments

* address review comments
2026-01-12 19:21:34 +02:00
..
2026-01-01 18:01:43 +01:00
2025-12-10 15:28:59 +01:00
2025-12-10 15:28:59 +01:00
2025-05-30 16:25:45 +03:00
2025-12-10 15:28:59 +01:00
2025-12-10 15:28:59 +01:00