From 8be5073c51b2c3341151e0e8645deda06135c5c1 Mon Sep 17 00:00:00 2001 From: Martin Garton Date: Wed, 6 Aug 2025 18:02:38 +0100 Subject: [PATCH] Fix typo (#223) [skip ci] Fix typo `lama-swap` -> `llama-swap` --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 3338634..a43288e 100644 --- a/README.md +++ b/README.md @@ -36,7 +36,7 @@ Written in golang, it is very easy to install (single binary with no dependencie ## How does llama-swap work? -When a request is made to an OpenAI compatible endpoint, lama-swap will extract the `model` value and load the appropriate server configuration to serve it. If the wrong upstream server is running, it will be replaced with the correct one. This is where the "swap" part comes in. The upstream server is automatically swapped to the correct one to serve the request. +When a request is made to an OpenAI compatible endpoint, llama-swap will extract the `model` value and load the appropriate server configuration to serve it. If the wrong upstream server is running, it will be replaced with the correct one. This is where the "swap" part comes in. The upstream server is automatically swapped to the correct one to serve the request. In the most basic configuration llama-swap handles one model at a time. For more advanced use cases, the `groups` feature allows multiple models to be loaded at the same time. You have complete control over how your system resources are used.