From 4d02ccd26a4a913b640c0cb548dc9fc884f5f442 Mon Sep 17 00:00:00 2001 From: Benson Wong Date: Fri, 30 May 2025 09:38:45 -0700 Subject: [PATCH] Update README.md [skip ci] --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 870a23e..791c6c4 100644 --- a/README.md +++ b/README.md @@ -76,7 +76,7 @@ Check the [configuration documentation](https://github.com/mostlygeek/llama-swap Docker is the quickest way to try out llama-swap: ```shell -# use CPU inference +# use CPU inference comes with the example config above $ docker run -it --rm -p 9292:8080 ghcr.io/mostlygeek/llama-swap:cpu # qwen2.5 0.5B @@ -95,7 +95,7 @@ $ curl -s http://localhost:9292/v1/chat/completions \ ```
-Docker images are built nightly ... +Docker images are built nightly for cuda, intel, vulcan, etc ... They include: