Dechunk HTTP requests by default (#11)

ProxyManager already has all the Request body's data. There is no never
a need to use chunked transfer encoding to the upstream process.
This commit is contained in:
Benson Wong
2024-11-19 09:40:44 -08:00
parent 5021e0f299
commit 7eec51f3f2
3 changed files with 10 additions and 4 deletions

View File

@@ -7,7 +7,7 @@ models:
cmd: >
models/llama-server-osx
--port 8999
-m models/Llama-3.2-1B-Instruct-Q4_K_M.gguf
-m models/Llama-3.2-1B-Instruct-Q4_0.gguf
proxy: http://127.0.0.1:8999
# list of model name aliases this llama.cpp instance can serve
@@ -18,7 +18,7 @@ models:
checkEndpoint: /health
"qwen":
cmd: models/llama-server-osx --port 8999 -m models/Qwen2.5-1.5B-Instruct-Q4_K_M.gguf
cmd: models/llama-server-osx --port 8999 -m models/qwen2.5-0.5b-instruct-q8_0.gguf
proxy: http://127.0.0.1:8999
aliases:
- gpt-3.5-turbo
@@ -40,5 +40,5 @@ models:
cmd: models/llama-server-osx --port 8999 -m models/doesnotexist.gguf
proxy: http://127.0.0.1:8999
"broken_timeout":
cmd: models/llama-server-osx --port 8999 -m models/Qwen2.5-1.5B-Instruct-Q4_K_M.gguf
cmd: models/llama-server-osx --port 8999 -m models/qwen2.5-0.5b-instruct-q8_0.gguf
proxy: http://127.0.0.1:9000