support TTS /v1/audio/speech (#36)
This commit is contained in:
@@ -21,7 +21,12 @@ Any OpenAI compatible server would work. llama-swap was originally designed for
|
|||||||
- ✅ Easy to config: single yaml file
|
- ✅ Easy to config: single yaml file
|
||||||
- ✅ On-demand model switching
|
- ✅ On-demand model switching
|
||||||
- ✅ Full control over server settings per model
|
- ✅ Full control over server settings per model
|
||||||
- ✅ OpenAI API support (`v1/completions`, `v1/chat/completions`, `v1/embeddings` and `v1/rerank`)
|
- ✅ OpenAI API supported endpoints:
|
||||||
|
- `v1/completions`
|
||||||
|
- `v1/chat/completions`
|
||||||
|
- `v1/embeddings`
|
||||||
|
- `v1/rerank`
|
||||||
|
- `v1/audio/speech`
|
||||||
- ✅ Multiple GPU support
|
- ✅ Multiple GPU support
|
||||||
- ✅ Run multiple models at once with `profiles`
|
- ✅ Run multiple models at once with `profiles`
|
||||||
- ✅ Remote log monitoring at `/log`
|
- ✅ Remote log monitoring at `/log`
|
||||||
|
|||||||
@@ -78,6 +78,9 @@ func New(config *Config) *ProxyManager {
|
|||||||
pm.ginEngine.POST("/v1/embeddings", pm.proxyOAIHandler)
|
pm.ginEngine.POST("/v1/embeddings", pm.proxyOAIHandler)
|
||||||
pm.ginEngine.POST("/v1/rerank", pm.proxyOAIHandler)
|
pm.ginEngine.POST("/v1/rerank", pm.proxyOAIHandler)
|
||||||
|
|
||||||
|
// Support audio/speech endpoint
|
||||||
|
pm.ginEngine.POST("/v1/audio/speech", pm.proxyOAIHandler)
|
||||||
|
|
||||||
pm.ginEngine.GET("/v1/models", pm.listModelsHandler)
|
pm.ginEngine.GET("/v1/models", pm.listModelsHandler)
|
||||||
|
|
||||||
// in proxymanager_loghandlers.go
|
// in proxymanager_loghandlers.go
|
||||||
|
|||||||
Reference in New Issue
Block a user