-
Notifications
You must be signed in to change notification settings - Fork 28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[feature request] rewrite the model name passed to upstream #69
Comments
I've been considering this feature and with the recent changes it would be fairly easy to implement. I am curious, what is your use case for llama-swap+ollama vs llama-swap+llama-server? |
Thanks for your quick answer.
Why not both? I'm using llama-swap to not only swap models but also to swap inference engines depending on the constraints i have (RAM, VRAM, time, inputs...). For example:
I'm using llama-swap to select the best upstream backend depending on the chosen model. |
I use it for that too! Makes it a lot easier to swap between engines for capabilities. In this case, the “:” used for profiles is conflicting with ollama’s naming conventions. And “upstream_name” is an override so the model name can be set to anything. |
Fixed in #71 and released in v95! Example of usage: models:
"qwq":
proxy: http://127.0.0.1:11434
cmd: my-server
# use this new configuration parameter to override what's in the request
useModelName: "qwen:qwq" |
I would like to configure the model name passed to upstream server.
From my understanding, right now the proxy sends the model name configured in the yaml file of llama-swap.
Use case:
Function to modify:
I think a good place to insert the change is proxyOAIHandler() in proxymanager.go
Example config file:
The text was updated successfully, but these errors were encountered: