Replies: 2 comments
-
The path to llama-server should be to the one in the container. Looks like the paths to the binary and models are ones on the host machine. |
Beta Was this translation helpful? Give feedback.
0 replies
-
This issue is stale because it has been open for 2 weeks with no activity. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Describe the bug
I can't run any model even though I verified that the paths I provided are correct.
Expected behaviour
Model running.
Operating system and version
My Configuration
Proxy Logs
Upstream Logs
Here is the command I launched llama-swap with :
sudo docker run -it --rm -p 9292:8080 -v /home/grand/llm/llama_swap_config.yaml:/app/config.yaml ghcr.io/mostlygeek/llama-swap:cuda
I removed the -runtime nvidia because it was not running at all with it.
Beta Was this translation helpful? Give feedback.
All reactions