Replies: 1 comment 3 replies
-
whatever is in your llama-swap config is crashing on startup. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I would like to share my hacky setup.
I wanted to keep llama-swap fully containerized, but I also wanted to use it with docker containers for TTS and ASR. I was already using Komodo, which has an API to start or stop stacks (among other functions, see https://docs.rs/komodo_client/latest/komodo_client/api/execute/index.html), so I figured I could simply integrate llama-swap into Komodo instead of implementing a redundant docker-in-docker setup.
Here is my llama-swap config:
Using the
curl
command without the&& sleep 1
, llama-swap would successfully start the appropriate stack, but it wouldn't recognize that the model is loaded, showing the following error:My understanding is that
curl
is instantaneous and llama-swap is expecting a process, so that's why I use&& sleep 1
to work around that limitation.I do not add
&& sleep 1
after thecurl
command incmdStop
, however, because with it, the model doesn't switch fromstopping
tostopped
until thehealthCheckTimeout
value is reached.My understanding is that without
&& sleep 1
, llama-swap is "skipping graceful stop" (which is being handled by Komodo anyway), so the model is recognized as stopped at the expected time.I'm sorry for the numerous edits to this post, and I hope this can help some of you!
Beta Was this translation helpful? Give feedback.
All reactions