Skip to content

Commit 2ed1164

Browse files
committed
Readme
1 parent 6d3b32a commit 2ed1164

File tree

1 file changed

+31
-0
lines changed

1 file changed

+31
-0
lines changed

README.md

Lines changed: 31 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,31 @@
1+
# podman-ai-lab-stack
2+
3+
An external provider for [Llama Stack](https://github.com/meta-llama/llama-stack) allowing for the use of [Podman AI Lab](https://github.com/containers/podman-desktop-extension-ai-lab) for inference.
4+
5+
## Usage
6+
7+
1. Ensure Podman Desktop is running and AI Lab extension is loaded
8+
9+
2. Run the Podman AI Lab external provider inside of a container via [Podman](https://podman.io/):
10+
11+
```bash
12+
podman run -p 8321:8321 ghcr.io/feloy/podman-ai-lab-stack:nightly
13+
```
14+
15+
This will start a Llama Stack server which will use port 8321 by default. You can test this works by using the Llama Stack Client:
16+
17+
```bash
18+
llama-stack-client models list
19+
20+
llama-stack-client models register <id-of-model-loaded-in-podman-ai-lab>
21+
22+
llama-stack-client inference chat-completion --message "tell me a joke" --stream
23+
```
24+
25+
## Configuration
26+
27+
By default, Podman AI Lab listens on port `10434`, and the Podman AI Lab external provider is configured to access this port by default. If you want to provide another address/port, you can pass the `PODMAN_AI_LAB_URL` environment variable to the provider, for example:
28+
29+
```bash
30+
podman run -p 8321:8321 --env PODMAN_AI_LAB_URL=http://host.containers.internal:10435 ghcr.io/feloy/podman-ai-lab-stack:nightly
31+
```

0 commit comments

Comments
 (0)