-
Notifications
You must be signed in to change notification settings - Fork 844
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Showing
2 changed files
with
34 additions
and
1 deletion.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,33 @@ | ||
# Meta Reference Distribution | ||
|
||
The `llamastack/distribution-meta-reference-gpu` distribution consists of the following provider configurations. | ||
|
||
|
||
| **API** | **Inference** | **Agents** | **Memory** | **Safety** | **Telemetry** | | ||
|----------------- |--------------- |---------------- |-------------------------------------------------- |---------------- |---------------- | | ||
| **Provider(s)** | meta-reference | meta-reference | meta-reference, remote::pgvector, remote::chroma | meta-reference | meta-reference | | ||
|
||
|
||
### Start the Distribution (Single Node GPU) | ||
|
||
> [!NOTE] | ||
> This assumes you have access to GPU to start a TGI server with access to your GPU. | ||
> [!NOTE] | ||
> For GPU inference, you need to set these environment variables for specifying local directory containing your model checkpoints, and enable GPU inference to start running docker container. | ||
``` | ||
export LLAMA_CHECKPOINT_DIR=~/.llama | ||
``` | ||
|
||
> [!NOTE] | ||
> `~/.llama` should be the path containing downloaded weights of Llama models. | ||
|
||
To download and start running a pre-built docker container, you may use the following commands: | ||
|
||
``` | ||
docker run -it -p 5000:5000 -v ~/.llama:/root/.llama --gpus=all llamastack/llamastack-local-gpu | ||
``` | ||
|
||
### Alternative (Build and start distribution locally via conda) | ||
- You may checkout the [Getting Started](../../docs/getting_started.md) for more details on starting up a meta-reference distribution. |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters