Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama stack distributions / templates / docker refactor #266

Merged
merged 32 commits into from
Oct 21, 2024
Merged

Conversation

yanxi0830
Copy link
Contributor

@yanxi0830 yanxi0830 commented Oct 18, 2024

Test

--list-templates

image

ollama docker

llama-stack/llama_stack/distribution/docker/ollama$ ls
compose.yaml  ollama-run.yaml

llama-stack/llama_stack/distribution/docker/ollama$ docker compose up
image image

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Meta Open Source bot. label Oct 18, 2024
@yanxi0830 yanxi0830 marked this pull request as ready for review October 18, 2024 02:38
@yanxi0830 yanxi0830 changed the title docker compose ollama scripts llama stack distributions / templates / docker refactor Oct 19, 2024
@terrytangyuan
Copy link
Contributor

terrytangyuan commented Oct 19, 2024

Is it possible to get #178 merged before further refactoring? Otherwise we'll end up fixing a lot of conflicts again. cc @ashwinb

@ashwinb
Copy link
Contributor

ashwinb commented Oct 21, 2024

Is it possible to get #178 merged before further refactoring? Otherwise we'll end up fixing a lot of conflicts again. cc @ashwinb

@terrytangyuan Done! I just merged #178.

ashwinb and others added 8 commits October 21, 2024 10:46
…ce (#270)

PR #201 had made several changes while trying to fix issues with getting the stream=False branches of inference and agents API working. As part of this, it made a change which was slightly gratuitous. Namely, making chat_completion() and brethren "def" instead of "async def".

The rationale was that this allowed the user (within llama-stack) of this to use it as:

```
async for chunk in api.chat_completion(params)
```

However, it causes unnecessary confusion for several folks. Given that clients (e.g., llama-stack-apps) anyway use the SDK methods (which are completely isolated) this choice was not ideal. Let's revert back so the call now looks like:

```
async for chunk in await api.chat_completion(params)
```

Bonus: Added a completion() implementation for the meta-reference provider. Technically should have been another PR :)
This PR adds vLLM inference provider for OpenAI compatible vLLM server.
Trying out readthedocs
@yanxi0830 yanxi0830 merged commit 23210e8 into main Oct 21, 2024
4 checks passed
@yanxi0830 yanxi0830 deleted the ollama_docker branch October 29, 2024 17:05
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Meta Open Source bot.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

7 participants