Skip to content

Commit

Permalink
[pre-commit.ci] auto fixes from pre-commit.com hooks
Browse files Browse the repository at this point in the history
for more information, see https://pre-commit.ci
  • Loading branch information
pre-commit-ci[bot] committed Nov 11, 2024
1 parent 0dbae69 commit fa46865
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion comps/llms/faq-generation/vllm/langchain/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ This microservice interacts with the vLLM server to generate FAQs from Input Tex

If you start an LLM microservice with docker, the `docker_compose_llm.yaml` file will automatically start a VLLM service with docker.

To setup or build the vLLM image follow the instructions provided in [vLLM Gaudi](https://github.com/opea-project/GenAIComps/tree/main/comps/llms/text-generation/vllm/langchain#22-vllm-on-gaudi)
To setup or build the vLLM image follow the instructions provided in [vLLM Gaudi](https://github.com/opea-project/GenAIComps/tree/main/comps/llms/text-generation/vllm/langchain#22-vllm-on-gaudi)

### 1.1 Setup Environment Variables

Expand Down

0 comments on commit fa46865

Please sign in to comment.