From c2cfd29c21c381382ba5f660465503c30e947e1c Mon Sep 17 00:00:00 2001 From: Shangyin Tan Date: Tue, 6 Aug 2024 22:57:50 -0700 Subject: [PATCH] Update HFClientVLLM.mdx vllm needs to be started with OAI compatible server. --- .../language_model_clients/local_models/HFClientVLLM.mdx | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/deep-dive/language_model_clients/local_models/HFClientVLLM.mdx b/docs/deep-dive/language_model_clients/local_models/HFClientVLLM.mdx index 059c2a9..bdce7c8 100644 --- a/docs/deep-dive/language_model_clients/local_models/HFClientVLLM.mdx +++ b/docs/deep-dive/language_model_clients/local_models/HFClientVLLM.mdx @@ -9,7 +9,7 @@ Refer to the [vLLM Server API](/api/local_language_model_clients/vLLM) for setti ```bash #Example vLLM Server Launch - python -m vllm.entrypoints.api_server --model meta-llama/Llama-2-7b-hf --port 8080 + python -m vllm.entrypoints.openai.api_server --model meta-llama/Llama-2-7b-hf --port 8080 ``` This command will start the server and make it accessible at `http://localhost:8080`. @@ -79,4 +79,4 @@ print(response) *** - \ No newline at end of file +