From ed4dcafb55d2cdec47dfb9f453d21f13088ede99 Mon Sep 17 00:00:00 2001 From: sekyonda <127536312+sekyondaMeta@users.noreply.github.com> Date: Wed, 19 Jul 2023 11:05:51 -0400 Subject: [PATCH] Update inference.md --- inference/inference.md | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/inference/inference.md b/inference/inference.md index 9c858684..d30ca86b 100644 --- a/inference/inference.md +++ b/inference/inference.md @@ -43,4 +43,10 @@ Alternate inference options include: To use vLLM you will need to install it using the instructions [here](https://vllm.readthedocs.io/en/latest/getting_started/installation.html#installation). Once installed, you can use the vLLM_ineference.py script provided [here](vLLM_inference.py). -[**TGI**](https://github.com/huggingface/text-generation-inference): Text Generation Inference (TGI) is another inference option available to you. For more information on how to set up and use TGI see [here](https://github.com/huggingface/text-generation-inference). +Below is an example of how to run the vLLM_inference.py script found within the inference folder. + +``` bash +python vLLM_inference.py --model_name <PATH/TO/LLAMA/7B> +``` + +[**TGI**](https://github.com/huggingface/text-generation-inference): Text Generation Inference (TGI) is another inference option available to you. For more information on how to set up and use TGI see [here](hf-text-generation-inference/README.md). -- GitLab