From 4c7ecfd5b90f3fcf7e8f55f4013615574021a9fa Mon Sep 17 00:00:00 2001 From: root Date: Fri, 20 Oct 2023 19:45:21 -0600 Subject: [PATCH] Add llm-vscode-inference-server --- src/LLM.tex | 18 ++++++++++++++++++ 1 file changed, 18 insertions(+) diff --git a/src/LLM.tex b/src/LLM.tex index 5119f58..88f4741 100644 --- a/src/LLM.tex +++ b/src/LLM.tex @@ -540,3 +540,21 @@ pip install torch==1.13.1+cu116 --extra-index-url https://download.pytorch.org/w pip install -r requirements.txt python main.py --port 8080 --host 0.0.0.0 --pretrained "Phind/Phind-CodeLlama-34B-v2" \end{minted} + +\subsection{llm-vscode Inference Server} +Perhaps. +Uses vllm-project. + +\begin{minted}{sh} +mkdir -p ~/devel/wangcx18 +cd ~/devel/wangcx18/ +git clone --recursive https://github.com/wangcx18/llm-vscode-inference-server +cd llm-vscode-inference-server/ +virtualenv venv +source venv/bin/activate +pip install -U pip setuptools wheel +#pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 +pip install torch torchvision torchaudio +pip install -r requirements.txt +python api_server.py --host 0.0.0.0 --port 8080 +\end{minted}