@@ -51,15 +51,20 @@ function start_vllm_services() {
5151 pip install -U huggingface_hub
5252 huggingface-cli download $LLM_MODEL --local-dir " ${MODEL_PATH} /${LLM_MODEL} "
5353 fi
54+ HF_CACHE=" ${HOME} /.cache"
55+ if [ ! -d " ${HF_CACHE} " ]; then
56+ mkdir -p " ${HF_CACHE} "
57+ echo " Created directory: ${HF_CACHE} "
58+ fi
5459 echo " give permission to related path..."
5560 sudo chown 1000:1000 ${MODEL_PATH} ${DOC_PATH} ${TMPFILE_PATH}
56- sudo chown -R 1000:1000 ${HOME} /.cache/huggingface
61+ sudo chown -R 1000:1000 ${HF_CACHE}
5762 HF_ENDPOINT=https://hf-mirror.com
5863 # vllm ENV
5964 export NGINX_PORT=8086
6065 export vLLM_ENDPOINT=" http://${HOST_IP} :${NGINX_PORT} "
6166 TENSOR_PARALLEL_SIZE=$( get_user_input " your tp size" 1)
62- SELECTED_XPU_0= $( get_user_input " selected GPU " " 0 " )
67+ read -p " selected GPU [ $( seq -s, 0 $(( TENSOR_PARALLEL_SIZE - 1 )) ) ] " SELECTED_XPU_0 ; SELECTED_XPU_0= ${SELECTED_XPU_0 :- $(seq -s, 0 $((TENSOR_PARALLEL_SIZE - 1)))}
6368 DP_NUM=$( get_user_input " DP number(how many containers to run vLLM)" 1)
6469 for (( x= 0 ; x< DP_NUM; x++ ))
6570 do
@@ -138,9 +143,14 @@ function start_services() {
138143 exit 0
139144 fi
140145 fi
146+ HF_CACHE=" ${HOME} /.cache"
147+ if [ ! -d " ${HF_CACHE} " ]; then
148+ mkdir -p " ${HF_CACHE} "
149+ echo " Created directory: ${HF_CACHE} "
150+ fi
141151 echo " give permission to related path..."
142152 sudo chown 1000:1000 ${MODEL_PATH} ${DOC_PATH} ${TMPFILE_PATH}
143- sudo chown -R 1000:1000 ${HOME} /.cache/huggingface
153+ sudo chown -R 1000:1000 ${HF_CACHE}
144154 HF_ENDPOINT=https://hf-mirror.com
145155
146156 # export ENV
0 commit comments