#!/bin/bash VLLM_VERSION="v0.8.4" BASE_MODEL_NAME="Qwen/Qwen2.5-3B-Instruct" ADAPTER_NAME="id4thomas/emotion-predictor-Qwen2.5-3B-Instruct" ADAPTER_DIR="..." # Path to the directory containing the LoRA adapter docker run --runtime nvidia --gpus all \ -v ${ADAPTER_DIR}:/vllm-workspace/adapter \ -v ./cache:/root/.cache/huggingface \ -p 8010:8000 \ --ipc=host \ vllm/vllm-openai:${VLLM_VERSION} \ --model "Qwen/Qwen2.5-3B-Instruct" \ --lora-modules '{"name": "${ADAPTER_NAME}", "path": "/vllm-workspace/adapter", "base_model_name": "${BASE_MODEL_NAME}"}' \ --enable-lora \ --max-lora-rank 16 \ --served-model-name ${BASE_MODEL_NAME} \ --gpu-memory-utilization=0.5