Created
February 23, 2025 22:14
-
-
Save dipta007/6e8bf085a77f3d3d476477ca0f5a2885 to your computer and use it in GitHub Desktop.
Deploy VLLM with udocker
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# pip install udocker | |
udocker pull vllm/vllm-openai:latest | |
udocker create --name=vllm vllm/vllm-openai:latest | |
udocker setup --nvidia --force vllm | |
udocker run \ | |
--volume="/${PWD}:/workspace" \ | |
--env "HF_TOKEN=TOKEN" \ | |
vllm \ | |
--model "Qwen/Qwen2.5-0.5B-Instruct" | |
# you can add any arguments from: https://docs.vllm.ai/en/latest/serving/engine_args.html |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment