[docker] Update stable image to vllm==0.15.1#5221
[docker] Update stable image to vllm==0.15.1#5221Begunner wants to merge 11 commits intoverl-project:mainfrom
Conversation
There was a problem hiding this comment.
Code Review
This pull request updates the stable Docker images to use vllm==0.15.1, simplifying the installation process by using pip instead of building from source. The changes also include pinning the transformers library version for better reproducibility and updating tests to accommodate the new dependencies. My main concern is a potential issue in docker/Dockerfile.stable.vllm where installing vllm might override the specified torch version, leading to a build failure in subsequent steps. I've provided a suggestion to address this.
|
|
||
| # use torch2.9.1+cudnn9.16 to avoid conflict | ||
| RUN pip install torch==2.9.1 torchvision torchaudio --index-url https://download.pytorch.org/whl/cu129 | ||
| RUN pip install vllm==0.15.1 |
There was a problem hiding this comment.
Installing vllm without the --no-deps flag may lead to its dependencies, including torch, being installed. This could potentially override the specific torch==2.9.1 version installed in the preceding step. If the torch version changes, the sed command on line 35 will fail as its path is hardcoded for torch-2.9.1+cu129.dist-info. The previous installation method for vllm used --no-deps, suggesting that dependencies are managed elsewhere. To prevent the build from breaking, you should add the --no-deps flag to this installation command.
RUN pip install --no-deps vllm==0.15.1
There was a problem hiding this comment.
+1 here, it's usually safer to fully determine vllm 0.15.1 dependencies, install those first, and install vllm with --no-deps.
You can figure out vllm deps using something like
uv venv --python 3.12 --seed
source .venv/bin/activate
uv pip install vllm==0.15.1
uv pip show vllm
and can then pin the specific version shown by uv pip list
thvasilo
left a comment
There was a problem hiding this comment.
Thanks for the update, as confirmation I've tested vllm 0.15.0 myself and am able to run jobs with current dev branch of verl.
| git \ | ||
| wget \ | ||
| cmake \ | ||
| vim \ |
There was a problem hiding this comment.
I think this can be removed as it not all users need vim and adds weight to iamge?
What does this PR do?
Checklist Before Starting
[{modules}] {type}: {description}(This will be checked by the CI){modules}includefsdp,megatron,veomni,sglang,vllm,rollout,trainer,ci,training_utils,recipe,hardware,deployment,ray,worker,single_controller,misc,perf,model,algo,env,tool,ckpt,doc,data,cfg,reward,like[megatron, fsdp, doc]{type}is infeat,fix,refactor,chore,test[BREAKING]to the beginning of the title.[BREAKING][fsdp, megatron] feat: dynamic batchingTest
API and Usage Example
# Add code snippet or script demonstrating how to use thisDesign & Code Changes
Checklist Before Submitting
Important
Please check all the following items before requesting a review, otherwise the reviewer might deprioritize this PR for review.
pre-commit install && pre-commit run --all-files --show-diff-on-failure --color=alwaysci-requestchannel in theverlSlack workspace. (If not accessible, please try the Feishu group (飞书群).)recipesubmodule, please also update the reference to the submodule commit viagit submodule update --remoteorcd recipe && git pull origin main.