-
Notifications
You must be signed in to change notification settings - Fork 2k
chore: set default device to cpu on Multimodal models #5994
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
/bot run |
|
PR_Github #11771 [ run ] triggered by Bot |
|
PR_Github #11771 [ run ] completed with state |
c27625d to
0b4f1ce
Compare
|
/bot run |
|
PR_Github #11795 [ run ] triggered by Bot |
|
PR_Github #11795 [ run ] completed with state |
|
Non-blocking comment: Is this change not needed for Gemma3, @yechank-nvidia? |
|
Hi @brb-nv, double-checked on the H100, seeing one keyword difference output, so changing again. Thx for pointing out. |
0b4f1ce to
339c195
Compare
|
/bot run |
|
PR_Github #11882 [ run ] triggered by Bot |
|
PR_Github #11882 [ run ] completed with state |
339c195 to
574afdf
Compare
|
/bot run |
|
PR_Github #12491 [ run ] triggered by Bot |
|
PR_Github #12491 [ run ] completed with state |
|
/bot run |
1. Change use_fast=True for Qwen2/2.5-VL models 2. Change the test keywords accordingly Signed-off-by: yechank <[email protected]>
Signed-off-by: yechank <[email protected]>
Signed-off-by: yechank <[email protected]>
574afdf to
85255e0
Compare
WalkthroughThe changes update device defaults from "cuda" to "cpu" for multimodal input processing and model inference across utility functions, example scripts, and model input processors. Test data for multimodal model outputs is also revised, updating expected keyword lists for various models and modalities. No new features or control flow modifications are introduced. Changes
Sequence Diagram(s)sequenceDiagram
participant User
participant ExampleScript
participant Utils
participant ModelInputProcessor
participant Model
User->>ExampleScript: Run multimodal example
ExampleScript->>Utils: Load image/video/audio (device="cpu")
Utils-->>ExampleScript: Return processed tensor(s) on CPU
ExampleScript->>ModelInputProcessor: Preprocess inputs (no explicit device)
ModelInputProcessor-->>ExampleScript: Return processed inputs
ExampleScript->>Model: Run inference (inputs on CPU)
Model-->>ExampleScript: Output results
Estimated code review effort2 (~15 minutes) Possibly related PRs
Suggested reviewers
Poem
📜 Recent review detailsConfiguration used: .coderabbit.yaml 📒 Files selected for processing (5)
💤 Files with no reviewable changes (1)
🔇 Additional comments (15)
✨ Finishing Touches
🪧 TipsChatThere are 3 ways to chat with CodeRabbit:
SupportNeed help? Create a ticket on our support page for assistance with any issues or questions. Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments. CodeRabbit Commands (Invoked using PR comments)
Other keywords and placeholders
Documentation and Community
|
|
/bot run |
|
PR_Github #12548 [ run ] triggered by Bot |
|
PR_Github #12548 [ run ] completed with state |
|
/bot run |
|
PR_Github #12571 [ run ] triggered by Bot |
|
PR_Github #12571 [ run ] completed with state |
Signed-off-by: yechank <[email protected]> Signed-off-by: Shreyas Misra <[email protected]>
Signed-off-by: yechank <[email protected]> Signed-off-by: Ransiki Zhang <[email protected]>
Signed-off-by: yechank <[email protected]> Signed-off-by: Lanyu Liao <[email protected]>
This PR changes the default device type to be cpu on the main-process so that we are avoiding the interference for worker process GPU job.
Also,
Summary by CodeRabbit
Bug Fixes
Tests