[https://nvbugs/6037654][fix] Set DeepEP low-latency token limit for qwen3 CI to prevent OOM#13484
[https://nvbugs/6037654][fix] Set DeepEP low-latency token limit for qwen3 CI to prevent OOM#13484byshiue wants to merge 1 commit intoNVIDIA:mainfrom
Conversation
…ghput_latency] to prevent OOM Signed-off-by: bhsueh <11360707+byshiue@users.noreply.github.com>
📝 WalkthroughWalkthroughThe Changes
Estimated code review effort🎯 2 (Simple) | ⏱️ ~5 minutes 🚥 Pre-merge checks | ✅ 3 | ❌ 2❌ Failed checks (2 warnings)
✅ Passed checks (3 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing Touches🧪 Generate unit tests (beta)
Comment |
There was a problem hiding this comment.
Actionable comments posted: 1
🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.
Inline comments:
In `@tests/integration/defs/accuracy/test_llm_api_pytorch.py`:
- Around line 4532-4541: The environment override setting
TRTLLM_DEEP_EP_TOKEN_LIMIT=256 is currently applied unconditionally around the
LLM context and thus affects both latency and throughput_latency cases; restrict
this patch to only the throughput case by moving the mock.patch.dict(...) so it
only wraps the throughput_latency branch (or conditionally apply it when the
test variable/mode equals "throughput_latency"), leaving the latency branch
unchanged; locate the current mock.patch.dict block that wraps the LLM(...)
context and instead apply it only when creating the LLM for the throughput test
case (referencing the LLM(...) call and local variables like attention_dp and
kv_cache_config to find the correct instantiation).
🪄 Autofix (Beta)
Fix all unresolved CodeRabbit comments on this PR:
- Push a commit to this branch (recommended)
- Create a new PR with the fixes
ℹ️ Review info
⚙️ Run configuration
Configuration used: Path: .coderabbit.yaml
Review profile: CHILL
Plan: Enterprise
Run ID: e6a3711a-5079-4389-9008-e8d4de25be07
📒 Files selected for processing (2)
tests/integration/defs/accuracy/test_llm_api_pytorch.pytests/integration/test_lists/waives.txt
💤 Files with no reviewable changes (1)
- tests/integration/test_lists/waives.txt
|
/bot skip --comment "The fixed test is not covered by l0 pre-merge CI" |
|
PR_Github #45665 [ skip ] triggered by Bot. Commit: |
|
PR_Github #45665 [ skip ] completed with state |
…ghput_latency] to prevent OOM
Summary by CodeRabbit
test_fp8[throughput_latency]test to run for the Qwen3-235B-A22B model.test_fp8accuracy test configuration with adjusted runtime parameters.Description
Test Coverage
PR Checklist
Please review the following before submitting your PR:
PR description clearly explains what and why. If using CodeRabbit's summary, please make sure it makes sense.
PR Follows TRT-LLM CODING GUIDELINES to the best of your knowledge.
Test cases are provided for new code paths (see test instructions)
Any new dependencies have been scanned for license and vulnerabilities
CODEOWNERS updated if ownership changes
Documentation updated as needed
Update tava architecture diagram if there is a significant design change in PR.
The reviewers assigned automatically/manually are appropriate for the PR.
Please check this after reviewing the above items as appropriate for this PR.
GitHub Bot Help
To see a list of available CI bot commands, please comment
/bot help.