Skip to content

[NV] DSR1 FP8 B300 Dynamo TRT (both MTP & STP)#638

Open
jthomson04 wants to merge 6 commits intomainfrom
nv/dsr1-fp8-b300-dynamo-trt
Open

[NV] DSR1 FP8 B300 Dynamo TRT (both MTP & STP)#638
jthomson04 wants to merge 6 commits intomainfrom
nv/dsr1-fp8-b300-dynamo-trt

Conversation

@jthomson04
Copy link
Collaborator

No description provided.

Signed-off-by: jthomson04 <jothomson@nvidia.com>
@jthomson04 jthomson04 changed the title [NV] DSR1 FP8 B300 Dynamo TRT [NV] DSR1 FP8 B300 Dynamo TRT (8k1k and 1k1k) Feb 5, 2026
@jthomson04 jthomson04 changed the title [NV] DSR1 FP8 B300 Dynamo TRT (8k1k and 1k1k) [NV] DSR1 FP8 B300 Dynamo TRT Feb 5, 2026
Signed-off-by: jthomson04 <jothomson@nvidia.com>
Signed-off-by: jthomson04 <jothomson@nvidia.com>
@jthomson04 jthomson04 marked this pull request as ready for review February 5, 2026 18:43
@jthomson04 jthomson04 requested a review from a team as a code owner February 5, 2026 18:43
@claude
Copy link
Contributor

claude bot commented Feb 5, 2026

Claude finished @jthomson04's task in 2m 1s —— View job


PR Review: [NV] DSR1 FP8 B300 Dynamo TRT

  • Gather context and read changed files
  • Validate master config + perf-changelog
  • Validate container image accessibility
  • Validate enroot import in launch script
  • Validate model prefixes
  • Check for bugs and logic errors
  • Post review summary

Summary

This PR adds FP8 precision support for DSR1 on B300 with Dynamo TRT, including 27 benchmark configurations (1k1k and 8k1k, both MTP and STP variants), launch script updates to handle FP8 model paths, and a perf-changelog entry.

Validations passed:

  • perf-changelog.yaml updated alongside nvidia-master.yaml
  • Image nvcr.io/nvidia/ai-dynamo/tensorrtllm-runtime:0.8.1.post1 is publicly accessible on NGC ✓
  • enroot import docker://$IMAGE pattern present in launch script ✓
  • Model prefix dsr1 is valid ✓
  • Launch script FP8 handling mirrors the existing GB200 pattern (SRT_SLURM_MODEL_PREFIX, model path, served model name) ✓

One issue found — see inline comment:

  • 🟡 All 27 CONFIG_FILE paths use recipies/ (typo) while existing B300 FP4 configs use recipes/. If the srt-slurm repo directory is actually named recipes/, all benchmark runs will fail to locate their config files. Please verify the actual directory name.

No other blocking issues found.

Signed-off-by: jthomson04 <jothomson@nvidia.com>
Signed-off-by: jthomson04 <jothomson@nvidia.com>
@functionstackx functionstackx changed the title [NV] DSR1 FP8 B300 Dynamo TRT [NV] DSR1 FP8 B300 Dynamo TRT (both MTP & STP) Feb 5, 2026
@functionstackx
Copy link
Contributor

@jthomson04 @kedarpotdar-nv is it fine to stop this test and merge it?

seems like 2 1k1k and 4 8k1k of it has passed already? any concern that we need to test the full sweep?

image

@functionstackx
Copy link
Contributor

@pr-claude review plz

@claude
Copy link
Contributor

claude bot commented Feb 5, 2026

Claude finished @functionstackx's task in 1m 40s —— View job


PR Review: [NV] DSR1 FP8 B300 Dynamo TRT

  • Gather context and read changed files
  • Validate master config + perf-changelog
  • Validate container image accessibility
  • Validate enroot import in launch script
  • Validate model prefixes
  • Check for bugs and logic errors
  • Post review summary

Summary

LGTM - no blocking issues found.

The previous review's issues have both been fixed:

  • recipies/recipes/ typo corrected in commit 28151c8
  • Model prefix fixed in commit 489fbfb

All validations pass:

  • perf-changelog.yaml updated alongside nvidia-master.yaml
  • Image nvcr.io/nvidia/ai-dynamo/tensorrtllm-runtime:0.8.1.post1 is publicly accessible on NGC ✓
  • enroot import docker://$IMAGE pattern present in launch script ✓
  • Model prefix dsr1 is valid across all 27 configs ✓
  • All CONFIG_FILE paths use recipes/ (no typo remnants) ✓
  • No duplicate CONFIG_FILE paths ✓
  • Launch script FP8 handling (SRT_SLURM_MODEL_PREFIX, model path, served model name) mirrors the existing GB200 pattern exactly ✓

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

Status: No status

Development

Successfully merging this pull request may close these issues.

3 participants