feat(inference): add Nemotron 3 Nano Omni to CLOUD_MODEL_OPTIONS#2628
Conversation
|
Note Reviews pausedIt looks like this branch is under active development. To avoid overwhelming you with review comments due to an influx of new commits, CodeRabbit has automatically paused this review. You can configure this behavior by changing the Use the following commands to manage reviews:
Use the checkboxes below for quick actions:
📝 WalkthroughWalkthroughAdded ChangesCloud model options and tests
Estimated code review effort🎯 2 (Simple) | ⏱️ ~10 minutes Poem
🚥 Pre-merge checks | ✅ 5✅ Passed checks (5 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing Touches🧪 Generate unit tests (beta)
Comment |
Adds `nvidia/nemotron-3-nano-omni-reasoning-30b-a3b` to the curated cloud model picker and its matching test entry. Super 120B remains the default. Motivation: the multimodal hermes-omni-demo cookbook in brevdev/nemoclaw-demos currently has to do a post-onboard `openshell inference set` to switch the gateway from Super to Omni, because the wizard only exposes Super. Adding Omni here lets users select it during `nemoclaw onboard` directly and lets cookbooks drop the manual swap step. Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com> Signed-off-by: Patrick Moorhead <pmoorhead@nvidia.com>
ba24257 to
4b52d13
Compare
|
✨ Thanks for submitting this PR that adds Nemotron 3 Nano Omni to the curated cloud model picker, allowing users to select it during onboard directly. This change aims to improve the user experience by eliminating the need for a manual model swap step in multimodal cookbooks. Related open issues: |
Signed-off-by: Aaron Erickson <aerickson@nvidia.com>
Signed-off-by: Aaron Erickson <aerickson@nvidia.com>
Signed-off-by: Aaron Erickson <aerickson@nvidia.com>
ericksoa
left a comment
There was a problem hiding this comment.
Approved after fixing the NVIDIA Build model ID to nvidia/nemotron-3-nano-omni-30b-a3b-reasoning, resolving the main-branch conflict, updating the shifted model-picker tests, and verifying the focused/local test coverage plus the completed PR checks.
Summary
Adds
nvidia/nemotron-3-nano-omni-30b-a3b-reasoning(label: Nemotron 3 Nano Omni 30B) to the curated cloud model picker. Super 120B remains the default.Motivation
The multimodal
hermes-omni-democookbook in brevdev/nemoclaw-demos currently has to do a post-onboardopenshell inference set --model private/nvidia/nemotron-3-nano-omni-30b-a3b-reasoningto switch the gateway from Super to Omni, because the wizard only exposes Super. The current cookbook frames this as "You picked Super 120B during onboarding because that's what the menu offers, but this cookbook needs Omni..." — that workaround is awkward and reviewers in brevdev/nemoclaw-demos#23 have called it out as an abuse of the installer.Adding Omni here lets users select it during
nemoclaw onboarddirectly and lets multimodal cookbooks drop the manual swap step entirely.Test plan
inference-config.test.tsupdated to include the new model id in the expected list — runs as part ofnpm testnemoclaw onboard --agent hermesshows Omni as option in the model pickerModel: private/nvidia/nemotron-3-nano-omni-30b-a3b-reasoninginnemoclaw <name> statusFiles changed
src/lib/inference-config.ts— one new entry inCLOUD_MODEL_OPTIONSsrc/lib/inference-config.test.ts— matching expected-list update🤖 Generated with Claude Code
Summary by CodeRabbit
New Features
Tests
Signed-off-by: Aaron Erickson aerickson@nvidia.com