Switch Inference Models at Runtime
Change the active inference model while the sandbox is running. No restart is required.
Prerequisites
- A running NemoClaw sandbox.
- The OpenShell CLI on your
PATH.
Switch to a Different Model
Set the provider to nvidia-nim and specify a model from build.nvidia.com:
$ openshell inference set --provider nvidia-nim --model nvidia/nemotron-3-super-120b-a12b
This requires the NVIDIA_API_KEY environment variable.
The nemoclaw onboard command stores this key in ~/.nemoclaw/credentials.json on first run.
Verify the Active Model
Run the status command to confirm the change:
$ openclaw nemoclaw status
Add the --json flag for machine-readable output:
$ openclaw nemoclaw status --json
The output includes the active provider, model, and endpoint.
Available Models
The following table lists the models registered with the nvidia-nim provider.
You can switch to any of these models at runtime.
| Model ID | Label | Context Window | Max Output |
|---|---|---|---|
nvidia/nemotron-3-super-120b-a12b | Nemotron 3 Super 120B | 131,072 | 8,192 |
nvidia/llama-3.1-nemotron-ultra-253b-v1 | Nemotron Ultra 253B | 131,072 | 4,096 |
nvidia/llama-3.3-nemotron-super-49b-v1.5 | Nemotron Super 49B v1.5 | 131,072 | 4,096 |
nvidia/nemotron-3-nano-30b-a3b | Nemotron 3 Nano 30B | 131,072 | 4,096 |
Related Topics
- Inference Profiles for full profile configuration details.