Switch Inference Models at Runtime

Change the active inference model while the sandbox is running. No restart is required.

Prerequisites

  • A running NemoClaw sandbox.
  • The OpenShell CLI on your PATH.

Switch to a Different Model

Set the provider to nvidia-nim and specify a model from build.nvidia.com:

$ openshell inference set --provider nvidia-nim --model nvidia/nemotron-3-super-120b-a12b

This requires the NVIDIA_API_KEY environment variable. The nemoclaw onboard command stores this key in ~/.nemoclaw/credentials.json on first run.

Verify the Active Model

Run the status command to confirm the change:

$ openclaw nemoclaw status

Add the --json flag for machine-readable output:

$ openclaw nemoclaw status --json

The output includes the active provider, model, and endpoint.

Available Models

The following table lists the models registered with the nvidia-nim provider. You can switch to any of these models at runtime.

Model IDLabelContext WindowMax Output
nvidia/nemotron-3-super-120b-a12bNemotron 3 Super 120B131,0728,192
nvidia/llama-3.1-nemotron-ultra-253b-v1Nemotron Ultra 253B131,0724,096
nvidia/llama-3.3-nemotron-super-49b-v1.5Nemotron Super 49B v1.5131,0724,096
nvidia/nemotron-3-nano-30b-a3bNemotron 3 Nano 30B131,0724,096