Back to Nemoclaw

NVIDIA NemoClaw: OpenClaw Plugin for OpenShell

README.md

latest12.6 KB
Original Source

NVIDIA NemoClaw: OpenClaw Plugin for OpenShell

<!-- start-badges -->

<!-- end-badges -->

NVIDIA NemoClaw is an open source reference stack that simplifies running OpenClaw always-on assistants safely. It installs the NVIDIA OpenShell runtime, part of NVIDIA Agent Toolkit, a secure environment for running autonomous agents, and open source models such as NVIDIA Nemotron.

Alpha software

NemoClaw is available in early preview starting March 16, 2026. Interfaces, APIs, and behavior may change without notice as we iterate on the design. The project is shared to gather feedback and enable early experimentation. We welcome issues and discussion from the community while the project evolves.


Quick Start

Follow these steps to get started with NemoClaw and your first sandboxed OpenClaw agent.

ℹ️ Note

NemoClaw creates a fresh OpenClaw instance inside the sandbox during onboarding.

<!-- start-quickstart-guide -->

Prerequisites

Check the prerequisites before you start to ensure you have the necessary software and hardware to run NemoClaw.

Hardware

ResourceMinimumRecommended
CPU4 vCPU4+ vCPU
RAM8 GB16 GB
Disk20 GB free40 GB free

The sandbox image is approximately 2.4 GB compressed. During image push, the Docker daemon, k3s, and the OpenShell gateway run alongside the export pipeline, which buffers decompressed layers in memory. On machines with less than 8 GB of RAM, this combined usage can trigger the OOM killer. If you cannot add memory, configuring at least 8 GB of swap can work around the issue at the cost of slower performance.

Software

DependencyVersion
LinuxUbuntu 22.04 LTS or later
Node.js20 or later
npm10 or later
Container runtimeSupported runtime installed and running
OpenShellInstalled

Container Runtime Support

PlatformSupported runtimesNotes
LinuxDockerPrimary supported path today
macOS (Apple Silicon)Colima, Docker DesktopRecommended runtimes for supported macOS setups
macOSPodmanNot supported yet. NemoClaw currently depends on OpenShell support for Podman on macOS.
Windows WSLDocker Desktop (WSL backend)Supported target path

💡 Tip

For DGX Spark, follow the DGX Spark setup guide. It covers Spark-specific prerequisites, such as cgroup v2 and Docker configuration, before running the standard installer.

Install NemoClaw and Onboard OpenClaw Agent

Download and run the installer script. The script installs Node.js if it is not already present, then runs the guided onboard wizard to create a sandbox, configure inference, and apply security policies.

bash
curl -fsSL https://www.nvidia.com/nemoclaw.sh | bash

If you use nvm or fnm to manage Node.js, the installer may not update your current shell's PATH. If nemoclaw is not found after install, run source ~/.bashrc (or source ~/.zshrc for zsh) or open a new terminal.

When the install completes, a summary confirms the running environment:

──────────────────────────────────────────────────
Sandbox      my-assistant (Landlock + seccomp + netns)
Model        nvidia/nemotron-3-super-120b-a12b (NVIDIA Cloud API)
──────────────────────────────────────────────────
Run:         nemoclaw my-assistant connect
Status:      nemoclaw my-assistant status
Logs:        nemoclaw my-assistant logs --follow
──────────────────────────────────────────────────

[INFO]  === Installation complete ===

Chat with the Agent

Connect to the sandbox, then chat with the agent through the TUI or the CLI.

Connect to the Sandbox

Run the following command to connect to the sandbox:

bash
nemoclaw my-assistant connect

This connects you to the sandbox shell sandbox@my-assistant:~$ where you can run openclaw commands.

OpenClaw TUI

In the sandbox shell, run the following command to open the OpenClaw TUI, which opens an interactive chat interface.

bash
openclaw tui

Send a test message to the agent and verify you receive a response.

ℹ️ Note

The TUI is best for interactive back-and-forth. If you need the full text of a long response such as a large code generation output, use the CLI instead.

OpenClaw CLI

In the sandbox shell, run the following command to send a single message and print the response:

bash
openclaw agent --agent main --local -m "hello" --session-id test

This prints the complete response directly in the terminal and avoids relying on the TUI view for long output.

Uninstall

To remove NemoClaw and all resources created during setup, in the terminal outside the sandbox, run:

bash
curl -fsSL https://raw.githubusercontent.com/NVIDIA/NemoClaw/refs/heads/main/uninstall.sh | bash

The script removes sandboxes, the NemoClaw gateway and providers, related Docker images and containers, local state directories, and the global nemoclaw npm package. It does not remove shared system tooling such as Docker, Node.js, npm, or Ollama.

FlagEffect
--yesSkip the confirmation prompt.
--keep-openshellLeave the openshell binary installed.
--delete-modelsAlso remove NemoClaw-pulled Ollama models.

For example, to skip the confirmation prompt:

bash
curl -fsSL https://raw.githubusercontent.com/NVIDIA/NemoClaw/refs/heads/main/uninstall.sh | bash -s -- --yes
<!-- end-quickstart-guide -->

How It Works

NemoClaw installs the NVIDIA OpenShell runtime and Nemotron models, then uses a versioned blueprint to create a sandboxed environment where every network request, file access, and inference call is governed by declarative policy. The nemoclaw CLI orchestrates the full stack: OpenShell gateway, sandbox, inference provider, and network policy.

ComponentRole
PluginTypeScript CLI commands for launch, connect, status, and logs.
BlueprintVersioned Python artifact that orchestrates sandbox creation, policy, and inference setup.
SandboxIsolated OpenShell container running OpenClaw with policy-enforced egress and filesystem.
InferenceNVIDIA cloud model calls, routed through the OpenShell gateway, transparent to the agent.

The blueprint lifecycle follows four stages: resolve the artifact, verify its digest, plan the resources, and apply through the OpenShell CLI.

When something goes wrong, errors may originate from either NemoClaw or the OpenShell layer underneath. Run nemoclaw <name> status for NemoClaw-level health and openshell sandbox list to check the underlying sandbox state.


Inference

Inference requests from the agent never leave the sandbox directly. OpenShell intercepts every call and routes it to the NVIDIA cloud provider.

ProviderModelUse Case
NVIDIA cloudnvidia/nemotron-3-super-120b-a12bProduction. Requires an NVIDIA API key.

Get an API key from build.nvidia.com. The nemoclaw onboard command prompts for this key during setup.

Local inference options such as Ollama and vLLM are still experimental. On macOS, they also depend on OpenShell host-routing support in addition to the local service itself being reachable on the host.


Protection Layers

The sandbox starts with a strict baseline policy that controls network egress and filesystem access:

LayerWhat it protectsWhen it applies
NetworkBlocks unauthorized outbound connections.Hot-reloadable at runtime.
FilesystemPrevents reads/writes outside /sandbox and /tmp.Locked at sandbox creation.
ProcessBlocks privilege escalation and dangerous syscalls.Locked at sandbox creation.
InferenceReroutes model API calls to controlled backends.Hot-reloadable at runtime.

When the agent tries to reach an unlisted host, OpenShell blocks the request and surfaces it in the TUI for operator approval.


Key Commands

Host commands (nemoclaw)

Run these on the host to set up, connect to, and manage sandboxes.

CommandDescription
nemoclaw onboardInteractive setup wizard: gateway, providers, sandbox.
nemoclaw <name> connectOpen an interactive shell inside the sandbox.
openshell termLaunch the OpenShell TUI for monitoring and approvals.
nemoclaw start / stop / statusManage auxiliary services (Telegram bridge, tunnel).

Plugin commands (openclaw nemoclaw)

Run these inside the OpenClaw CLI. These commands are under active development and may not all be functional yet.

CommandDescription
openclaw nemoclaw launch [--profile ...]Bootstrap OpenClaw inside an OpenShell sandbox.
openclaw nemoclaw statusShow sandbox health, blueprint state, and inference.
openclaw nemoclaw logs [-f]Stream blueprint execution and sandbox logs.

See the full CLI reference for all commands, flags, and options.

Known limitations:

  • The openclaw nemoclaw plugin commands are under active development. Use the nemoclaw host CLI as the primary interface.
  • Setup may require manual workarounds on some platforms. File an issue if you encounter blockers.

Learn More

Refer to the documentation for more information on NemoClaw.

  • Overview: Learn what NemoClaw does and how it fits together.
  • How It Works: Learn about the plugin, blueprint, and sandbox lifecycle.
  • Architecture: Learn about the plugin structure, blueprint lifecycle, and sandbox environment.
  • Inference Profiles: Learn about the NVIDIA cloud inference configuration.
  • Network Policies: Learn about egress control and policy customization.
  • CLI Commands: Learn about the full command reference.
  • Troubleshooting: Troubleshoot common issues and resolution steps.
  • Discord: Join the community for questions and discussion.

License

This project is licensed under the Apache License 2.0.