Ollama + NVIDIA: safer, faster OpenClaw with NemoClaw and Nemotron 3 Super
https://www.nvidia.com/en-us/ai/nemoclaw/ Ollama and NVIDIA have teamed up to make OpenClaw faster and safer: Nemotron 3 Super is available to run with Ollama – a new open model from NVIDIA that ranks #1 on PinchBench (the benchmark measuring OpenClaw effectiveness), with 5x more throughput than previous models. NemoClaw is now available with built-in Ollama support to provide a safe environment for your OpenClaw assistant with privacy and security guardrails. OPENCLAW WITH NEMOTRON 3 SUPER Start by downloading Ollama: Download Ollama Next, run OpenClaw with Nemotron 3 Super via Ollama's cloud: ollama launch openclaw --model nemotron-3-super:cloud Nemotron 3 Super can be used on Ollama's cloud for free. For challenging tasks, Ollama's Pro and Max plans enable your OpenClaw assistant to run multiple sub-agents and background tasks at the same time. On machines with 96GB of VRAM or more, Nemotron 3 Super can also be run locally using Ollama. NEMOCLAW: RUN OPENCLAW MORE SAFELY NVIDIA's NemoClaw is an open source stack that installs OpenClaw with added privacy and security controls, including pre-configured support for Ollama. curl -fsSL https://nvidia.com/nemoclaw.sh | bash This will install NemoClaw and run its Installer. Enter 2 to select Ollama when prompted: When prompted for a model, select nemotron-3-super:cloud (or choose another model you may wish to run). Once OpenClaw is running, connect to it: nemoclaw my-assistant connect For more information on NemoClaw, see NVIDIA's documentation. OPENSHELL: A SAFE RUNTIME FOR OTHER AGENTS. NVIDIA's OpenShell is a private runtime for agents and assistants that brings the safety of NemoClaw to other agents such as Claude Code, Codex, OpenCode and more. To get started with OpenShell, first download and install it. Next, create a new sandbox with Ollama: openshell sandbox create --from ollama Finally, run Ollama to launch an agent: