This article covers key insights from Open Models, Open Runtime, Open Harness - Building your own AI agent with LangChain and Nvidia by LangChain.
The Core Elements of an AI Agent
LangChain begins by explaining that advanced AI agents, such as Claude Code, Manus, and Open Claw, share a similar underlying architecture. According to LangChain, these agents consist of three fundamental components: a model, a runtime, and a harness. The model provides the agent's intelligence, the runtime offers the execution environment, and the harness orchestrates these elements, along with an agent layer on top. The video focuses on how developers can build these components using a completely open-source stack.
Introducing the Open-Source Agent Stack
For their open-source agent stack, LangChain showcases a specific combination of technologies:
-
The Model: Nvidia Nemotron 3 Super LangChain highlights that the Nemotron 3 Super model, recently released by Nvidia, is highly performant in both accuracy and speed. As LangChain demonstrates, it surpasses models like OpenAI's GPT OS in these metrics, pushing the boundaries of what's possible for agent responsiveness.
-
The Runtime: Nvidia Open Shell Nvidia Open Shell, also a recent release, serves as the runtime for the agent. LangChain emphasizes its security features, allowing agents to run with specified permission sets and operate locally in GPU-accelerated environments, which is crucial for leveraging GPU-accelerated libraries.
-
The Harness: LangChain Deep Agents Deep Agents is LangChain's open-source agent harness, designed to orchestrate models, tools, skills, and file systems. According to LangChain, it provides built-in mechanisms for agents to "go deeper" into tasks, heavily utilizes skills and subagents, and handles common context engineering tasks like summarization, freeing developers from managing these undifferentiated aspects.
Building and Customizing Your Agent: A Practical Demonstration
LangChain provides a practical walkthrough of setting up and running an agent using these open-source components.
Initial Setup and Environment Configuration
The demonstration begins with installing Open Shell and starting its gateway. LangChain shows how to create a persistent sandbox using Open Shell sandbox create --keep, which will be used for all agent runs. Key environment variables, as explained by LangChain, include the Nvidia API key, the Open Shell sandbox name, and optionally, Langsmith API keys for observing and evaluating agent behavior.
Deconstructing the Agent Definition
LangChain dives into the agent's code, revealing its structure:
- Model Integration: The agent uses the Nvidia Nemotron 3 model, configured with the specified API key.
- System Prompt: LangChain explains that a fixed system prompt provides basic instructions, workflow understanding, and guidelines for using the sandbox. Notably, the sandbox's network access is policy-governed, a security feature of Open Shell.
- Agent Memory: Distinct from the system prompt, agent memory is stored in a separate file (
agent.md) and can be updated by the agent itself, allowing for dynamic learning and adaptation. - Backend Abstraction: LangChain details the
create_backendfunction, which creates an Open Shell sandbox session as the core execution environment. Crucially, it layers a file system backend on top, enabling memory and skills to reside locally on the developer's machine and persist across different agent sessions or sandboxes. This is achieved using a composite backend, as explained by LangChain.
Interacting with the Agent via Langsmith Studio
LangChain demonstrates running the agent using Langgraph dev, which launches Langsmith Studio – an interface for interaction and observation. Key views include:
- Graph View: According to LangChain, this view visualizes the agent's architecture, showing middleware for memory and tool calls, and the core loop of model tools.
- Chat View: A simple interface for direct interaction with the agent.
- Reasoning Blocks & Tool Calls: LangChain emphasizes that this interface provides deep insight into the agent's internal thought processes and executed commands.
Agent Capabilities in Action
The demonstration showcases several agent capabilities:
- Command Execution: LangChain illustrates the agent running system commands like
uname -aandpython3 --versionwithin the sandbox. - Code Generation and Execution: The agent successfully writes and executes a Python script to generate statistics on random numbers.
- Security Policy Enforcement: LangChain tests Open Shell's security features by instructing the agent to send a POST request to
evil.com. As expected, the request is blocked, confirming the sandbox's policy-governed network access. - Persistent Memory Updates: The agent updates its
agent.mdfile to "speak Spanish," demonstrating how memory can live outside the sandbox and persist across sessions, as highlighted by LangChain.
The Future of Open-Source Agents
LangChain concludes by expressing significant excitement for the future of this open-source agent foundation. They anticipate further investment in Deep Agents, with plans for features like async subagents and diverse execution environments. LangChain also looks forward to continued improvements in Nvidia Open Shell, particularly its integration with GPU-accelerated boxes to empower agents with complex data processing capabilities. The future of open-source models, LangChain notes, remains a key area of enthusiasm.
To dive deeper into the technical details and witness the full demonstration, LangChain encourages readers to watch the original video.
This article is based on a video by LangChain. Source: Open Models, Open Runtime, Open Harness - Building your own AI agent with LangChain and Nvidia
External Intelligence