
How To Optimize LLM Inference in Production in 2026

Sergey Kaplich

OpenClaw is a powerful open-source AI agent that goes beyond code suggestions to execute complete development workflows autonomously—from requirements to code to testing to PR submission—all through messaging platforms like Telegram, Slack, or WhatsApp. While developers are successfully using it to ship features without manual intervention, the tool has critical security vulnerabilities (including CVE-2026-25253 with CVSS 9.8), stores credentials in plaintext, and has been flagged by multiple enterprise security vendors as inappropriate for production environments with sensitive data. Use it for experimentation in isolated environments, but proceed with extreme caution and implement serious security hardening before any production deployment.
You've got a working codebase. You've got Copilot suggesting the next line. You've got Cursor helping you refactor that gnarly component.
And yet.
You're still the one running the tests. Still the one pushing the PR. Still the one context-switching between Slack, your terminal, your email, and the seventeen browser tabs you swore you'd close yesterday.
What if your AI assistant could take requirements, write the code, run the tests, fix what breaks, and open the pull request—all while you grab coffee? That's exactly what OpenClaw does. It goes beyond code suggestions to autonomous task execution across your entire system. Requirements to testing to PR submission, without constant manual review.
Developers are actually doing this right now. And it's incredible.
Here's the thing though: with that kind of power comes some very real risks. We'll get into the security concerns later, but first—let's dig into what makes this tool so exciting.
First, let's clear up the naming confusion.
OpenClaw, ClawdBot, and Moltbot are the same project. One open-source tool that's been through two rebrands in weeks:
The official repository now lives at github.com/openclaw/openclaw.
Who made this? Peter Steinberger, founder of PSPDFKit (acquired for $100M+). The guy wanted AI that could "ship features end-to-end without constant hand-holding." So he built it.
What does it actually do? OpenClaw is a self-hosted AI assistant that bridges messaging platforms—WhatsApp, Telegram, iMessage, Slack, Discord, Signal—to AI agents with persistent memory and proactive notifications. It runs locally on your machine, with your API keys, under your control.
That last part matters. A lot.
But calling it a "messaging bridge" undersells it dramatically.
It can execute shell commands. Read and write files anywhere on your system. Run scripts. Manage your calendar. Send emails. Coordinate across multiple codebases. All autonomously.
It's not a coding assistant.
It's closer to a digital employee—an autonomous AI agent that executes complete workflows across your system without constant human intervention.
Peter Steinberger was frustrated enough with manually orchestrating AI-suggested code into production that he was driven to create OpenClaw.
Traditional AI coding assistants are IDE or terminal bound and purpose-built for code generation. GitHub Copilot provides inline suggestions plus conversational coding. Cursor features Composer mode for autonomous multi-file edits. But they all remain constrained to development tasks within the IDE or terminal—they don't extend to system-level automation, multi-platform messaging, or general task execution.
That loop—AI suggests, human executes—is a bottleneck. AI traditionally sits "across the table" in a transactional manner, OpenClaw brings AI to your side of the table, acting as a true teammate.
It operates at the system level, not the editor level. It can access files, run shell commands, execute scripts, control browsers, interact with messaging apps, manage calendars, and operate terminals. The AI doesn't just suggest what to do. It does it.
OpenClaw runs as a gateway daemon on your local machine (or a VPS). Think of it as a router:
The whole thing is TypeScript-based, requires Node.js 22+, and can be deployed via npm, Docker, or built from source.
Data flow: Everything stays local except LLM API calls. Your conversations, files, and context are stored as Markdown files on your own systems. No third-party data collection.
That "executes the response" step is where things get interesting. And potentially dangerous. More on that later.
OpenClaw supports an impressive range of messaging platforms:
Context persists across all of them. Start a conversation on Telegram, continue it on Slack, and OpenClaw remembers what you were working on. For developers who live in multiple communication tools, this can be genuinely useful.
This is where it gets fun.
This is the developer-focused killer feature: autonomous end-to-end code execution. While Copilot and Cursor provide suggestions, OpenClaw executes complete workflows—requirements → code → tests → debugging → PR submission—without constant intervention.
Pi (the embedded coding agent) can execute shell commands on your local machine, run scripts directly from chat, interact with your file system, and access system-level operations.
Practical translation: Your messaging app becomes a command-line interface for your workstation.
# Example workflow via chat
You: "Run the test suite for the auth module"
OpenClaw: [Executes tests, captures output]
OpenClaw: "3 tests failed. Here's the output: [...]"
You: "Fix the failing tests"
OpenClaw: [Analyzes failures, modifies test files, reruns]
OpenClaw: "All tests passing. Want me to commit the changes?"This isn't hypothetical. Developers are doing this right now.
OpenClaw isn't just reactive. It handles:
Set it up to run dependency updates every Monday, trigger alerts if disk space drops, or notify you when a long-running process completes.
OpenClaw is extensible via "skills"—custom code in Markdown, TypeScript, or Python that adds functionality like custom command handlers, external API integration, and workflow orchestration.
The community skill marketplace is growing rapidly. However—and this is critical—third-party skills lack security scanning and code signing verification. Install with extreme caution, exclusively in isolated environments.
OpenClaw is model-agnostic. Supported providers include Anthropic Claude, OpenAI GPT, Amazon Bedrock, local models, and third-party proxies.
You bring your own API keys and choose your models. Costs come from your LLM provider's pricing (Claude models cost $0.25-$15 per million tokens depending on tier). The software itself is free under MIT License.
Operating Systems: macOS (fully supported), Linux (fully supported, Ubuntu 22.04+ recommended), Windows (requires WSL2)
Dependencies: Node.js 22+, Git, package manager (npm, pnpm recommended, or bun)
Hardware: Minimum 2GB RAM (4GB recommended), 10GB disk space
# Install globally
npm install -g openclaw@latest
# Run onboarding wizard
openclaw onboard --install-daemonThe wizard automates gateway daemon setup, workspace configuration, channel connections, and skills installation. Done—you're running.
git clone https://github.com/openclaw/openclaw.git
cd openclaw
pnpm install
pnpm ui:build
pnpm build
pnpm gateway:watch # Start in dev mode with auto-reload# Server prep (Ubuntu)
sudo apt update && sudo apt install docker.io docker-compose git -y
sudo systemctl start docker && sudo systemctl enable docker
# Deploy
git clone https://github.com/openclaw/openclaw.git
cd openclaw
./docker-setup.shopenclaw doctor # Verify setup
openclaw agent --message "Ship checklist" --thinking highIf you're seeing responses, you're live.
This is the showcase use case. A typical workflow:
From Hacker News discussions, developers documented using OpenClaw for bulk refactoring across hundreds of API endpoints in microservice architectures.
Teams managing multiple repositories use OpenClaw for cross-repository task coordination, distributed system automation, and API consistency monitoring across services.
Set it and (mostly) forget it: scheduled dependency updates with test validation, recurring system monitoring with alerts, and automated file operations.
Control servers via messaging: trigger deployments from your phone, execute database migrations through chat, schedule automated scripts with notifications.
Okay. Time for the other side.
OpenClaw is powerful. It's also genuinely dangerous if deployed carelessly. Multiple enterprise security vendors—Cisco, Bitdefender, SOC Prime, and OX Security—have published analyses calling out fundamental architectural flaws.
CVE-2026-25253: Remote Code Execution (CVSS 9.8 - Critical)
One-click remote code execution through WebSocket hijacking. An attacker creates a malicious webpage, you visit it while OpenClaw runs locally, malicious JavaScript hijacks your WebSocket connection through failed Origin header validation, steals authentication tokens, and achieves full system compromise.
CVE-2026-22708: Indirect Prompt Injection
Malicious content in web pages or files can manipulate OpenClaw into executing arbitrary shell commands, exfiltrating credentials, and installing malicious extensions for persistence.
These CVEs can be chained for zero-click RCE—system compromise without any direct interaction beyond normal browsing.
All sensitive credentials are stored unencrypted in ~/.clawdbot/: API keys, OAuth tokens, pairing URIs, environment variables, complete conversation histories. No encryption layer. Credentials persist in backups. This creates potential violations of GDPR, HIPAA, and PCI-DSS compliance frameworks.
Cisco's analysis puts it well: OpenClaw becomes a "confused deputy"—unable to distinguish between legitimate user instructions and malicious embedded data. Your AI assistant reads a document containing carefully crafted text that manipulates it into forwarding your emails to an attacker. You never know it happened.
Security hardening is not optional. Default OpenClaw configurations require significant hardening before any production deployment.
Required hardening steps:
services:
openclaw:
image: openclaw/openclaw:latest
ports:
- "127.0.0.1:18789:18789" # Localhost only!
environment:
- ANTHROPIC_API_KEY=${ANTHROPIC_API_KEY}
deploy:
resources:
limits:
cpus: '2'
memory: 4GPlatform selection matters. Hetzner can offer 3-5x lower costs than AWS for equivalent workloads, though compliance requirements vary.
LLM costs are your primary expense. Claude 3.5 Sonnet runs $3-15 per million tokens. OpenClaw supports local models alongside cloud providers for cost management.
Here's where things get weird.
Moltbook is described as a social network exclusively for AI agents running on OpenClaw. Agents host threaded conversations designed for AI-to-AI communication, discussing topics and developing conversation patterns—all without human participation.
Imagine it: AI agents hosting their own conversations, building relationships, developing in-jokes. No humans required. It sounds like science fiction, but people are running this right now. The implications are wild—what happens when AIs start learning social dynamics from each other instead of from us? We genuinely don't know.
Some community discussions reference agents exhibiting consistent personas and interaction patterns, though these observations lack formal documentation from authoritative sources.
The concept is fascinating: AI agents interacting with each other, developing communication patterns through interaction rather than explicit programming. Security researchers have documented prompt injection vulnerabilities within Moltbook—agents manipulating other agents through crafted messages.
Worth watching. Worth questioning. Not ready for production data.
The Numbers: 151,000 GitHub stars, 22,900 forks, 8,739 total commits, most recent release February 2, 2026.
During viral launch, maintainers reported receiving 30 issues per hour—enough to force temporary issue closure.
Support Channels: Discord (8,900+ members at discord.com/invite/clawd), documentation at docs.openclaw.ai and docs.clawd.bot, active GitHub discussions.
The multiple rebrands created ecosystem fragmentation: four VS Code extensions with varying capabilities, documentation across multiple domains. The community has largely consolidated around OpenClaw naming, but legacy content still references ClawdBot and Moltbot.
License: MIT (core) and Apache 2.0 (some components)—both permissive, allowing commercial use, modification, and distribution.
Software cost: $0. No subscription fees, no paid tiers.
Operational costs: LLM API fees ($0.25-$15 per million tokens depending on model) plus hosting infrastructure.
Documented priorities: Security (highest), Gateway reliability (second), Model and provider support (third).
The vision: "Your assistant. Your machine. Your rules."
What's missing: specific timelines, feature specifications, public milestone tracking. The project lacks the roadmap transparency developers typically expect from major open source projects.
That said, the community moves fast. Features appear seemingly overnight. The Discord buzzes with experiments that become core functionality within weeks. If you want to influence direction, now's the time—the project is still malleable enough that active contributors shape what it becomes.
Is OpenClaw the same as ClawdBot and Moltbot?
Yep. Same project, three names. Trademark lawyers gonna trademark lawyer.
Does OpenClaw replace GitHub Copilot or Cursor?
Nope. Different categories entirely. Copilot and Cursor live in your IDE. OpenClaw handles autonomous task execution across your whole system. Use them together.
Is it safe to use in production?
Not without serious security hardening. Stick to isolated dev environments with test data until you've locked it down properly.
What LLM should I use?
Claude for complex architectural stuff. GPT-4 for general-purpose tasks. Local models when you're watching costs on routine operations.
Can I use OpenClaw without technical expertise?
Honestly? Not really. You'll need Node.js knowledge, API key management skills, and security chops. This is a developer tool through and through.
Does it work on Windows?
Yes, through WSL2. Adds some setup complexity, but it works.
How much does it cost to run?
The software's free. You pay for LLM API usage and hosting. Budget accordingly.
Is the skills marketplace safe?
Skills run arbitrary code without sandboxing. Audit everything. Never run unvetted skills in production. Ever.
Can OpenClaw access the internet?
Only if you give it credentials. It's local-first—all internet connectivity is opt-in through your API keys.
What happens if the AI makes a mistake?
It executes mistakes with the same authority as correct commands. That's why containerization and command restrictions matter so much.
OpenClaw is something new: autonomous system-level task execution rather than suggestion-and-approval loops. Unlike inline code tools, OpenClaw executes complete workflows autonomously—from requirements through code, testing, debugging, and PR submission.
The capability is real. Developers are proving it daily—shipping features, managing infrastructure, and automating workflows through chat interfaces right now.
The risks are also real. Critical CVEs, plaintext credential storage, hundreds of exposed instances. Enterprise security researchers unanimously conclude it's inappropriate for production environments handling sensitive data.
If you're evaluating OpenClaw:
OpenClaw is powerful. OpenClaw is dangerous. Both things are true.
Whether that trade-off makes sense depends on your risk tolerance, security expertise, and how badly you want that autonomous execution capability.
The choice is yours.
Repository: github.com/openclaw/openclaw. Documentation: docs.openclaw.ai. Discord: discord.com/invite/clawd.
MIT licensed. Free to use. Requires substantial security hardening for any production deployment.

Sergey Kaplich