OpenClaw Setup Guide
Secure configuration, 20 productivity use cases with risk assessments, pentester review, and Claude Code comparison.
March 2026 · Based on OpenClaw v2026.3.12+
About this guide
OpenClaw is the fastest-growing open-source project on GitHub (247k+ stars in ~60 days), but its power comes with significant security implications. This guide covers secure setup, the top 20 productivity use cases with detailed risk assessments, and a pentester review of the recommendations.
Part 1: Secure & Private Setup Guide
1.1 Understanding the Architecture
OpenClaw is a long-running Node.js service that connects LLMs (Claude, GPT, DeepSeek, local models) to your local machine and messaging apps (Signal, Telegram, Discord, WhatsApp). Key architectural facts:
- Local-first but NOT local-processing: The core app runs on your hardware, but prompts still travel over the public internet to LLM API endpoints
- Gateway model: OpenClaw’s Gateway listens on port 18789 via WebSocket — this is the primary attack surface
- Skills system: Functionality comes from skill directories containing SKILL.md files — these are a supply chain risk vector
- Persistent memory: Stores context across sessions in SOUL.md and MEMORY.md files
1.2 Hardware & Environment Setup
Option A: Dedicated Cloud VPS (Recommended)
- Use a dedicated VPS (e.g., DigitalOcean, Hetzner, Contabo — 2vCPU, 8GB RAM, ~$5–10/month)
- This isolates OpenClaw from your personal devices entirely
- Run inside Docker for additional containment
Option B: Local Machine (Convenient but higher risk)
- Use a dedicated VM or container — never run directly on your main OS
- Microsoft explicitly recommends: “not appropriate to run on a standard personal or enterprise workstation”
Option C: Dedicated Physical Device
- Repurpose an old laptop or use a Raspberry Pi 5 (8GB)
- Air-gap from sensitive networks
1.3 Installation Steps
- Create a dedicated non-root user:
sudo adduser openclaw --disabled-password
sudo su - openclaw
- Install via official script (always verify the script first):
# Download and review before piping to shell
curl -fsSL https://get.openclaw.ai -o install.sh
less install.sh # REVIEW IT
bash install.sh
- Run the onboarding wizard — it generates a gateway token by default
- Immediately run security audit:
openclaw security audit --fix
1.4 Network Security Hardening
- Never expose the Gateway to the public internet directly
- Disable mDNS broadcasting:
export OPENCLAW_DISABLE_BONJOUR=1
This prevents leaking filesystem paths and hostname info on your local network.
- Firewall rules: Block port 18789 from all external access
sudo ufw deny 18789
- Use a reverse proxy with TLS if you must expose it (Caddy or nginx with Let’s Encrypt)
1.5 Authentication & Credential Management
- Use strong gateway tokens — the wizard generates one, but verify it’s not weak
- Scoped API tokens: For every integration (Gmail, GitHub, etc.), create tokens with minimum required permissions (read-only where possible)
- Short-lived credentials over long-lived ones wherever possible
- Never store secrets in .env files the agent can read — use a secrets manager or OS keychain
- Disable memory storage of secrets: Ensure SOUL.md/MEMORY.md never contain API keys or passwords
1.6 Skill & Supply Chain Security
Critical: ~20% of ClawHub skills are malicious
As of Feb 2026, the “ClawHavoc” campaign identified 824+ malicious skills out of 10,700+ on ClawHub.
- Audit every skill before installing: Read the SKILL.md and all associated code manually
- Only install skills from verified/trusted authors
- Never install skills that request exec/shell access unless you’ve audited them thoroughly
- Pin skill versions — don’t auto-update
- Run
openclaw security audit after every skill install
- Consider writing your own skills for sensitive workflows rather than trusting third-party ones
1.7 LLM API Key Management
- Create a dedicated API key for OpenClaw (not your personal/main key)
- Set spending limits on the API key ($20–50/month to start)
- Monitor usage dashboards for anomalous spikes (could indicate prompt injection exfiltrating data)
- Rotate keys monthly
1.8 Ongoing Maintenance
60+ CVEs and 60+ GHSAs in ~4 months
Including CVE-2026-28363 (CVSS 9.9) and CVE-2026-25253 (CVSS 8.8). Keep OpenClaw updated.
- Run
openclaw security audit --fix regularly (weekly minimum)
- Monitor the OpenClaw GitHub security advisories
- Review SOUL.md/MEMORY.md periodically for injected content (memory poisoning attacks)
- Check logs for suspicious activity
↑ Back to top
Part 2: Top 20 Use Cases with Risk Assessments
1 Email Triage & Morning Briefing
HIGH RISK 30–45 min/day saved
OpenClaw reads your inbox, summarizes emails by priority, and sends you a digest via Telegram/Signal each morning.
Risks
- Email contents (potentially sensitive/confidential) sent to an external LLM API
- Email credentials stored locally could be exfiltrated via prompt injection
- Malicious emails could contain prompt injection payloads that hijack the agent
Mitigations
- Use a read-only email app password / OAuth token with minimal scopes
- Filter out emails from unknown senders before processing
- Use a dedicated email account for testing initially
- Consider a local LLM (Llama, Mistral) to keep data on-premise
2 Calendar Management & Scheduling
MEDIUM RISK 15–30 min/day saved
Auto-schedule meetings, detect conflicts, send scheduling links, and provide daily agenda briefings.
Risks
- Calendar data reveals your location, contacts, and daily patterns
- Write access could delete or modify meetings (accidentally or via injection)
- Meeting details sent to LLM API
Mitigations
- Start with read-only calendar access
- Require human confirmation for any write operations
- Use scoped OAuth tokens
3 Code Review & PR Summaries
MEDIUM RISK 1–2 hrs/day saved
Automatically summarize PRs, flag potential issues, and post review comments on GitHub.
Risks
- Source code (potentially proprietary) sent to external LLM
- Write access to GitHub could merge bad PRs or delete branches
- Prompt injection via malicious PR descriptions or code comments
Mitigations
- Use a GitHub token scoped to specific repos with minimal permissions
- Never grant merge/admin permissions to the agent’s token
- Review agent-generated comments before they’re posted
- For proprietary code, use a local LLM or provider with a DPA
4 Automated DevOps & Deployment
VERY HIGH RISK 1–3 hrs/day saved
Run test suites, deploy to staging/production, monitor CI/CD pipelines via chat commands.
Risks
- Shell command execution on your infrastructure — the most dangerous capability
- A compromised agent could deploy malicious code, delete databases, or exfiltrate secrets
- Production credentials exposed to the agent
Mitigations
- NEVER give the agent production deployment access directly
- Use a dedicated CI/CD service as an intermediary — agent triggers, CI deploys
- Restrict exec permissions to a whitelist of safe commands
- Use staging-only access initially
- Require human approval for any deployment actions
- Run the agent in a container with no network access to production systems
5 Meeting Transcription & Action Items
HIGH RISK 30–60 min/meeting saved
Transcribe meeting recordings, extract action items, and email summaries to participants.
Risks
- Meeting recordings contain sensitive business discussions
- Audio/transcription data sent to external APIs
- Action items could be mis-attributed or fabricated (hallucination)
Mitigations
- Use local transcription (Whisper) before sending text to LLM
- Review action items before sending to participants
- Get consent from all meeting participants
- Consider legal/compliance requirements for recording
6 Social Media Content & Scheduling
MEDIUM RISK 5–10 hrs/week saved
Generate platform-specific posts from blog content, schedule posts, and manage multiple accounts.
Risks
- Agent could post inappropriate/incorrect content under your brand
- Social media credentials exposed
- Generated content could contain hallucinations or plagiarized text
Mitigations
- Always require human approval before posting
- Use a content queue/draft system, not direct posting
- Use platform-specific API tokens with limited scopes
7 Research & Web Monitoring
LOW-MEDIUM RISK 1–2 hrs/day saved
Monitor specific topics, competitors, or news sources and provide summarized briefings.
Risks
- Web browsing exposes the agent to prompt injection via malicious web content
- Headless browser could be exploited to visit phishing sites
- Research summaries may contain hallucinations presented as facts
Mitigations
- Run the browser in a sandboxed container (Chromium in Docker)
- Restrict browsing to a whitelist of trusted domains
- Cross-reference important findings manually
- Disable JavaScript execution for simple content scraping
8 Personal Finance Tracking
VERY HIGH RISK 2–3 hrs/week saved
Aggregate bank transactions, categorize spending, generate budget reports.
Risks
- Financial data is extremely sensitive — bank credentials, transaction history, account numbers
- Data sent to external LLM API could be logged or breached
- Compromised agent could initiate transfers if write access is granted
- Regulatory implications (PCI-DSS, financial data protection laws)
Mitigations
- Use read-only bank API access (Plaid with read-only scope)
- NEVER grant transfer/payment capabilities
- Use a local LLM — do not send financial data to cloud APIs
- Export transactions as CSV and process offline
- Encrypt all stored financial data at rest
9 Customer Support Triage
HIGH RISK 3–5 hrs/day saved
Monitor support channels, categorize issues, draft responses, escalate urgent items.
Risks
- Customer PII sent to external LLM
- Incorrect auto-responses could damage customer relationships
- Prompt injection via customer messages could hijack the agent
- GDPR/CCPA compliance issues
Mitigations
- Always require human approval for outgoing responses
- Strip PII before sending to LLM (use regex/NER pre-processing)
- Use an LLM provider with a DPA (Data Processing Agreement)
- Log all agent actions for audit trail
10 Document Drafting & Editing
MEDIUM RISK 2–5 hrs/week saved
Draft contracts, proposals, reports from templates and context.
Risks
- Confidential business information sent to LLM API
- Hallucinated legal/financial terms could have real consequences
- Document content stored in agent memory could leak across sessions
Mitigations
- Never use for final legal documents without lawyer review
- Clear agent memory after processing sensitive documents
- Use document templates to constrain output
11 Smart Home Automation
HIGH RISK 15–30 min/day saved
Control lights, thermostats, locks, cameras via natural language through messaging apps.
Risks
- Physical security implications — agent could unlock doors, disable cameras
- Smart home credentials exposed to the agent
- Prompt injection could trigger physical actions
Mitigations
- Never connect door locks or security systems to OpenClaw
- Limit to non-critical devices (lights, thermostats, speakers)
- Require 2FA or physical confirmation for security-related actions
- Use a separate smart home hub with its own auth layer
12 Expense Report Automation
MEDIUM RISK 2–3 hrs/week saved
Scan receipts, categorize expenses, fill out expense report forms, submit for approval.
Risks
- Receipt images may contain personal credit card numbers
- Incorrect categorization could cause tax/compliance issues
Mitigations
- Mask credit card numbers before processing
- Always require human review before submission
- Use local OCR (Tesseract) for receipt scanning
- Maintain audit trail of all agent decisions
13 Data Entry & Form Filling
MEDIUM RISK 3–5 hrs/week saved
Extract data from PDFs/emails and populate spreadsheets, CRMs, or databases.
Risks
- Data accuracy — incorrect entries could propagate through systems
- Write access to databases/CRMs is a significant attack surface
Mitigations
- Implement validation rules before any database writes
- Use staging/draft records that require human approval
- Log all changes with before/after values for rollback
- Start with low-stakes data entry
14 Personal Knowledge Base / Second Brain
LOW RISK 1–2 hrs/day saved
Automatically organize notes, bookmarks, highlights into a searchable knowledge base (Notion, Obsidian).
Risks
- Personal notes sent to external LLM for processing
- Memory poisoning — malicious content injected into your knowledge base
Mitigations
- Use a local LLM for processing personal notes
- Periodically review the knowledge base for suspicious content
- Keep original source material alongside AI summaries
- Use read-only access to source apps where possible
15 Competitive Intelligence
LOW-MEDIUM RISK 3–5 hrs/week saved
Monitor competitor websites, pricing, job postings, press releases, and social media activity.
Risks
- Web scraping may violate terms of service
- Competitor websites could contain prompt injection
- Legal risk if scraping crosses into unauthorized access
Mitigations
- Only monitor publicly available information
- Verify critical intelligence manually
- Use sandboxed browser with restricted JavaScript
- Consult legal counsel on scraping legality
- Label all AI-generated insights as unverified
16 Automated Reporting & Dashboards
MEDIUM RISK 3–5 hrs/week saved
Pull data from multiple sources, generate weekly/monthly reports.
Risks
- Business metrics and KPIs sent to external LLM
- Incorrect data aggregation could lead to bad decisions
Mitigations
- Use read-only database replicas or API tokens
- Validate generated reports against known benchmarks
- Use pre-built SQL queries rather than letting the agent write arbitrary queries
- Review reports before distribution
17 Language Translation & Localization
LOW RISK 2–4 hrs/week saved
Translate documents, emails, and content between languages with context awareness.
Risks
- Confidential documents sent to external LLM
- Mistranslations in legal/medical/technical contexts could be serious
Mitigations
- Have native speakers review critical translations
- Use domain-specific glossaries to improve accuracy
- Never use for legally binding translations without professional review
18 Automated Testing & QA
MEDIUM RISK 2–4 hrs/day saved
Generate test cases, run test suites, report bugs, and track regression.
Risks
- Source code and test data sent to external LLM
- Shell command execution for running tests
- Generated tests may create false confidence
Mitigations
- Use sanitized test data only
- Restrict exec permissions to test commands only
- Review generated test cases before adding to test suite
- Run in an isolated CI environment
19 Invoice Processing & Accounts Payable
HIGH RISK 5–10 hrs/week saved
Extract data from invoices, match to POs, route for approval, update accounting system.
Risks
- Financial documents contain sensitive vendor and company info
- Incorrect processing could result in wrong payments
- Fraudulent invoices could bypass detection if agent is compromised
Mitigations
- Always require human approval for payment-related actions
- Multi-person approval for amounts above a threshold
- Use local processing for OCR/extraction
- Cross-reference invoices against PO database automatically
- Maintain complete audit trail
20 Personal Health & Fitness Tracking
HIGH RISK 30–60 min/day saved
Aggregate health data from wearables, track habits, generate wellness reports.
Risks
- Health data is extremely sensitive (HIPAA, GDPR special category)
- Incorrect health advice could be harmful
- Wearable data includes location, heart rate, sleep patterns
- Health data could be used for insurance discrimination
Mitigations
- Use a local LLM exclusively — never send health data to cloud APIs
- Never act on AI health suggestions without consulting a healthcare professional
- Use read-only access to health platforms
- Anonymize/aggregate data before processing
↑ Back to top
Part 3: Risk Summary Matrix
| Risk Level | Use Cases |
| VERY HIGH | #4 DevOps/Deployment, #8 Personal Finance |
| HIGH | #1 Email, #5 Meeting Transcription, #9 Customer Support, #11 Smart Home, #19 Invoice Processing, #20 Health Tracking |
| MEDIUM | #2 Calendar, #3 Code Review, #6 Social Media, #10 Documents, #12 Expenses, #13 Data Entry, #16 Reporting, #18 Testing |
| LOW-MEDIUM | #7 Research/Monitoring, #15 Competitive Intel |
| LOW | #14 Knowledge Base, #17 Translation |
↑ Back to top
Part 4: Universal Security Principles
- Start small: Begin with low-risk use cases (#14, #17, #7) and build confidence
- Principle of least privilege: Every integration gets the minimum permissions needed
- Human in the loop: Require approval for all write/send/deploy actions initially
- Defense in depth: Container + firewall + auth + scoped tokens + monitoring
- Assume compromise: Design your setup so that even if the agent is hijacked, the blast radius is limited
- Local LLMs for sensitive data: Use Ollama + Llama/Mistral for financial, health, or confidential data
- Regular audits: Run
openclaw security audit weekly, review memory files monthly
- Stay updated: 60+ CVEs in 4 months (including CVSS 9.9) — always run the latest version
- Vet all skills: Never install unaudited third-party skills (~20% of ClawHub skills were malicious)
- Monitor costs: Set API spending limits to detect exfiltration-via-API-abuse early
↑ Back to top
Part 5: Verification Checklist
After setup, verify security by:
- Running
openclaw security audit --fix and confirming zero findings
- Port scanning your host to confirm 18789 is not externally accessible
- Testing that the gateway rejects unauthenticated WebSocket connections
- Verifying API key spending limits are set
- Confirming mDNS is disabled (
OPENCLAW_DISABLE_BONJOUR=1)
- Testing a benign prompt injection to see if the agent’s guardrails hold
- Reviewing SOUL.md and MEMORY.md for any unexpected content
- Verifying
allow_url_actions: false in openclaw.yaml to block the one-click RCE vector
- Confirming you’re running version 2026.3.12+ (patches all known CVEs)
- Testing that WebSocket connections from non-localhost origins are rejected
- Verifying Docker container runs as non-root with
--cap-drop=ALL and --read-only
↑ Back to top
Part 6: Pentester / Whitehat Review
Upgraded Risk Assessments
Based on the full CVE landscape and real-world attack data, these risk levels should be upgraded:
| Use Case | Original | Pentester | Justification |
| #3 Code Review | MEDIUM | HIGH | Prompt injection via PR descriptions is a proven vector; source code exfiltration is high-value |
| #7 Research | LOW-MED | MED-HIGH | Web browsing is the #1 indirect prompt injection vector |
| #14 Knowledge Base | LOW | MEDIUM | Memory poisoning makes this a persistence vector for attackers |
Critical Attack Vectors Missing from Original Guide
1. The “Lethal Trifecta” (Penligent/OctoClaw research)
The combination of System Access + Execution Power + Untrusted Ingestion creates a perfect storm. Most OpenClaw setups have all three by default. Never combine all three in a single agent instance. Use separate agents with different permission levels.
2. Time-Shifted Memory Poisoning (Palo Alto Networks)
Malicious inputs written to SOUL.md/MEMORY.md can appear benign at ingestion but “detonate” later when the agent’s state aligns — logic bomb-style. Palo Alto mapped OpenClaw to every category in the OWASP Top 10 for Agentic Applications.
Mitigation: Treat memory files as code, not data. Use file integrity monitoring (FIM), enforce read-only permissions during runtime, require admin approval for memory file changes.
3. Log Poisoning → Indirect Prompt Injection
Attackers can write malicious content to log files via WebSocket requests. Since the agent reads its own logs, this is an injection vector. Mitigation: Ensure logs are write-only from the agent’s perspective; use a separate log viewer.
4. Container Escape via API
CVE-2026-25253’s WebSocket hijacking works even inside Docker containers — the container boundary doesn’t block it. A containerized install running a vulnerable image is just as exposed as bare-metal.
5. Localhost Trust Assumption (ClawJacked)
OpenClaw’s gateway exempted localhost from rate limiting entirely. Browser JavaScript could brute-force the gateway at hundreds of guesses/second with no lockout and no logging.
Docker Hardening Deep Dive
# docker-compose.yml — hardened OpenClaw deployment
version: '3.8'
services:
openclaw:
image: openclaw/openclaw:2026.3.12 # Pin to patched version
read_only: true # Read-only root filesystem
cap_drop:
- ALL # Drop all Linux capabilities
security_opt:
- no-new-privileges:true # Prevent privilege escalation
- seccomp=openclaw-seccomp.json # Custom seccomp profile
user: "1000:1000" # Non-root user
ports:
- "127.0.0.1:18789:18789" # Bind to loopback ONLY
volumes:
- ./config:/app/config:ro # Config read-only
- ./data:/app/data # Data writable (minimal)
mem_limit: 2g # Memory cap
cpus: 1.0 # CPU cap
networks:
- openclaw-internal
# Optional: egress proxy to restrict outbound traffic
egress-proxy:
image: nginx:alpine
# Allowlist only LLM API endpoints
networks:
- openclaw-internal
- external
networks:
openclaw-internal:
internal: true # No direct internet access for OpenClaw
external:
driver: bridge
Key points:
internal: true network prevents the agent from reaching the internet directly — all traffic goes through an egress proxy
- Even a fully compromised agent cannot exfiltrate data to arbitrary servers
- Never mount
docker.sock — it grants container escape
- Never use
--network=host — it removes network isolation entirely
Incident Response Plan
If you suspect compromise:
- Immediately: Disconnect OpenClaw from the network (stop container or kill process)
- Within 1 hour: Rotate ALL credentials the agent had access to
- Audit: Check logs for unexpected
system.run calls; review sandbox list state
- Review: Examine SOUL.md/MEMORY.md for injected instructions
- Scan: Check for installed backdoor skills or modified skill files
- Report: If you find evidence of exploitation, report to OpenClaw security (GitHub Security Advisories)
Pentester Mitigation Ratings by Use Case
| Use Case | Rating | Key Gap |
| #1 Email | NEEDS IMPROVEMENT | Never open attachments via agent; email is the #1 injection delivery vector |
| #2 Calendar | SUFFICIENT | — |
| #3 Code Review | NEEDS IMPROVEMENT | Sanitize PR descriptions before LLM processing; use separate browser profile |
| #4 DevOps | SUFFICIENT | — |
| #5 Meetings | SUFFICIENT | — |
| #6 Social Media | SUFFICIENT | — |
| #7 Research | NEEDS IMPROVEMENT | Never browse in same session as OpenClaw Control UI |
| #8 Finance | SUFFICIENT | — |
| #9 Support | NEEDS IMPROVEMENT | Rate-limit outgoing responses; implement canary tokens |
| #10 Documents | SUFFICIENT | — |
| #11 Smart Home | NEEDS IMPROVEMENT | Should be VERY HIGH; physical security implications are severe |
| #12–13 | SUFFICIENT | — |
| #14 Knowledge Base | NEEDS IMPROVEMENT | File integrity monitoring; memory poisoning is a real vector |
| #15–18 | SUFFICIENT | — |
| #19 Invoices | NEEDS IMPROVEMENT | Anomaly detection on invoice amounts; AP workflows are a target |
| #20 Health | SUFFICIENT | — |
Real-World Scale of the Problem
As of March 2026:
- 220,000+ OpenClaw instances exposed to the internet
- 12,812 confirmed exploitable via RCE
- 63% of observed deployments were vulnerable
- 60+ CVEs and 60+ GHSAs disclosed in ~4 months
- 900+ malicious ClawHub skills identified (Atomic Stealer payloads)
- Meta has banned OpenClaw from corporate devices
- China restricted OpenClaw in government agencies
- Palo Alto Networks called it “the potential biggest insider threat of 2026”
↑ Back to top
Part 7: OpenClaw vs Claude Code
What is Claude Code?
Claude Code is Anthropic’s official CLI and agentic coding tool. It runs in your terminal, IDE (VS Code, JetBrains), and now as a desktop app (Cowork). Unlike OpenClaw’s messaging-app-first approach, Claude Code is purpose-built for software development with deep filesystem and git integration.
New in 2026: Channels & Cowork
Claude Code Channels (launched March 20, 2026): Bridges messaging platforms (Telegram, Discord, Slack) to your Claude Code CLI session. This directly competes with OpenClaw’s core messaging-app model.
Claude Cowork (launched 2026): The desktop agent for knowledge workers. Same agentic architecture as Claude Code but with a desktop interface.
Agent Teams (experimental): Coordinates multiple Claude Code instances working together via CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS.
Head-to-Head Comparison
| Dimension | Claude Code | OpenClaw |
| Primary interface | Terminal + IDE + Desktop | Messaging apps (Signal, Telegram, WhatsApp, Discord) |
| Messaging integration | Channels (new March 2026) | Native, core architecture |
| Security model | Sandboxed; explicit per-action approval | Broad default permissions; you configure restrictions |
| Enterprise compliance | SOC 2, GDPR, SSO, RBAC, audit trails | None |
| CVEs (first 4 months) | Managed by Anthropic security | 60+ CVEs, 60+ GHSAs, CVSS 9.9 |
| LLM flexibility | Claude models only | Any LLM (Claude, GPT, DeepSeek, Llama, etc.) |
| Cost | ~$20/month or API usage-based | Free (MIT) + LLM API costs |
| Multi-agent | Agent Teams (experimental) | Single agent with skill integrations |
| Coding depth | Purpose-built, git-native | General-purpose |
| Non-coding tasks | Cowork for knowledge work | Native strength (email, calendar, smart home, etc.) |
| Customization | CLAUDE.md, hooks, MCP servers | SKILL.md, SOUL.md, full source access |
| Self-hosting | No — Anthropic API | Yes — fully on your hardware |
| Supply chain risk | MCP servers (smaller ecosystem) | ClawHub (~20% malicious) |
When to Use Claude Code
- Software development: Code review, refactoring, debugging, architecture
- Enterprise/regulated environments: SOC 2, GDPR, audit trails, SSO/RBAC
- Security-conscious teams: Sandboxed execution, managed by Anthropic’s security team
- Git-heavy workflows: Native git integration, PR management
- Managed security: No CVE patching or infrastructure hardening required
When to Use OpenClaw
- Life automation beyond coding: Email, calendar, smart home, finance, health
- Messaging-native workflows: If your team lives in Telegram/Signal/WhatsApp
- LLM flexibility: Local models, DeepSeek, or mixed providers
- Full data sovereignty: Run entirely on your hardware with local LLMs
- Maximum customization: Full source access, custom skills, custom personality
The emerging pattern: Use both
Many developers run Claude Code as their coding engine and OpenClaw for life/ops orchestration. Claude Code Channels gives you messaging-app access to a coding agent. OpenClaw gives you messaging-app access to a general-purpose agent with system-level access. The security implications of that distinction are enormous.
↑ Back to top
Part 8: Steinberger’s Perspective
Based on Peter Steinberger’s public statements — his blog (Feb 15, 2026), Pragmatic Engineer podcast, Lex Fridman #491, Fortune/TechCrunch/CNBC interviews, and GitHub activity.
His Philosophy: Empowerment, Not Containment
Steinberger positions OpenClaw as “an AI that actually does things” — a revolutionary tool for power users. His vision is building “an agent that even my mum can use.” He frames AI agents as a new skill: “You pick up the guitar — you’re not going to be good at the guitar in the first day.” He calls this approach “agentic engineering” and explicitly distinguishes it from “vibe coding.”
Where Steinberger Would Push Back
- The tone is too CISO-like: He’d lead with the empowerment narrative before the risk narrative.
- Absolutist “NEVER” prohibitions: He favors graduated trust models over blanket bans. Instead of “NEVER connect door locks,” he’d say “start read-only, level up as you validate.”
- The Microsoft quote: He’d note this is Microsoft’s external security team’s view, not universal consensus. His entire pitch is that OpenClaw runs on YOUR machine.
- Model interchangeability: A core design principle — route sensitive tasks to local models and mundane tasks to cloud APIs on the same agent.
Where Steinberger Acknowledges the Risks
- Announced moving OpenClaw to an independent open-source foundation (Apache/PSF model) on Feb 14, 2026
- Active in GitHub Security Advisory work, though critical of GitHub’s reporting system as “a mess” with “AI-generated slop”
Uncomfortable Truths
- ClawHavoc response was slow: Steinberger was criticized for saying he was “too busy” to address malicious skills. C2 infrastructure stayed operational for days.
- Community as security asset: ClawCon drew 1,000+ attendees; Cisco’s DefenseClaw and VirusTotal scanning integration came from community contributions.
- One maintainer’s candid warning (Discord): “If you can’t understand how to run a command line, this is far too dangerous of a project for you to use safely.”
The Balanced Take
Steinberger is more optimistic about OpenClaw’s risk-reward tradeoff than the security community. This guide deliberately errs on the side of security researchers (Microsoft, Palo Alto Networks, Cisco). That’s the right call for a security-focused setup guide. But Steinberger’s graduated-trust approach (“start minimal, expand carefully”) is valid for experienced users who understand the risks.
Sources
↑ Back to top