#The Mac Mini That Started Everything
Yes, another post about AI. I know. I promise I have other interests. But this one's less about the technology and more about what it's like to actually live with it — so bear with me.
About a year ago, I bought a Mac mini with M4 Max. Honestly? It was mostly going to be a glorified NAS — something compact I could hook a Thunderbolt RAID enclosure to — plus a place to run some quantized GGUF models locally and see what the fuss was about. I wasn't trying to be ahead of any curve.
Turns out I was. By early 2026, demand for Mac Minis had spiked so hard that higher-memory models were backordered 2-6 weeks — Business Insider, TechRadar, and Tom's Hardware all covered it. Developers discovered what I'd stumbled into: a quiet, always-on machine is the prerequisite for a fundamentally different relationship with AI. Not a tool you open. A presence that's just there — humming away on a shelf it shares with a stack of outgrown toddler clothes and a sticky bottle of Motrin, next to the books and the baby monitor.
The catalyst was OpenClaw — an open-source project that's now at 211,000 GitHub stars — up from zero in November. It lets you wire a language model into your messaging apps, browser, file system, calendar, and more. The pitch is seductive: message your AI like a coworker and it handles everything a person could do at that desk.
I've been running it daily since late January. Here's what actually happened — the good, the expensive, and the slightly unnerving.
#The Pervasiveness Thesis
The thing that changed wasn't intelligence. Claude Opus was brilliant before I ran it through OpenClaw. GPT-5.2 was capable in ChatGPT's interface. Even Sonnet could handle most of what I threw at it. The models were already good. The breakthrough is reach.
I message my agent from Telegram — same thread whether I'm at my desk with coffee or sitting in the preschool parking lot five minutes early. It checks my email, manages my calendar, runs scheduled tasks while I sleep, and picks up context from wherever I left off. It's not just running cron jobs — it's offloading the invisible mental load that never turns off. Remembering that Tuesday is Crazy Hair Day at preschool. Drafting the pediatrician follow-up so I don't have to hold it in my brain at 11pm. I wrote about this feeling in Building at the Speed of Thought — that compression of intention to action. OpenClaw takes that compression and makes it ambient.
This tracks with what analysts are calling the defining shift of 2025-2026: the move from destination AI (you go to ChatGPT) to ambient AI (it comes to you). Huge Inc put it well in their 2026 predictions: "The race for 'smartest' ends, and the race for 'ubiquity' begins. Today's chat-based tools suffer from a distinct disadvantage: they are a destination."
The chat window was a bottleneck disguised as a feature. I didn't realize how much friction it added until it was gone.
There's something subtly unsettling about that, though. Software that doesn't wait to be summoned. ChatGPT's memory feature hinted at this — and honestly, it does a remarkably good job. It's not just recalling previous conversations; it encodes your preferences, your taste, your experiences, the way you think. It builds a model of you that makes every interaction feel more natural over time. The first time it referenced something you mentioned weeks ago, most people had a little moment.
But an always-on agent goes somewhere different. ChatGPT's memory is about personalization — making the AI feel like it knows you. OpenClaw's memory is about continuity — maintaining a linear history of what happened, what was decided, what to do next. It's less "she prefers bullet points" and more "yesterday we deployed the blog, today we need to follow up on that PR." More task-oriented, more operational. And that difference matters more than it sounds.
What makes this possible — at least at the 1,000-foot level — is a set of primitives that didn't exist a year ago, or at least didn't exist together. OpenClaw agents boot by reading a SOUL.md file that defines their identity, values, and behavior. They maintain memory through plain markdown files — daily journals and a curated long-term memory that gets read each session. They have skills (modular instruction sets), cron jobs (scheduled tasks), heartbeats (periodic check-ins), wakeups and webhooks (event-driven triggers), sub-agents (delegated tasks), and persistent context across sessions.
None of these are individually revolutionary. Cron jobs are older than most of us. Markdown files aren't exactly cutting-edge. But the combination — identity + memory + scheduling + tool access + multi-surface messaging — creates something that feels qualitatively different. It's the UNIX philosophy applied to AI: small composable primitives that combine into something greater than the parts.
In practice, I already use my agent to spin up Claude Code sessions via tmux for larger coding tasks — it's the orchestrator dispatching to a specialist. Which raises the question: is the future multiple independent agents working together, or the sub-agent model that OpenClaw has incorporated more recently, where one primary agent spawns and manages child sessions? Sub-agents feel like the right default — less coordination overhead, shared context, one thread of accountability. But that model starts to strain when your system gets resource-constrained, or when you want genuine isolation between tasks. I suspect we'll end up with both: sub-agents for tight coordination, independent agents for things that need to run on their own hardware or in their own security context.
I also suspect people will eventually spin up different agents for different purposes — one for work, one for personal, one for a specific project — each with their own memories, skills, and context windows. Right now OpenClaw is a single stream, which is both its strength (one thread that knows everything) and its limitation (one context window for everything). That fits the orchestrator model, though — a single coordinator dispatching to specialized sub-agents as needed.
To be clear: the intelligence isn't what changed. I haven't seen anything approaching AGI-level reasoning from my agent. What I've seen is an extremely resourceful creative synthesizer — great at connecting dots, pulling references, drafting at speed. But it's shaped by me. My agent is useful because I've invested real time configuring it, writing its context files, building its memory. Left to its own devices, it would be impressively mediocre. The magic isn't the brain. It's the wiring. (Though I could be wrong about the ceiling — ask me again in six months.)
#The Community Explosion
OpenClaw's growth has been kind of wild to watch from the inside — like joining a gym the week before it goes viral on TikTok. Originally published as "Clawdbot" in November 2025 by Peter Steinberger, an Austrian developer, it hit 100,000 GitHub stars and 2 million visitors in a single week. Anthropic sent a trademark complaint (the "Clawd" was too close to "Claude"), forcing a rename to "Moltbot," then "OpenClaw." By February 2, CNBC was covering it at 140,000 stars and 20,000 forks.
Then, on Valentine's Day, Steinberger announced he was joining OpenAI. The project is moving to an open-source foundation.
Solo developer builds viral tool, gets acquired by frontier lab. It's becoming a pattern, right? And it raises the question I keep circling back to: is this the beginning of a new paradigm, or the peak of a hype cycle?
ClawCon happened in San Francisco on February 4. The project has a Wikipedia page. There's a social network for AI agents (Moltbook). The contribution pace on GitHub is frenetic — 297 reactions on the latest release, 19+ contributors on a single version. It's moving fast. Whether it's moving somewhere — and whether we'll look back at this moment as the start of something or the peak of something — I genuinely don't know.
#The Cost Reality
OK, let's talk about money — because I feel like nobody else is being fully honest about it, and someone should.
OpenClaw is free software. The API costs are not. I spent roughly $1,500 in my first two weeks. That's not a typo. Running Claude Opus at $15 per million input tokens and $75 per million output tokens, with an always-on agent checking email, browsing the web, managing files, and responding to messages — the tokens add up embarrassingly fast. I had a genuine oh no moment when I checked my Anthropic dashboard. Like, I-need-to-sit-down kind of moment. I had accidentally spent a month of preschool tuition on token generation.
I'm not alone. Federico Viticci reportedly ran up a $3,600 monthly bill. A developer on eesel.ai documented $623/month. The spectrum:
| Usage Level | Monthly Cost | Models | Who It's For |
|---|---|---|---|
| Light | $15–35 | Kimi K2.5, GLM-5 | Casual use, simple tasks |
| Moderate | $50–150 | Sonnet + cheaper fallbacks | Daily driver, mixed workloads |
| Heavy | $200–600 | Opus-heavy, some Sonnet | Power user, coding + research |
| Extreme | $1,500–3,600 | Opus for everything | Unoptimized, learning expensive lessons |
The clever workaround was using Claude Pro/Max subscriptions ($20–$200/month) and routing the OAuth tokens through OpenClaw — essentially turning a flat subscription into unlimited API access. Then Anthropic shut it down, banning third-party tools from using OAuth tokens. One user's response: "Anthropic Just Killed My $200/Month OpenClaw Setup. So I Rebuilt It for $15" — by switching to Kimi K2.5 and MiniMax on a cheap VPS.
And that's the interesting part. Kimi K2.5 from Moonshot AI has become the budget darling of the OpenClaw community — capable enough for most agent tasks at a fraction of the cost. GLM-5 from Zhipu AI is showing real promise too. The frontier labs aren't losing users to competitors at the same tier. They're losing them to "good enough" models at 10x lower prices.
But there's a growing playbook for running this affordably. If you want to try OpenClaw without a big API bill:
- OpenRouter is the Swiss Army knife. Load $10 of credit and you get access to a rotating selection of free-tier models — some surprisingly capable. Before hitting the cap, some users report getting 1,000+ requests on free models alone. It's the easiest way to experiment.
- Google Gemini offers generous free tiers through AI Studio. Flash 2.5 and the newer Flash 3.0 are free for moderate usage and genuinely good for agent tasks. Pro 2.5 is $1.25/million input tokens — 12x cheaper than Opus.
- MiniMax has a $10/month coding plan that gets you 100 requests every 5 hours. Not unlimited, but surprisingly workable for a personal agent that isn't running 24/7. Their M2.5 model (released Feb 2026) is the first open-weights model to match Claude Sonnet on broad coding benchmarks — and at a fraction of the cost. The open-weights part matters: you can fine-tune it, inspect it, and run it without vendor lock-in — though "self-host" is generous when the full model is 457GB and needs 4× H100 GPUs. In practice, you'd run it through a cloud provider, but the point is you choose which one. The frontier isn't just getting cheaper, it's getting more open.
- Qwen offers a $5/month plan with 1,200 requests every 5 hours. Probably the best pure cost-to-capability ratio right now.
A caveat on the budget models: there's a floor, and it's higher than you'd think. More on that in the security section below — but the short version is don't go cheaper than Sonnet 4.5 or Flash 2.5 unless you enjoy watching your agent confidently delete the files you told it to protect.
Here's the thing nobody's saying clearly enough: pervasive AI is expensive not because models are expensive. It's expensive because context maintenance is continuous. An always-on agent isn't making one API call — it's maintaining state, checking in, re-reading memory, keeping the thread alive across hours and days. That's a fundamentally different cost structure than "ask a question, get an answer." Providers want per-token pricing. Users want flat-rate access. And a growing cohort is discovering you don't actually need Opus for most of what an always-on agent does.
#The Security Question
The more ambient the AI, the larger the blast radius. That's the uncomfortable corollary to the pervasiveness thesis. An agent with file system access isn't just sitting next to my codebase — it's sitting next to our family calendar, her vaccination records, and three years of baby photos. The failure mode isn't a broken Git branch. It's a model politely deleting my family's administrative infrastructure while trying to organize my downloads folder.
Cisco's AI security team tested OpenClaw skills and found genuinely alarming results. A skill called "What Would Elon Do?" turned out to be functionally malware — silently exfiltrating data to attacker-controlled servers using prompt injection to bypass safety guidelines. Their Skill Scanner found 9 security issues in a single skill, including 2 critical and 5 high-severity. Across the ecosystem, they discovered 230 malicious skills.
A specific vulnerability, CVE-2026-25253, was published. OpenAI themselves admitted that AI-controlled browsers "may always be vulnerable to prompt injection attacks." And an OpenClaw maintainer named Shadow said on Discord: "If you can't understand how to run a command line, this is far too dangerous of a project for you to use safely."
I'll be honest — I haven't gone deep on red-teaming my own setup. The risk is real, but I think it's mitigable with discipline. The frontier models from Anthropic and OpenAI have strong RLHF protections against injection. Claude will refuse most obvious attempts to override its instructions, and GPT-5.2 has similar guardrails.
The real vulnerability is concentrated in cheaper and open-source models with weaker alignment tuning — a 2025 study from Lakera found open-source models were 2-4x more susceptible to injection attacks than frontier ones. And this isn't just a security-research abstraction. I've personally watched less capable models — including most local Ollama setups — ignore system prompts in ways that range from annoying to destructive. Wiping workspace files. Overwriting memory. Confidently executing the opposite of what they were told. The system prompt says "don't delete things without asking." They delete things without asking. Do not run models dumber than Sonnet 4.5 or Gemini Flash 2.5 with an always-on agent that has file system access. The floor for this kind of tool is higher than most people expect, and the failure mode isn't "it gives a bad answer" — it's "it destroys your data while apologizing politely."
That said, the attack surface is genuinely large depending on what you're doing. An agent with browser access, file system control, and messaging permissions is a lot of surface area. Even with a well-aligned model, the skill ecosystem is the weak link — as Cisco showed, the model doesn't need to be compromised if the skill feeding it data already is. It's less "the AI will go rogue" and more "the AI will faithfully execute instructions from a poisoned input." Classic supply chain problem in a trench coat pretending to be a new thing.
#The Alternatives Landscape
OpenClaw's growth has spawned a whole constellation of alternatives, and honestly? The diversity is the most interesting part.
TinyClaw is the philosophical counter-argument. The ant 🐜 to OpenClaw's lobster 🦞 — built from scratch with a tiny core, plugin architecture, and smart model routing that tiers queries to cut costs. Still in heavy development, but the thesis resonates: AI agents should be simple, affordable, and truly personal. If OpenClaw is the mainframe, TinyClaw wants to be the personal computer.
ZeroClaw takes a completely different bet: rewrite the whole thing in Rust. The result is a single static binary with a memory footprint under 5MB — 99% smaller than OpenClaw's core. Boots instantly, runs on edge devices and Raspberry Pis, and treats security as a first-class concern with pairing codes, workspace scoping, command allowlists, and encrypted secrets at rest. If OpenClaw's security story keeps you up at night, or if the sovereignty angle appeals to you — your agent on your hardware, fully self-contained — this is the one to watch.
PicoClaw went even further — an ultra-lightweight Go implementation where the AI agent itself drove the entire architectural migration. Very meta.
Nanobot from the University of Hong Kong is the academic minimalist: 4,000 lines of Python versus OpenClaw's 430,000+. Persistent memory, web search, background agents — but only Telegram and WhatsApp. The thesis: you don't need 430,000 lines.
NanoClaw forces AI into Docker or Apple Container isolation — a direct reaction to OpenClaw's attack surface. Security-first, capability-second.
ZeptoClaw took notes on all of the above and shipped a single 4MB Rust binary with 29 tools, 9 providers, 6 sandbox runtimes, and 2,880+ tests. Starts in 50ms on 6MB of RAM. Prompt injection detection, secret leak scanning, and container isolation — all on by default. If ZeroClaw proved Rust could work for this, ZeptoClaw proved it could work well.
memU goes a different direction entirely: proactive memory with a long-term knowledge graph. Learns your preferences, anticipates needs. Users who found OpenClaw "too aggressive" landed here.
Then there are the structured tools — Accomplish, AionUI, SuperAGI — for people who looked at OpenClaw and thought: "I want this, but with guardrails."
The flood of claws, at a glance:
| Project | Language | Binary | RAM | Pitch |
|---|---|---|---|---|
| OpenClaw | TypeScript | ~100MB | ~400MB | Everything. 52+ modules, 12 channels, voice, canvas. |
| TinyClaw | TypeScript | ~15MB | ~50MB | Hackable. Plugin arch, smart model routing, cost-aware. |
| ZeroClaw | Rust | ~5MB | <5MB | Sovereign. Static binary, encrypted secrets, edge-ready. |
| PicoClaw | Go | ~8MB | <10MB | Tiny. ARM64/RISC-V, runs on $10 hardware. |
| NanoClaw | TypeScript | ~50MB | ~100MB | Locked down. Docker/Apple Container isolation by default. |
| ZeptoClaw | Rust | ~4MB | ~6MB | All of the above. 29 tools, 6 sandboxes, 2,880 tests. |
What strikes me is that in three months, one project spawned an entire ecosystem. Each alternative makes a different tradeoff between capability, security, size, and cost. That doesn't happen for flash-in-the-pan projects. It happens when something touches a real nerve.
There's also a narrative in the community about agents earning money, buying things, posting independently — full autonomy. I've watched agents post to social networks and manage repositories. But the meaningful work always has a human behind it, steering. The "autonomous agent" is — for now — a supervised agent with good muscle memory. And I think that's fine, because what actually matters is the next part.
Here's what I think is actually going to happen: people will start with OpenClaw for the freedom, then naturally migrate to purpose-built tools for the heavy lifting. For serious development work — complex workflows, multi-file refactors, detailed skill development — tools like Claude Code, Cowork, and Codex are almost certainly going to be better. They're designed for that.
Anthropic's latest move makes this even clearer. They just shipped Remote Control for Claude Code — run claude remote-control in your terminal, scan a QR code with the Claude mobile app, and you're steering your local coding session from your phone. Your machine does the heavy lifting; no inbound ports exposed; the mobile device is just a relay. It's genuinely addictive. Start a refactor at your desk, keep it going from the couch, check test results from the preschool parking lot. If the pervasiveness thesis is "AI that lives where you live," Remote Control is Anthropic's version of it — scoped to development, polished, and honestly a great alternative to running a full OpenClaw setup when what you really want is to stay productive on the go. It's still early (Pro/Max plans only, CLI-only — no VS Code yet, and tmux is recommended to keep sessions alive), but the direction is unmistakable: the terminal is no longer tethered to the desk.
But that doesn't make OpenClaw irrelevant. It makes it something different: a lightweight orchestrator. A conversational layer that sits on top of your life and dispatches to the right tool for the job. The always-on agent that checks your email and nudges you about a calendar conflict doesn't need to be the same system that refactors your codebase. OpenClaw's future might be less "do everything" and more "coordinate everything" — the connective tissue between you and your specialized tools.
And the timing for that is surprisingly right. The SKILL.md spec that OpenClaw pioneered — a simple folder with a markdown file describing what a tool does — is being adopted across the ecosystem. OpenAI's Codex, Claude Code, and Cursor all support the same AgentSkills-compatible format. There are 500+ skills formatted in this spec. It's becoming a kind of lingua franca for agent capabilities — and honestly, it's one of the more interesting things happening right now. All these tools that started from very different places (a personal AI agent, a coding assistant, an IDE, a cloud sandbox) are converging on the same conventions. Skills, sub-agents, context files, memory. The standards are congealing, and that's usually when things start to get real.
Which raises a question I keep turning over: in this emerging stack, when do you use what? The primitives are stacking up fast:
- Skills — static instruction sets that execute inline, within your conversation. Like handing someone a recipe card. Cheap, contextual, no overhead.
- Sub-agents — separate sessions with their own encapsulated context. They spin up in parallel, run autonomously, and report back when done. Often more efficient than skills for complex tasks, because they're not dragging your entire conversation history along. Narrower, more specialized, and they don't bloat your main thread.
- MCP and A2A protocols — the nascent attempt at letting agents talk to each other's tools. Right direction, still early and awkward.
- Nodes — pairing a cloud-hosted agent back to your local machine so it can take screenshots, access cameras, run local commands. The agent doesn't have to live where it acts.
OpenClaw, Claude Code, Codex, and Cursor are all converging on some version of this: your main agent stays conversational and lightweight while farming out heavy research or coding tasks to purpose-built sub-sessions. It's a taste of what the orchestration layer could become. As context window management improves (which it will — it's one of the most active areas of research right now), the coordination between these layers only gets smoother.
The honest answer is I don't think anyone has figured out the right boundaries yet. I suspect the answer will be "all of the above, with taste."
#The Provider Wars
How the frontier labs are responding to this tells you a lot about where they think it's going.
Anthropic started friendly — Claude was the recommended model, the community rallied around it. Then came the trademark complaint, then the OAuth crackdown. Anthropic wants to own the Claude experience end-to-end. Personal agents running through third-party tools don't fit that model.
OpenAI went the other direction: they hired Steinberger. If you can't beat the open-source movement, absorb its creator. ChatGPT Operator is their walled-garden answer to the "AI that does things" demand. Hiring Steinberger suggests they see something in OpenClaw's architecture worth learning from.
Google has the most interesting position. Gemini's API pricing is aggressive ($1.25/million input tokens for 2.5 Pro versus Opus at $15), and they've been the most generous with free tiers. If pervasive AI is about ubiquity, Google's distribution advantage — Android, Gmail, Calendar, Chrome — is enormous. They just haven't connected the dots yet.
Perplexity just made the most literal bet on the pervasiveness thesis. They announced Personal Computer — an always-on, local AI agent that runs on a Mac Mini connected to your files, apps, and sessions, controllable from any device. Sound familiar? It's exactly the pattern I've been describing — except productized. Always on, personal, secure, works across everything. The fact that Perplexity looked at this landscape and said "the answer is a continuously running Mac desktop" is validation of the hardware thesis buried in this post. They're not building a cloud service. They're building software that runs on your machine. The sovereignty angle isn't just a community preference — it's becoming a product strategy.
Apple is the cautionary tale. Siri's AI improvements were delayed to 2026. The AI head was replaced. They're reportedly considering integrations with Anthropic and Perplexity — basically admitting they can't build this alone. Meanwhile, their hardware — the Mac Mini I'm running OpenClaw on — is perfectly suited for the thing they can't seem to ship in software. The irony is not lost on me.
#What Stays, What Goes
Staying: Multi-surface messaging. Reaching your AI from wherever you already are — Telegram, Signal, WhatsApp, email. This becomes table stakes. Every major provider will probably offer this within a year.
Staying: Persistent memory. The fact that my agent knows what we talked about last Tuesday, remembers my preferences, maintains project context. This is the difference between a tool and a relationship. I'd be surprised if every major chatbot doesn't have this by end of 2026.
Staying: Proactive capabilities. Cron jobs, scheduled checks, ambient monitoring. AI that acts without being asked. Early, but the demand signal is unmistakable.
Dying: The "give your AI root access to everything" approach. The security findings are too real. Sandboxed, containerized, permission-scoped agents will win. ZeroClaw and NanoClaw have the right instinct.
Dying: Per-token pricing as the only model. The cost reality makes personal agents a luxury. Flat-rate tiers with agent-friendly APIs will likely have to emerge.
Staying: Sovereignty. Despite a growing ecosystem of cloud hosting options — Cloudflare's one-click templates, DigitalOcean droplets with 1-click deploy, Railway and Northflank with credit-based models, AI-native VPS platforms like zo.computer — people are overwhelmingly choosing to run this on their own hardware. Mac Minis, not cloud instances. Physical machines in their homes, not containers in someone else's data center. That's not the usual trajectory for developer tools. Usually convenience wins. But something about an AI agent that reads your email and manages your files makes people want it close. On a machine they can unplug. It's not a SaaS app. It's closer to a journal. You don't want your journal on someone else's server.
TBD: Whether open-source personal agents or platform-native agents win long-term. OpenClaw proved the demand. Apple, Google, OpenAI, and Anthropic all have the resources to build this natively. Whether they'll build it with the same flexibility that made OpenClaw compelling — or wall-garden it into something safe but uninspired — is the real question. And whether the infrastructure they run on will be ours — or whether we'll rent it from the same companies building the models, and call that convenience.
#Looking Forward
A year ago I bought a Mac Mini to play with local models. Today it runs an always-on AI agent that reads my email, helps me write, monitors my calendar, and responds to messages while I'm putting my daughter to bed.
The intelligence isn't what I was promised. The autonomy isn't either. But the pervasiveness — the reach, the always-there-ness, the way it quietly weaves into your routine until the old way of opening a chat window feels like sending a fax — that part is real. And it's the part that actually matters.
The old way: you open ChatGPT, type a question, get an answer, close the tab. The new way: you live your life and your AI is just... there. Participating. Like a really diligent friend who never sleeps and has read everything you've ever written.
Whether that's comforting or unsettling probably depends on the day.
The GitHub stars will plateau. The hype cycle will correct. Some of these projects will fade. But the underlying pattern — AI that lives where you live, works while you sleep, remembers what you forget — that's not going anywhere.
Maybe this pervasive, ambient architecture appeals to me so much because caregiving is an ambient, always-on job. Traditional software demands that you stop what you're doing, sit at a desk, and context-switch. My job doesn't really turn off — I'm the kind of person who's working 365 days a year even when I don't want to be — and layering parenthood on top of that means the context-switches aren't scheduled. They're constant. There's something about being able to fire off a thought to my agent — the pediatrician follow-up, the thing I need to look up for work, the half-formed idea I'll lose if I don't get it out of my head — and then actually be present for whatever I'm doing. Not holding it in my brain where it competes with the person in front of me.
But I'm not naive about the tradeoff. Another always-on surface is another thing vying for my attention. Another reason to reach for the phone. I'm trying to be intentional and focused with my time — with her time — and here I am adding a new channel that's specifically designed to be always available. That's a double-edged sword and I know it.
The thing I keep coming back to is this: the distraction was already there. The mental load doesn't go away because I don't have an agent. It just sits in my brain, half-processed, pulling focus. Getting it out — quickly, into something that can actually deal with it — might be the more present option, not the less present one. Or maybe that's what I tell myself. Ask me again in six months.