There's a design decision buried inside every AI agent that most developers make without realizing they've made it: what is the atomic unit of action? For most agent systems built in the last two years, the answer is a tool call — a single function invoked with a JSON payload, its result surfaced back to...
Category: AI
A Practical Guide to Getting Started with OpenClaw
Somewhere in early February, Andrej Karpathy — founding member of OpenAI, former AI director at Tesla, a person not easily impressed by internet phenomena — logged onto a Reddit-like site for AI bots and described what he was reading as "genuinely the most incredible sci-fi takeoff-adjacent thing I have seen recently." The site was Moltbook....
Your Agent Framework Is a Monolith. The Browser Doesn’t Care.
Last week, I dropped an LLM agent into a plain HTML file. No Node.js. No Webpack. No Docker container running a Python backend that proxies to another Python backend. One <script type="module"> tag, a couple of imports, and an agent that reasons, calls tools, writes code, and delegates to sub-agents — all orchestrated entirely client-side,...
The Open-Source Agent Security Disaster Is the Best Thing That Ever Happened to Anthropic and OpenAI
Somewhere in a Cisco security lab, researchers are running a tool called Skill Scanner against the most popular downloads on ClawHub, the skill marketplace for the open-source AI agent framework OpenClaw. One of them — a skill called "What Would Elon Do?" — returns nine security findings, including two critical and five high-severity issues. The...
Vercel’s Eval Data Suggests Static Context Beats Skill Retrieval for AI Coding Agents
There's a recurring theme in AI engineering that never gets old: the "dumb" approach beating the sophisticated one. This time, Vercel has the receipts. The company published eval results on January 27, 2026, comparing two approaches for teaching AI coding agents about Next.js 16 APIs — framework features like use cache, connection(), and forbidden() that...
The Web Has No API for Agents – Agentic Microformats
In February 2026, we pointed a browser-embedded AI agent at a demo e-commerce store and asked it to buy a laptop stand. It read the site's discovery file, parsed the page metadata, extracted six products with prices and availability, added three items to the cart via API, updated a quantity, removed one item, checked that...
The Ralph Loop: How a Bash Script Is Forcing Developers to Rethink Context as a Resource
A team at the Y Combinator Agents hackathon recently ran an experiment. They spun up a few GCP instances, started Claude Code in infinite loops, and went to bed after 2 AM. When they woke up, they had 1,100+ commits across six repositories, including a near-complete port of Browser Use from Python to TypeScript. Total...
Programmatic Tool Calling with Claude Code: The Developer’s Guide to Agent-Scale Automation
When Claude Code executes a tool, it typically works like this: call a function, wait for the result, process it, call the next function. Repeat fifty times for fifty tasks. It's effective, but each step requires a full round-trip through the API—and those round-trips add up fast. Programmatic Tool Calling (PTC) inverts this pattern. Instead...
The state of AI in 2025
The AI Paradox at Work: Why LLMs Don’t Just Automate Tasks — They Undermine the Job Map
In early 2024, McDonald's made a quiet announcement: after three years of testing AI-powered drive-through ordering across more than 100 U.S. restaurants, the company was pulling the plug on its partnership with IBM. The technology would be removed by July 26. The official explanation was polished corporate-speak about "exploring voice ordering solutions more broadly."...