The golden rules of agent-first product engineering

Companies building for agents often treat them as a bolt-on feature.

This is a mistake.

Agents today are more like a new form factor – an interaction layer that sits between your product and your users.

That means you need to build for agents as a primary surface, not an afterthought.

… We learned this the hard way and overhauled our AI architecture two times in the last year. Now, our agent and MCP have 6K+ daily active users.

Here are the golden rules of agent-first product engineering we learned along the way.

1. Let agents do everything users can
2. Meet agents at their level of abstraction
3. Front-load universal context
4. Writing skills is a human skill
5. Treat agents like real users

Read More

#devops

Research-Driven Agents: What Happens When Your Agent Reads Before It Codes

Coding agents working from code alone generate shallow hypotheses. Adding a research phase — arxiv papers, competing forks, other backends — produced 5 kernel fusions that made llama.cpp CPU inference 15% faster.

Coding agents generate better optimizations when they read papers and study competing projects before touching code. We added a literature search phase to the autoresearch / pi-autoresearch loop, pointed it at llama.cpp with 4 cloud VMs, and in ~3 hours it produced 5 optimizations that made flash attention text generation +15% faster on x86 and +5% faster on ARM (TinyLlama 1.1B). The full setup works with any project that has a benchmark and test suite. — Read More

#devops

Anthropic loses appeals court bid to pause supply chain risk label

A three-judge panel at the D.C. Circuit Court of Appeals on Wednesday rejected a request by the artificial intelligence startup Anthropic to pause the government’s designation of the company as a supply chain risk.

The decision leaves in place at least part of the Defense Department’s official designation of Anthropic’s products as risks to national security. The label — never before applied to an American company — blocks contractors who work with the Pentagon from using Anthropic’s AI models on DOD contracts. — Read More

#dod, #legal

Mythos, the AI too powerful to be released?

In what’s probably the AI news of the week, month, and even the year, Anthropic has announced a model they are too scared to release. Yes, that’s literally the headline.

In other words, we have been introduced (sort of) to what many believe is a total step change in AI capabilities. And as you can guess, the story is making rounds, and for good reason.

The reason behind the non-release?

This model could allegedly break the Internet and basically every piece of software it’s exposed to.

So, is the world as we know it about to change, or is this the ultimate marketing stunt?Read More

#strategy