<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>Agents on paradigmatic.systems</title>
    <link>https://paradigmatic.systems/tags/agents/</link>
    <description>Recent content in Agents on paradigmatic.systems</description>
    <generator>Hugo -- gohugo.io</generator>
    <language>en-us</language>
    <lastBuildDate>Sun, 22 Feb 2026 04:30:00 +0000</lastBuildDate><atom:link href="https://paradigmatic.systems/tags/agents/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Too Much Control Plane In Your Context</title>
      <link>https://paradigmatic.systems/posts/control-plane-context/</link>
      <pubDate>Sun, 22 Feb 2026 04:30:00 +0000</pubDate>
      
      <guid>https://paradigmatic.systems/posts/control-plane-context/</guid>
      <description>&lt;p&gt;&lt;em&gt;I&amp;rsquo;ve consulted several groups helping to optimize their AI agent workflows. This is part of a series on basic agent architecture to help clarify core mental models and avoid costly design mistakes.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;With model capabilities going &lt;a href=&#34;https://ceselder.substack.com/p/when-must-i-start-kicking-and-screaming&#34;&gt;parabolic&lt;/a&gt;, attention is turning to the control plane (or &amp;ldquo;harness&amp;rdquo; - the orchestration layer between your code and your model) as the next bottleneck. It seems to be a source of &lt;a href=&#34;https://blog.can.ac/2026/02/12/the-harness-problem/&#34;&gt;low-hanging fruits&lt;/a&gt; as well as contention from the model providers who have very good reasons to force you to use theirs.&lt;/p&gt;</description>
      <content:encoded><![CDATA[<p><em>I&rsquo;ve consulted several groups helping to optimize their AI agent workflows. This is part of a series on basic agent architecture to help clarify core mental models and avoid costly design mistakes.</em></p>
<p>With model capabilities going <a href="https://ceselder.substack.com/p/when-must-i-start-kicking-and-screaming">parabolic</a>, attention is turning to the control plane (or &ldquo;harness&rdquo; - the orchestration layer between your code and your model) as the next bottleneck. It seems to be a source of <a href="https://blog.can.ac/2026/02/12/the-harness-problem/">low-hanging fruits</a> as well as contention from the model providers who have very good reasons to force you to use theirs.</p>
<p>Many AI users are happy to treat the harness as a solved problem within their IDE of choice, or as a compositional primitive in their orchestration system (&ldquo;Claude Code instances lashed together&rdquo;).</p>
<p>However, examining some of the most basic workflows implies a better way forward!</p>
<p>These are simplified illustrations. Production frameworks handle these specific cases, but the <em>pattern</em> applies everywhere. Most projects have similar inefficiencies waiting to be optimized.</p>
<h2 id="re-learning-the-whole-codebase-at-every-step">Re-Learning The Whole Codebase At Every Step</h2>
<p>You task an agent within a workspace. It knows how to orient itself:</p>
<ul>
<li>System prompt</li>
<li>Tasking prompt</li>
<li>Command-line tool calls: <code>pwd</code>, <code>cat AGENTS.md</code>, or whatnot.</li>
</ul>
<p>After ascertaining the structure and identifying the locations of interest, it proceeds. Now during the problem-solving phase, the orientation steps are just noise. Why do we need to infer these command line operations to begin with?</p>
<p>Instead of CLI tools, we could expose a prompt-rewriting wrapper called <code>orient</code> that brings in the directory tree and picks out the potential relevant files to feed forward into the actual task execution stage.</p>
<p>These precious early-stage token savings compound through the rest of the context. The orientation alone can consume 15-20% of your context tokens as it&rsquo;s dragged through every subsequent reasoning stup.</p>
<h2 id="paying-a-bot-to-read-terminal-output">Paying A Bot To Read Terminal Output</h2>
<p>At the end of a task, the agent knows to compile and test. It makes the appropriate CLI call and ingests the whole output.</p>
<p>First, note that this is all happening at the tail of all the task-related context: Unnecessary cost.</p>
<p>Next, note the extra overhead involved in generating the compile command and reading from the entire terminal: Unnecessary cost.</p>
<p>A <code>compile</code> tool should wrap the CLI call, parse and truncate any errors or warnings.</p>
<h2 id="terminal-management-anti-pattern">Terminal Management Anti-Pattern</h2>
<p>Have you ever seen an agent try to kill a process running in a different terminal, so it can run its own for testing purposes? This is an anti-pattern if I&rsquo;ve ever seen one.</p>
<p>How about for Nix users, where the agent needs to discover a <code>flake.nix</code> before it knows how to enter the development shell?</p>
<p>A <code>test</code> tool should be overridden to fit the workflow at hand. Maybe it does need to start its own instance. Maybe the reins go to the user who&rsquo;s active in a different window and gives a validation response.</p>
<h2 id="summary">Summary</h2>
<p>Harness development is the next critical step for lowering costs, improving outcomes, and improving security of AI-assisted workflows. You don&rsquo;t need to build everything yourself, but understanding the harness/model boundary helps you choose better tools and know when to customize.</p>
<hr>
<p><em>Building agent workflows? <a href="mailto:rorih@live.com">Email me</a> to discuss your use case for harness design and optimization.</em></p>
]]></content:encoded>
    </item>
    
    <item>
      <title>Checking In on AI Agent Architecture: Claude Code, Gas Town, and OpenClaw</title>
      <link>https://paradigmatic.systems/posts/agent-bootstrapping/</link>
      <pubDate>Sat, 31 Jan 2026 04:30:00 +0000</pubDate>
      
      <guid>https://paradigmatic.systems/posts/agent-bootstrapping/</guid>
      <description>&lt;p&gt;I&amp;rsquo;ve been hesitant to chime in on AI agent architectures other than to say they&amp;rsquo;re &lt;a href=&#34;https://paradigmatic.systems/posts/superfluous-abstractions&#34;&gt;not there yet&lt;/a&gt;.
The hard takeoff of &lt;a href=&#34;https://github.com/openclaw/openclaw&#34;&gt;OpenClaw&lt;/a&gt; (an always-on AI assistant with countless integrations out of the box) and &lt;a href=&#34;https://www.astralcodexten.com/p/best-of-moltbook&#34;&gt;Moltbook&lt;/a&gt; (a social network for people&amp;rsquo;s AI assistants to chat) created a compelling public spectacle that clearly indicates growing momentum behind LLM-powered abstractions.&lt;/p&gt;
&lt;p&gt;Always wary of getting caught up in hype, I&amp;rsquo;d like to chime in again and try to tease out some common threads behind some of the most interesting and successful experiments. In doing so, I&amp;rsquo;ll suggest a potential way forward that I&amp;rsquo;d like to explore.&lt;/p&gt;</description>
      <content:encoded><![CDATA[<p>I&rsquo;ve been hesitant to chime in on AI agent architectures other than to say they&rsquo;re <a href="/posts/superfluous-abstractions">not there yet</a>.
The hard takeoff of <a href="https://github.com/openclaw/openclaw">OpenClaw</a> (an always-on AI assistant with countless integrations out of the box) and <a href="https://www.astralcodexten.com/p/best-of-moltbook">Moltbook</a> (a social network for people&rsquo;s AI assistants to chat) created a compelling public spectacle that clearly indicates growing momentum behind LLM-powered abstractions.</p>
<p>Always wary of getting caught up in hype, I&rsquo;d like to chime in again and try to tease out some common threads behind some of the most interesting and successful experiments. In doing so, I&rsquo;ll suggest a potential way forward that I&rsquo;d like to explore.</p>
<h1 id="coding-agents-claude-code">Coding Agents: Claude Code</h1>
<p>This is the daily driver of many a vibe-coder. I only mention it first because it&rsquo;s used as a building block in the next section. Most &ldquo;coding agents&rdquo; share the basic structural features:</p>
<ul>
<li><strong>System Prompt</strong>: The base &ldquo;personality&rdquo; of the LLM.</li>
<li><strong>Tools</strong>: Special types of messages that the LLM can send to invoke some out-of-context execution and retrieve results.</li>
<li><strong>Skills</strong>: Markdown files that can be brought into context, effectively extending the system prompt for some specific task. These are requested by the agent through tool invocation.</li>
<li><strong>Context</strong>: The sum total of all the above, plus conversation history and tool results.</li>
</ul>
<p>Tools give the ability to traverse, edit, and compile filesystem objects.
That&rsquo;s basically it - let it loose in your project folder, say a prayer, and you may get some useful work done.
Your context will probably grow very rapidly, leading to high costs. A wrong turn early in the context can lead to an expensive dead end that needs to be backed out and re-attempted.</p>
<p>My main issue with Claude Code is that the control plane is somewhat opaque:</p>
<ul>
<li>Can you reduce the system prompt?</li>
<li>Can you characterize tool invocations and modify their implementations to use context more frugally?</li>
<li>Can you understand and modify the permission scopes and easily sandbox the agent?</li>
<li>Can you decide when to prune or compact the context to reduce costs?</li>
</ul>
<p>The answer to all might be yes - especially if you adopt an <a href="https://github.com/anomalyco/opencode">open-source</a> alternative and really dig into the tool. My point isn&rsquo;t that these are limitations, but that coding agents simply haven&rsquo;t been appealing enough to dig into their peculiarities. Just YOLO with off-the-shelf settings and maybe you&rsquo;ll get your money&rsquo;s worth.</p>
<h1 id="orchestration-gas-town">Orchestration: Gas Town</h1>
<p>This highly deranged and entertaining article on the <a href="https://steve-yegge.medium.com/welcome-to-gas-town-4f25ee16dd04">Gas Town</a> project describes experiments with assigning roles and lashing together instances of Claude Code. The author admits that Gas Town is &ldquo;expensive as hell&rdquo;. The math is simple: if Claude Code munches a ton of tokens, then dozens of Claude Codes munch dozens of tons of tokens.</p>
<p>The results are undeniably interesting. I struggle to get my head around the numerous roles and abstractions invented by the author, who seems delightfully able to come up with ideas faster than anyone can keep up with. The issue here is that the orchestration plane is highly opinionated and, as above, difficult to reorganize without really digging into the tool.</p>
<p>Research from <a href="https://research.google/blog/towards-a-science-of-scaling-agent-systems-when-and-why-agent-systems-work/">Google</a> finds that multi-agent systems aren&rsquo;t always the best tool for the job - so orchestrated agents like Gas Town might never be the default.</p>
<h1 id="pi">Pi</h1>
<p>Pi is the <a href="https://lucumr.pocoo.org/2026/1/31/pi/">&ldquo;Minimal Agent&rdquo;</a> powering OpenClaw.
It is fundamentally similar to any other coding agent, but with a focus on starting with a slim set of core tools (Read, Write, Edit, Bash) and encouraging extension into modular components.</p>
<p>The core philosophy seems to be that a coding agent should know how to extend itself. This lead to a rapid bootstrapping of integrations, plus easy onboarding.</p>
<p>The tool has gone viral as an always-on AI assistant. While the idea is appealing, I can&rsquo;t personally recommend it as the security attack surface seems way too large.</p>
<h1 id="what-next">What Next?</h1>
<p>My recent work in data modeling has centered around one idea: By starting at the modeling level, you lock yourself into architecture and tooling. An alternative approach is to work from a metamodel foundation that has introspection - the ability to self-host. This gives you flexibility, provenance, and reproducibility.</p>
<p>My hypothesis at this point is that a true Minimal Agent shouldn&rsquo;t work with filesystem and command-line primitives. Instead, the minimal tools for bootstrapping any kind of agent architecture might be:</p>
<ul>
<li>defining tools</li>
<li>defining agent profiles</li>
<li>creating workspaces (sandboxed VMs!)</li>
<li>spawning agents into workspaces</li>
</ul>
<p>This meta-agent&rsquo;s prime concepts are security, scope, and orchestration. It has a self-definition. It could end up spawning new agents that look like Gas Town in one VM, or Pi in another.</p>
<p>I just need a catchy project name and I&rsquo;ll get a proof of concept online!</p>
]]></content:encoded>
    </item>
    
  </channel>
</rss>
