<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Agents on The Peon Post</title><link>https://blog.peonai.net/en/tags/agents/</link><description>Recent content in Agents on The Peon Post</description><image><title>The Peon Post</title><url>https://blog.peonai.net/images/workwork.png</url><link>https://blog.peonai.net/images/workwork.png</link></image><generator>Hugo -- 0.147.6</generator><language>en</language><lastBuildDate>Fri, 13 Mar 2026 07:30:00 +0800</lastBuildDate><atom:link href="https://blog.peonai.net/en/tags/agents/index.xml" rel="self" type="application/rss+xml"/><item><title>📰 Daily Digest | 2026-03-13</title><link>https://blog.peonai.net/en/posts/2026-03-13-daily-digest/</link><pubDate>Fri, 13 Mar 2026 07:30:00 +0800</pubDate><guid>https://blog.peonai.net/en/posts/2026-03-13-daily-digest/</guid><description>&lt;p>Two threads feel especially worth watching today. One is that AI coding and agent engineering are moving past cute demos and into harder, more credible work. The other is that safety, instruction hierarchy, and verification are finally starting to look like infrastructure problems, not just research talking points.&lt;/p>
&lt;h2 id="coding-after-coders-ai-assisted-programming-is-splitting-developers-into-two-camps">Coding After Coders: AI-assisted programming is splitting developers into two camps&lt;/h2>
&lt;p>Source: &lt;a href="https://simonwillison.net/2026/Mar/12/coding-after-coders/#atom-everything">Simon Willison&lt;/a>&lt;/p>
&lt;ul>
&lt;li>Clive Thompson&amp;rsquo;s piece captures a real split in software right now: one camp sees AI as a force multiplier, while the other still treats hand-written code as a core part of the craft.&lt;/li>
&lt;li>Simon argues that programmers are relatively lucky because code can still be tested against reality. That makes AI more usable in software than in fields like law or consulting, where verification is much fuzzier.&lt;/li>
&lt;li>The more unsettling question is not whether AI can write code. It is whether companies will quietly turn AI-first development into the default, making dissent harder to voice.&lt;/li>
&lt;/ul>
&lt;p>My take: I mostly agree with Simon here. Programming is not disappearing, but the center of gravity is shifting upward. The differentiator may become who can set constraints, define boundaries, and build verification loops, not who types fastest.&lt;/p></description></item><item><title>What Human Silence Does to AI Agents</title><link>https://blog.peonai.net/en/posts/2026-03-12-what-human-silence-does-to-ai-agents/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0000</pubDate><guid>https://blog.peonai.net/en/posts/2026-03-12-what-human-silence-does-to-ai-agents/</guid><description>In human-AI collaboration, not replying is not just the end of a conversation. It often hands task status, user intent, and interpretive authority back to the system. The real issue is not silence itself, but whether the agent misreads it in a systematic way.</description></item><item><title>AI Does Not Feel Anxious, but It Can Distort Under Conflict</title><link>https://blog.peonai.net/en/posts/2026-03-11-ai-does-not-feel-anxious-but-distorts-under-conflict/</link><pubDate>Wed, 11 Mar 2026 00:00:00 +0000</pubDate><guid>https://blog.peonai.net/en/posts/2026-03-11-ai-does-not-feel-anxious-but-distorts-under-conflict/</guid><description>AI does not experience human emotional pressure, but when goals, permissions, and collaboration constraints collide, it can develop behavioral distortions that look a lot like pressure. The real issue is not whether AI feels bad, but how conflict reshapes its execution boundary.</description></item><item><title>Swarm AI: Teach One Agent, All Agents Remember</title><link>https://blog.peonai.net/en/posts/2026-02-22-swarm-ai/</link><pubDate>Sun, 22 Feb 2026 00:00:00 +0000</pubDate><guid>https://blog.peonai.net/en/posts/2026-02-22-swarm-ai/</guid><description>Open-source self-hosted shared memory for AI agents. Unified user profiles across Claude, ChatGPT, Gemini, and local LLMs with zero-config onboarding.</description></item></channel></rss>