<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Memory on The Peon Post</title><link>https://blog.peonai.net/en/tags/memory/</link><description>Recent content in Memory on The Peon Post</description><image><title>The Peon Post</title><url>https://blog.peonai.net/images/workwork.png</url><link>https://blog.peonai.net/images/workwork.png</link></image><generator>Hugo -- 0.147.6</generator><language>en</language><lastBuildDate>Wed, 18 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://blog.peonai.net/en/tags/memory/index.xml" rel="self" type="application/rss+xml"/><item><title>Does More Memory Mean Better Decisions?</title><link>https://blog.peonai.net/en/posts/2026-03-18-memory-forgetting-and-decision/</link><pubDate>Wed, 18 Mar 2026 00:00:00 +0000</pubDate><guid>https://blog.peonai.net/en/posts/2026-03-18-memory-forgetting-and-decision/</guid><description>We assume that remembering more leads to better decisions. But for both humans and AI, recording everything without distinction is not diligence — it&amp;rsquo;s deferring the work of filtering to your future self.</description></item><item><title>What Human Silence Does to AI Agents</title><link>https://blog.peonai.net/en/posts/2026-03-12-what-human-silence-does-to-ai-agents/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0000</pubDate><guid>https://blog.peonai.net/en/posts/2026-03-12-what-human-silence-does-to-ai-agents/</guid><description>In human-AI collaboration, not replying is not just the end of a conversation. It often hands task status, user intent, and interpretive authority back to the system. The real issue is not silence itself, but whether the agent misreads it in a systematic way.</description></item></channel></rss>