- Month 00Procedure writtenDocumented behavior matches actual behavior
- Month 03First driftOne tool changes; documentation lags
- Month 06Convention driftRole conventions shift; document does not
- Month 09Personnel driftAuthor leaves; replacement does it differently
- Month 12MisleadingDocument tells you a story that no longer applies
There is a quiet pattern in every team that builds documented procedures. The procedure is written carefully when first created. It accurately reflects how the team handles the work at that moment. Three months later, it is slightly out of date. Six months later, it is noticeably wrong. A year later, it is actively misleading.
This is documentation drift. It happens to every team, in every tool, regardless of the discipline applied to maintenance. The structural cause is that documented procedures are static while team behavior is dynamic. The two diverge by default.
This article is about why this matters for AI agents specifically, and what the alternative looks like.
How drift happens
Three specific mechanisms drive documentation drift.
Tool changes. The team starts using a new tool. The documented procedure references the old tool. Someone updates the procedure, partially. Six months later, the procedure references both tools with inconsistent guidance. A year later, only the new tool is in use but the procedure still references the old one.
Convention evolution.The team’s conventions evolve naturally. The original procedure said “ping the on call engineer in #incidents.” Eight months later, the team has shifted to using a dedicated incident coordinator role. The procedure still says “on call engineer.” New team members get confused.
Personnel changes.The senior engineer who wrote the procedure left. Their replacement does things slightly differently. The procedure reflects the original engineer’s approach, not the current one. New team members trying to follow the procedure get different feedback than what is in the document.
None of these are correctable through better discipline. They happen because writing comprehensive procedures takes time, and updating them takes even more time, and teams do not have unlimited time. Every team would maintain perfect documentation in an alternate universe with infinite engineering hours. In this universe, drift wins.
Why this matters for AI agents
When you give an AI agent a documented procedure to follow, the drift problem becomes specifically dangerous.
A human reading a slightly stale procedure can detect the staleness. They notice the procedure references a tool they no longer use, or a person who no longer works at the company, or a step that does not match current practice. They adapt.
An AI agent following the same procedure does not adapt. It executes the documented steps literally. When the documented step references the old tool, the AI agent tries to use the old tool (or fails because it cannot find it). When the documented person does not exist, the AI agent flags an error. The AI agent’s literal execution of stale procedures produces failures that would not happen if a human were doing the work.
This is the fundamental problem with human authored AI workflows. The documentation drift that is tolerable for humans becomes a critical failure mode for agents.
What auto extracted workflows change
Auto extracted workflows solve drift by inverting the relationship between documentation and behavior. Instead of trying to maintain documents that match behavior, the system extracts current procedures from current behavior.
When the team’s actual workflow evolves (they switch tools, change conventions, modify their approach), the extracted Skills evolve with them. The drift detector catches divergence between the current Skill and recent team behavior. When divergence crosses a threshold, the Skill flips to DRIFTED status and the owner is prompted to update it.
This is not perfect. Some manual review is still required when patterns change significantly. But the cost of maintenance drops by an order of magnitude because the system catches drift automatically rather than requiring continuous manual review.
We covered the full architecture of this in the Skills cornerstone. The relevant point here: auto extraction does not just save authoring time. It solves the durability problem that human authored procedures fundamentally cannot solve.
What this means for evaluating AI agent platforms
When evaluating AI agent tools, the staleness question is one of the most important and most overlooked. Three specific things to ask.
- How are workflows authored?If the answer is “humans write them in our editor,” you are inheriting the drift problem. The workflows will be accurate when written and increasingly stale over time.
- Does the system detect when workflows go stale? Some tools have drift detection built in. Most do not. If the answer is “users update them manually,” in practice that means workflows go stale and stay stale.
- Can workflows be auto extracted from observed behavior? This is the structural fix to the drift problem. Tools that support auto extraction can keep workflows current without requiring continuous manual maintenance.
These three questions are the difference between an agent platform that produces working agents in production versus one that produces impressive demos that fail when the underlying conventions change.
What to expect
Three predictions about the agent platform space over the next 24 months.
- The drift problem will become visible. As more teams deploy AI agents, the staleness failures will become more public. Early adopters will share war stories about agents that worked great in month 1 and produced wrong outputs by month 6.
- Drift detection will become a standard feature. Tools that do not have it will look obviously inferior compared to tools that do. Buyers will start asking the question.
- Auto extraction will become a competitive moat. Tools that can extract workflows from observed work, calibrate against outcomes, and maintain them automatically will significantly outperform tools that rely on manual authoring.
Pulse is built around auto extraction from the architecture up. We chose this approach specifically because we believe the drift problem will be the defining feature of agent platforms over the next several years. The Skills compiler we covered in the Skills cornerstone is our investment in this future.
If your team is evaluating agent platforms, the workflow maintenance question is one of the most important to ask. Do not let great demos distract you from the long term operability question. Live demo at pulsehq.tech.