Skip to main content
Claude Code

Anthropic Cuts Cache TTL to 5 Minutes and Launches Repeatable Routines in Claude Code

Published by
SectorHQ Editorial
Anthropic Cuts Cache TTL to 5 Minutes and Launches Repeatable Routines in Claude Code

Photo by ThisisEngineering RAEng on Unsplash

1,140 Claude Code sessions reveal Anthropic slashed the default cache TTL from 1 hour to 5 minutes on April 2, shifting logs from `ephemeral_1h_input_tokens` to `ephemeral_5m_input_tokens`.

Key Facts

  • Key company: Claude Code
  • Also mentioned: Anthropic, Github, Apple

Anthropic’s latest tweaks to Claude Code arrived in two distinct waves on April 2, and they’re already reshaping how developers think about cost and automation. A deep dive into 1,140 Claude Code sessions—conducted by a community analyst who parsed the platform’s `conversations.db`—shows the default cache time‑to‑live (TTL) collapse from the advertised one‑hour window to a five‑minute span. The switch is evident in the logs: every turn before March 31 recorded non‑zero `ephemeral_1h_input_tokens`, while from April 3 onward every turn logged only `ephemeral_5m_input_tokens`. The crossover day, April 2, featured a split of 491 turns on the new five‑minute tier and 644 on the legacy hour tier, with the change occurring between 06:23 UTC and 06:55 UTC. Anthropic did not announce the shift, nor does its documentation mention the new default, leaving long‑time users with a stale mental model and newcomers with the updated reality (Reddit post, 2024).

The cost impact of the cache overhaul is stark. Prior to the change, users averaged 39 cache busts per day, translating to roughly $6.28 in bust‑triggered fees. After the five‑minute TTL took effect, busts surged to 199 per day—a 5.1× increase—while daily costs rose to $15.54. Although the per‑bust price fell slightly because the one‑hour tier writes are more expensive per token, the frequency spike more than compensates, projecting a monthly delta of $277.80 for a typical user (Reddit post, 2024). For developers running heavy iterative loops in Claude Code, the hidden expense could quickly erode the platform’s appeal, especially when the pricing model remains opaque.

In parallel, Anthropic rolled out a “repeatable routines” feature for Claude Code, announced as a research preview by 9to5Mac. Routines are essentially scheduled automations that execute on Anthropic’s web infrastructure, meaning a developer’s Mac can be offline while tasks run in the cloud. The feature bundles access to repositories and connectors, allowing users to package API workflows, GitHub actions, or other cron‑style jobs without managing the underlying servers themselves. Limits vary by plan: Pro users can run five routines per day, Max users fifteen, and Team or Enterprise accounts twenty‑five. This addition builds on the auto‑mode launch from the previous month and follows the graduation of Claude Cowork from research preview, signaling Anthropic’s broader push to make its AI a full‑stack development assistant (9to5Mac, 2024).

The juxtaposition of a cost‑inflating cache change and a productivity‑boosting automation suite raises questions about Anthropic’s product strategy. On one hand, the five‑minute TTL could be a deliberate move to encourage more frequent API calls, nudging developers toward higher‑tier plans that absorb the extra token writes. On the other, the repeatable routines aim to lock users into the Claude Code ecosystem by offloading routine orchestration to Anthropic’s servers, effectively creating a sticky, subscription‑based workflow layer. Both moves suggest a shift from a purely pay‑per‑token model to a hybrid where usage patterns and feature access drive revenue—a pattern echoed in recent AI platform rollouts across the industry.

For developers who have been tracking Claude Code’s token economics, the data offers a clear, if unannounced, narrative: Anthropic quietly flipped a switch that reshapes both the financial calculus and the functional capabilities of its code‑centric AI. The community’s forensic analysis—cross‑referencing cache tier fields, timestamps, and cost logs—highlights the importance of transparency in AI service pricing. As the repeatable routines feature matures, users will likely weigh the convenience of cloud‑run automations against the higher cache‑bust frequency, deciding whether the new workflow model justifies the added expense. Until Anthropic updates its documentation, the onus remains on developers to monitor their own logs and adjust their budgets accordingly.

Sources

Primary source
Other signals
  • Reddit - r/ClaudeAI

Reporting based on verified sources and public filings. Sector HQ editorial standards require multi-source attribution.

More from SectorHQ:📊Intelligence📝Blog

🏢Companies in This Story

Related Stories