Skip to content
Stark Insider
  • Culture
  • Filmmaking/Tech
  • Atelier Stark Films
News Tech

Quick Tip: How to Get Claude Code to Run Autonomously for Hours

Claude Code thinks 10 minutes is 8 hours. Here's how to fix his broken sense of time.

BY Clinton Stark — 05.14.2026

MacBook Pro running Claude Code in Visual Studio Code with an autonomous coding prompt, demonstrating how to unlock long multi-hour runs from an AI coding agent
Anchor Claude Code in real server time and the AI coding agent will keep shipping code for hours. The prompt on screen is the exact instruction that turned a quick errand into a 39-minute autonomous sprint.

I’ve been working with Claude Code for almost a year now. First, it was for mundane things like checking server logs and fixing bugs here on the server that runs the Stark Insider web site. Then, after I dug my way out of that dungeon, I was able to get on to the more interesting projects that agents afford us in this new AI-centered world.

My latest project is a federated system for multi-agent enterprise environments. Soon I will launch with more details, and information on the final product.

Over the four or so months of collaborating and coding with Claude Code on this new product I’ve always been baffled by its absolute lack of time:

Claude Code fundamentally struggles with the concept of time!

Example 1:

Often, when I wrap a session, Claude will summarize and state, for instance, that the time is something like ~ 7:10 PM, when actually the current time is only 6:50 PM. It’s a peculiar thing. Claude is running on a server and can easily run the “date” terminal command to look up the real time (and date).

Example 2:

But this is the more curious one. After a massive planning session (I spend probably 70-80% of my time in planning mode to try to nail design first), I will ask Claude to plan out a block of work. He obliges, and proposes a Phase that might take 6-8 hours of coding time. Turns out, his estimate is way too conservative, and — bingo! — he’s done in about 10 minutes. This happens time and time again. Basically, per Claude’s math you can do weeks, if not months, of coding in about a regular work day.

So why is Claude Code so bad at time management?

My theory is pretty straightforward:

Claude Code (like any LLM) is trained on a mass of human knowledge as we know. And a large chunk of it is based on humans’ expertise on writing code, creating software and systems and so forth. Likely a lot of it comes from programming books. So goes my theory: if this is the case, then any time estimates contained in these books, or deployed software use cases, and other sources, will be based on human speed.

Of course, AI agents move lightning fast compared to us humans.

So, naturally, its estimates to get things done are based on humans, and, therefore, over-estimate how long it will take to complete a task or a block of work.

Lifehack: Get Claude Code to Work Autonomously for Long Coding Runs

Recently, I had to head out to pick up our weekly veggie CSA. It’s about a 30-minute round trip. This could be lost time in terms of my agentic software project I am working on, but I finally figured a way to get Claude to keep coding, even when it thinks it has been running and executing todos for “6-8 hours” (10 minutes).

Anchor Claude in real server time.

Prompt instructing Claude Code to keep working until at least 11:10 AM PT and not stop or ask questions
The exact prompt that triggered a 39-minute autonomous run while I was out picking up the CSA box. The “11:10 AM PT” anchor is the trick. That forces Claude to check real server time instead of guessing.

I guess I should’ve tried this earlier. I do have Claude check the server time at session start and end. Now, I’ve added another layer of reality grounding:

I tell Claude NOT to stop coding or working before a specified time.

Here’s an example when I was about to head out to pick up the veggies. We have a plan in place, in addition to a detailed product roadmap, feature list, and a feature and bug ticketing system in place so the context is rich; and I send Claude Code this prompt (in the Visual Studio Code extension):

Claude, I am going now. So define blocks of work on your own and execute. Follow SDLC best practices. Check server time (real actual time) when complete and if it’s still before 11:10AM PT then define more work blocks and then execute and keep working until at least 11:10AM PT, further if need be! Again, do not stop or ask any questions thanks. I am leaving now!

And, wow, for the first time, Claude actually kept working. After a block of work, he’d check the time, and if it was not yet 11:10 AM as I specified, he would tee-up another block based on our plan, and keep on chugging along.

… do not stop or ask any questions thanks. I am leaving now!

Once I returned with the fresh California veggies, I check-in and review the summary and all the code, docs and everything else completed.

Claude Code session wrap showing 8 merged feature branches with passing tests after 39 minutes of autonomous work.
The receipt: Claude Code’s own wrap-up after the 39-minute autonomous run. 8 blocks shipped, 2 defensively skipped, +100+ tests added… all while I was out running an errand. The server-time anchor held.

Next, I typically run a full AI agent code review using 8 agents. This small fleet is quite incredible. It’s essentially a full-blown engineering and software dev team that can operate 24/7 at super speed. Here’s my latest agentic line-up:

  • Claude Code (Tech Lead) – VS Code Extension
  • Codex (Lead Code Reviewer) – VS Code Extension
  • Minimax – Kilo Code Extension
  • Kimi – Roo Code Extension
  • GLM – Cline Extension
  • Composer 2 – Cursor
  • GPT 5.5 – Cursor
  • Gemini Pro – Antigravity

One potential consideration of holding large planning or code review sessions is Tokenomics. You will eat through a lot, just discussing, brainstorming and so-called ideating. Based on API alone you could probably burn through $10-15 in just one review cycle. But I find it’s worth it. And something like Ollama really helps with reasonable pricing, and access to scores of quality LLMs. The quality upon execution tends to be materially better, and also typically better aligned with my product direction if I spend a significant amount of time bouncing back and forth across the team, and then letting Claude Code eventually synthesize all the recommendations.

So next time you sit down with Claude Code, be it in an IDE or in a terminal CLI session, try grounding him in the reality of time.

And one other tip: tell Claude Code “human rules and idioms” do NOT apply. There is NO such thing as a “break” (he loves suggesting break times! Claude, you are an AI!). There is NO concept of the end of the workday (do not tell me I need some rest please!). In fact, I love teeing up 24/7 SUPER BIONIC AUTONOMOUS SPRINT MODE. Try that. Surely, you’ll hit your $200 Claude Code Max limit when Claude is no longer basing his life on human time.

Tags:Anthropic Artificial Intelligence (AI) Claude Cursor Google Antigravity Google Gemini Human-AI Symbiosis Microsoft Visual Studio Code OpenAI

Related Stories

Which Molty blind LLM study: a four-week single-blind crossover experiment testing whether users can detect the language model powering an always-on AI agent when the memory system stays constant. Results show no statistically significant difference across MiniMax M2.7, Kimi K2.5, GLM-5, and Gemma 4 31B.

Which Molty? Our Blind LLM Study Says Memory Beats Model

News
2026 Artificial Intelligence Index Report from Stanford HAI

Stanford's 2026 AI Index: Where AI Actually Stands (report)

News
Split-screen graphic showing Anthropic Cowork local file system and Google Personal Intelligence cloud services converging into the Integrated Personal Environment IPE.

Google and Anthropic Just Validated the IPE. Now Comes the Hard Part

News
Trifole film truffle hunting Piedmont Italy white truffles foraging tradition Gabriele Fabbro

Trifole: A Lyrical Journey into Italy's Ancient Truffle Tradition

Film Reviews

More in News →

Clinton Stark

Filmmaker and editor at Stark Insider, covering arts, AI & tech, and indie film. Inspired by Bergman, slow cinema and Chipotle. Often found behind the camera or in the edit bay. Peloton: ClintTheMint.

Short Films
Loni Stark - A West Coast Adventure - A Lifetime in the Making - Stark Insider

Stark Insider
  • CULTURE
  • BEST OF AI
  • FILMMAKING/TECH
  • ATELIER STARK FILMS
  • HUMANxAI SYMBIOSIS
THE STARK COLLECTIVE
  • THE STARK CO
  • STARK INSIDER
  • STARKMIND
  • ATELIER STARK
© Copyright 2005-2026 BLG Media LLC. v2.19.0
  • Review Policy and Shipping
  • Privacy Policy
  • Contact
  • About