Your progress syncs automatically across all devices. Sign in, or sign up today and get started!
Your subscription is inactive or does not include access to this content. Reactivate or upgrade your plan to continue.
View Plans & Enroll Sign outSomething powerful is happening inside these models — and most people have no idea how it actually works. This is your briefing. We teach it like a detective investigation: each Case File is a real case to crack, not a textbook chapter to memorize. That's why everything here is labeled as Cases, Evidence, and Debriefs — you're the detective, AI is the case. 8 cases. No jargon walls. By the end, you'll know how to use AI confidently, spot when it's lying, and understand what everyone else in the room is missing.
You gave your agent a task. It started working — searching, reading, writing.
Then it stopped. Mid-task. And now it's asking you to press 'Continue.'
Nothing crashed. Nothing went wrong. This is a feature — once you understand why it exists.
When an agent takes an action — searching the web, reading a file, calling an API, running code, sending a message — that action is called a tool call. The model reasons about what to do, then executes a tool. The tool returns a result. The model reasons again. That's one cycle of the ReAct loop.
Each of those tool executions is a tool call. A single research task might involve: 5 web searches, 8 document reads, 3 data extractions, 1 file write. That's 17 tool calls — and the model hasn't written the summary yet.
Every model has a per-response limit on how many tool calls can occur before it must stop and return control. This is not a bug or a cost-cutting measure. The reasons are deliberate:
When you press Continue, you're starting a new response turn. The agent receives a summary of where it left off, plus any new context from the tools it already called, and resumes execution. It's not starting over — it's picking up from a checkpoint.
Some platforms handle this automatically with a 'compaction' step — the model summarizes its progress, compresses the conversation history to save context space, and continues without requiring a human click. Claude Code uses this approach for long coding tasks.
The limit isn't a constraint to fight — it's a design parameter to work with. Well-designed agents treat each response turn as a logical phase:
If each phase fits within one response's tool call budget, the agent progresses cleanly through human-reviewable checkpoints. If you try to compress all phases into one run, you'll hit the limit mid-task and get a partial result.
// Practical rules for working with tool call limits
When designing an agent task: estimate how many tool calls it will require. Break it into phases.
If your agent stops unexpectedly: press Continue — don't restart. The work so far is preserved.
If the same agent keeps stopping in the same place: that step is too complex. Break it into smaller steps.
For critical tasks: review progress at each Continue checkpoint rather than running through blindly.
For fully automated pipelines: build compaction and continuation logic into the workflow so humans aren't needed at every checkpoint.
// Case Closed: The Continue Button
Tool calls are actions an agent takes using external capabilities.
All models have a per-response tool call limit — a mandatory safety checkpoint.
The Continue button resumes execution from where the agent paused, preserving prior work.
Design agentic workflows in phases that each fit within one response's tool call budget.
The limit is a feature, not a bug — it keeps humans in the loop on long-running autonomous tasks.