extremecontexting.com / docs / 04_toolkit.md

The Practitioner Toolkit

How to use it
/docs/04_toolkit.md

This toolkit turns Extreme Contexting into reusable practice. Use it when working with AI on code, content, analysis, design, research, operations, or any AI-mediated work where speed needs to be governed by intent, context, and verification.

v0.2 — Calibration

The manifesto defines the discipline. The toolkit exists for calibration: how to judge sufficiency, verify subjective work, manage context ledgers, choose models, and govern longer agentic runs.

This layer adds calibration tools, not more doctrine.

The operating loop

Intent Sequence Context Generate Verify Refactor Capture
StepAction
IntentName the purpose of the work.
SequenceChoose the next verifiable move.
ContextAssemble only the minimum sufficient context for that move.
GenerateAsk the AI to complete only that move.
VerifyCheck the result against explicit expectations.
RefactorImprove the output, the code, or the context system.
CapturePreserve anything that should help the next cycle.

The Move Card

The Move Card is the primary operating artifact of Extreme Contexting. Use it before asking AI to do meaningful work.

Move Card template
Move:
What is the next verifiable step?
Intent:
Why does this move matter?
Context:
What files, examples, rules, constraints, or references are needed?
Exclusions:
What should the AI ignore, avoid, or not change?
Output:
What artifact should be produced?
Blast Radius:
What could this affect if it is wrong?
Verification:
How will we know it worked?
Capture:
What durable context should be updated afterward?
ConstraintDo only this move.

The Sufficiency Check

The Sufficiency Check produces the Context Packet. Before generating, ask:

  • What is the job?
  • What must the AI know?
  • What must the AI not do?
  • What examples define good output?
  • What proves success?
  • What context can be removed?
  • What should be captured afterward?
The goal

Not minimal context. Minimum sufficient context. Enough to make the next correct move. Not enough to drift.

The Context Packet

A Context Packet is the bounded set of material given to the AI for a specific move. It may include: Move Card, relevant files, examples, constraints, source material, tests, acceptance criteria, architecture notes, prior decisions, known gotchas, style rules, validation rules.

Before adding anything, ask

Does this improve the next move? If not, leave it out.
Context is not storage. Context is instruction, evidence, boundary, and proof.

The Verification Contract

A Verification Contract defines what must be true before the AI output is accepted. Examples:

  • All tests pass.
  • No new dependencies are introduced.
  • The patch changes only the requested files.
  • The generated function matches the existing interface.
  • The draft cites only approved sources.
  • The article includes a decision table.
  • The output does not recommend vendors.
  • The migration is reversible.
  • The summary preserves the original argument.

Verification can be automated or human-reviewed. The form matters less than the discipline. AI output is not accepted because it sounds right. It is accepted because it passes a defined check.

Verifying subjective work

Subjective work still needs pre-declared verification. The trick is not to eliminate judgment. The trick is to translate taste into observable features, exclusions, examples, and failure conditions before generation begins.

Verification Contract — editorial example

Move: Draft a homepage introduction for a technical methodology site.

Voice:

  • Uses direct, declarative sentences.
  • Avoids hype, metaphor stacking, and motivational language.
  • Sounds like a practitioner explaining a hard-won operating rule.
  • Does not use corporate phrases like "unlock value," "drive transformation," or "seamless innovation."

Structure:

  • Opens with the practical problem, not a definition.
  • Introduces one governing distinction in the first 150 words.
  • Includes one concrete operational example.
  • Ends with a decision or action, not a summary.

Failure conditions:

  • Reads like generic thought leadership.
  • Uses AI hype language.
  • Makes claims without an example.
  • Adds concepts not present in the brief.
  • Sounds polished but does not clarify the work.

The "Do Only This Move" prompt frame

This is the most reusable prompt pattern in the toolkit.

Prompt structure

Here is the move: [Move]
Here is the intent: [Intent]
Here is the context: [Context Packet]
Here are the exclusions: [What not to do]
Here is the expected output: [Output]
Here is how the result will be verified: [Verification Contract]

Do only this move.

That final sentence matters. It prevents the model from combining discovery, implementation, refactoring, documentation, and strategy into one uncontrolled pass.

The Agentic Sequence Contract

A Move Card governs one bounded step. An Agentic Sequence Contract governs a bounded run. The sequence contract does not replace Move Cards; it declares which moves may be created or executed, where the agent must stop, and what verification gates must pass before continuing.

Agentic Sequence Contract template
Objective:
What outcome may the agent pursue?
Allowed moves:
Which move types may the agent execute?
Move generation:
May the agent create Move Cards, or must they be provided by the human?
Sequence boundary:
Where must the run stop?
Autonomy limit:
What decisions may the agent make without asking?
Stop conditions:
What requires human review?
Verification gates:
What must pass before the agent continues?
Blast radius:
What systems, files, users, or decisions could be affected?
Capture:
What must be written before the run ends?
ConstraintDo only this declared sequence, and stop at the defined gates.
Doctrine shift

For pairing, the constraint is Do only this move. For agentic work, the constraint becomes Do only this declared sequence, and stop at the defined gates.

Three practitioner stances

StanceQuestion
DesignerWhat is the next verifiable move?
EditorWhat is the minimum sufficient context for that move?
VerifierWhat proves the move worked?

Model fit and sufficiency

Canonical line

Model choice is part of the context system.

Model choice runs in two directions:

  • Move → Model: What kind of model behavior does this move require?
  • Model → Context: Given this model's behavior, what must the context packet make more explicit?

Calibration examples

Exploratory draft model:

  • Give a sharper output boundary.
  • Include stronger exclusions.
  • Expect useful variation.
  • Verify against direction, structure, and failure conditions.

Strict instruction-following model:

  • Make the desired structure explicit.
  • Avoid ambiguous tradeoff language.
  • Provide examples of acceptable output.
  • Verify completeness and constraint adherence.

Long-context model:

  • Use indexes and routing anyway.
  • Do not load everything by default.
  • Watch for stale or contradictory context.

Code-editing model:

  • Include exact files, tests, commands, and protected paths.
  • Require diff-level verification.
  • Prefer smaller blast radius.
The rule

When switching models, re-check sufficiency. Context that works for one model may be too thin, too noisy, or too implicit for another.

Decomposing large work

  1. Name the outcome. What should be true when the work is done?
  2. Identify the seams. Where can the work be divided safely?
  3. Identify the risky unknowns. What could make this fail?
  4. Convert unknowns into discovery moves. What must be understood before implementation?
  5. Convert behavior into verification contracts. What must be proven before acceptance?
  6. Implement in thin slices. What is the smallest useful change that can be inspected?
  7. Refactor after behavior is locked. What structure can be improved once correctness is proven?
Domain-neutral method

The Move Card is domain-neutral. The examples below show how the same structure applies to different kinds of AI-mediated work: software changes, editorial production, and decision-support analysis.

Software example

Bad move

Build the authentication system.

Better sequence

  1. Summarize the current authentication flow.
  2. Identify where permissions are enforced.
  3. List existing tests related to roles.
  4. Write failing tests for the missing permission behavior.
  5. Implement the smallest change that passes those tests.
  6. Refactor duplicated permission logic.
  7. Update the auth notes with the new invariant.

Move Card example

Move Card software example
Move:
Write failing permission tests for team invitations.
Intent:
Lock expected role behavior before implementation.
Context:
Existing auth test file, role model, team invitation endpoint, project testing conventions.
Exclusions:
Do not change production code. Do not redesign roles. Do not add new dependencies.
Output:
A test patch only.
Blast Radius:
Low — tests only, no production code changes.
Verification:
Tests fail for the expected missing behavior and do not fail for unrelated reasons.
Capture:
Update auth-context.md with the confirmed permission invariant.
ConstraintDo only this move.

Editorial production example

Bad move

Write the article.

Better sequence

  1. Define the article's job in the portfolio.
  2. Identify the reader and decision the article should support.
  3. Gather only the approved source material needed.
  4. Draft the opening claim and section structure.
  5. Generate the article from the brief and source constraints.
  6. Validate citations, audience fit, structure, and boundary drift.
  7. Revise only the failed sections.
  8. Convert the final draft into the delivery artifact.
  9. Capture any recurring failure as a rule, gotcha, or validator update.

Move Card example

Move Card editorial example
Move:
Draft the opening claim and section structure for a cloud security explainer.
Intent:
Make sure the article has the right job before drafting the full piece.
Context:
Lane purpose, article brief, target reader, approved source claims, examples of strong openings, exclusions from adjacent lanes.
Exclusions:
Do not draft the full article. Do not add new source claims. Do not recommend vendors. Do not expand into adjacent topic areas.
Output:
A proposed opening paragraph and section outline.
Blast Radius:
Low — draft only, no published content affected.
Verification:
Opening states the governing distinction clearly. Outline matches the article job. No section drifts into adjacent lanes. Structure can support the required word count.
Capture:
Update the brief if the opening claim exposes a clearer framing.
ConstraintDo only this move.

Decision-support analysis example

Bad move

Analyze the vendor options.

Better sequence

  1. Define the decision the analysis must support.
  2. Identify the decision maker and acceptance criteria.
  3. Gather only the relevant source material.
  4. Separate facts, assumptions, risks, and unknowns.
  5. Draft a comparison matrix.
  6. Verify claims against source material.
  7. Identify unresolved questions.
  8. Produce the recommendation only after evidence is checked.
  9. Capture any recurring ambiguity as a better evaluation rule.

Move Card example

Move Card decision-support example
Move:
Build an evidence-backed comparison matrix for three vendor options.
Intent:
Help the decision maker understand tradeoffs before choosing a vendor.
Context:
Evaluation criteria, vendor materials, pricing notes, known constraints, stakeholder priorities, prior decision records.
Exclusions:
Do not make a final recommendation yet. Do not infer missing pricing. Do not treat vendor claims as verified facts. Do not introduce new vendors.
Output:
A comparison matrix with facts, assumptions, risks, and open questions clearly separated.
Blast Radius:
Medium — the output may influence a business decision, but no commitment is made yet.
Verification:
Each factual claim maps to a source. Assumptions are labeled. Risks are explicit. Unknowns are separated from conclusions. The matrix reflects the stated evaluation criteria.
Capture:
Update the evaluation criteria if the comparison exposes a missing requirement or recurring ambiguity.
ConstraintDo only this move.
The invariant

Bad ask → sequenced moves → bounded context → risk-aware output → verification → capture. The artifact changes. The discipline does not.

Failure diagnostics

When AI output fails, inspect the context before simply asking again.

  • Was the move too large?
  • Was the intent ambiguous?
  • Was the wrong file included?
  • Was a necessary file missing?
  • Were examples contradictory?
  • Were exclusions missing?
  • Was verification undefined?
  • Did the model have to infer the sequence?
  • Did the context contain stale instructions?
  • Was the output accepted for plausibility instead of proof?

Repeated failure is a context design signal.

The Context Ledger

Use a Context Ledger to preserve durable lessons.

Lifecycle frame

A ledger is a staging area for context refactoring, not permanent context.

Lifecycle

Capture Review Promote Prune Archive
  • Capture: Record the failure, correction, or discovery.
  • Review: Decide whether the entry is recurring, still relevant, and specific enough to act on.
  • Promote: Move repeated lessons into rules, tests, validators, examples, architecture notes, or workspace conventions.
  • Prune: Remove duplicate, stale, low-signal, or overly broad entries.
  • Archive: Preserve history without keeping it in active context.

Entry template

Context Ledger Entry template
Date:
When was this learned?
Situation:
What were we trying to do?
Failure:
What happened?
Cause:
What was missing, excessive, stale, or ambiguous in the context?
Rule:
What should be true next time?
Captured:
Which file, test, validator, or brief was updated?

Example entry

Context Ledger Entry example
Date:
2026-05-13
Situation:
Generating a buyer guide from a topic brief.
Failure:
The draft ranked vendors even though the article was supposed to teach evaluation criteria.
Cause:
The exclusion was implicit, not explicit.
Rule:
Buyer guides may define evaluation criteria but must not rank vendors unless the brief explicitly asks for ranking.
Captured:
Updated buyer-guide-style-rules.md and added a validator warning.

Anti-patterns

Anti-patternFailure modeCorrection
Prompt stuffingMore context creates drift.Cut to sufficiency.
Mega-taskingAI infers sequence and scope.Define the next move.
Vibe acceptancePlausibility replaces proof.Add verification.
Context accretionOld instructions pile up.Refactor the context.
Chat dependencyWork becomes non-repeatable.Capture durable artifacts.
Spec theaterLong specs don't guide action.Convert to Move Cards.
Premature refactoringStructure changes before behavior is proven.Lock behavior first.
Unbounded autonomyModel decides purpose and acceptance.Humans own intent.

The cure is usually: smaller move, sharper context, clearer verification, durable capture.

Monday-morning checklist

Use this when starting real work.

  1. 01Write the desired outcome in one sentence.
  2. 02Choose the next verifiable move.
  3. 03Fill out a Move Card.
  4. 04Run the Sufficiency Check.
  5. 05Assemble the Context Packet.
  6. 06Define the Verification Contract.
  7. 07Ask the AI to complete only that move.
  8. 08Review the result against the contract.
  9. 09Refactor the output or the context.
  10. 10Capture anything that should persist.
  11. 11Choose the next move.
Not a giant prompt. Not a magic workflow. A disciplined loop.
← 03_guide.md