EXAMPLE INTAKE: SCORE 12 — STANDARD DISCOVERY

Stop guessing how much discovery
a problem needs.

DD6 structures the work. CIRK governs the execution.

A proposed standard that classifies problem complexity across six dimensions to determine how much discovery is needed before an AI agent touches it.

IIntentHow clear is what the requester wants?
DDomainHow much specialist knowledge is needed?
SStakeholdersHow many perspectives must align?
TTestabilityCan an agent verify the outcome?
PPrecedentHave we done something similar before?
BBoundaryAre the problem boundaries defined?

Intake: add multi-tenant isolation to the API

I: 2 | D: 3 | S: 2 | T: 2 | P: 3 | B: 2

→ Score 14 → Standard Discovery (2–4 structured sessions)

Understand the problem before the agent touches it.

One-size-fits-all asks

"Same process for everything?"

DD6 asks

"How much does this problem need?"

The shift

AI changed what matters.

AI agents can write code in seconds. But they cannot understand vague problems. What matters now is not how fast the code is written. It is:

DD6 classifies that reality.

Discovery is not overhead. It is error reduction.

The quality of the input determines the quality of the output.

The model

Six dimensions. One discovery depth.

Each intake is scored from 1 to 3 across six dimensions. The vector defines how much discovery is needed.

[ I, D, S, T, P, B ] → Discovery Depth
I

Intent Clarity

How clear is what the requester actually wants?

I1Crystal clear, spec-ready
I2Partially clear — goal understood but specifics need refinement
I3Vague or ambiguous — cannot articulate the desired outcome
D

Domain Depth

How much specialist knowledge is needed?

D1Standard domain — general engineering knowledge sufficient
D2Moderate specialization — specific business rules or patterns
D3Deep expertise — security, compliance, performance, domain science
S

Stakeholder Convergence

How many perspectives need to align?

S1Single stakeholder defines the problem
S2Two to three stakeholders — engineering + product or design
S3Multiple stakeholders — cross-functional alignment required
T

Testability

Can an AI agent verify the outcome is correct?

T1Easily testable — clear metrics, deterministic outcomes
T2Partially testable — some aspects require judgment
T3Hard to test — subjective outcomes, requires human evaluation
P

Precedent

Have we done something similar before?

P1Strong precedent — many similar examples, established patterns
P2Partial precedent — related work exists but new aspects
P3No precedent — completely new territory
B

Boundary Clarity

Are the boundaries of the problem clear?

B1Clear boundaries — well-defined scope, explicit in/out
B2Partial boundaries — general scope understood, edges fuzzy
B3Unclear — open-ended, scope is part of what needs discovery

Composite score = I + D + S + T + P + B

6–8 Skip No discovery needed — skip straight to spec
9–11 Shallow 1–2 focused sessions — clarify the dominant dimension
12–14 Standard 2–4 structured sessions — intent, scope, risk progression
15–17 Deep 4+ iterative sessions — hypothesis testing and validation
18 Emergency Stabilize first — triage, then classify and discover

"Score the discovery reality, not the political preference."

DD6 Scoring Guidance

What DD6 is not

× An effort estimate
× A risk assessment (that is CIRK)
× A sprint planning tool
× A replacement for discovery

DD6 governs how much discovery to do — not the discovery itself.

DD6 in 60 seconds

Four steps.

01

Pick an intake

Any unit of work entering your backlog: a feature, a bug, a refactor, a migration.

02

Score it

Assign I, D, S, T, P, B values from 1 to 3. Ask: how clear? how deep? how many stakeholders? how testable? how precedented? how bounded?

03

Apply discovery depth

The vector maps to a discovery mode — skip, focused, structured, iterative, or stabilize. The depth is proportional to the complexity.

04

Use it consistently

Teams gain a shared language: "This is high D." "Low P, we need exploration." "I3 — clarify intent first."

Rule of thumb

High I→ more intent clarification needed
High D→ bring in domain experts
High S→ schedule alignment sessions
High T→ define acceptance criteria rigorously
Low P→ more exploration, less assumption
Low B→ scope definition is the first priority

Discovery policies

DD6 as policy-as-code.

The vector defines what happens — not just how complex the problem is.

Skip
Score 6–8 · all ≤ 2

No discovery session needed. Generate spec directly from intake using a template.

Focused
Score 9–11

1–2 focused sessions. Clarify the dominant dimension. Structured Q&A, not open-ended exploration.

Structured
Score 12–14

2–4 sessions with defined phases. Intent → scope → risk progression. Exit criteria per phase.

Iterative
Score 15–17

4+ sessions. Hypothesis testing and validation. Architectural exploration. Multi-stakeholder workshops.

Stabilize
Score 18 · incident

Human-led stabilization. No structured discovery during crisis. Triage → stabilize → then classify.

Policy rules by dimension signal

I3Intent clarification is the first priority
D3Domain expert must participate in discovery
S3Multi-stakeholder alignment session required
T3Define acceptance criteria rigorously before spec
P3Exploration and prototyping recommended
B3Scope definition is the first phase of discovery

Examples

DD6 in practice.

Each vector maps to a concrete discovery depth, with reasoning for each dimension.

I1 D1 S1 T1 P1 B1 → 6 Skip

Add dark mode toggle to settings page

I1 crystal clear intent D1 standard engineering S1 single stakeholder T1 easily testable P1 strong precedent B1 clear boundaries
✓ Skip discovery ✓ Generate spec from template
I2 D2 S2 T2 P2 B2 → 12 Standard

Add real-time notifications for governance violations

I2 partially clear intent D2 moderate domain knowledge S2 engineering + product alignment T2 partially testable P2 partial precedent B2 fuzzy edges
⚠ 2–4 structured sessions ⚠ Stakeholder alignment needed
I3 D3 S1 T2 P3 B3 → 15 Deep

Intermittent data corruption in batch processing

I3 vague — cannot articulate cause D3 deep expertise required S1 single team owns it T2 partially testable P3 no precedent B3 boundaries completely unclear
× 4+ iterative sessions × Hypothesis testing required × Domain expert required

Design principles

Built to be adopted.

Open

MIT licensed. No vendor lock-in. Any team, tool, or platform can implement it.

Simple

Six dimensions, scored 1–3. No certifications, no training required.

Proportional

Discovery depth matches problem complexity. Simple problems skip. Complex problems get what they need.

Composable

Works standalone or as upstream input to CIRK. Problem → DD6 → Discovery → Spec → CIRK → Execution.

FAQ

Common questions.

Is DD6 a replacement for discovery sessions? +

No. DD6 governs how much discovery to do. It does not replace the sessions themselves.

Does DD6 estimate time? +

Not directly. DD6 maps to discovery depth (number of sessions), which has time implications. But duration is derived from depth, not scored directly.

Can DD6 be used without CIRK? +

Yes. DD6 is a standalone standard for problem classification. However, the full value emerges when DD6 (upstream) feeds into CIRK (downstream) as a unified framework.

Is DD6 tied to Orbit618? +

No. Orbit618 is one possible implementation environment for DD6, but DD6 is designed as a standalone open standard.

Is DD6 a product? +

No. DD6 is a standard and a shared classification language. Products and platforms may implement it, but the model itself is implementation-agnostic.

Why six dimensions instead of four (like CIRK)? +

CIRK has four dimensions because execution risk can be captured in four independent axes. Problem complexity requires more axes because the problem space has more independent variables: intent, domain, stakeholders, testability, precedent, and boundaries each vary independently and affect discovery depth differently.

Why not use Cynefin directly? +

Cynefin classifies problems into domains (Clear, Complicated, Complex, Chaotic) but does not provide a scoring model or dimensions specific to software. DD6 uses Cynefin's domains as depth labels but adds the six-dimensional scoring that makes classification systematic and auditable.

Can I score DD6 with AI? +

Yes. DD6 supports heuristic scoring (rules-based) and AI-assisted scoring (a model evaluates the intake and assigns scores). AI-assisted scoring is especially useful for high-volume intake triage. See scoring guidance for practical advice and examples.

What is the most important dimension? +

It depends on your context. For teams with many stakeholders, S is often the strongest signal. For teams with novel domains, D and P dominate. For vague requirements, I and B matter most.

Open questions

Where DD6 might be wrong.

These are tensions we are still debating. If you have answers or counterexamples, we want to hear them.

1

Should DD6 include a "risk" dimension?

DD6 measures problem complexity, not execution risk (that is CIRK). But some teams argue that risk awareness during discovery changes how much exploration is needed. Should risk be a seventh dimension, or does CIRK already cover it?

2

Can AI reliably score DD6 without human calibration?

AI-assisted scoring works for high-volume triage, but accuracy depends on context the model may not have. Is human calibration always needed for the first pass, or can teams trust AI scoring from day one?

3

How does DD6 interact with existing triage processes?

Most teams already have intake workflows: Jira triage, backlog grooming, planning poker. DD6 adds a classification layer. Can it slot in without disrupting those flows, or does it require process change?

Have a perspective? Join the discussion on GitHub →

Try DD6 with your team.

MIT licensed. No dependencies. Works with anything. We want to see where it fails.