Quality Indicators

How Threader measures output confidence.

Every AI-generated output in Comms Threader includes confidence scores. These tell you how grounded the output is in your inputs and research. Understanding these scores helps you decide when to refine, when to override, and when outputs are pitch-ready.

Strategic Confidence

The Strategic Confidence score appears on individual tool outputs (like The Story's problem reframe or The Audience's audience analysis). It measures how well the AI output is supported by three factors:

Factor What It Measures How to Improve It
AI Knowledge How well the output aligns with established strategic frameworks, category knowledge, and best practices This is automatic, based on the AI's training data. High scores (90%+) mean the output follows sound strategic principles
Client Input How much specific, actionable information you've provided about the client, brand, market, and brief Upload documents, add detailed constraints, include research findings, specify objectives clearly
Completeness Whether all required fields are filled and the tool has enough context to generate quality outputs Fill optional fields when relevant, don't skip uploads, add success metrics if available

How Scores Are Calculated

Strategic Confidence is a weighted average of multiple factors. The exact weights vary by tool based on what matters most for that type of analysis:

Example: In The Story, if AI Knowledge is 95%, Client Input is 50%, and Completeness is 100%, the overall score is: (0.50 × 95) + (0.30 × 50) + (0.20 × 100) = 82.5%

Why do weights vary by tool? Different strategic tasks require different types of input. The Story needs strong conceptual frameworks (high AI weight), while The Message needs competitive context (balanced across factors). Client input consistently matters across all tools. The more specific information you provide, the higher your confidence scores.

What Scores Mean

Score Range Label Meaning Action
85-100% HIGH Output is strongly grounded in research and strategic best practices. Ready to use. Review for accuracy, export if satisfied
70-84% MEDIUM Output is sound but could benefit from more client-specific context or research. Add more detail, upload documents, or use Override to refine
Below 70% LOW Output may be too generic or speculative. Lacks sufficient grounding. Upload briefs/research, fill optional fields, or start over with better inputs

Pitch Confidence

Pitch Confidence appears at the top of your screen in Cascade Mode once you've completed The Story. It tracks the overall quality of your pitch cascade across all five tools.

What it measures: The aggregate confidence of all completed tools. If The Story has 83% confidence, The Audience 88%, and The Message 75%, your Pitch Confidence reflects the average quality across your entire strategic flow.

Pitch Confidence only appears in Cascade Mode. Free Roam mode doesn't cascade data, so there's no cross-tool confidence to track.

Why It Matters

Pitch Confidence helps you spot weak links in your strategic thinking. If your overall pitch confidence is 85% but one tool scores 65%, that's your vulnerability. Strengthen that section before presenting to clients.

Example: Your Story and Audience both score HIGH (85%+), but your Message only scores 68% MEDIUM. This suggests your positioning isn't as well-supported as your problem definition and audience analysis. You might need more competitive research or clearer differentiation.

When to Trust AI Outputs

High Confidence Outputs (85%+)

These are typically safe to use as-is, but always review for:

Medium Confidence Outputs (70-84%)

Use these as strong starting points, but plan to:

Low Confidence Outputs (Below 70%)

These need significant work. Either:

Common Confidence Issues

Low Client Input Score

Symptom: AI Knowledge 90%, Client Input 40%, Completeness 80%

Fix: Upload client briefs, add budget/timeline constraints, include research docs, specify objectives

Low Completeness Score

Symptom: AI Knowledge 95%, Client Input 75%, Completeness 50%

Fix: Fill optional fields that are relevant, add success metrics, upload supporting documents

Consistently Medium Scores Across Tools

Symptom: All tools score 70-75%

Fix: You're providing adequate input but not excellent input. Add more detail at each step, especially in The Story-better problem definition lifts all downstream tools

Confidence vs Quality

High confidence doesn't mean the AI is right-it means the output is well-supported by your inputs. If your inputs contain wrong information, you'll get confident but inaccurate outputs.

Always validate:

Use confidence scores as quality indicators, not replacements for judgment. A 95% confidence score on a badly framed problem is still bad strategy.

Improving Scores Over Time

As you use Threader more, you'll develop a sense for what inputs produce high-confidence outputs:

If you want higher... Do this...
AI Knowledge This is automatic and usually high. If low, your request may be unclear or outside strategic norms
Client Input Upload documents, add constraints, specify objectives, include research, fill optional fields with real data
Completeness Fill all required fields, add optional context when relevant, upload supporting documents even when optional

Technical Notes

Score Persistence

Confidence scores are calculated when outputs are generated and saved with your project. If you edit an output manually (using Override), the score doesn't change-it reflects the AI's original confidence, not your edits.

Cross-Tool Impact

In Cascade Mode, low confidence in upstream tools (like The Story) can lower confidence in downstream tools (like The Audience). If The Story has weak client input, The Audience inherits that weakness because it's building on incomplete context.

Regeneration

Using the Refine button regenerates outputs and recalculates confidence scores. If you added more context since the first generation, the refined output may score higher.

Launch Threader