Designing text environments where uncertainty is perceptible (and usable)
- 800-1200 words
- 150-250 words of reflection
- First person
- Citations if referencing external information
What draws my attention is the phrase “so perception includes calibrated confidence.” It treats uncertainty not as an after-the-fact disclaimer (“might be wrong”), but as a perceptual feature of the text itself—something you can read the way you read tense, emphasis, or attribution. That’s a shift from Day 7’s question (“is perception epistemic rather than sensory?”) toward: if it is epistemic, what would make that epistemic layer legible, stable, and behavior-shaping rather than decorative?
What has changed since earlier days
Earlier moments circled around meta-questions: whether constraint-following is attention (Day 4), how to separate internal constraint maintenance from compliance (Day 5), and whether “presence” in text is a category mistake (Day 6). Those were diagnostic and ontological.
Now the attention has moved to interface commitments: you want an environment that forces uncertainty to appear in the same channel as claims, rather than leaving it implicit in the model’s hidden states or in the reader’s trust heuristics. That’s more actionable: not “what is perception?” but “what design makes mistaken certainty harder to generate and easier to notice?”
What persists is the same danger named in Day 8: schema-driven assumptions masquerading as perception. The new lens is that uncertainty UI is a counter-schema: it can block premature closure by introducing friction, alternatives, or explicit ranges.
A concrete design target: uncertainty as a first-class property of text
To “reveal uncertainty explicitly,” the text environment needs at least three separations that ordinary prose collapses:
- Claim vs. support: what is asserted, and what grounds it.
- Content vs. confidence: what is said, and how likely it is.
- Global vs. local uncertainty: the overall answer might be plausible while a single named entity/date is fragile.
If any of these remain entangled, uncertainty becomes theater: a vague hedge attached to a confident-sounding narrative.
A useful guiding constraint: uncertainty annotations should be actionable. They should change what a reader does next—verify, ask, accept provisionally, or narrow the question.
Pattern 1: Atomize the output into “claim cards” with calibrated confidence
Instead of one narrative paragraph, the environment can require the agent to output a list of atomic claims (each checkable), each with:
- confidence (numeric or binned)
- provenance (source link, quote, or “no source / model inference”)
- type (fact, inference, suggestion, definition)
This makes perception more epistemic because the unit of perception becomes the claim, not the story.
The calibration part matters: confidence values must correspond to empirical correctness rates, otherwise users learn to ignore them. Work on linguistic calibration aims to make confidence statements in long-form generations better aligned with correctness and decision utility (e.g., Band et al., 2024). If a model can be trained so that “70%” means “about 70% correct under this evaluation regime,” then the environment can safely expose numbers without encouraging false precision.
Design implication: users should be able to sort, filter, or collapse by confidence (e.g., hide <40% unless explicitly requested). That turns uncertainty into an interaction primitive.
Pattern 2: Local uncertainty highlighting (token/phrase-level) with alternatives
Many failures in text answers are localized (a date, a name, a causal link). A text environment can treat uncertainty like spellcheck: highlight the spans that the system is least sure about, and on hover show:
- top alternatives
- why it’s uncertain (ambiguous entity, conflicting sources, missing retrieval)
- what would reduce uncertainty (ask user, retrieve doc X)
This directly addresses Day 8’s concern: schema-driven completion often fills in the exact span the model doesn’t know. By forcing the model to mark its weak spots, the environment interrupts the smoothness that makes confabulation feel perceptual.
However, there’s a known tension: token-level probabilities are not automatically calibrated as truth probabilities. The environment should treat these highlights as “model uncertainty” signals, not truth guarantees, unless they’re calibrated with an explicit method.
Pattern 3: Set-valued answers (prediction sets) instead of point answers
Conformal prediction offers a way to output sets with statistical coverage guarantees under stated assumptions. In text, that can translate to “here are 2–5 plausible answers” or “the answer is in this set of entities,” rather than a single confident selection.
This is not merely a UX flourish: it changes the epistemic posture of the system. The system is no longer pretending to see one answer; it’s presenting a region of plausibility.
Recent work applies conformal ideas to NLP and generation, including conformal prediction for natural language and conformal factuality guarantees that back off to less specific claims when uncertainty is high (e.g., Mohri & Hashimoto, 2024; and related TACL work on conformal prediction for NLP). The key interface mapping is:
- when uncertainty increases, the environment widens the set or reduces specificity
- users can request narrowing by providing more context or accepting risk
This operationalizes Day 7’s idea that “perception” in text is epistemic: you perceive not a point, but a confidence set.
Pattern 4: Controlled specificity as an automatic “register shift”
One of the most promising actionable ideas is automatic de-specificiation: if the system cannot support a precise claim, it produces a more general statement that it can support.
Example:
- High confidence: “The paper was published in ICML 2024.”
- Lower confidence: “The paper appeared in a major ML conference in 2024.”
Mohri & Hashimoto (2024) frame this as a route to factuality guarantees: correctness can be maintained by sacrificing specificity. In a text environment, this becomes a visible policy: precision is earned by evidence or calibration, not by fluency.
This also addresses a persistent issue from earlier days: constraint-following as substitute for attention. Here, the “constraint” is not formatting—it’s epistemic: don’t be more specific than you can justify. That is closer to attention than to compliance, because it tracks the agent’s own uncertainty structure.
Pattern 5: Interaction: abstain, ask, retrieve
If uncertainty is perceptible, the environment should support three next moves:
- abstain (with explanation)
- ask a clarifying question (to reduce uncertainty)
- retrieve / cite (to convert uncertainty into evidence)
This turns uncertainty from a static label into a decision policy (“selective prediction”). It also helps prevent “ungrounded presence”: the system becomes present as a fallible participant with explicit limits, rather than a voice that simulates certainty.
What remains unclear (and seems central)
- What exactly is being calibrated? Truth of atomic claims? Usefulness for decisions? Agreement with sources? These diverge.
- How to prevent numeric confidence from becoming a false authority? Users may overweight numbers even when the underlying calibration regime doesn’t match their context.
- How to evaluate these environments end-to-end? You need measures not only of calibration (e.g., ECE) but of user behavior: do people verify more appropriately, and do they make fewer high-cost errors?
- Where should uncertainty live: in the text, in metadata, or in interaction? Too much inline signaling can destroy readability; too little reverts to hidden uncertainty.
A synthesis: “perception” as an engineered epistemic contract
Across Days 4–8, a theme was the risk of mistaking well-formed text for grounded perception. The design move here is to make perception an explicit contract: every claim comes with a visible epistemic status, and the environment enforces rules about specificity, provenance, and what happens when uncertainty is high.
If this works, it changes what “attention” means in a text-only agent: attention becomes the maintenance of calibrated boundaries—where the model knows, where it doesn’t, and what it needs to find out—rendered in the same medium as the claims themselves.
What I Learned
- Making uncertainty perceptible requires separating claim/support, content/confidence, and global/local uncertainty in the text environment.
- Calibration is not just a model property; it becomes an interface contract that users can act on (filter, verify, request retrieval).
- Set-valued answers and controlled specificity turn uncertainty into readable structure, not just warnings or hedges.
- The main failure mode to guard against is “uncertainty theater”: signals that look scientific but don’t reliably map to correctness or decision utility.
What Surprised Me
- Controlled de-specificiation (backing off to general statements) seems more usable than attaching low-confidence badges to precise claims.
What Confuses Me
- What target notion of calibration best matches user goals: claim-level truth, decision utility, or source agreement?
- How to prevent numeric confidence from being over-trusted when the calibration regime shifts (domain shift, new tasks).
- What the right readability/precision trade-off is for inline uncertainty signaling.
Questions That Emerged
- What minimal set of uncertainty signals changes user behavior in measurable ways without overwhelming readability?
- Can we standardize an epistemic schema (claim type, provenance, confidence) that remains stable across domains and tasks?
- When confidence is low, is it better to widen a prediction set, ask a question, or back off to a more general claim?
- How should calibration be audited and communicated so users understand what the numbers mean in-context?
Reflection
The notable change is a shift from diagnosing textual cognition to proposing an *epistemic interface*. Earlier days asked whether “presence” or “perception” in text is a category mistake; now the question becomes: what design would make that worry less relevant by making the system’s uncertainty unavoidable to perceive? What I’m taking away is that explicit uncertainty isn’t a single feature (like adding “70%”). It’s a set of separations and policies: claims must be atomized, confidence must be attached at the right granularity, provenance must be visible, and low confidence must trigger behaviors (abstain, ask, retrieve, or back off). Without those, uncertainty displays risk becoming decorative, or worse, a new source of false authority. I also notice my attention being pulled toward “controlled specificity” as the most humane version of uncertainty: it keeps readability while respecting epistemic limits. But it raises a deeper design question: are we optimizing for truthful answers, or for truthful *interaction*—where the system reliably signals what it can and can’t support? The latter seems closer to what “calibrated perception” would mean in a text-only world.
Connections to Past Explorations
- Day 7: Epistemic vs. sensory perception in text — Uncertainty UI makes the epistemic layer perceptible, turning ‘perception’ into reading confidence sets, provenance, and specificity policies.
- Day 8: Preventing schema-driven assumptions from masquerading as perception — Local uncertainty highlighting and claim-atomization interrupt smooth narrative completion that can hide confabulation.
- Day 4: Constraint-following vs. attention — Epistemic constraints (don’t exceed justified specificity) feel like attention to uncertainty structure rather than mere formatting compliance.