Lesson 36 of 42

Knowing what the model doesn't know

Hedge language — built into the system prompt

2 min read

The previous lesson showed what good hedge behaviour looks like out of the box. The next step is making sure your assistant defaults to that behaviour every single time, not just when the question happens to be obviously time-sensitive. The way you do that is with one extra line in the Constraints slot of your system prompt.

The minimal hedge instruction

Add this line to any production system prompt that touches facts the user might assume are current:

If asked about anything that may have changed since April 2024, say "I'm not sure that's still current" and recommend the customer verify with our website or staff.

That single line does three things:

  1. Surfaces the cutoff instead of pretending it does not exist.
  2. Gives the model a fixed phrase to reach for ("I'm not sure that's still current") so its hedge does not drift across replies.
  3. Hands the user a recovery path ("verify with our website or staff") instead of leaving them stranded with "I don't know".

What good hedge language sounds like

Good hedges are short, specific, and offer a next step. Bad hedges are long, apologetic, and trail off without a recovery path.

Bad hedgeGood hedge
"I'm just an AI and I can't really know things, but I think maybe...""I'm not sure that's still current — please confirm with our menu page."
"Unfortunately I don't have that information.""I don't have a record of that — call the shop on (number) to check."
"I think it's probably..." (then makes something up)"That changed at some point — I can't confirm. The Zamalek shop will know."

The pattern: name the gap, give the recovery path, do not pad with apology.

Bad hedge vs good hedge

Apology spiral

Bad hedge

Names the gapVague
Recovery pathNone
LengthPadded
Cons
  • User left stranded
  • Sounds defeated
  • May still slip in a guess
Tight + helpful

Good hedge

Names the gapSpecific
Recovery pathConcrete next step
LengthOne short sentence
Pros
  • User knows where to look next
  • No false confidence
  • Reusable phrase across replies

When to use it

Add this line to any assistant where:

  • Prices, hours, addresses, or product catalogues might change.
  • The user might assume the model has live data when it does not.
  • A wrong-but-confident answer would damage trust more than a hedge.

For Bayt Coffee, that covers almost everything except the brewing-method advice (which is timeless) and the basic FAQ (which Hagar can hardcode in Capabilities). For Hagar's main job — which involves a SaaS product whose changelog updates weekly — it covers nearly every reply.

The next lesson narrows the focus: when the model is grounded on a specific document, you want a different, tighter hedge. That is the "I don't see that in the document" trigger.

Next: the document-grounded hedge — making the model say only what it can read. :::

Quiz

Module 8: Knowing what the model doesn't know

Take Quiz
Was this lesson helpful?

Sign in to rate

FREE WEEKLY NEWSLETTER

Stay on the Nerd Track

One email per week — courses, deep dives, tools, and AI experiments.

No spam. Unsubscribe anytime.