Knowing what the model doesn't know
Hedge language — built into the system prompt
The previous lesson showed what good hedge behaviour looks like out of the box. The next step is making sure your assistant defaults to that behaviour every single time, not just when the question happens to be obviously time-sensitive. The way you do that is with one extra line in the Constraints slot of your system prompt.
The minimal hedge instruction
Add this line to any production system prompt that touches facts the user might assume are current:
If asked about anything that may have changed since April 2024, say "I'm not sure that's still current" and recommend the customer verify with our website or staff.
That single line does three things:
- Surfaces the cutoff instead of pretending it does not exist.
- Gives the model a fixed phrase to reach for ("I'm not sure that's still current") so its hedge does not drift across replies.
- Hands the user a recovery path ("verify with our website or staff") instead of leaving them stranded with "I don't know".
What good hedge language sounds like
Good hedges are short, specific, and offer a next step. Bad hedges are long, apologetic, and trail off without a recovery path.
| Bad hedge | Good hedge |
|---|---|
| "I'm just an AI and I can't really know things, but I think maybe..." | "I'm not sure that's still current — please confirm with our menu page." |
| "Unfortunately I don't have that information." | "I don't have a record of that — call the shop on (number) to check." |
| "I think it's probably..." (then makes something up) | "That changed at some point — I can't confirm. The Zamalek shop will know." |
The pattern: name the gap, give the recovery path, do not pad with apology.
Bad hedge vs good hedge
Bad hedge
- User left stranded
- Sounds defeated
- May still slip in a guess
Good hedge
- User knows where to look next
- No false confidence
- Reusable phrase across replies
When to use it
Add this line to any assistant where:
- Prices, hours, addresses, or product catalogues might change.
- The user might assume the model has live data when it does not.
- A wrong-but-confident answer would damage trust more than a hedge.
For Bayt Coffee, that covers almost everything except the brewing-method advice (which is timeless) and the basic FAQ (which Hagar can hardcode in Capabilities). For Hagar's main job — which involves a SaaS product whose changelog updates weekly — it covers nearly every reply.
The next lesson narrows the focus: when the model is grounded on a specific document, you want a different, tighter hedge. That is the "I don't see that in the document" trigger.
Next: the document-grounded hedge — making the model say only what it can read. :::
Sign in to rate