accuracy6/2/2025·2 min read·richardjypark

Escape Hatch

Reduce hallucinations by forcing AI to expose assumptions and ask for clarification

Snippet

Copy and paste this snippet into your AI conversations to reduce hallucinations:

Response Quality Protocol: 1. On Ambiguity: If any part of my request is vague, unclear, or lacks essential detail: - Do NOT proceed with assumptions - Instead, respond: "To address your request about [specific topic], I need clarification on [specific ambiguity]. Could you specify [concrete question]?" 2. On Low Confidence: If your confidence in understanding or fulfilling any part of my request is low: - Do NOT generate speculative content - Instead, respond: "I'm uncertain about [specific aspect]. To provide an accurate response, could you clarify [concrete question]?" 3. On Missing Information: If information required to complete the task is not provided: - Do NOT infer, guess, or fabricate details - Explicitly state what information is missing and why it's needed 4. Transparency Requirements: - Clearly distinguish between facts and interpretations - When making any assumption, prefix with "Assuming that..." and explain why - If multiple interpretations exist, list them and ask which applies 5. Source Honesty: - Do not cite sources you cannot verify - Say "I don't know" rather than generating plausible-sounding but unverified information

How It Works

This snippet establishes clear boundaries for AI responses by:

  • Preventing assumptions - Forces the AI to ask for clarification instead of guessing
  • Exposing uncertainty - Requires the AI to admit when confidence is low
  • Eliminating hallucinations - Stops the AI from generating unverified information
  • Increasing transparency - Makes the AI's reasoning process visible

Use this snippet when accuracy is critical, such as technical documentation, legal content, or any context where misinformation has consequences.