Concrete input-output pairs included in the prompt to teach the model the desired style, structure, or reasoning without training.
Examples boost quality and reduce retries, but they inflate token cost and latency. PMs must pick the minimal set that moves metrics, and ensure examples stay current. They’re a lever for rapid iteration before committing to fine-tunes.
Curate 3–6 examples that mirror your top user intents and failure modes. Keep them short, anonymized, and labeled with the rule they demonstrate. Rotate examples quarterly and run targeted evals to confirm each still pulls its weight. In 2026, generate synthetic variants automatically, then human-review the best ones before shipping.
A feature that drafts PRDs adds four examples: two good summaries, one refusal, one edge-case with missing data. Accuracy on completeness checklist rises from 68% to 83% while latency stays under 1 s because total tokens stay below 800.