Techniques

Meta-Prompting

Using a language model to write or improve prompts that another model (or the same one) will execute -- treating prompt engineering itself as a prompt engineering task.

First published April 14, 2026

Meta-prompting turns a model into your prompt author. You describe the task you want to solve, and the model generates a structured prompt (with instructions, constraints, examples) you can then use in production.

It works because frontier models have absorbed enormous amounts of prompt engineering content during training. They know the patterns. The practical move: draft v0 yourself, meta-prompt an improvement pass, iterate 2-3 times. Diminishing returns kick in fast; a human with domain knowledge usually beats meta-prompting for the final 10%.

Example Prompt

I need to classify customer support tickets into priority levels (low, medium, high, critical). Write a system prompt that:
- Defines each priority level with clear criteria
- Tells the model to output only the label
- Handles ambiguous cases by defaulting to medium
- Resists prompt injection from ticket content
Return only the system prompt, nothing else.

When to use it

  • You're stuck on a prompt and want a fresh structure
  • You need to generate many similar prompts at scale
  • Bootstrapping a new task without a reference implementation

When NOT to use it

  • You already know the domain better than the model
  • The task is trivial and a human-written prompt is faster
  • Production critical paths where you can't verify the generated prompt