jknavigate Enteropen Ctrl+dCtrl+uscroll ggscroll top Gscroll bottom

Adversarial Prompting

Adversarial prompting is the technique of framing prompts so one AI session works against another—or against its own prior output after clearing context.

Examples

Why It Works

The key is removing collaborative framing. When you clear context and prompt with an adversarial frame, the model produces output that simulates opposition rather than agreement. It’s not defending work it just helped create.

The Pattern

  1. Complete the initial work (write code, generate tests, etc.)
  2. Clear the context or switch models
  3. Prompt with an adversarial frame (“find issues”, “break this”, “what would fail”)
  4. Use the critical output to improve the original work