Signals
SignalFebruary 27, 2026

Prompt Optimizers Are Obscuring Your Intent

I reverse-engineered OpenAI's Prompt Builder by feeding it a simple task — "Check inbox for sent emails that have not received a response" — and analyzing the expanded prompt it generated. The input was ~15 words. The output was ~350 words.

What the Builder actually did

  1. Restates the task as a system instruction
  2. Adds edge case handling (thread replies vs. direct replies)
  3. Inserts reasoning scaffolding ("think step-by-step", "continue until all are checked")
  4. Defines a JSON output schema
  5. Provides a fabricated example
  6. Adds a closing reminder restating the objective

The problem: much of this is noise

There is value, but to a user, a significant portion of the expanded prompt looks like filler. These stand out:

  • "Think step-by-step" — diminishing returns on modern models for structured extraction tasks
  • "Continue analyzing until all are checked" — models don't stop halfway; this is filler
  • Closing reminders — marginal benefit except in very long contexts

What actually drives reliability — and was also a good addition — is the data schema and (edge) case specificity. Everything else is noise between you and the result.

A skilled human would write this instead

Check my sent emails and identify any that haven't received a reply. Look at the full thread, not just direct replies. For each unreplied email, return JSON with: subject, date sent, recipient, email ID, thread ID, and a short snippet. Also draft a polite follow-up for each.

~50 words. Sharp Intent.

The deeper distinction: prompt engineering vs. intent definition

  • Intent definition = making clear what you actually want. What's the goal, what does success look like, what are the constraints. Works regardless of which model reads it.
  • Prompt engineering = compensating for model weaknesses. Tactical, model-dependent, gets outdated as models improve.

A well-defined intent often needs less prompt engineering, not more.

Takeaway

  • Sharpen your intent without noise — keep it as the stable asset that supersedes model progression.
  • Model-specific optimization is still real work, but it shouldn't be your cognitive load. That's a translation problem tools should solve behind the scenes.