← All posts

LLM Prompt Engineering AI

Prompting Is Not Magic

Reading The Prompt Report made me rethink what "good prompting" actually means


"The real skill is not finding the perfect prompt. It is building a process for improving prompts over time."

I recently read The Prompt Report: A Systematic Survey of Prompt Engineering Techniques, and what I appreciated most about it was its lack of mysticism.

A lot of writing about prompting still sounds like folklore. Somewhere out there, we are told, there exists a perfect phrase — a secret arrangement of instructions that will unlock the model's full intelligence. The prompt becomes a kind of spell. If it fails, the assumption is that we simply have not found the right wording yet.

This paper pushes in the opposite direction. Instead of treating prompting as a collection of clever tricks, it treats it as a field that can be organized, analyzed, and improved systematically. That shift sounds small, but I think it matters. It changes prompting from something you perform into something you engineer.

And that is the main reason I found the paper useful.

The paper's biggest contribution is not a trick, but a map

At a high level, The Prompt Report tries to bring order to a space that has grown faster than its vocabulary. It proposes a structured taxonomy of prompting methods, clarifies terminology, and surveys a large body of prior work rather than focusing on a single headline technique.

What that means in practice is that the paper is less interested in selling you one "best" method than in helping you understand the landscape: what kinds of prompting techniques exist, how they differ, where they are used, and what assumptions they rely on.

That alone makes it more valuable than most prompt advice floating around online. The internet is full of prompt templates. What is much rarer is a framework for thinking about prompts.

The most important idea: prompting is iterative

The biggest mental shift I took from the paper is this: prompt engineering is not about writing a brilliant instruction once. It is about iterating on a task interface.

That framing is simple, but it cuts through a lot of confusion. A prompt is not just a sentence. It is a working surface between a human intention and a model behavior. And like most interfaces, it rarely comes out right on the first try.

  • You test it.
  • You inspect failures.
  • You tighten the wording.
  • You restructure the output.
  • You try examples.
  • You compare variants.
  • You repeat.

This sounds closer to product design or experimental workflow tuning than to writing. And I think that is exactly right.

Once you see prompting that way, "being good at prompts" starts to mean something different. It is no longer mainly about sounding clever or authoritative. It becomes a matter of being systematic: defining the task clearly, constraining ambiguity, evaluating outputs, and making small changes deliberately. That feels much closer to reality than the usual mythology.

Most prompt problems are not language problems

One of the most useful distinctions in the paper is between prompt engineering and answer engineering.

Prompt engineering is about how you ask. Answer engineering is about how the model answers — and, more importantly, whether that answer is actually usable.

This is an underappreciated point. In many workflows, the failure is not that the model misunderstood the task. The failure is that it returned something messy, inconsistent, or difficult to parse. The model may be directionally correct and still practically unusable.

That is why output design matters so much:

  • If you need one label, ask for one label.
  • If you need structured data, ask for structured data.
  • If only three values are allowed, define the set explicitly.
  • If a downstream system needs predictable formatting, do not leave the shape of the answer open-ended.

This seems obvious in retrospect, but many people start by trying to make the model "smarter" when the real issue is that the output contract is too loose. In that sense, a surprising amount of prompt engineering is really interface discipline.

Small changes matter more than they should

Another recurring theme in the paper is prompt sensitivity. Models can be strangely responsive to details that humans would consider minor: formatting, ordering, delimiters, wording, the placement of examples, even stylistic inconsistencies. That makes prompting feel unstable in a way that traditional software often does not.

You can have two prompts that mean nearly the same thing to a human reader and still produce noticeably different results. That has two implications.

First, prompt design deserves more rigor than people often give it. If details matter, then versioning, evaluation, and careful comparison matter too.

Second, we should be cautious about over-explaining success. When a prompt works, it is tempting to build a story about why it works. Sometimes that story is right. Sometimes the model is responding to something much more local and idiosyncratic than we realize. In other words: prompting rewards experimentation, but it punishes overconfidence.

"Advanced" prompting is not automatically better

The paper also includes benchmarking across different prompting strategies, and one of the healthier lessons is that more elaborate methods do not always win.

That should not be surprising, but it is a useful corrective. Once a technique gets a recognizable name, it tends to acquire status. People start to treat it as an upgrade by default. Add chain-of-thought. Add self-consistency. Add examples. Add decomposition. Add another layer.

But sophistication is not the same thing as fitness for purpose. Some tasks benefit from extra structure. Some do not. Some prompting methods help only under specific conditions. Some improve one dimension of performance while making another worse. The right comparison is not between "simple" and "advanced" in the abstract — it is between alternatives on the task you actually care about.

The practical lesson is straightforward: treat prompting techniques like experimental variables, not prestige markers.

The case study is the most honest part of the paper

The section I found most memorable was the real-world case study on a difficult classification problem involving mental-health-related language.

What makes that section so strong is that it does not pretend prompt engineering is clean. The paper shows an actual iterative process: attempts, revisions, false starts, output-format problems, model behavior that is hard to explain, and gradual improvement rather than a magical breakthrough. At one point, a duplicated piece of text in the prompt appeared to help performance in a way that was not easy to justify. That detail stayed with me because it captures the current state of the field better than any elegant diagram could.

Prompting today is partly engineering and partly empirical negotiation with a system that is not fully transparent. That does not make it useless. It just means the right mindset is not certainty. It is disciplined iteration.

What I changed after reading it

After reading The Prompt Report, I came away with a much simpler view of what good prompting practice should look like:

  1. Start with the task, not the wording. Before you optimize the phrasing, define what success actually is. Are you classifying, extracting, summarizing, rewriting, or generating? A vague task usually leads to a vague prompt.
  2. Constrain the output early. A clean answer format often matters more than a clever instruction. If the output needs to be reliable, specify its shape as early as possible.
  3. Build a baseline before adding techniques. Try the simplest workable version first. Without a baseline, you cannot tell whether examples, reasoning steps, or extra scaffolding actually improved anything.
  4. Treat formatting as a real variable. Delimiters, ordering, and examples are not cosmetic. They can change behavior in meaningful ways.
  5. Version your prompts. A prompt that works today may drift tomorrow, especially when the underlying model changes. Good prompts are maintained, not merely written.
  6. Bring domain knowledge into the loop. Prompt quality is not just about language. It depends on understanding the task well enough to define what the model should and should not do.
  7. Do not confuse one good output with a robust workflow. An impressive demo is not the same thing as a stable system. Confidence should come from repeated evaluation.

The bigger takeaway

If I had to reduce the paper to one sentence, it would be this: prompting is becoming less like prompt writing and more like behavior design.

That is a good thing. It means the field is maturing. It means we can stop talking about prompts as if they are magical artifacts and start treating them as components in a broader system: task framing, output design, evaluation, monitoring, and iteration.

That is also why I think this paper is worth reading even for people who already use language models every day. Not because it contains a secret method, but because it helps replace superstition with method. And right now, that is probably more valuable.