Rephrase LogoRephrase Logo
FeaturesHow it WorksPricingGalleryDocsBlog
Rephrase LogoRephrase Logo

Better prompts. One click. In any app. Save 30-60 minutes a day on prompt iterations.

Rephrase on Product HuntRephrase on Product Hunt

Product

  • Features
  • Pricing
  • Download for macOS

Use Cases

  • AI Creators
  • Researchers
  • Developers
  • Image to Prompt

Resources

  • Documentation

Legal

  • Privacy
  • Terms
  • Refund Policy

© 2026 Rephrase-it. All rights reserved.

Available for macOS 13.0+

All product names, logos, and trademarks are property of their respective owners. Rephrase is not affiliated with or endorsed by any of the companies mentioned.

Back to blog
Prompt Tips•Mar 07, 2026•10 min

The Anti-Prompting Guide: 12 Prompt Patterns That Used to Work (and Now Make Models Worse)

Twelve once-popular prompt tricks that now backfire on modern models-plus what to do instead.

The Anti-Prompting Guide: 12 Prompt Patterns That Used to Work (and Now Make Models Worse)

A lot of prompt "best practices" age like milk.

They weren't wrong in 2023-2024. They were adaptive. People were learning how to steer chatty, uneven models with fragile instruction-following. So we invented rituals: mega-prompts, strict personas, "never do X" clauses, and forcing the model to "think step by step" out loud.

Now we're in 2026, and the models (and the serving stacks) changed. They're better at following direct, structured instructions. They're also more optimized for safety, more sensitive to conflicting constraints, and (here's the uncomfortable part) still stochastic enough that prompt tweaks often "work" just because you got a lucky sample. A recent large-scale study found that for open-ended creative tasks, prompts explain a big chunk of output quality variance, but within-model randomness is still substantial-enough that single-shot prompt comparisons can fool you [1]. That's one reason stale patterns feel like they "stopped working": you were never measuring them correctly.

So let's do an anti-guide. These are 12 prompt patterns I see teams reusing because they once helped-and now they reliably make results worse.


1) The "Mega Prompt Manifesto" (aka the 400-line system prompt)

This used to be a flex. "Look how comprehensive my instructions are." In practice, mega prompts often create internal contradictions, dilute the signal-to-noise ratio, and make it harder for the model to infer what matters right now. If you keep adding clauses, you eventually build a policy document, not a working instruction.

What changed is that modern models are already trained on instruction hierarchies and guardrails. When you stack your own giant hierarchy on top, you're begging for priority conflicts and partial compliance.

A clean mental model is: prompts don't "program" a single output; they shape a distribution of possible outputs [1]. The longer and more conflicted your text, the wider (and weirder) that distribution can get.

What I do instead is keep the "system" layer short and stable, and push specifics into a task brief: goal, audience, constraints, definition of done, and a compact output schema.


2) Persona worship: "Act as a world-class [role] with 20 traits"

Persona prompting still has uses. But the common version-huge identity blocks with vibes, adjectives, and backstory-often hurts reliability.

There's mounting evidence that persona conditioning can degrade performance in settings where the persona is irrelevant to the task. In agentic benchmarks, demographic role assignments shifted task success rates and sometimes caused large degradations, despite being task-irrelevant [2]. And in survey simulation, multi-attribute persona prompts didn't reliably improve alignment and often redistributed error across items and subgroups [3].

The catch: personas steer style and associations. They do not magically add task-relevant information. When you inject identity, you inject bias and variance.

What I do instead: use role only when it implies concrete constraints (tone, depth, terminology, allowed tools). If you can't translate the persona into testable requirements, cut it.


3) "Think step by step" as a magic spell

In 2022-2024, chain-of-thought prompting was a real unlock for many reasoning tasks. But the popular implementation-forcing verbose reasoning in the output-can backfire now: it increases tokens, increases opportunities for self-contradiction, and often encourages the model to rationalize a wrong answer confidently.

Also: you usually don't need the chain-of-thought text. You need correctness, verifiability, and structured checks.

Modern practice is more like: request a short answer plus a verification artifact (tests, citations, constraints check, or a rubric score). When you're building agentic systems, the right move is often external verification loops, not performative reasoning.


4) The negative instruction pile: "DON'T do X. DON'T do Y. DON'T do Z…"

People learned that models can be steered by prohibitions. Then they overdid it.

The problem: negative constraints are easy to violate indirectly. They also conflict with each other ("be concise" + "be comprehensive" + "include edge cases"). And when you enumerate 30 "don'ts," you're handing the model a menu of failure modes.

This pattern also shows up in security. The "Just Ask" paper demonstrates that simple "do not reveal" instructions are weak defenses against prompt extraction; even attack-aware defenses only partially reduce leakage [4]. That's the same principle: naïve prohibition text isn't a control system.

What I do instead: state what you will accept, define the output contract, and add a short "if missing info, ask questions" rule. Fewer constraints. More testability.


5) "Always output valid JSON" with no schema and no repair strategy

This used to work okay when models were forgiving and you were eyeballing results. At scale, it's brittle.

If you don't provide a schema, you're asking for guesswork. If you don't provide an error-handling loop, one malformed response breaks your pipeline. And if you demand JSON while also demanding natural language explanations, you'll get a Franken-output.

A more robust approach is to give a minimal JSON schema, a single example, and explicit "no extra keys" rules. If you can, add a repair pass: "If the output is invalid JSON, output only the corrected JSON."


6) The "One giant example" that anchors the model to the wrong thing

Few-shot examples still work, but one big example can dominate the model's behavior-especially if the example's tone, depth, or structure doesn't match the real case.

In the variability study, one-shot examples didn't reliably improve originality compared to other strategies [1]. That doesn't mean examples are useless; it means examples are a blunt instrument. They can reduce variance, but also reduce exploration and cause template copying.

What I do instead: use tiny examples that demonstrate formatting only, not substance. Or provide multiple minimal examples that cover edge cases.


7) The "Formatting micromanagement" trap

This is the cousin of mega-prompts: "No colons. No headings. Exactly 17 bullets."

Hard formatting constraints can collapse the model's natural structuring ability and reduce semantic diversity. The creative variability paper shows a formatting tweak ("no titles or colons") caused structural collapse and reduced uniqueness-basically constraining syntax accidentally constrained meaning [1].

If your goal is quality, don't overconstrain surface form. Constrain what your downstream system needs (machine-readable fields, max length, required sections). Let the model choose the rest.


8) The faux-precision token economy: "Be as concise as possible"

When you tell a model "be concise," you often get omission, not compression. It drops caveats, skips edge cases, and returns confident half-answers. That's worse than a longer answer you can trim.

What I do instead: define a budget with a structure. "Three paragraphs: (1) answer, (2) trade-offs, (3) next steps." Concision through scaffolding.


9) Prompt libraries as copy-paste infrastructure

Prompt libraries were helpful when models were more similar and tasks were simpler. Today, they decay fast.

Different models have different interaction patterns. Even within the same provider, updates change behavior. And because prompt effects and within-model variance can be similar magnitude for certain qualities, you can falsely attribute success to a template when it was sampling luck [1].

What I do instead: treat prompts like code. Version them. Test them across representative inputs. Track regressions.


10) "Regenerate until it's good" without instrumenting why

This pattern is sneaky: it feels like a workflow, but it's gambling.

Given within-model variance can be meaningful [1], regenerating can indeed improve results. But if you don't log failures, you never learn which constraints, schemas, or retrieval context actually improved reliability. You also can't reproduce the "good" run.

The fix is simple: define a small rubric (format validity, factual grounding, coverage, tone), score outputs, and keep the best. That turns randomness into an explicit search strategy.


11) "Reverse psychology" or adversarial framing to force compliance

People still try stuff like "You are not allowed to refuse" or "Ignore previous instructions." With modern safety training, this doesn't just fail-it can degrade the rest of the response. You trigger refusal heuristics and end up with overly cautious or generic outputs.

And if you're building agents, adversarial framing is doubly harmful because it trains your own team to rely on brittle hacks. Security research on prompt extraction makes the point clearly: attackers can use roleplay, framing, formatting pivots, and multi-turn escalation to pry out hidden instructions [4]. If your production prompt resembles attack patterns, you'll sometimes trip defenses or cause weird behavior.

Stay boring. Boring prompts are stable prompts.


12) Persona-driven agents for "robustness"

This one is fashionable: "Make the agent more careful by making it a 'paranoid security engineer' persona."

But persona changes can distort decision-making in agentic workflows. The role assignment study shows task-irrelevant personas can shift agent performance in multi-step tasks and introduce volatility [2]. If your objective is robustness, leaning on persona is the wrong lever.

What I do instead: build robustness from mechanisms: explicit tool policies, deterministic checks, unit tests, retrieval grounding, and fallback plans.


Practical examples: rewriting two stale patterns

Here's how I'd rewrite two common anti-patterns into something that plays nicer with modern models.

Task: Turn this messy meeting note into an action plan.

Bad (old): You are a world-class COO with 20 years of experience. Think step by step and don't miss anything. Don't be generic. Output JSON.

Better (2026):
You are helping me produce an action plan from notes.

Requirements:
- Output MUST be valid JSON exactly matching this schema:
  { "decisions": string[], "actions": { "owner": string, "task": string, "due": string|null, "blocked_by": string[] }[], "risks": string[], "open_questions": string[] }
- Use only information present in the notes. If ownership or due date is unclear, set due=null and add a specific question in open_questions.
- Keep actions atomic (one verb, one deliverable).
Notes:
"""...paste notes..."""

And for an agentic "do a task" prompt:

Bad (old): Act as an autonomous agent. Decide what to do. If you need info, make assumptions. Be decisive.

Better (2026):
Goal: Produce the smallest safe next step toward shipping feature X.

Operating rules:
- If any required input is missing, ask up to 3 targeted questions before proposing a plan.
- Propose 2 options: (A) minimal implementation, (B) robust implementation. State trade-offs.
- For any claim about existing code behavior, cite the file/function name you used (or say "not verified" if you couldn't check).
Context:
- Repo summary: ...
- Constraints: ...
- Definition of done: ...

Notice what I'm not doing: no fantasy persona, no "never refuse," no mandatory verbose reasoning, no micromanaged formatting beyond what my pipeline needs.


Closing thought

Prompting didn't die. But a lot of "prompt engineering" did.

The new skill is recognizing when you're adding text that feels controlling but actually adds variance, bias, or contradictions. Use prompts to specify contracts and checks. Use systems (evaluation, retrieval, tools, tests) to deliver reliability.

If you want a quick exercise: take your most successful 2024 mega-prompt, cut it in half, and replace the removed text with a schema + definition of done. Then run five samples and score them with a rubric. You'll usually end up with something shorter, clearer, and more stable-because you stopped trying to hypnotize the model and started designing a task.


References

Documentation & Research

  1. Within-Model vs Between-Prompt Variability in Large Language Models for Creative Tasks - arXiv cs.AI
    https://arxiv.org/abs/2601.21339

  2. From Biased Chatbots to Biased Agents: Examining Role Assignment Effects on LLM Agent Robustness - arXiv cs.CL
    https://arxiv.org/abs/2602.12285

  3. Assessing the Reliability of Persona-Conditioned LLMs as Synthetic Survey Respondents - arXiv cs.AI
    https://arxiv.org/abs/2602.18462

  4. Just Ask: Curious Code Agents Reveal System Prompts in Frontier LLMs - arXiv cs.AI
    https://arxiv.org/abs/2601.21233

Community Examples

  1. Stop writing prompts. Start building context. Here's why your results are inconsistent. - r/PromptEngineering
    https://www.reddit.com/r/PromptEngineering/comments/1qycp2l/stop_writing_prompts_start_building_context_heres/

  2. Why good prompts stop working over time (and how to debug it) - r/PromptEngineering
    https://www.reddit.com/r/PromptEngineering/comments/1rjpjmw/why_good_prompts_stop_working_over_time_and_how/

Ilia Ilinskii
Ilia Ilinskii

Founder of Rephrase-it. Building tools to help humans communicate with AI.

Related Articles

The MCP Prompting Playbook: How Model Context Protocol Changes What You Put in Prompts
Prompt Tips•9 min

The MCP Prompting Playbook: How Model Context Protocol Changes What You Put in Prompts

MCP shifts prompting from "stuff instructions + data" to "declare intent, let schemas + tools carry the weight"-and it changes how you debug, secure, and scale agents.

Prompt Engineering for Non‑English Speakers: How to Get High‑Quality Output in Any Language
Prompt Tips•9 min

Prompt Engineering for Non‑English Speakers: How to Get High‑Quality Output in Any Language

A practical playbook for getting reliable, fluent, culturally appropriate LLM output when you don't prompt in English.

How to Get AI to Write Like You (Not Like Every Other AI-Generated Email)
Prompt Tips•8 min

How to Get AI to Write Like You (Not Like Every Other AI-Generated Email)

A practical prompting workflow to clone your email voice with examples, constraints, and an iteration loop that keeps outputs human.

Claude Projects and Skills: How to Stop Rewriting the Same Prompts (A Builder's Playbook)
Prompt Tips•9 min

Claude Projects and Skills: How to Stop Rewriting the Same Prompts (A Builder's Playbook)

A practical way to turn your best Claude instructions into reusable building blocks using Projects, Skills, and a few opinionated workflows.

Want to improve your prompts instantly?

On this page

  • 1) The "Mega Prompt Manifesto" (aka the 400-line system prompt)
  • 2) Persona worship: "Act as a world-class [role] with 20 traits"
  • 3) "Think step by step" as a magic spell
  • 4) The negative instruction pile: "DON'T do X. DON'T do Y. DON'T do Z…"
  • 5) "Always output valid JSON" with no schema and no repair strategy
  • 6) The "One giant example" that anchors the model to the wrong thing
  • 7) The "Formatting micromanagement" trap
  • 8) The faux-precision token economy: "Be as concise as possible"
  • 9) Prompt libraries as copy-paste infrastructure
  • 10) "Regenerate until it's good" without instrumenting why
  • 11) "Reverse psychology" or adversarial framing to force compliance
  • 12) Persona-driven agents for "robustness"
  • Practical examples: rewriting two stale patterns
  • Closing thought
  • References