I prompt AI ChatGPT and Claude all day long. Writing, refining, debugging, repeating.
And like anyone working with these systems daily, I fall into patterns—shortcuts that used to work, habits I don’t always question.
So I stepped back and did an audit: What’s still effective in Summer 2025? What’s outdated?
The models have changed. The defaults have shifted. And while prompting has gotten easier, getting consistent, accurate results at scale still depends on how you ask.
This post is a little geeky—but it could significantly improve your results with relatively little effort.
Here are five techniques that continue to deliver—and the ones that are no longer worth your tokens.


The right hires make the difference.
Scale your AI capabilities with vetted engineers, product managers, and builders—delivered with enterprise rigor.
AI-powered candidate matching + human vetting.
Deep talent pools across LatAm, Africa, SEA.
Zero upfront fees—pay only when you hire.


Five Prompting Methods Every AI User Should Know
How to fine-tune your prompts for consistent, higher-quality AI outputs
Mastering Prompt Engineering in 2025
Prompt engineering isn't going away. Despite improvements in model intelligence, well-constructed prompts remain critical—especially when precision, control, or scale is required. In real-world systems, a prompt can make or break the product.
With the right structure, you can increase task accuracy by up to 70%. This guide details five techniques that continue to deliver results—along with real examples and the studies that validate them.
1. Few-Shot Prompting
Show the model what good looks like. Instead of describing your desired output, provide a few representative examples of input–output pairs.
Models are pattern recognizers. The more concrete examples you provide, the better the model understands the task.
Check out this guide on how to get better at few-shot prompting from my friends at Prompt Hub.
Example:
When writing a custom code generation tool for internal APIs, a team used three sample API documentation-to-code conversions. The few-shot prompt produced syntactically correct and secure Python output, replacing a brittle template engine. Results outperformed zero-shot prompts by a wide margin, as PromptHub benchmarks show. This was for a production system, but the same tactic works just as well in chatbot interactions.
How To Use:
Example 1:
Input: "Summarize our Q2 revenue meeting"
Output: "Revenue up 12%, driven by US enterprise renewals..."
Example 2:
Input: "Summarize team offsite debrief"
Output: "Key takeaways: async-first habits, sales enablement gaps..."
[Now Prompt]
Input: "Summarize last Friday's roadmap discussion"
2. Decomposition
Overview: Break down complex tasks into subproblems. First, ask the model to list the steps it needs to solve the task, then guide it through solving each one sequentially.
LLMs often falter on compound reasoning tasks. Structuring the process improves accuracy and interpretability.
Example:
A car dealership chatbot needed to process vague return requests. The prompt was refactored to ask the model:
Is this a valid customer?
What car did they rent?
What’s the return window policy?
Do they qualify?
This decomposition flow increased resolution accuracy by 40% in live deployment.
Academic Evidence:
The Plan-and-Solve Plus framework outperformed zero-shot and chain-of-thought prompting on complex math problems.
How To Use:
Task: Can this customer return the car they rented?
Prompt: Before answering, list the subproblems you need to solve.
Response:
1. Identify the customer
2. Check the rental date
3. Determine the car's condition
4. Match return window policy
Then solve each step in order and summarize whether the customer is eligible for a return.
3. Self-Criticism
Ask the model to evaluate its own output before finalizing. Prompt it to identify weaknesses, then revise the response based on that feedback.
Self-reflection activates reasoning capabilities the model might not otherwise apply in a single-pass generation.
Example:
A fintech tool used GPT-4o to draft investment memos. After the initial output, a follow-up prompt asked, "Critique this memo for clarity, logical gaps, and unsupported claims.” Revisions based on that feedback increased analyst satisfaction scores by 33%.
Academic Evidence:
Chain-of-thought with self-consistency improved reasoning on logic and math datasets.
How To Use:
Initial Prompt: "Write a 300-word memo explaining the benefits of our Q4 restructuring."
Follow-Up: "Review your response. Identify any weaknesses, unclear claims, or missing data."
Final Step: "Revise the memo based on your critique."
4. Additional Information (a.k.a. Context Injection)
Provide relevant background at the top of your prompt. This can include definitions, business context, or task-specific data.
LLMs don’t know your internal goals, product strategy, or workflows—unless you tell them.
Example:
A startup doing clinical NLP analysis embedded research definitions of "entrapment" (a psychological condition) into the prompt. Without that background, GPT-4 misclassified the symptom. Including two paragraphs of academic context improved classification accuracy from 51% to 87%.
Academic Evidence:
IBM’s RAG + few-shot study confirmed that preloaded context consistently improved classification and summarization outcomes.
Example-as-Prompt (EaP) showed that well-curated examples retrieved from memory boosted e-commerce performance metrics across search, QA, and classification.
How To Use:
[CONTEXT]
Entrapment is defined as a persistent belief of being trapped in a situation with no perceived way to escape...
[INSTRUCTION]
Classify the following Reddit post as expressing entrapment or not.
5. Ensembling
Overview: Query multiple models (or multiple prompts) and select the most consistent or accurate response. Think of it as consulting a panel of advisors rather than relying on a single one-shot response.
Why It Works: No prompt is perfect. But when multiple prompts converge on the same answer, accuracy increases.
Enterprise Example:
Walmart ensembled LLaMA 2, GPT‑4, and PaLM‑2 to extract product attributes from descriptions. The ensemble outperformed every individual model and drove measurable improvements in click-through and conversion rates.
Academic Evidence:
Boosted prompt ensembles showed ensembling lifted scores on complex math tasks like GSM8k and AQuA.
How To Use:
Prompt A: Use chain-of-thought to solve this math problem.
Prompt B: Solve the same problem directly.
Prompt C: Ask it in the style of a tutor.
Collect responses → Apply majority vote or scoring criteria.
What to Stop Doing
Role Prompting (for accuracy tasks)
Telling the model “You are a math professor” or “You are a world-class researcher” doesn’t improve accuracy. It may affect tone but not performance on objective tasks.
PromptHub has another great overview of why and when role or persona prompting might have a positive effect on outcomes.
Chain-of-Thought Prompting
This comes with a caveat: many of the most powerful models, like OpenAI O3 or Gemini 2.5, already use internal reasoning so this is a redundant step (I still use this with GPT‑4o, but not with O3 or GPT‑4 Turbo). But if you are using a smaller model that doesn’t have reasoning this can be helpful. Use explicit reasoning prompts like "Let's think step by step." Often built into newer models, but still helpful in edge cases.
When It Helps:
Non-reasoning models like Claude 3 or early GPT-4o
When consistency matters across large batches
When LLM skips intermediate reasoning steps
How To Use:
Prompt: “Solve the following problem. Let’s think step by step: [Insert task here]”
Reward/Threat Prompting
Phrasing like “This is critical to my job” or “You’ll get a tip” doesn’t improve model output. These are not part of the model’s training process and won’t influence its behavior.
Just wasted tokens. Strip them out.
Summary Table: What to Use, When

Bottom Line
Prompting is infrastructure. Not fluff.
If you're building LLM-powered tools, how you write prompts determines accuracy, safety, and utility. Treat prompts like product code—version them, test them, optimize them—especially in product environments where you don’t see every output.

I appreciate your support.

Your AI Sherpa,
Mark R. Hinkle
Publisher, The AIE Network
Connect with me on LinkedIn
Follow Me on Twitter