Prompting Detail

 

Prompt Engineering: The Core Skill

Prompt Engineering is not about tricks — it’s about clear thinking.

You are shaping how the model interprets the world, what role it assumes, and how it delivers output. Prompting is design thinking applied to language. The rest of this breakdown branches out from that central principle.


🎯 Prompting Modes: Zero Shot, One Shot, Few Shot

Zero Shot

What it is: You give the model a task with no examples, just a clear instruction.
Best for: Simple requests, factual queries, summarization, transformation tasks (e.g., “Turn this into bullet points”).

πŸ”Ή Focus on:

  • Clear instruction ("Summarize this like I'm 12.")

  • Strong constraints ("Max 100 words.")

  • Defined audience or purpose ("Write as if you’re speaking to a board of directors.")

πŸ”Έ Common Pitfall: Too vague = too generic. Add detail.


One Shot

What it is: You show the model a single example of the task before asking it to do the same.
Best for: Style transfer, formatting, or logic demonstration.

πŸ”Ή Use when:

  • Task benefits from a “model” or sample format

  • You want the model to copy tone, voice, or structure

  • Clarity is more important than variation

πŸ”Έ Focus on:

  • Example + Instruction pairing: “Here’s one version. Now make another like it.”

  • Consistency and structure


Few Shot

What it is: You provide 2–5 examples of how you want the model to respond.
Best for: Complex reasoning, classification, structured output, creative variation.

πŸ”Ή Use when:

  • You want generalization based on patterns

  • The task is nuanced or has multiple correct outputs

πŸ”Έ Focus on:

  • Pattern clarity — make your few examples varied enough to expose the logic

  • Consistency — format examples identically

πŸ”Έ Watch for:

  • Hitting token limits

  • The model mimicking only the most recent example


πŸ”§ Prompt Construction Blocks (Middle Layer of Your Chart)

These blocks define the clarity and control of your prompt — use them to fine-tune performance across Zero/One/Few Shot strategies:

🟑 Context

Give the model the backstory or assumptions. It’s like giving it the first few pages of a script.

Example: “You’re an HR manager reviewing candidate resumes for a customer service role…”

🟑 Instruction

Clear, direct task description.

Bad: “Fix this.”
Good: “Rewrite the following paragraph for a 10th grade reading level, keeping it under 75 words.”

🟑 Specificity

Include detail about what matters — audience, tone, format, length, scope.

Use phrases like:

  • “Avoid industry jargon”

  • “Respond in exactly 3 bullet points”

  • “Use persuasive tone, but no hype”

🟑 Examples

Demonstrate what you want. Anchor the model with clear references.

Pair with:

  • “Follow the structure below.”

  • “Use a similar tone to this example.”

🟑 Restrictions

Set limits: “Do not include personal opinion.” “Avoid repeating the prompt.”
This prevents drift, especially in longer completions.

🟑 Questions

Models behave best when responding to precise, purposeful questions.

Structure matters:

  • Binary (“Should we proceed?”)

  • Choice (“Which of these is better, and why?”)

  • Exploratory (“What are the potential risks of this plan?”)


πŸ” Troubleshooting + T/S Error Path

Linked to your ERRORS → Troubleshooting → T/S Error path in the diagram, here’s how to spot, categorize, and fix bad outputs:

A. Types of Errors

  • Hallucination → Add constraints or evidence prompts (“Only cite what’s in the text”)

  • Format Fail → Respecify with layout prompts or use structured examples

  • Incoherence → Try Few-Shot to give clearer expectations

  • Underperformance → Reframe question; possibly switch to Chain-of-Thought prompting

B. Debug Flow

  1. Start simple. Strip the task back to its basics.

  2. Isolate failure. Try changing just one thing.

  3. Use critique prompts. “What might be wrong with your answer?”

  4. Restart thread if memory confusion.


🧠 Mindset Anchors: Think Like a Designer, Not Just a User

Models don’t “think.” They extrapolate from patterns. They don’t know truth — they know text. So:

  • Prompt like you’re building scaffolding

  • Iterate like you’re tuning a machine

  • Validate like you’re editing a draft


4. Chain-of-Thought (CoT) Prompting

Definition: Encouraging the model to generate intermediate reasoning steps before arriving at a final answer.

Use Cases:

  • Mathematical problem-solving.

  • Logical reasoning tasks.

Example:

"Question: If you have 3 apples and buy 2 more, how many apples do you have?
Let's think step by step."en.wikipedia.orglinkedin.com+4learnprompting.org+4arxiv.org+4


5. Self-Consistency

Definition: Generating multiple reasoning paths and selecting the most consistent answer among them.

Use Cases:

Example:

"Ask the model the same question multiple times and choose the most frequent answer."


6. Tree of Thoughts (ToT)

Definition: Structuring the model's reasoning as a tree, exploring multiple potential paths before selecting the best one.

Use Cases:

  • Strategic decision-making.

  • Complex problem-solving.arxiv.org

Example:

"For a given problem, generate multiple solution paths, evaluate each, and choose the most effective one."


7. Meta Prompting

Definition: Focusing on the structure and syntax of tasks rather than specific content details.

Use Cases:

  • Abstracting tasks to their structural components.

  • Improving generalization across tasks.

Example:

"Given a task description, generate a prompt that would effectively instruct a model to perform it."en.wikipedia.org


8. ReAct (Reasoning and Acting)

Definition: Combining reasoning steps with actions, allowing the model to interact with tools or environments during the reasoning process.

Use Cases:

  • Tasks requiring both thought and interaction, such as using calculators or databases.

Example:

"Question: What is the capital of France?
Thought: I need to look up the capital of France.
Action: Search for 'Capital of France'
Observation: The capital of France is Paris.
Answer: Paris."en.wikipedia.org


9. Automatic Reasoning and Tool-use (ART)

Definition: Enabling the model to autonomously decide when and how to use external tools to aid in reasoning.

Use Cases:

  • Complex tasks that benefit from external computations or data retrieval.

Example:

"For a math problem, the model decides to use a calculator API to compute the result."


10. Least-to-Most Prompting

Definition: Breaking down complex problems into simpler subproblems and solving them sequentially.

Use Cases:

  • Tasks that can be decomposed into smaller, manageable parts.

Example:

"To solve a complex equation, first solve for one variable, then substitute and solve for the next."


11. Progressive-Hint Prompting (PHP)

Definition: Providing the model with incremental hints based on its previous responses to guide it toward the correct answer.

Use Cases:

  • Improving accuracy in reasoning tasks through iterative guidance.

Example:

"If the model's initial answer is incorrect, provide a hint and ask it to try again."en.wikipedia.org


12. Automatic Prompt Engineer (APE)

Definition: Using the model to generate and refine its own prompts for improved performance.

Use Cases:

  • Automating prompt creation for various tasks.

Example:

"Given a task description, the model generates a suitable prompt to perform the task effectively."linkedin.com+16en.wikipedia.org+16promptingguide.ai+16


13. Active-Prompt

Definition: Using feedback or interactivity to adapt and evolve prompts during execution.

Use Cases:

  • Iterative tasks where the prompt needs to adjust based on previous outputs.

Example:

"After each model response, evaluate its accuracy and adjust the next prompt accordingly."


14. Directional Stimulus Prompting

Definition: Using specific cues to guide model behavior, such as tone or bias direction.

Use Cases:

  • Creative writing.

  • Sentiment control.

Example:

"Write a story in a humorous tone about a day at the beach."


15. Program-Aided Language Models (PAL)

Definition: Integrating code execution with language models to enhance reasoning capabilities.

Use Cases:

  • Tasks requiring precise calculations or data manipulation.

Example:

"For a data analysis task, the model writes and executes a Python script to process the data."


16. Reflexion

Definition: Allowing the model to evaluate and reflect on its own outputs, leading to improved responses.

Use Cases:

  • Error correction.

  • Iterative improvement.

Example:

"After generating an answer, the model reviews it for potential errors and revises if necessary."


17. Multimodal Chain-of-Thought (Multimodal CoT)

Definition: Performing step-by-step reasoning across multiple modalities, such as text and images.

Use Cases:

  • Visual question answering.

  • Diagram interpretation.

Example:

"Given an image of a graph and a related question, the model analyzes the image and provides a detailed answer."


18. Graph Prompting

Definition: Utilizing graph-based representations in prompts to capture relationships and structures.

Use Cases:

  • Knowledge graph construction.

  • Social network analysis.

Example:

"Represent the relationships between characters in a novel as a graph and analyze their interactions.


🧰 Summary Table: When to Use What

Prompt TypeBest ForKey Add-OnsExample Use Case
Zero ShotStraightforward tasksInstruction + Constraints“Write a job description for an electrician.”
One ShotFormat mimicry, structure replicationOne example + task“Here’s one quote email. Make another.”
Few ShotPatterned reasoning, creative synthesis2–5 examples + format hints“Classify these customer reviews as Positive/Neutral/Negative.”
Contextual PromptingComplex tasks needing setupRole + scenario“You’re a compliance officer reviewing a contract…”
Chain-of-ThoughtReasoning-heavy questionsStep-by-step prompts“Let’s solve this one piece at a time…”
Critique/IterativeQuality controlSelf-review prompts“What’s weak about your last answer?”


Comments

Popular posts from this blog

Prompting - Unit 9: Automatic Prompt Engineer (APE)

Intro to Prompting