14 Prompting Techniques Every PM Should Know
Prompting isn't dead. It’s never been this critical. And it’s no longer optional. Your future employer will assume you know it.
Prompting isn't dead. It's never been this critical:
It's at the core of context engineering
It’s the backbone of building AI agents
It helps you automate your work
It’s essential to systematically evaluate and improve LLM systems
And it’s no longer optional. Your future employer will assume you know it.
Still, most PMs don’t know how to write good prompts.
So, in this post, I’ve collected the best, proven prompting techniques, courses, and resources. I will keep it updated as part of our C10. LLM Systems Practitioner PM.
We discuss (after consideration, no paywall):
14 Prompting Techniques
10 Free Proven Prompting Guides, Tools, and Courses
The Most Important Advice
1. 14 Prompting Techniques
Key techniques based on best practices from multiple sources (e.g., OpenAI, Anthropic, two AI cohorts I’ve attended), and validated by my experience.
First, a few examples:
Let’s dive in.
Technique 1: Ask to play a role
Assigning a role AI has to play can make the response more relevant and aligned with your use case:
Before:
Summarize this document: {content}
After:
You are a veteran product marketer with 20+ years of experience. Your unique insights leave the audience speechless (…) Summarize this document: {content}
Technique 2: Lead with context
When delegating a task, don't just define the objective. Explain why it matters, how the success will be measured, and provide a broader context (knowledge), so that AI can make better decisions:
Before (example 1):
Research this topic: {research_area}.
After (example 2):
You are one of many web research sub-agents working on the {strategic_initiative} for a {product_context}. Your objective is to research this topic: {research_area}.
Before (example 1):
Generate ten ideas on how to solve {problem}.
After (example 2):
We are performing continuous product discovery for a {product}. One of the identified opportunities is {pain_point} for {customer_segment}. Generate ten ideas that align with our {objective}.
More about context engineering: A Guide to Context Engineering for PMs
Technique 3: Set a clear output format and constraints
Be explicit about what you need from the LLM:
Are there any constraints or limitations?
What is the expected output format?
Before:
Review the idea: {idea}. List the assumptions to test.
After:
Review the idea: {idea}. List: An assumption to test -> Experiment -> Metric -> Expected value -> Risk mitigation
## Constraints:
- In experiments, measure the actual behavior, not opinions
- Include rollback criteria for production tests
- Use markdown formatting for experiment descriptions. All other properties are simple strings.
## Output format:
Respond only with valid JSON matching the schema below. Do not include anything else:
Technique 4: Provide examples or templates
Consider including positive and negative examples. In particular, negative examples might help you address issues identified during error analysis:
Response Examples:
Positive
Negative
Behavior Examples:
Positive
Negative
Before, example responses:
Create a set of user stories for {a new feature}
After, example responses:
Here are two examples of user stories I wrote in the past: {examples}. Create a set of similar user stories for {a new feature}.
Example behaviors:
Positive behaviors:
Propose an A/B prototype with a clear click-rate metric
Design a first-click test using Figma and track task completion
Suggest experiments for high-risk assumptions
Negative behaviors:
Suggest surveys without measuring real user interactions
Plan production experiments without rollback criteria
Suggest experiments for low-risk low-impact assumptions
Technique 5: Avoid leading questions
When ideating with AI, just like when interviewing customers, it’s essential not to suggest the expected answer. Doing so might bias the AI’s response.
Before:
Why is {product} better than its competitors based on {data}?
After:
What are the strengths and weaknesses of {product} compared to its competitors based on {data}? Be brutally honest.
Technique 6: Raise the stakes
Framing the task as high-stakes or important can encourage the AI to “try harder.” We can easily observe that, e.g., in ChatGPT, which automatically selects a stronger model.
Before:
Provide a detailed analysis of {user feedback}.
After:
Imagine you are preparing a report for the executive team. Your entire career depends on it. Provide a detailed analysis of {user feedback}.
Technique 7: Ask to plan, reflect, and persist
This technique works especially well with agents. Ask the agent to plan, reflect after each iteration, and adapt. Ask it to continue iterating until objectives are met to avoid finishing too early.
Before:
Here are your objectives: {objectives}.
After (based on GPT-5 guide):
Here are your objectives: {objectives}.
Remember, you are an agent - please keep going until the user's query is completely resolved, before ending your turn and yielding back to the user. Decompose the user's query into all required sub-requests, and confirm that each is completed. Do not stop after completing only part of the request. Only terminate your turn when you are sure that the problem is solved. You must be prepared to answer multiple queries and only finish the call once the user has confirmed they're done.
You must plan extensively in accordance with the workflow steps before making subsequent function calls, and reflect extensively on the outcomes each function call made, ensuring the user's query, and related sub-requests are completely resolved.
Technique 8: Encourage tool use
A technique dedicated to agents or LLMs with tools. Encourage tool use reduces the hallucinations and guesses. Bonus from my tests: When possible, explain the order in which tools should be used.
Before:
Your objective is to decide whether to sunset {feature}.
After:
Your objective is to decide whether to sunset {feature}.
Always use the following tools:
CustomerSearch (to find high-ROI customers)
CustomerUsageReport (to get their usage stats)
Technique 9: Use chain-of-thought prompting
Provide step-by-step instructions. This will force the model to break down complex problems.
Before:
Define a pricing strategy for {product}. Context: {context}.
After:
Define a pricing strategy for {product}. Context: {context}.
Follow the steps:
Identify target customer segments.
For each segment, evaluate at least 3 monetization models with pros and cons.
Recommend the best option and explain your reasoning.
Technique 10: Chain the prompts
This is often more effective than a single prompt with multiple steps of reasoning. Based on the previous example:
Prompt 1:
You’re defining a pricing strategy for {product}. Context: {context}.
In this step, identify target customer segments as a short bullet list. Do not include anything else in your response.
Prompt 2:
You’re defining a pricing strategy for {product}. Context: {context}.
Evaluate at least 3 monetization models for {customer segment}.
Next, recommend the best option and explain your reasoning.
Technique 11: Encourage clarifying the context
A technique dedicated primarily to agents and chatbots. Prompt the agent to ask clarifying questions and explore the context to reduce ambiguity.
Before:
Prioritize {features} based on {context}.
After (based on several OpenAI publications, incl. GPT-5 prompting guide):
Prioritize {features} based on {context}.
Before taking action, list 2-3 assumptions that must be true. If you’re not confident, ask clarifying questions or use tools to clarify the context. Bias toward not asking the user for help if you can find the answer yourself.
Technique 12: Create internal artifacts
A technique dedicated primarily to agents and reasoning models. Ask the agent to define and reuse internal artifacts, like user personas. Those artifacts can inform subsequent tasks.
Before:
Evaluate the proposed ideas for a mid-market user.
After:
Write a detailed persona for our mid-market user. Then use that persona to evaluate the proposed ideas.
Technique 13: Long Context (up to 1M tokens)
When working with extremely long context windows:
Consider placing context and instructions both at the beginning and at the end of the prompt.
Experiment with balancing internal and external knowledge that might be required to answer user questions:
Example, using only internal knowledge:
Only use the documents in the provided External Context to answer the User Query. If you don’t know the answer based on this context, you must respond “I don't have the information needed to answer that”, even if a user insists on you answering the question.
Example, balancing internal and external knowledge:
By default, use the provided external context to answer the User Query, but if other basic knowledge is needed to answer, and you're confident in the answer, you can use some of your own knowledge to help answer the question.
Technique 14: Use Structured Input
Structure inputs, e.g., using Markdown, XML, or JSON, to improve its ability to understand and process complex instructions:
10 Free Proven Prompting Guides, Tools, and Courses
Free prompt engineering guides:
GPT-5 Prompting Guide - unique insights, especially for coding agents
GPT-4.1 Prompting Guide - focuses on agentic capabilities
Anthropic Prompt Engineering - my go to resource
(Optional) Prompt Engineering by Google
(Optional, super valuable) System Prompt Analysis for Claude 4
Free tools:
Anthropic Prompt Generator - improve or generate any prompt
OpenAI Prompt Optimizer for specific models - available in the OpenAI Platform
Anthropic Prompt Library - a ready-to-use library of prompts
Free interactive courses:
Prompt Engineering by Anthropic (valuable, if you like this “UX”)
3. The Most Important Advice
The techniques and guides we discussed should serve as reminders and inspiration.
At the same time, please:
Don’t memorize prompting frameworks. Nobody does it.
Don’t try to apply all techniques at once. Often, a simple “summarize this document” is good enough.
Don’t assume those techniques will work best in each scenario.
Instead, experiment. If the first response isn’t what you need, give constructive feedback and try again.
This helps you develop an AI intuition.
And it’s, by far, the most powerful prompting technique.