·

Hands On Context Toolkit

Hands On Context Toolkit

Overview

This topic is where everything from Module 2 converges into a single, tangible output: your personal Context Toolkit. The preceding five topics have built your conceptual and technical foundation — you understand how LLMs process context, how to architect prompts, what context to include and at what layer, and how to iterate on outputs. Now it is time to apply all of that in a structured, hands-on exercise that produces a set of reusable artifacts you will use in your actual product work starting today.

A Context Toolkit is not a document library or a knowledge base. It is a collection of purpose-built, AI-ready context artifacts: context briefs, prompt templates, structured context blocks, and an organizational system that makes the right artifact immediately accessible when you need it. The difference between a toolkit and a collection of documents is that a toolkit is designed to be used — everything in it has a clear trigger (when to use it), a clear format (how to structure the context), and a clear ownership model (who keeps it current).

Building this toolkit during the course, with real examples from your actual product work, is significantly more valuable than building it later in a hypothetical exercise. The investment of 4–6 hours across this module's exercises yields compound returns for every AI-assisted task you run going forward. Teams that build and share a common toolkit see the most dramatic productivity and quality improvements from AI adoption — because consistency and shared context are the multipliers that individual prompt skill alone cannot provide.

This topic is structured as four sequential exercises: auditing your tasks and their context needs, building templates for your core task types, testing those templates against real scenarios, and organizing and sharing the final toolkit with your team. Each exercise is detailed and hands-on — you are expected to produce real artifacts by the end of each section, not just complete conceptual activities.


Audit Your Daily PM Tasks and Identify Their Context Needs

The foundation of a useful Context Toolkit is an accurate inventory of the tasks you actually run with AI assistance — not the tasks you think you should be running, or the tasks that sound impressive in a course, but the real, recurring things you do every day or every week that could be improved with AI support. Before you can build templates, you need to know what you are templating for.

A task audit also surfaces something important: the gap between the context you currently provide and the context those tasks actually require for high-quality AI output. Most product managers, when they audit honestly, find that their AI sessions are systematically under-contextualized — they are running complex tasks with minimal context and then wondering why the outputs are generic. The audit makes this visible and gives you a specific action list rather than a vague intention to "do better with context."

How to run the task audit:

The audit has three phases. Phase 1 is inventory: list every recurring PM task you have performed in the last 30 days where you used (or could have used) AI assistance. Be comprehensive — include the small tasks, not just the big analysis work. Phase 2 is classification: for each task, classify it by task type (generation, analysis, communication, planning, review) and by frequency (daily, weekly, monthly, ad-hoc). Phase 3 is context analysis: for each task, identify what context the AI would need to produce a high-quality output for your specific product situation — and compare that to what you have been providing.

The context gap analysis: For each task, rate the current context you provide on a scale of 1–3:
- 1 — Minimal context: You describe the task but provide little or no product-specific context
- 2 — Partial context: You provide some context (maybe the product area or user segment) but miss key elements (OKRs, constraints, current state)
- 3 — Full context: You consistently provide strategic, tactical, and operational context appropriate to the task

Most product managers will find that 60–70% of their recurring tasks are rated 1 or 2. These are your highest-leverage toolkit building targets — the tasks where investing in a reusable context template will produce the most dramatic improvement in output quality.

Hands-On Steps

  1. Open a spreadsheet or Notion table with the following columns: Task name | Task type | Frequency | Context currently provided (1–3) | Context needed (list the required fields) | Toolkit priority (High/Med/Low).

  2. Set a 20-minute timer. In that time, list every AI-assisted PM task you have run in the last 30 days. Aim for at least 15 items. If you have not been using AI much, list every task that you could use AI for — the audit applies equally to potential future tasks.

  3. For each task, fill in the Task type (choose from: User story writing / Acceptance criteria / Prioritization / Discovery synthesis / Stakeholder communication / Sprint planning / Retrospective analysis / Requirements review / Risk assessment / Research analysis / Meeting preparation / Other).

  4. Rate the context you currently provide for each task on the 1–3 scale. Be honest. If you typically paste a vague description and ask for help, that is a 1.

  5. For tasks rated 1 or 2, write the context fields that would make the output significantly better. Use the context stack framework from Topic 4: which strategic, tactical, and operational fields does this task need?

  6. Assign toolkit priority: High = tasks that are frequent AND currently under-contextualized (rated 1 or 2). These are your first-build targets. Medium = tasks that are frequent but well-contextualized, or infrequent but under-contextualized. Low = tasks that are infrequent and well-contextualized already.

  7. Sort by toolkit priority (High first). Circle the top 4–6 High-priority tasks. These are the tasks you will build toolkit templates for in the next exercise.

Prompt Examples

Prompt (using AI to assist the audit itself):

You are a senior product manager helping me audit my AI-assisted workflow.

I have listed the following recurring tasks from the past month. For each task, identify:
1. The task type (generation / analysis / communication / planning / review)
2. The minimum context fields required for a high-quality AI output on this task (use the categories: strategic context, user segment, current state, constraints, metrics/data, stakeholder context)
3. Whether this task is better served by a single-shot prompt or conversational iteration

Tasks:
1. Writing user stories for new features
2. Preparing weekly stakeholder status updates
3. Scoring backlog items using RICE
4. Synthesizing themes from user interview notes
5. Evaluating sprint scope at sprint planning
6. Drafting acceptance criteria for stories already written by the team
7. Preparing roadmap review presentations for exec team
8. Writing NPS survey analysis reports

Output format: Table with columns: Task | Type | Required Context Fields | Single-shot or Iterative

Expected output: A filled-in audit table identifying the context requirements for each task type, helping you see precisely what context fields are missing from your current practice — the foundation for building targeted templates.

Learning Tip: The most valuable discovery from the task audit is usually not what you expected to find. Look especially for high-frequency tasks you have been running with minimal context (rated 1) — those are your hidden quick wins. A single good template for a daily task you run with minimal context today will save you hours per month and dramatically improve the quality of outputs that drive real product decisions.


Build Context Templates for Discovery, User Stories, Stakeholder Updates, and Sprint Planning

With your audit complete and your high-priority tasks identified, you now build the actual toolkit templates. This section provides step-by-step guidance for constructing each of the four most common PM context template types: discovery, user stories, stakeholder updates, and sprint planning. Each template type has a specific structure optimized for its task requirements.

Template structure: Every toolkit template consists of four components:
1. Role frame: The specific role framing that works best for this task type
2. Context block: A fill-in-the-blank context template with [PLACEHOLDERS] for all variable fields
3. Task instruction: The specific instruction for this task type, with output format specification
4. Refinement sequence (if applicable): 1–2 follow-up prompts for tasks that benefit from iteration

Template Type 1: Discovery Synthesis Template

Discovery templates are used when synthesizing user research — interviews, survey data, usability test notes, support ticket themes — into product insights and opportunity statements. The context requirements are: strategic context (OKRs, product vision, current roadmap themes) + research data (pre-processed transcripts, verbatims, or theme summaries) + research question or focus area.

ROLE FRAME:
You are a senior product manager specializing in user research synthesis. You have deep expertise in identifying unmet user needs and ranking them by business impact.

CONTEXT BLOCK:
Product: [FILL IN: name + one-sentence description]
Target user segment for this research: [FILL IN: role + company type + size]
Current OKR this discovery is in service of: [FILL IN: objective + key result]
Research question: [FILL IN: the specific question this research was designed to answer]
Research data: [PASTE: pre-processed interview excerpts / verbatim summaries / survey themes]

TASK INSTRUCTION:
Synthesize the research data above into product insights. Provide:
1. The top [3–5] unmet needs identified, ranked by frequency and severity
2. For each unmet need: a one-sentence need statement, the evidence base (which participants mentioned it), and the estimated business impact if addressed
3. One "How Might We" problem statement for the highest-ranked unmet need
4. One user story hypothesis for the highest-ranked unmet need: "We believe [feature] will [outcome] for [user] because [reasoning]"

Output format: Numbered list for unmet needs, then HMW statement, then hypothesis. No bullet points — use numbered lists for all ranked items.

REFINEMENT SEQUENCE:
Turn 2: "Challenge the top-ranked unmet need. What are the two strongest counter-arguments that this is not worth solving? What evidence would change your ranking?"
Turn 3: "For the user story hypothesis, add a one-paragraph validation plan: What is the minimum viable test to confirm or refute this hypothesis in 2 weeks or less?"

Template Type 2: User Story Generation Template

User story templates are used to generate well-formed stories with acceptance criteria from feature briefs, discovery outputs, or stakeholder requests. The context requirements are: user segment (with behavioral characteristics) + current state + capability needed + constraints + success definition.

ROLE FRAME:
You are a senior business analyst with 8 years of experience writing requirements for enterprise SaaS products. Your stories are known for testability, engineering clarity, and complete edge case coverage.

CONTEXT BLOCK:
Product area: [FILL IN: module or feature area]
User: [FILL IN: role + 2 behavioral characteristics relevant to this story]
User story need: [FILL IN: what capability the user needs and why]
Current state: [FILL IN: what the user does today without this capability]
Solution constraints: [FILL IN: technical, UX, or business constraints the solution must respect]
Success definition: [FILL IN: what "done well" looks like — measurable or observable]
Explicitly out of scope: [FILL IN: what this story does NOT cover]

TASK INSTRUCTION:
Generate one production-ready user story. Output:
- User story: As a [role], I want [capability] so that [outcome]
- Rationale: 2 sentences on why this matters specifically for the described user
- Acceptance criteria: 5 Given/When/Then criteria
- Edge cases: 2 edge cases engineering should handle that are NOT covered by the above criteria
- Definition of Done items: 3 items beyond "acceptance criteria pass" (e.g., performance benchmark, accessibility requirement, analytics event)

Output format: Use the headers above as section labels.

REFINEMENT SEQUENCE:
Turn 2: "Review the 5 acceptance criteria against the INVEST criteria (Independent, Negotiable, Valuable, Estimable, Small, Testable). Flag any criterion that fails any INVEST check and suggest a rewrite."

Template Type 3: Stakeholder Update Template

Stakeholder update templates are used for producing executive summaries, status emails, product update Slack messages, and briefing documents for different audience types. The context requirements are: audience role + their primary concerns + key facts to communicate + desired action or response.

ROLE FRAME:
You are a product manager known for clear, audience-calibrated communication. You write updates that are always read to completion by busy executives.

CONTEXT BLOCK:
Audience: [FILL IN: role + their primary business concerns (2–3 bullet points)]
Communication vehicle: [FILL IN: email / Slack / slide / verbal]
Subject of update: [FILL IN: what this update is about]
Key facts:
  - Situation: [FILL IN: current state in one sentence]
  - Change/Decision: [FILL IN: what changed or was decided]
  - Business impact: [FILL IN: so what — why does this matter to the audience's concerns]
  - Timeline: [FILL IN: relevant dates]
  - Risks (if any): [FILL IN: what could go wrong and what is being done about it]
Action required from audience: [FILL IN: what you need from them, or "None — FYI only"]

TASK INSTRUCTION:
Write a [FILL IN: format] update for the stated audience. Requirements:
- Maximum [FILL IN: word count]
- Lead with business impact, not process or technical detail
- Use the language and concerns of the stated audience — not PM jargon
- If there are risks, do not bury them — state them clearly with mitigation
- End with a clear action item or "No action required from you"
- No passive voice

Output format: [FILL IN: format] only. No meta-commentary.

REFINEMENT SEQUENCE:
Turn 2: "Identify any jargon in the above update that the stated audience would not understand. Rewrite those phrases in plain language."
Turn 3: "List the top 2 questions the audience is most likely to ask after receiving this update. Write one-sentence answers to each."

Template Type 4: Sprint Planning Template

Sprint planning templates are used to evaluate sprint scope, generate sprint goals, and assess risks before committing to a sprint. The context requirements are: current OKR + tactical context (top backlog items + sizes + rationale) + operational context (capacity + dependencies + constraints).

ROLE FRAME:
You are a senior product manager facilitating a sprint planning session. You optimize sprint scope for OKR impact within hard capacity constraints.

CONTEXT BLOCK:
OKR this sprint serves: [FILL IN: objective + key result with current and target metrics]
Sprint capacity: [FILL IN: team size + availability % + story points available]
Stakeholder commitments: [FILL IN: any committed items that must be in scope regardless of scoring]
Dependencies: [FILL IN: external team dependencies that could affect delivery]
Tech debt risks: [FILL IN: known technical risks that affect specific backlog items]

Candidate stories for this sprint:
[FILL IN: numbered list — Story name | Size in points | OKR contribution (High/Med/Low) | Any dependency or risk flag]

TASK INSTRUCTION:
Recommend a sprint scope. For each candidate story:
- Include or Defer, with one-sentence rationale
- Flag any story that touches a stated dependency or tech debt risk
- Confirm the total included story points fit within stated capacity
- Generate a one-sentence sprint goal for the recommended scope

Output format: Recommendation table (Story | Decision | Rationale | Risk flags) followed by Sprint goal statement.

REFINEMENT SEQUENCE:
Turn 2: "For each 'Include' decision, identify the single most likely reason this story will not be completed as planned. What is the contingency?"
Turn 3: "Draft a 3-sentence sprint review narrative for this scope — assuming everything is delivered as planned — to be read to stakeholders at the end of the sprint."

Hands-On Steps

  1. For each of the four template types above, fill in at least one complete template with real context from your current product work. Use your actual product, actual OKRs, actual team capacity.

  2. For each completed template, run the primary task instruction with a real AI tool (Claude, ChatGPT, or Gemini). Evaluate the output quality on: specificity (references your context), accuracy (respects constraints), and usability (can you use it without editing?).

  3. Where the output falls short, identify the missing or inadequate context field and update the template's context block. Run again and compare.

  4. For each template, run at least one turn of the refinement sequence. Note whether the refinement turn produces meaningfully different output or whether the first-pass quality was sufficient.

  5. Customize each template for your specific product context: add product-specific context fields that are always relevant for your work but not covered in the generic template above (e.g., "Regulatory constraint: [your specific constraint]" or "Integration requirement: [your specific integration]").

  6. Time yourself using each template from a cold start (finding the template, filling in the [FILL IN] fields, running it, and evaluating the output). Note the time. Your goal is under 5 minutes from template open to usable output for straightforward tasks.

Prompt Examples

Prompt (discovery template filled in — ready to run):

You are a senior product manager specializing in user research synthesis. You have deep expertise in identifying unmet user needs and ranking them by business impact.

Product: ProjectFlow — B2B project management SaaS for SMB architecture and engineering firms (5–50 employees)
Target user segment: Principal architects and project managers at general contractor firms
Current OKR this discovery is in service of: Objective: Reduce time-to-first-value for new accounts. KR: Reduce average onboarding completion to 5 days (currently 11 days).
Research question: Why do new accounts take so long to complete onboarding, and what would make them complete it faster?

Research data (pre-processed — 5 interviews, signal moments only):

Participant 1 (Principal architect, 8-person firm, 4 days to complete onboarding):
[Pain] "I kept getting distracted — I'd start setting up the project and then get a call and lose my progress."
[Workaround] "I had to start the onboarding 3 separate times before I finished it."
[Unmet need] "I just needed it to save where I was and let me come back."

Participant 2 (Project manager, 22-person firm, did not complete onboarding):
[Pain] "Step 4 asked me to invite my whole team. I don't make that decision alone — I had to check with my boss."
[Workaround] "I skipped it and tried to come back but didn't know how to resume."
[Unmet need] "Let me skip steps I can't complete right now and finish them later."

Participant 3 (Office manager, 12-person firm, 7 days to complete):
[Pain] "The template options didn't match our type of work. We do renovation and restoration, not new construction."
[Workaround] "I picked the closest template and manually edited everything."
[Unmet need] "Templates for our specific work type."

Participant 4 (Principal, 6-person firm, 2 days to complete):
[Positive] "The project creation flow was fast. Once I had a real project set up it clicked."
[Unmet need] "I wish I had created a real project on day 1 instead of a fake demo one."

Participant 5 (Project coordinator, 18-person firm, 9 days to complete):
[Pain] "I'm not technical. Some of the settings screens looked complicated and I wasn't sure what they were for."
[Workaround] "I just left all the settings as default and hoped they were fine."
[Unmet need] "Plain language descriptions of what each setting does and which ones actually matter."

Synthesize the research data above into product insights. Provide:
1. The top 4 unmet needs, ranked by frequency and severity
2. For each unmet need: a one-sentence need statement, the evidence base, and the estimated business impact
3. One "How Might We" problem statement for the highest-ranked unmet need
4. One user story hypothesis for the highest-ranked unmet need

Expected output: Four ranked unmet needs drawn from the five specific interview excerpts (resume/save state, skippable steps, specific templates, plain-language settings), each with an evidence base that references the specific participants — not generic "improve onboarding" output.

Learning Tip: When filling in your template context blocks, write as if you are briefing a smart but completely uninformed colleague on your product for the first time. Every detail that you assume "goes without saying" should be written down. The templates that produce the best outputs are the ones that make the implicit explicit — the product domain knowledge, the team constraints, the political realities, the user behaviors that are obvious to you but invisible to the model.


Test Your Templates Against Real Product Scenarios

Templates are only useful if they actually work. The testing phase is where you validate each template against real scenarios from your product work — not hypothetical examples, but actual recent tasks where you can compare the AI-assisted output against what you or your team produced manually. This comparison is the most honest measure of template quality, and the gaps between AI output and manual output tell you exactly what context or instruction adjustments to make.

How to run a template test:

A template test has four steps:
1. Identify a real past task — a user story you wrote last month, a stakeholder update you drafted last week, a sprint scope decision you made last sprint. You have the real output to compare against.
2. Run the template with the context from that past task — reconstruct the context as it existed at the time (not as it exists now). Fill in all [FILL IN] placeholders with the real data from that past situation.
3. Compare AI output to your manual output — do not evaluate the AI output in isolation. Evaluate it specifically against what you actually produced. Be honest about both directions: where is the AI output better? Where is it worse? Where is it missing elements your manual work included?
4. Identify the gaps and update the template — every gap between AI output and manual output points to a missing context field, an inadequate output format instruction, or a refinement sequence that needs to be added.

What good looks like in a template test: A well-performing template should produce output that matches or exceeds your manual work quality in 70% or more of dimensions evaluated. It should do so in a fraction of the time. If the AI output is 70% as good as your manual work and takes 20% of the time, the template has strong productivity value even without further refinement. If it is below 50% as good, the template has significant context gaps that need addressing before it saves you time reliably.

Common gaps revealed by template testing:

  • Missing "current state" field: AI generates ideal-state solutions without knowing what the user does today. Fix: add explicit "current state" to the context block.
  • Missing stakeholder constraints: AI recommends things your stakeholders would never approve. Fix: add a "stakeholder constraints" section listing any known political or organizational constraints.
  • Output format doesn't match your deliverable requirements: AI produces a different format than what your team uses for this artifact. Fix: update the output format specification to match your team's actual format.
  • Too generic despite context: Some templates work for 80% of cases but fail for edge cases in your product domain. Fix: add domain-specific context fields and a "specifically for our product" caveat that highlights domain-specific conventions.

Hands-On Steps

  1. Select one past task from each of your four template types (discovery synthesis, user story, stakeholder update, sprint planning). Choose tasks from the last 2–4 weeks so the context is fresh.

  2. For each past task, reconstruct the context as it existed at the time: current OKRs, team capacity, stakeholder situation, user research data. Write it in the template [FILL IN] fields.

  3. Run each template with the reconstructed context. Do not look at your original manual output until you have the AI output in hand.

  4. Compare side by side. For each template, rate the AI output on: (a) Completeness — did it cover everything your manual output covered? (b) Specificity — was it as specific as your manual output? (c) Accuracy — were the constraints and context respected? (d) Usability — how much editing would you need before using it?

  5. Identify the top 2 gaps for each template. For each gap, write the template update needed: either a new context field to add, a constraint to make more explicit, or an output format clarification.

  6. Update all four templates with the identified improvements. Re-run each template with the same past context. Evaluate whether the gaps are resolved.

  7. Run each updated template one more time with a different (more recent) real scenario. This confirms the improvements generalize, not just that they fixed the specific case they were calibrated on.

Prompt Examples

Prompt (sprint planning template — test scenario):

[Template filled in with real past sprint context]

You are a senior product manager facilitating sprint planning.

OKR this sprint serves:
Objective: Improve 30-day retention from 44% to 58%
KR: Onboarding completion rate to 60% (currently 34%)
Current sprint: Sprint 14 (July 7–20)

Sprint capacity: 6 engineers at 75% availability. Velocity baseline: 42 points. Available: ~32 points.
Stakeholder commitments: Drawing upload must be in Sprint 15–16 (committed to Hartwell Construction)
Dependencies: Drawing upload requires Platform file storage API confirmed for Sprint 15 start.
Tech debt risks: Notification module memory leak — avoid notification-related work until Sprint 16.

Candidate stories:
1. Onboarding progress save/resume | 8 pts | High OKR contribution | No dependencies
2. Skip and return to onboarding steps | 6 pts | High OKR contribution | No dependencies
3. Industry-specific project templates (3 templates) | 13 pts | High OKR contribution | No dependencies
4. Drawing upload v1 (basic PDF only) | 7 pts | Low OKR contribution | Platform dependency, ready Sprint 15 start
5. In-app notification center | 9 pts | Medium OKR contribution | Tech debt risk (notification module)
6. Team bulk invite | 5 pts | Medium OKR contribution | No dependencies

Recommend sprint scope. For each candidate: Include or Defer with one-sentence rationale. Confirm total points fit 32-point capacity. Flag dependency/tech debt risks. Generate sprint goal.

Expected output: A sprint scope recommendation that includes the highest OKR-contributing items within 32 points (likely stories 1, 2, and 4 based on the committed drawing upload), defers the notification center due to tech debt risk, and generates a sprint goal tied to the onboarding completion OKR — matching what an experienced PM would recommend in this situation.

Learning Tip: The most valuable test is the one where the AI output diverges from what you manually produced — especially where the AI output is actually better. When the AI identifies something you missed, or structures an output more clearly than you did, study why. What context did you give the AI that surfaced that insight? Did the output format instruction force a structure that revealed a gap in your manual work? These divergences are your most valuable template learning moments.


Organize and Share Your Context Toolkit with Your Product Team

A context toolkit that lives only in your personal notes has limited leverage. The multiplier effect comes from sharing it with your product team — BAs, PMs, POs, and product designers — so that every team member starts AI sessions from a common, high-quality baseline. A shared toolkit reduces the variance in AI output quality across team members from significant to minimal, enables team members to build on each other's AI work, and creates a forcing function for the ongoing maintenance that keeps the toolkit relevant.

Recommended organization structure:

The ideal toolkit organization system is a Notion database (or equivalent in your team's tool: Confluence, Google Sheets, or a shared Google Doc). The database should have the following structure:

TEMPLATE LIBRARY DATABASE
Properties:
- Template name
- Task category (Discovery / Requirements / Communication / Planning / Review / Analysis)
- Task type (Generation / Analysis / Transformation / Planning)
- Context layers required (Strategic / Tactical / Operational — checkboxes)
- Prompt approach (Single-shot / Iterative)
- Last tested date
- Last updated date
- Owner (who is responsible for keeping this template current)
- Quality rating (1–5 stars, based on team usage feedback)
- Notes (any known limitations or specific usage guidance)

Pages linked to each entry:
- Full template text (with [FILL IN] placeholders)
- Sample output (from a real or illustrative use case)
- Refinement sequence prompts

Context document storage:

Alongside the template library, maintain a "living context documents" section:
- Strategic context card (updated quarterly by PM lead): Current product vision, OKRs, roadmap themes, market positioning
- Tactical context snapshot (updated weekly by product owner or PM): Current sprint goal, top backlog items, active stakeholder commitments
- Operational context brief (updated weekly by PM or tech lead): Team capacity, key dependencies, tech debt flags

Link these documents to the relevant templates so team members can paste the current context block without having to look up or reconstruct it.

Team onboarding for the toolkit:

A shared toolkit is only valuable if team members use it. Run a 45-minute onboarding session when you first share the toolkit:
1. Walk through the organizational structure (15 min)
2. Demonstrate a live template use with a real current task (10 min)
3. Have each team member run one template on a real task of their own during the session (15 min)
4. Address questions and gather initial feedback on missing templates (5 min)

After the session, assign a "toolkit buddy" role for the first month: two team members who check in on toolkit usage in weekly syncs and help each other troubleshoot templates that aren't working well.

Maintenance model:

A shared toolkit requires an explicit maintenance model or it will become stale and stop being used. Three maintenance practices:

  1. Quarterly review (60 min): Update the strategic context card with new OKRs, retire templates that are no longer relevant, add templates for new recurring task types that have emerged. Rotate the "toolkit owner" role each quarter.

  2. Sprint-boundary updates (15 min, weekly): Update the tactical context snapshot. Flag any templates that need adjustment based on sprint context changes.

  3. Continuous improvement ("fix before you close"): When a team member uses a template and finds a gap, they update the template before closing the session. This distributed maintenance model keeps the library high-quality without requiring a separate maintenance effort.

Hands-On Steps

  1. Set up your toolkit storage. Create a Notion database, Confluence page, or Google Sheet using the structure described above. Set up the properties and linked document sections. Even if you are starting with just your 4 templates from this module, the structure should support growth to 30+ templates.

  2. Enter your four templates into the database: discovery, user story, stakeholder update, and sprint planning. For each, fill in all properties, attach the template text with [FILL IN] placeholders, and add a sample output from your testing in the previous section.

  3. Create the three living context documents: strategic context card, tactical context snapshot, and operational context brief. Fill each one with current content. Link each to the relevant templates.

  4. Identify who on your team should have access and who should be a contributor (able to add and edit templates) vs. a viewer (can use templates but not edit). Set permissions accordingly.

  5. Schedule the toolkit onboarding session with your team for the next 1–2 weeks. Prepare a 10-minute live demonstration using a real current task.

  6. Designate the first toolkit owner and document the quarterly review cadence (first review date on the calendar now). Establish the "fix before you close" norm in the team — communicate it at the onboarding session.

  7. After 4 weeks of team usage, run a brief survey (5 questions, async): Which templates have you used? Which produced the best outputs? Which produced disappointing outputs? What templates are missing? Rate the overall toolkit value 1–5. Use the feedback to prioritize your Q2 toolkit improvements.

Prompt Examples

Prompt (team kickoff — toolkit demonstration):

[Live demonstration prompt — use this during your team onboarding session]

You are a senior product manager on a B2B SaaS team.

[Paste your team's current strategic context card here]

[Paste your team's current tactical context snapshot here]

Task: Using only the context above, generate a sprint goal statement for the next sprint and identify the top 2 risks to achieving it. For each risk: state the risk, rate it High/Med/Low, and suggest a mitigation or contingency.

Format: Sprint goal (one sentence) → Risk table (Risk | Likelihood | Impact | Mitigation)

Expected output: A contextually grounded sprint goal and risk table — run live in front of the team to demonstrate how the shared context card produces immediately useful output. The demonstration is most powerful when you show it with no additional explanation — the model's output speaks for itself.


Prompt (quarterly toolkit review — identifying gaps):

You are a senior product manager reviewing your team's AI toolkit after one quarter of use.

Team: 3 product managers, 2 business analysts, 1 product owner. B2B SaaS, construction vertical. 2-week sprints.

Current toolkit templates:
1. Discovery synthesis (used 8 times — avg quality rating: 4.2/5)
2. User story generation (used 31 times — avg quality rating: 3.8/5)
3. Stakeholder update (used 12 times — avg quality rating: 4.5/5)
4. Sprint planning scope evaluation (used 6 times — avg quality rating: 3.5/5)

Team feedback themes:
- "We need something for retrospective action items — we always end up doing that ad hoc"
- "The user story template is great but we need something specifically for epic-level story mapping"
- "We're spending a lot of time on requirements gap analysis before sprints — would love a template"
- "BA team uses the acceptance criteria review a lot but it's not in the toolkit yet"

Based on the usage data, quality ratings, and feedback themes above:
1. Identify the top 3 new templates to build for Q2 based on value opportunity
2. Identify 2 improvements to the existing templates based on the quality ratings
3. Recommend a prioritization for the Q2 toolkit development roadmap

Output format: Prioritized list for new templates, then improvement notes for existing templates, then a quarterly roadmap recommendation.

Expected output: A quarterly toolkit development roadmap with specific template priorities, improvements, and a sequencing rationale — treating the toolkit itself as a product and using AI to help manage its roadmap.

Learning Tip: Treat your Context Toolkit like a product, not a document. It has users (your team), a backlog (the template gaps and improvements you have identified), a quality metric (average output quality ratings from users), and a release cadence (quarterly reviews). When you apply product thinking to your own toolkit — with an owner, a roadmap, a feedback loop, and a maintenance model — it becomes a living, improving asset rather than a document that goes stale. This product mindset about your own workflows is one of the defining characteristics of a truly AI-fluent product manager.


Key Takeaways

  • A Context Toolkit is a collection of purpose-built, AI-ready context artifacts: context briefs, prompt templates, structured context blocks, and an organizational system — designed to be used, not just stored.
  • The task audit is the foundation of the toolkit: inventory your 10–15 most frequent AI-assisted tasks, classify them by type, rate your current context quality (1–3), and identify the highest-priority gaps. High-frequency tasks with low context quality (rated 1 or 2) are your first-build targets.
  • The four most universally valuable toolkit template types for product managers are: discovery synthesis, user story generation, stakeholder communication, and sprint planning. Each has a specific context structure optimized for its task requirements.
  • Template testing against real past scenarios is the quality gate before sharing. Compare AI output to your manual output; identify gaps; update templates to close them. A well-performing template should match or exceed manual quality in 70%+ of dimensions.
  • Shared team toolkits multiply the value of individual templates. A Notion database with structured properties, living context documents (strategic card + tactical snapshot + operational brief), and a clear maintenance model is the recommended organizational approach.
  • Maintain the toolkit like a product: quarterly reviews for OKR updates and template retirement, weekly updates for tactical context, and the "fix before you close" norm for continuous distributed improvement.
  • After one quarter of active toolkit use by a team of 3–5 PMs and BAs, expect 30–60 minutes of weekly time savings per person and a measurable improvement in AI output quality consistency across team members.
  • The context toolkit is the compound interest of your AI practice. The investment in building it pays returns for every subsequent session, for every team member who uses it, for as long as the toolkit is maintained.