We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track
PM Daily Digest
by avergin 100 sources
Curates essential product management insights including frameworks, best practices, case studies, and career advice from leading PM voices and publications
Sachin Rekhi
Big Ideas
1) AI is pushing product teams from "symphony" to "jazz"
Deb Liu argues that many product orgs were built like symphonies: defined roles, structured handoffs, detailed specifications, and careful preparation across product, engineering, design, analytics, research, and other functions . AI changes that operating model by making it possible to spin up a prototype quickly, blur role boundaries, and work inside guardrails with more real-time improvisation . Her leadership implication is equally important: leaders move from coordinating a fixed score to setting themes and guardrails for small, high-trust teams .
- Why it matters: The premium shifts from rigid orchestration toward faster adaptation and learning .
- How to apply: Ask whether your team is optimized for orchestration or adaptation, whether roles are too rigid, and whether PMs are waiting for engineering to build ideas that they could prototype first .
2) In B2B, product work naturally drifts away from business reality
"We have to force the work back into connection with reality."
Run the Business frames the gap between what product teams do and what the business needs as the most expensive dysfunction in B2B product companies, and argues that drift toward disconnection is the default . The fix is not just better execution; it is better connection. That means coherent strategy across legibility, synchronicity, composability, and affordability, plus a metric ladder that links growth, retention, and margins to strategic levers and then to team north stars .
- Why it matters: Without this ladder, teams slide into feature-factory behavior, reactive planning, or "south star" metrics that look good on a dashboard but pull the product in the wrong direction .
- How to apply: Check whether each team north star clearly maps to a strategic lever and then to business KPIs. Look for the known anti-patterns: direct revenue accountability for product teams, one-hop feature-to-revenue logic, and "orphan" teams that do not tie to any business outcome . The positive endpoint is creating users so successful they become advocates: "manufacturing champions" .
3) The strongest AI PM workflows are built on local context, not chatbot memory
Sachin Rekhi argues that Claude Code should be the primary AI productivity tool PMs focus on because it is built to generate PM artifacts such as documents and reports, works well with local markdown files, automates workflows through skills, agents, and commands, can run command-line tools, writes bespoke scripts, and keeps context portable instead of locked inside a proprietary chat history . He says he has already shifted the majority of his product work to it .
Dave Killeen shows what this looks like in practice: session-start hooks preload weekly priorities, quarterly goals, and working preferences; markdown files exist for each project, person, and company; meeting transcripts append automatically; mistakes get logged into a reusable file; and AI filters a large information diet down to novel, contrarian signals that matter .
- Why it matters: Context compounds. By day 30, the system can track relationships, commitments, and meeting history; by day 90, it can start surfacing patterns in the PM's own work .
- How to apply: Store reusable context locally, inject current priorities at session start, log repeated AI mistakes, and use AI to produce work products and workflows rather than only one-off answers .
4) AI raises the premium on direct user understanding
A recurring community theme is that AI gives PMs more data, but not necessarily more understanding. One poster argues that many AI tools act like dashboards: they explain what users did, not why they did it or what they actually need . Another makes the same point more directly: AI can summarize data, but it cannot replace the hard work of talking to users and validating what is real . In the strongest example, AI added value only after the research work was done, by helping identify a novel persona hidden in interview transcripts .
- Why it matters: Over-relying on AI as a substitute for research can create false confidence and widen the gap between PMs who do the work and PMs who look for a shortcut .
- How to apply: Keep user conversations human, then use AI on the back end for transcript summarization, clustering, and pattern-finding .
Tactical Playbook
1) Build a local-context PM operating system
A practical setup from the Claude Code examples:
- Create local markdown files for every active project, person, and company you need to track .
- Automatically append meeting transcripts into the right file so relationship history and commitments accumulate over time .
- Add session-start hooks that preload your weekly priorities, quarterly goals, and working preferences into every new session .
- Maintain a mistakes file and inject it at the start of future sessions so repeated AI errors become reusable guardrails .
- Point the system at newsletters, bookmarks, LinkedIn messages, and videos, and ask it to extract only the novel or contrarian signals that matter .
- Use the system to generate the artifacts PMs actually own: plans, reports, backlog outputs, and other deliverables .
Why this works: It turns context from something trapped in meetings and chat threads into a reusable operating layer. In Dave Killeen's example, that meant daily planning, backlog management, career tracking, and compressing 120 newsletters down to the 3% that mattered .
2) Use AI in discovery without outsourcing discovery
"If you don’t talk to your users yourself, you’ll have no idea what’s real."
A grounded AI-assisted discovery loop from the community discussion:
- Conduct the user conversations yourself .
- Use AI to summarize the transcript set and cluster themes .
- Ask AI to look for gaps, anomalies, or a missed persona in the material .
- Go back to the raw interviews to confirm what is real, why it happened, and what users actually need .
- Treat the model as an analysis aid, not a replacement for human interaction or judgment .
Why this works: It preserves respect for users and keeps the PM close to reality while still capturing some of AI's speed in synthesis .
3) Turn a difficult stakeholder into a predictable operating rhythm
Community advice on dealing with a hard-to-reach internal client converged on making access systematic instead of ad hoc:
- Set a recurring meeting cadence that matches how often you actually need decisions or clarification .
- Bring an agenda focused on feedback and requirements before the urgent moment arrives .
- Frame the meeting as a way to reduce last-minute requests and unplanned interruptions for the stakeholder, not just for you .
- Be explicit that you value their insight and clearly state what you need from the relationship .
- Keep a paper trail of outreach attempts so blockers are documented .
- If the stakeholder is critical, create face time through office visits or regular lunches when feasible .
- Escalate if non-response starts obstructing the work .
Why this works: It replaces reactive chasing with a clearer access model and gives you escalation cover if the relationship still fails.
Case Studies & Lessons
1) A non-coding PM turned a parked MCP idea into a production prototype
In sprint planning, a team dropped an MCP server idea for on-call troubleshooting because the estimate was 4-6 weeks of developer time . Later, a non-coding PM used GitHub Copilot with Opus 4.6 to generate a PRD, refine the design, scaffold the project, and work through five phases including testing, security audits, and bug fixes over 2-3 evenings. The team was impressed enough to start using the result in production .
- Lesson: Clear architecture patterns and first-principles decomposition can let PMs prototype ideas far earlier than a full delivery plan would suggest .
- Apply it: Use AI to produce a phased plan and rough implementation when an idea is strategically important but hard to fund upfront .
2) Dave Killeen turned Claude Code into a CPO operating layer
Dave Killeen uses Claude Code for daily planning, backlog management, and career tracking . The system is built from session hooks, auto-updating markdown files, a mistakes log, and AI-driven information filtering . One concrete outcome: he stopped reading 120 newsletters and instead reviewed the 3% that the system flagged as relevant .
- Lesson: The leverage came from durable context and continuous correction, not from a single prompt.
- Apply it: Treat AI setup as infrastructure. The more your files, hooks, and error logs improve, the better the next session starts.
3) A north-star metric can still point south
Run the Business highlights a Microsoft Windows Update example where a team targeted 90%+ of machines on the latest OS version. When the number stayed stuck around 80%, the team stopped asking users and forced updates without consent . The metric improved, but the user experience deteriorated into the familiar mid-sentence reboot .
- Lesson: A green metric can still represent bad product judgment if it is disconnected from customer value and business health .
- Apply it: Pressure-test every north star by asking what behavior it incentivizes, not just whether it is measurable.
Career Corner
1) Prototyping is becoming part of PM craft
Deb Liu's question is blunt: if you are a PM, are you waiting for engineering to build your ideas, or are you prototyping them yourself? Her broader argument is that PMs can now prototype and run research directly as role boundaries blur . The MCP example shows what that can look like in practice, even for a non-coding PM .
- Why it matters: The PM who can turn an idea into a testable first version creates faster learning and usually better conversations with engineering.
- How to apply: Use AI to draft phased plans, PRDs, and rough prototypes before asking the team for a full build commitment .
2) Learn an AI toolchain that produces real work, not just answers
Rekhi's case for Claude Code is career-relevant because it is about artifact generation, local context, automation, and portability, not just chatting with a model . He says he has moved the majority of his product work into that environment .
- Why it matters: PM leverage increasingly comes from compressing execution loops around documents, analysis, and workflow automation.
- How to apply: Build fluency with a toolchain that can operate on your files, preserve context, and generate deliverables PMs are accountable for .
3) Keep direct user understanding as your moat
The community view here is consistent: AI has not replaced the hard work. It can summarize data and surface patterns, but it should not replace direct conversations with users .
- Why it matters: In an environment where more PMs have access to the same models, first-hand understanding becomes more differentiating, not less .
- How to apply: Protect time for direct interviews and transcript review, then use AI after the fact to expand your analysis .
4) Before chasing a title, check whether the role gives you leverage
Run the Business argues that many "unimpactful" PMs never had control over any of the three real assets in a product company: product vision, R&D resources, or the P&L.
- Why it matters: Sometimes low impact is structural, not personal .
- How to apply: When evaluating roles, look past title and ask what decisions, resources, or business levers you will actually control.
Tools & Resources
- Claude Code for PMs — Rekhi's companion video to his argument that Claude Code is a better PM productivity layer than generic chatbots because it supports artifact generation, local markdown context, and workflow automation .
- pmskilltoolkit.com — a community-shared toolkit with 25 Claude/ChatGPT skills across discovery, roadmapping, competitive battle cards, win/loss debriefs, pricing strategy, and stakeholder politics. The creator says it draws on 1,500+ research sources and has reached 10k downloads.
- From Symphony to Jazz — useful if you are rethinking team shape, role boundaries, and leadership guardrails in AI-heavy product development .
- The Most Expensive Dysfunction in B2B Product Companies — a strong reference for reconnecting teams to business reality through KPI laddering and better north stars .
- Watchlist: real-time meeting copilots for internal knowledge retrieval — multiple community posts describe the same gap: typing questions into ChatGPT or Claude Code during technical meetings is too slow, conversations move on, and transcript review afterward often creates more follow-up questions .
Sachin Rekhi
Paul Graham
Teresa Torres
Big Ideas
1) Competitive advantage beats differentiation
“The most differentiated product? Might just be something nobody wants.”
Ravi Mehta’s framing is to shift strategy conversations from ‘what makes us different?’ to ‘what makes us better?’ Uniqueness only matters if customers actually want the unique thing; before PMF, similarity to products with proven demand can be an asset, while compounding advantages are what create monopoly-like outcomes . Uber’s edge over Lyft came from network effects that improved price and speed, while Spotify beat Tidal with more reliable streaming, not a more differentiated offer .
Why it matters: Differentiation can push teams toward features customers do not value; advantage keeps the conversation tied to customer benefit and compounding market position .
How to apply: In strategy reviews, ask three questions: What advantages do we already have relative to the market? How do those translate into a product customers experience as better? Which of those advantages can compound over time?
2) The spec has moved later in the product flow
Aakash Gupta describes an old flow of Idea → PRD → Design → Eng → QA → Ship taking 8–12 weeks, where the PRD acted as a permission document . The emerging flow is Idea → 5 prototypes → Evaluate → Kill 4 → Spec the survivor → Ship in 1–2 weeks, with the spec now serving as a decision record .
Company context still changes where documentation sits. Anthropic teams may skip PRDs and prototype directly, shipping 20–30 PRs a day; OpenAI still needs specs for products serving 800 million MAU with 15–25 labeled examples per feature; large enterprises still need docs for alignment across 5,000 people and three time zones .
Why it matters: A prototype shows what exists; the spec explains why it matters, how it will be measured, and when to stop. Gupta argues PMs who prototype first ship 5x more validated features .
How to apply: Create multiple fast variants before committing to one direction, then use the spec on the winning version to record the decision, success criteria, and pull-the-plug conditions. Adjust rigor based on stage and coordination needs, not habit .
3) Correct analysis is not enough if the underlying facts are wrong
“A report is a pull request against your organization’s knowledge.”
Leah Tharin’s example is a PM report built on a traffic spike that was real in the data but distorted by bot scraping; the processing was correct, but the baseline was polluted . Her fix is a verified context layer: a plain-text repository of leadership-verified facts that AI and analysts check before new analysis is accepted .
Why it matters: ‘Retrieval correctness’ is not the same as factual correctness. If known anomalies stay in people’s heads, wrong conclusions compound into new facts .
How to apply: After each review meeting, write the correction down in one line, make context files findable, require AI to read them before analysis, and force outputs to include methodology and sources .
4) Growth rate is often the cleaner prioritization metric than the number itself
Paul Graham argues that if you want a higher standard, graph the growth rate of the metric you care about rather than the absolute number; even keeping growth flat becomes difficult . The same lens helps teams spot promising product variants earlier. A new line of business making only a few thousand dollars a week can still be the important one if it is growing 10% weekly; if that rate is real and the market does not cap out, it compounds dramatically .
Why it matters: Absolute revenue can hide small but fast-growing tails that will eventually matter more .
How to apply: Review both absolute performance and growth rate for your core metric and for new variants. Do not pivot on one spike alone—first confirm the growth rate is sustained .
Tactical Playbook
1) Use AI to remove research production, not product judgment
Sachin Rekhi’s workflow for customer discovery is to spend less time on interview logistics and synthesis production, and more time mining research for insight . His sequence is:
- Read AI-synthesized findings
- Ask LLMs follow-up questions against the raw transcripts
- Dive into specific verbatims and watch interview videos yourself
- Derive the product implications personally
Why it matters: Used thoughtfully, AI can improve product intuition by freeing PM time for deeper engagement with customer evidence .
“At no point do I actually trust AI to come up with what to do in the product based on the research. That’s my job.”
2) Run a 60-minute pre-mortem before important launches
The pre-mortem reframes kickoff work by asking the team to imagine the project has failed six months from now and explain what happened . A simple agenda:
- Setup (5 min): Set the future-failure scenario
- Individual brainstorm (10 min): Write specific failure reasons, not vague ones
- Collect (15 min): Read and record each reason without debate
- Group and vote (15 min): Cluster ideas and pick the top 3–5 risks
- Prevention plan (10 min): Define preventive measures and early indicators
- Assign owners (5 min): Put names and deadlines on the actions
Why it matters: The method lowers social pressure, fights optimism bias, and gives quieter team members room to surface uncomfortable risks .
How to apply: Force specificity in the failure statements, and pair each top risk with the first measurable warning signal and an agreed decision trigger .
3) Turn corrections into a reusable context file
A practical version of Tharin’s verified context layer is lightweight: when a report is wrong because of seasonality, bots, renewals, or another known anomaly, write a one-line correction in plain text and keep it findable . Then tell your AI system to load the relevant file before analysis—for example, a timeline file for a specific year—and require a fixed output format with methodology and sources .
Why it matters: This catches wrong facts before they are merged into decision-making and repeated in later analysis .
4) Use autoresearch only when the task is objectively scorable
The autoresearch loop works when an agent can change one file, run an automated evaluation, and keep or revert the change based on a numeric score . The three hard requirements are:
- A clear metric, scored as a number rather than a feeling
- An evaluation harness that runs without a human in the loop
- One editable file for the agent to change
In that setup, the PM defines what ‘better’ means and the agent runs the iterations . Gupta reports about 12 experiments per hour and roughly 100 overnight; in one prompt-based skill, the score moved from 41% to 92% in four rounds .
How to apply: Start with one prompt, template, or skill that frustrates you, lock the evaluator, and let the agent commit only when the score improves .
Case Studies & Lessons
1) Medable built an agent platform for regulated, messy workflows
Medable’s Agent Studio is a no-code/low-code platform for configuring and deploying agents across the clinical trial lifecycle, with bring-your-own-model support, RAG knowledge, MCP connectors, workflow functionality, multiple trigger types, and versioned publishing . The platform approach came from two recurring problems: high cognitive load on humans working from 200-page protocols, and critical data spread across many systems .
Two examples show the pattern:
- ETMF agent: targets document classification across 80,000-plus documents per year, where users previously spent at least five minutes per document across 350 classifications. Medable started with human-in-the-loop validation and used a 2,000-document golden dataset to evaluate accuracy before launch .
- CRA agent: combines data from 13 clinical systems so clinical research associates can monitor quality and patient safety, and it adds recommended actions instead of only surfacing signals .
Lesson: This case depends on more than model choice: evals as a stage gate, post-launch monitoring, intent-to-design-to-test traceability for GXP compliance, ontology mapping across systems, and tool filtering to manage context bloat .
2) Autoresearch is already producing outcomes humans missed
Karpathy left an autonomous optimization loop running for two days and the agent found 20 improvements on code he had hand-tuned for months, including a missed bug, which stacked into an 11% speedup . Shopify CEO Tobi Lutke ran 37 experiments overnight and saw a 0.8B parameter model outperform a hand-tuned 1.6B model . The same pattern applied to Shopify’s Liquid templating engine produced 53% faster rendering and 61% fewer memory allocations from 93 automated commits .
Lesson: Autonomous iteration is most valuable where the evaluation function is clear and cheap to run; without that, the loop does not hold .
Career Corner
1) Open strong: give the arc, not the autobiography
When asked for a high-level overview, the hiring-manager advice here is simple: answer with the arc of your career, not a company-by-company retelling, and stop early enough to ask whether the interviewer wants more .
How to apply: Practice a two-minute version that covers transitions and themes, then pause with a check-in such as ‘Is that enough, or would you like more detail?’
2) In behavioral answers, lead with agency and strategy
Interviewers want to hear what you did, not just what the team shipped. That means using ‘I’ to explain the actions you took—running the numbers, creating context, surfacing a quiet voice—while still showing how those actions helped the team . The same video argues that strong PM answers also make strategy concrete: the goal, the market context, and the bets behind the work .
Why it matters: The role is framed here as maximizing ROI toward the company’s strategic goals, not being the team’s shield, glue, or requirements writer .
3) Show presence by handling ambiguity without blame
The same hiring manager warns against blaming founders, sales, engineers, or past bosses in interviews; it reads as victimhood rather than leadership . A better pattern is to ask for clarification, admit when you do not have the exact example, offer a related one, and stay calm under the question . That combination of honesty and composure is described as presence or credible confidence .
How to apply: Replace blame stories with reframing stories: explain how you understood the other party’s agenda, translated it for the team, or learned more before reacting .
Tools & Resources
- Autoresearch setup: Gupta’s quick-start suggestion is to install Claude Code, clone karpathy/autoresearch, and start with the prompt or skill that frustrates you most .
- Pre-mortem checklist: A lightweight template for running the launch exercise is linked here: blog.promarkia.com.
- Interview practice: A community-posted free AI-powered PM interview practice tool for product sense, execution, and leadership is here: interview-prep-master-shaiabadi.replit.app.
Paul Graham
scott belsky
Elena Verna
Big Ideas
1) PM work in AI is shifting from specs to evaluation systems
At Shopify, the move to AI products like Sidekick changed the PM job from hardening every requirement in a spec to defining 'what good looks like' and encoding that judgment in an internal LLM-based evaluator called Ace. PMs review fake or sanitized real conversations, score them across dimensions such as accuracy, grounding, merchant sentiment, goal fulfillment, execution quality, errors, language quality, and safety, and then approve aligned examples into a ground-truth set used to create a judge. The same pattern is applied to generated workflows in Flow, not just chat experiences.
Why it matters: Non-deterministic products cannot be fully managed through traditional specs alone; quality depends on whether the team can repeatedly evaluate open-ended outputs.
How to apply: Treat the eval stack as part of the roadmap. Define dimensions of quality early, create annotated examples, require agreement across raters on subjective calls, and reuse the approach anywhere your product generates content, queries, or workflows.
2) Outcomes are a better organizing unit than outputs
"When we measure success by what you ship, you get a feature factory. When you measure success by the impact of what you ship, you get a product team that learns, adapts, and creates real value."
Teresa Torres draws a clean distinction: an output is the thing you build, while an outcome is the customer or business change it creates. Teams given outputs have little latitude to explore; teams given outcomes have more freedom and more responsibility to find the best path.
Kate Tarling's service-organization guidance pushes the same logic up a level: list services from the outside in, set measures for what good looks like, map current work against those services, and reject the assumption that the status quo is a neutral or risk-free baseline.
Why it matters: This is not just a metrics tweak. It changes portfolio decisions, intake discipline, and how teams defend capacity.
How to apply: Rewrite incoming requests in problem-and-outcome language, choose metrics tied to value creation rather than shipment, and expect to iterate on the first version of the metric. If the work sits inside a broader service, map it to the service outcome before you prioritize it.
3) Growth is moving away from funnel tuning and toward access, trust, and community
Elena Verna argues that growth used to be an optimization problem focused on acquisition, activation, and monetization, but is now shifting toward strategic product giveaways, outcome-aligned monetization, and deliberate community building. She also argues that feature differentiation is weakening, making trust, brand, and human connection more central to adoption.
Her distribution view is similar: paid marketing is harder, organic app-store style discovery is crowded, and the clearest current channel is organic, creator-led social. At Lovable, that shows up as employee-led social, building in public, and feeding social feedback directly back into the product.
Why it matters: PMs can no longer assume that better funnel copy or pricing-page tweaks will do most of the work. Growth is increasingly tied to lowering barriers, delighting users, and creating visible feedback loops.
How to apply: Test barrier-removal moves, instrument how users discover and talk about the product, and align monetization with customer outcomes rather than short-term extraction. Also protect room for innovation: Verna warns that private companies can damage long-term competitiveness when quarterly pressure crowds out experimentation.
4) Agentic products and AI-native teams still need visible human judgment
Scott Belsky's product strategy point is simple: enterprise products that help humans get credit for work done by supervised agents may adopt better than products that simply do the work instead of humans, because credit affects ego, adoption, and accountability.
That complements a warning from Shreyas Doshi: founders and executives are seeing AI-drafted proposals collapse under pushback because the person presenting them has not thought them through deeply enough. AI-prepped ideas can fail the moment an unexpected question appears.
Why it matters: The failure mode is not just bad output. It is weak ownership—products that obscure human credit and teams that outsource reasoning.
How to apply: Design human review, attribution, and decision checkpoints into agent workflows. Internally, use AI for drafts, but require the owner to defend trade-offs, assumptions, and edge cases without the model in the room.
Tactical Playbook
1) Build an AI quality loop before you scale the feature
A practical loop from Shopify looks like this: define quality dimensions, assemble fake or sanitized real examples, have multiple PMs annotate them, approve aligned examples into a ground-truth set, and use that set to create or refine a judge. The same structure can be reused across different generative surfaces, including workflow generation in Flow.
- Define what good looks like in dimensions, not just one score.
- Review examples that are close to real usage, including sanitized live conversations where possible.
- Have at least two PMs rate the same examples to surface subjectivity early.
- Approve only the aligned examples into ground truth.
- Use the approved set to create or update the judge.
2) Interpret metrics with orthogonal context, not one impressive number
Julie Zhuo's example is a useful habit check: 5M MAU and 80% DAU/MAU looks strong on its own, but a 30-second average daily session length can make the same product look shallow—until you learn it is a payments app, where a short session may be exactly right. Her recommendation is to add orthogonal context: independent facts that reduce ambiguity from different directions.
Use this sequence:
- Start with the headline metric.
- Add a behavioral counterweight, such as session length or task-completion pattern.
- Add use-case or business-model context that changes what 'healthy' should mean.
- Delay conclusions until you have multiple independent facts pointing in the same direction.
3) Run a deep-dive sprint when you lack enough context to help
Vanessa Lee describes an unorthodox but practical leadership move: if something feels off, run a sprint with the team and meet daily. In her example with Shopify's inventory team, the work included examining current bugs, data structures, UX, and next-day experiments until she had a much better understanding of the domain and the team had a clearer shared aim.
Why it matters: Leaders often try to guide work from too far away. A short period of intense involvement can replace vague opinions with specific judgment.
How to apply: Use a time-boxed daily cadence, stay close to the real artifacts, and push one concrete next step each day. The goal is not permanent micromanagement; it is to earn enough context to give better guidance later with less involvement.
4) For hard stakeholder debates, bring proof and a lightweight artifact
The API versioning story at Shopify is a strong template for hard influence. The PM had spent 12 months working with teams that were inadvertently breaking the API, entered the meeting with four concrete examples, and brought a hand-drawn proposal for quarterly versioning with one year of backward compatibility. The drawing was deliberately simple so the CEO could engage with the idea rather than react to a polished deck.
Why it matters: Senior stakeholders are easier to move when you combine superior domain knowledge with a proposal that still leaves room for their input.
How to apply: Do the slow homework before the meeting, arrive with evidence instead of opinions, and use artifacts that clarify the decision without signaling that every implementation detail is already fixed.
5) Triage new work by service and outcome before you discuss priority
In complex organizations, Tarling recommends mapping existing work to services, comparing each item to what good looks like, and sorting work into strong fit, needs investigation, or candidate to stop. For new requests, a triage function should require the requester to describe the problem and the outcome before the work gets any further.
Why it matters: This moves prioritization away from who asked loudest and toward evidence, overlap, and strategic fit.
How to apply: Start in one cross-functional service area rather than changing the whole organization at once, learn where the real barriers are, and then expand the operating model.
Case Studies & Lessons
1) Shopify Sidekick and Flow: product judgment became infrastructure
Shopify's response to AI product quality was not just to improve prompts. The team built an internal evaluation workflow where PMs define the dimensions, score conversations, align across raters, and create ground-truth data that powers an internal judge. The same logic is used beyond Sidekick conversations and applied to generated flows in Flow.
Lesson: For AI products, your evaluation system is part of the product. If it is weak, quality work stays anecdotal.
2) API versioning at Shopify: courage works best when it is evidence-backed
The starting point was operational pain: teams were breaking an unversioned API. After a year of working directly with those teams, a more junior PM entered a meeting with the CEO knowing the domain in depth, carrying multiple proof points, and proposing quarterly versions with one year of backward compatibility. The CEO started from a 'not doing that' position, but the decision changed in a 45-minute meeting.
Lesson: Courage alone is not enough. The persuasive sequence was domain mastery, concrete evidence, and a proposal simple enough to discuss collaboratively.
3) Lovable Free Day: removing friction can be a growth event
Lovable made its platform free on March 8 to lower barriers to experimenting with AI and tech. Verna says the result was more than 100,000 new users and 120 events across 40 countries. Her broader framing is that growth is increasingly about giving access, building community, and creating delightful experiences that later connect to monetary outcomes.
The operating model around that is equally notable: build in public, encourage employees to post organically, watch social feedback closely, and turn that feedback into rapid product changes—she describes fixes going live in 20 minutes.
Lesson: Sometimes the best growth move is not another funnel optimization. It is a visible barrier-removal moment backed by fast feedback loops and a clear product story.
Career Corner
1) If AI tools are widely available, your edge shifts to judgment, strategy, and domain depth
Sachin Rekhi's answer to the 'how do PMs differentiate?' question is a useful checklist: stay at the frontier of AI fluency, keep high standards and taste, deepen domain expertise, invest in product strategy, and keep building design skill—especially interaction design, where he says AI still lags world-class practitioners.
How to apply: Pick one domain to know unusually well, use AI aggressively but discard low-quality output, and make sure you are still practicing strategy and design decisions rather than only prompting for them.
2) Do not let AI do your thinking for you
Shreyas Doshi's warning is career-relevant: competent executives are repeatedly seeing people bring AI-assisted proposals they cannot defend once the conversation gets messy. The failure shows up when a client pushes back, an investor asks an unplanned question, or an unexpected condition breaks the logic.
How to apply: Before circulating a doc or proposal, pressure-test it yourself. List what would change your mind, what edge case breaks the recommendation, and what question you hope nobody asks. If you cannot answer those without the model, the work is not ready.
3) Career signal: proof and courage compound faster than credentials
"PMs are the, I call it courage as a service to the team."
The Shopify interview frames PMs as the people who make hard calls, surface broken team dynamics, and sometimes kill months of work when it is wrong. The same conversation also notes a preference for people who have been founders for at least a year, and describes Shopify as a place where strong storytelling and decision-making can create real autonomy at scale. Paul Graham makes the adjacent point from startups: users care whether they like the product, not what is on your resume; a resume is only a predictor of performance.
How to apply: Build a track record of visible decisions, not just shipped tickets. Keep artifacts that show how you framed a problem, changed course, or defended a tough call; those are better career signals than generic claims of ownership.
Tools & Resources
- Shopify VP of Product on Building a $100B+ Agent-Led Commerce — strong on AI evaluation systems, leadership sprints, and stakeholder management under disagreement.
- How to fix broken systems - Kate Tarling (CEO, The Service Group) — practical for service mapping, portfolio cleanup, and problem-and-outcome intake triage.
- Teresa Torres on outputs vs outcomes — short refresher on choosing outcome metrics and iterating when the first version is wrong.
- Lovable Free Day, 100K New Users, Here's Why It Worked — useful for PMs rethinking growth around access, community, and building in public.
Product Management
Sachin Rekhi
Big Ideas
1) Production-ready GenAI is a systems problem, not a feature problem
The strongest framework this week comes from a Product School talk by an Amazon AI product leader, who cites Gartner’s estimate that 85% of AI projects never make it to production and argues that the gap is usually caused by missing system design, not missing features . The proposed five-pillar framework is:
- User-centric design grounded in real pain points and jobs-to-be-done
- Robust evaluation across trust, usefulness, adoption, and business impact—not accuracy alone
- Governance and safety with guardrails, transparency, and compliance built in from the start
- Scalable architecture for performance, cost, reliability, and extensibility
- Adoption strategy with pilots, enablement, community, and feedback loops
"AI products succeed when PMs design systems and not features"
Why it matters: This reframes the PM job for GenAI from shipping a capability to designing the full operating system around that capability—evaluation, trust, scale, and adoption included .
How to apply: Before calling a GenAI initiative “ready,” force a launch review that answers five questions: what user job is being solved, how success will be measured, what guardrails exist, how the system scales, and how adoption will be driven after launch . The same speaker’s guidance is to move fast, but build right, with 3–6 months achievable when teams do the upfront work that avoids re-architecture later .
2) The PM operating system is shifting from meetings and docs to loops, logs, and simulations
Andrew Chen argues that in an agentic world, the product role splits into two jobs: organizing humans and organizing agents. In his framing, standups become anomaly and run-log reviews, OKRs become continuous agent-based grading, PRDs give way to living agentic loops, and product reviews become simulations that test agent behavior under different constraints .
This future-facing view also fits the more current GenAI PM description from the Amazon talk: PMs already bridge AI capabilities to user problems, shape ethical and trustworthy AI use, and align technical and non-technical stakeholders .
Why it matters: The PM surface area is expanding from persuasion and coordination into instrumentation—prompts, evals, workflows, feedback loops, and behavior review .
How to apply: Treat agent behavior as something that needs product management, not just engineering. Build explicit prompts and evals, review deltas and failures, and make simulation or scenario testing part of pre-launch review for agentic systems .
3) As AI speeds up engineering, discovery is becoming the bottleneck
Sachin Rekhi’s concise diagnosis: engineering velocity has 10x’d with AI coding tools, but customer discovery hasn’t kept pace, so PMs are increasingly the constraint in deciding what to build, how to design it, and how to validate it before shipping .
He responds with 10 AI-powered discovery workflows spanning surveys, feedback streams, interview scripting, interview synthesis, AI-moderated interviews, prototype-based discovery, metrics analysis, and automated metric analysis .
Why it matters: Faster build loops create more pressure on PMs to improve discovery throughput and decision quality, not just documentation quality .
How to apply: Audit your current discovery flow and identify the slowest step. Then add AI support there first—survey analysis, interview synthesis, prototype discovery, or metric analysis—rather than trying to automate everything at once .
4) Messy docs can be a strength—if you design a clean interface to the organization
The Beautiful Mess argues that many high-performing product teams rely on freeform, manually migrated documents filled with links, flags, checklists, copied data, comments, and repeated context . The point is not formal structure; it is externalizing working memory so teams can reason through customer signals, hypotheses, dependencies, and half-formed ideas together .
The tension is that teams need local emergence, while organizations still need legibility about progress, risks, and focus . The preferred answer in the essay is not to eliminate the mess, but to design intentional interfaces: the smallest shared routines, objects, and language that let the rest of the org understand what is happening without crushing the frontline work .
Why it matters: PM teams often over-rotate toward official artifacts and lose the sense-making layer where important work actually happens .
How to apply: Let teams keep their working scratchpad, but define a minimal interface outward: a small set of recurring rituals, a few shared objects, and consistent language for status, risks, and decisions .
Tactical Playbook
1) Turn AI discovery into a repeatable four-stage workflow
A practical way to apply Rekhi’s 10 workflows is to map them into four stages:
- Collect signals: analyze customer surveys, automate survey programs, and automate feedback rivers
- Run interviews faster: generate interview scripts, conduct AI-moderated interviews, and synthesize interview feedback
- Test concepts: use prototypes for discovery and, where useful, generate synthetic user feedback
- Close the loop with numbers: analyze metrics and automate metric analysis
Why it matters: This creates a discovery pipeline that can better match faster engineering cycles .
How to apply: Start with one workflow per stage. For example, automate survey analysis, draft interview guides with AI, test concepts via prototypes, and then automate recurring metric readouts .
2) When a metric drops, require diagnosis before brainstorming
A useful community workflow for analytics triage starts with a strict rule: the agent does not get to brainstorm until it can identify where the delta is coming from—specifically the step, segment, and time period involved . Only after that first pass does it move to generating 2–3 experiments and a tracking checklist, with every idea mapped to a measurable metric .
Why it matters: It prevents the common PM pattern of spending 30 minutes in dashboards without notes, structure, or a clear next step .
How to apply: Standardize a three-step triage:
- First answer: what changed, where, and since when
- Then propose: 2–3 experiments tied to the diagnosed step or segment
- Then create: a tracking checklist so engineering gets a concrete handoff and each idea is measurable
The same thread raises two good discipline questions for teams to adopt: what are your first three checks when a metric drops, and do you document what you ruled out or rediscover it every time ?
3) Plan adoption before launch, not after it
The Amazon talk is especially strong on adoption mechanics. The suggested sequence is:
- Run pilots with a defined population and gather real feedback before global launch
- Build success stories so launch materials show concrete use cases, not just product claims
- Invest in documentation, tutorials, and training so users can self-serve and leaders understand the rollout
- Create a community where users can share tips, ask questions, and report issues
- Maintain a transparent roadmap and keep feedback loops active after launch
Why it matters: The speaker explicitly argues that building is only half the battle; without discoverability, enablement, and change management, even strong AI products fail to get used .
How to apply: Add adoption work to the launch checklist itself—pilots, champions, docs, training, community, and roadmap visibility—rather than treating them as marketing extras .
Case Studies & Lessons
1) Amazon collaboration spaces: a full-stack GenAI rollout
Problem: Teams across Amazon needed AI systems with their own knowledge bases, documents, settings, and tools; generic systems did not understand team-specific context .
Product decision: The team built collaboration spaces where teams could upload documents, customize prompts, integrate with other Amazon tools, and control access and permissions . They validated the concept with user research before writing code, built evaluation in from day one, treated governance and safety as core features, architected for scale, and paired the product with pilots, documentation, and community .
Outcomes: The rollout went from an initial 12–18 month timeline to six months from concept to global launch. Reported results included 40–50% faster prompt creation, 3x higher engagement for role-specific content, 2x retention on repeat user rates, and five major feature announcements in the first two months post-launch because the architecture was extensible .
Key takeaway: Enterprise GenAI speed came from doing more product work upfront, not less—especially on evaluation, governance, architecture, and adoption .
2) LennyRPG: how a non-technical product designer used AI to build a real product
Ben Shi, a non-technical product designer at Miro, built LennyRPG, a Pokémon-style RPG based on Lenny’s Podcast, as an AI-assisted product build . The process is notable because it mirrors classic product development more than “prompt and pray” building:
- Define the core idea and visualize it for the AI when the product is highly visual
- Create a PRD by having the AI interview the creator, then synthesize answers and artifacts into a single source of truth
- Build a POC around the core loop first
- Pivot fast when the stack is wrong—from RPG-JS to Phaser when the framework fought the quiz-based design
- Systematize repetitive work with CLI tools for quiz generation and avatar creation across hundreds of episodes
- Polish and ship with AI-assisted QA and UI cleanup
Two lessons stand out. First, Shi says getting the core idea and PRD right determines 80% of how smooth the rest of the build will be. Second, the early validation was intentionally lightweight: he shared the POC internally to see whether people understood what to do, whether the core loop made sense, and whether it felt fun rather than like work .
Key takeaway: AI can accelerate implementation and batch work, but the hard product choices—concept clarity, framework fit, game balance, and what “good” feels like—still required deliberate PM judgment .
Career Corner
1) Discovery is becoming a career-defining PM skill
If engineering velocity is increasing much faster than discovery velocity, PM leverage shifts toward faster learning, not just faster execution . Rekhi’s 10-workflow list is a useful skills map for PMs who want to stay ahead: survey analysis, feedback automation, interview design, interview synthesis, prototype discovery, and metrics automation .
How to apply: Pick one discovery workflow you do repeatedly and learn how to speed it up with AI this quarter .
2) Open-ended PM interviews are testing structured thinking under ambiguity
One candidate described repeatedly failing the brainstorming stage of PM interviews despite positive feedback on energy and bias to action . The examples were deliberately broad: propose three products after a data breach, explain how Spotify recommendations work, or organize a folder so others can navigate it easily . The thread’s core question was whether experienced PMs rely on a specific framework or thought process in these situations .
What to take from it: These rounds appear to reward legible reasoning and repeatable structure, not just raw creativity .
How to apply: Practice unfamiliar prompts and focus on making your reasoning easy to follow—problem framing, assumptions, options, and trade-offs—rather than trying to sound instantly brilliant .
3) Community signal: freelance PM work may be easier to win as concrete delivery work
In one Product Management thread, a PM with strong 0→1, 1→10, and AI prototype-building experience is exploring freelancing while building a portfolio of small AI projects and apps . A reply says Upwork still has some good opportunities, but few are true freelance PM roles; more are narrow tasks such as analytics configuration or effectively full-time work routed through the platform .
What to take from it: The clearer the deliverable, the easier the market fit may be for freelance PM work in today’s environment .
How to apply: If you are testing freelance PM work, package yourself around concrete outcomes—MVPs, prototypes, analytics setup, or specific product problem-solving—rather than a generic “fractional PM” label .
Tools & Resources
- Lean Product Meetup recording — Sachin Rekhi’s walkthrough of 10 AI-powered discovery workflows
- Building Production-Ready GenAI Products — five-pillar framework plus enterprise GenAI examples from Amazon
- How I built LennyRPG — written breakdown of the AI-assisted six-step build process
- LennyRPG video walkthrough — concise tool stack: Miro, ChatGPT, Claude Code, Cursor, GPT-4, Phaser, Supabase, and Vercel
- ClawRapid analytics skills directory — analytics connectors referenced in the metric-drop workflow
The community for ventures designed to scale rapidly | Read our rules before posting ❤️
Product Management
Big Ideas
1) AI has compressed prototyping time, not the need for PM judgment
Product School frames the PM bottleneck as time to build and time to learn, and defines vibe coding as using AI tools to turn natural language into a runnable prototype that users can react to . The gain is faster movement from idea to evidence, not permission to lower the shipping bar: the speaker is explicit that vibe coding is not production code, does not replace engineering, and does not remove security, privacy, reliability, accessibility, or review requirements .
“AI compresses execution. The writing, the code, the analysis. What it can’t compress: knowing what to build. Knowing what to cut. Taste. Judgment. Intent.”
Hiten Shah makes the same distinction directly: AI has closed the gap between I can build this and shipping speed, but the question of whether something is worth building still belongs to the product team .
Why it matters: As execution gets cheaper, product judgment becomes more important, not less .
How to apply: Use AI to shorten discovery loops, but keep production handoff, review, and safety standards unchanged .
2) The next PM tools are proactive systems, not just chat interfaces
OpenClaw is positioned as proactive, model-agnostic, and local: it can run cron jobs, scan channels, monitor websites, generate reports, and post to Slack while you sleep; it can also switch models by use case and keep data on your machine . The Product Compass case study describes a parallel pattern on the knowledge side: a file-based system with a brain file (CLAUDE.md), a router (knowledge/INDEX.md), domain folders, and progressive disclosure so only relevant context loads for a task .
Together, they point to a broader shift: PM leverage is moving toward persistent systems that store rules, memory, workflows, and hypotheses instead of relying on one-off prompts .
Why it matters: Repetitive PM work like standups, competitive monitoring, customer synthesis, and knowledge retrieval can compound when the system keeps structure across sessions .
How to apply: Start with one recurring workflow, externalize its rules and memory into files, and route the system only to the context it needs for that task .
3) AI may collapse role boundaries, but it raises the premium on customer empathy
In YC’s profile of an AI startup, one speaker says a single person can increasingly do combined PM, design, and engineering work, and that some work previously done by five or six people can now be done by one engineer or one PM in internal settings . At the same time, the company requires everyone to talk to customers once or twice a week and rotate through customer support, even with a 12-person engineering team, because it helped build customer empathy from day zero .
Why it matters: AI expands functional range, but customer contact still anchors prioritization and product judgment .
How to apply: Use AI to widen your prototyping and execution surface area, but protect direct customer conversations as a weekly habit rather than delegating all learning to dashboards or prompts .
Tactical Playbook
1) A four-step vibe coding loop for discovery
- Write a one-sentence job statement with situation, need, and outcome .
- Define the must-be-true assumption as the riskiest measurable condition that would make the idea worth building .
- Build the smallest believable demo with real prompts, real outputs, minimal UI, and realistic edge cases so users react with real behavior .
- Capture learnings including what worked, what broke, and whether to build, change, or kill the idea .
Why it matters: This speeds up evidence gathering without confusing a prototype with a shippable product .
How to apply: Keep the demo lightweight and disposable, test with real users in controlled conditions, and hand off anything real to engineering .
2) Make capacity trade-offs explicit by queue, not implicit by politics
A community discussion around Capacity Is the Roadmap argues that different work types—client work, technical debt, and maintenance—should sit in explicit queues because they compete for the same developer bandwidth . One commenter says discussions become more straightforward once you are negotiating developer bandwidth directly .
How to apply:
- Define queues by work type or business line .
- Add non-feature work like technical debt and maintenance, not just customer requests .
- Make shared bandwidth the explicit constraint in roadmap conversations .
- Ask which queue gets bandwidth this cycle before debating individual items .
Why it matters: It surfaces the real trade-off instead of letting some work stay invisible .
3) Run executive reviews with a no-surprises script
One consumer product team described quarterly ops reviews covering benchmarks, sentiment, channel dynamics, KPIs, supplier health, portfolio expectations, and roadmaps . The practical advice from replies was consistent:
- Hold pre-reviews with leaders whose support you need; do not introduce major issues for the first time in the room .
- Use a no-surprises approach with overlapping stakeholders and pre-assign allies on sensitive topics .
- Rehearse hard questions from the audience’s perspective and keep backup notes ready for detail .
- Make the review itself boring so time goes to decisions rather than explainers .
Why it matters: The meeting is not where alignment starts; it is where pre-work gets tested .
How to apply: Treat the deck as the last step in stakeholder management, not the first .
4) When a product has wow factor, validate utility with paid repeat behavior
A founder building an immersive desktop product asked how to separate visual impressiveness from genuine usefulness, and the signals raised were retention, repeat usage, and willingness to pay . One concise answer from the thread: charge people and see if they keep paying while continuing to use the product .
Why it matters: Novelty can create strong first reactions without creating durable value .
How to apply: For early tests, track whether users return and pay, not just whether they say the experience looks impressive .
Case Studies & Lessons
1) Twitch AI agent: prototype in the real environment, but keep the blast radius small
The Twitch product leader’s job was to keep chat fun, safe, and engaging without constant context switching during streams . The must-be-true assumption was strict: the AI could not say anything outside community guidelines or channel rules, which led to a second AI moderator role . The prototype ran in the speaker’s own channel with real viewers and real messages, but in a controlled environment where the service could be killed immediately from the desktop if anything got weird .
The learnings were captured and shared with the team, and the speaker is explicit that the goal was to learn what the prototype could and could not do—not to claim it was ready to ship .
Key takeaway: Real user behavior is more informative than polished mockups, as long as safety boundaries are explicit and reversible .
2) OpenClaw shows where PM automation is getting practical—and where it can go wrong
OpenClaw’s tested PM workflows include a Slack knowledge base, stand-up summaries before the first meeting, a competitive intelligence pipeline, voice-of-customer reports, and bug routing by customer tier . One example: a cron job can monitor competitor websites every 30 minutes and capture a pricing-page change that appears at 1 a.m. and is overwritten by morning . The system exposes its behavior through local markdown files such as soul.md, agents.md, memory.md, and heartbeat.md, plus a gateway dashboard at 127.0.0.1:18789.
But the same testing also surfaced sharp edges: one bot read personal files it should not have accessed, and another sent pairing codes to every WhatsApp contact on a phone .
Key takeaway: Persistent, local agents can automate real PM workflows, but setup and permissions are part of the product evaluation—not an afterthought .
3) A self-improving Claude system turned ad hoc analysis into compounding knowledge
The Product Compass author says the system started with raw curiosity—pasting screenshots into Claude and asking what made posts work . Over time, Claude suggested a knowledge hierarchy, built a cheaper data-fetching script, and began proposing edits to its own knowledge base . The resulting system now tracks 26 content templates, 13 active hypotheses, 50+ catalogued false beliefs, and 7 topic lanes with energy tracking .
The author keeps editorial control over what to post, what to kill, which angle to take, and which facts need extra checking, while Claude handles research, verification, structural options, and pattern matching against the knowledge base .
Key takeaway: Start messy, formalize only after patterns emerge, and keep final judgment human even as the system compounds .
Career Corner
1) Build your own case study if you want better interview material
In a community thread about PM training options, the strongest advice was to build your own product case study instead of relying on courses alone, because it gives you something concrete to walk through in interviews and iterate over time . The original poster was explicitly considering making a case study from online sources and using that to apply for interviews .
Why it matters: A self-built case study shows how you think, not just what you completed .
How to apply: Pick one product problem, build a lightweight case around it, and be ready to explain the decisions, trade-offs, and iterations you made .
2) The AI-era PM skill stack is widening, but judgment is still the moat
Across the sources, the pattern is consistent: AI lets PMs get closer to prototyping and cross-functional execution , but it does not answer whether something is worth building . It also does not remove the need for production standards or customer empathy .
One useful nuance from the YC discussion: even a fairly technical PM can still feel intimidated by raw JSON, which is a reminder that being effective in AI-native product work is not the same thing as being comfortable with every implementation artifact .
Why it matters: The PM advantage is moving toward problem framing, evidence gathering, customer contact, and judgment under faster execution cycles .
How to apply: Learn one prototyping workflow well, join customer conversations regularly, and measure yourself by the quality of decisions you enable—not by how much raw code you can tolerate .
Tools & Resources
- Vibe Coding as a PM: Hands-On AI Prototyping | Twitch Director of Product — a practical framework for job statements, must-be-true assumptions, smallest believable demos, and learning capture .
- The Complete Guide to OpenClaw for PMs [EXCLUSIVE] — a walkthrough of installation, Slack setup, local files, and five PM automations including standups, competitive monitoring, VoC, and bug routing .
- What I Learned Building a Self-Improving Agentic System with Claude — a detailed repo pattern built around
CLAUDE.md,knowledge/INDEX.md, progressive disclosure, and a PM-friendly adaptation of discovery, stakeholders, and channels . - The What & Why of Continuous Discovery — a free webinar on April 7, 2026 introducing a structured and sustainable approach to continuous discovery .
- Product Discovery Fundamentals — Teresa Torres’ hands-on program running April 28 to June 2, 2026 .
- On-demand courses worth bookmarking:
Lenny Rachitsky
Product Management
Big Ideas
1) Agent-aware execution is becoming a PM systems problem
Traditional project boards assume a human picks a ticket, works it, updates status, and ships code. Teams experimenting with MCP-connected coding agents reported duplicate work, unreliable progress, and difficulty knowing what actually shipped without checking commits and PRs . A detailed reply framed the fix as an orchestration layer that handles task locking, retries, run tracking, and status inference from repo, CI, and deployment signals .
Why it matters: once agents can start work automatically and run concurrently, the board can stop being a trustworthy source of truth .
How to apply: treat the board as derived state from commits, PRs, tests, and deployments, and add task leases or locking so two agents do not start the same work .
2) Many messy meetings are really two modes colliding
In one product design discussion, chaotic meetings were traced to a mix of diverging questions like "What if we tried this?" and converging questions about timelines, trade-offs, and choices happening at the same time . The suggested fix was simple: diverge first, then converge .
Why it matters: when exploration and decision-making happen simultaneously, ideas get interrupted and decisions stall even when the team broadly agrees .
How to apply: explicitly label which phase the group is in, explore options first, then move into evaluation, prioritization, and decision .
3) PMs can use their product toolkit on their own careers
Jacob Warwick argues that product people should approach interviews and negotiation the way they approach product work: identify the buyer, understand their needs, ask discovery questions, and remove friction . He also says product leaders, engineers, and designers tend to negotiate worse than more extroverted roles .
Why it matters: job descriptions rarely capture the real scope, so early compensation anchors can backfire once the role expands .
How to apply: before discussing numbers, ask why you are in the room, what problem the company is trying to solve, what has already been tried, and what success looks like six months in .
Tactical Playbook
1) Make AI work visible from engineering signals
- Name the failure modes first: task contention, status drift, and ship detection gaps .
- Put an orchestrator between agent, repo, and board to manage task locking, run lifecycle, retries, and telemetry .
- Map status to observable events: PR opened -> in progress; tests passing -> ready for review; PR merged -> done; deployment succeeded -> shipped .
- Use task leases so a crashed agent releases the work back to the pool after expiry .
- Keep ownership clear: one commenter argued developers or the developer team should own agents and their output; another pointed to team views based on commits, LOC, and activity frequency as a complementary tracking layer .
Why it matters: this shifts the board from self-reported activity to observable execution .
How to apply: start with the workflows already showing duplicate agent work or manual ship checks, then instrument repo, CI, and deployment events before adding more board automation .
2) Run decision meetings in two explicit passes
- Diverge: invite alternatives, questions, and possibilities .
- Converge: switch to trade-offs, prioritization, and decision .
- Make the phase change explicit so people stop solving different problems in the same conversation .
- Judge the meeting by decision quality, not airtime: the reported benefit was better conversation quality once the shift was explicit .
Why it matters: it reduces the repeated arguments and stalled decisions described in the original post .
How to apply: if a meeting starts feeling muddy, pause and ask whether the group is still generating options or is ready to choose .
Case Studies & Lessons
1) MCP-connected boards looked automated, but not trustworthy
A small engineering team connected a coding agent directly to its project board via MCP. Automatic updates looked promising at first, but multiple agents started on the same task, progress became unreliable, and the team still had to inspect commits and PRs to confirm what shipped .
Why it matters: better automation at the card level does not solve observability if the underlying workflow assumes a human execution loop .
How to apply: if your board says "done" but your team still checks repo and deployment events manually, treat that as a signal that board state and delivery state have drifted .
2) Separating divergence and convergence improved startup discussions
In a startup context, one team found that meetings on features, strategy, and product direction felt confusing not because people strongly disagreed, but because some were expanding the option set while others were trying to narrow it. Making the shift from divergence to convergence explicit improved the quality of the conversation .
Why it matters: teams can waste time diagnosing alignment problems when the real issue is mixed cognitive modes .
How to apply: split discussion guides, agendas, or facilitation prompts into an exploration section and a decision section .
3) Role scope can move more than base pay
Warwick described cases where roles initially in the $185k-$285k range ended at $1.1M, and two roles originally comped at $600k ended at $1.1M and $1.2M after the level shifted from senior director to VP .
Why it matters: the biggest negotiation lever may be role level and scope, not just a marginal change to the offer .
How to apply: keep testing whether the company is actually hiring for a bigger role than the job description suggests before settling on a number .
Career Corner
1) Delay the number until you understand the job
Warwick says almost nobody is doing only what is written in the job description, and interviews often reveal extra scope that was not documented upfront . That is why he argues against anchoring too early: a role that starts as a senior PM search can become something closer to director-level responsibility once the team reveals the real need .
"Be you, your authentic you and apply it to what you already know. You know how to do this in product. Design it for your career."
Why it matters: early numbers become harder to unwind once the company starts using your original anchor against a larger job .
How to apply: run the interview like discovery. Ask why they are excited about you, what challenge they need solved, what has and has not worked, and what a better future looks like for the hiring manager or leadership team .
2) Use a simple pushback, and do it live
Warwick says the line below often creates about a 20% improvement across levels, and that well-run negotiations average about 40% movement . He also recommends video or in-person conversations over email so tone and body language are part of the negotiation .
"What's the chance there's a little bit more here?"
Why it matters: Lenny's summary of the episode says many product people leave at least 20% on the table because they are afraid to ask the question at all .
How to apply:
- Start with gratitude and enthusiasm for the offer
- Take time to review instead of responding immediately
- Say the package feels lighter than expected, then ask about the range or the top end
- If paperwork comes back inconsistent with prior agreement, ask "Was that a mistake?" instead of automatically splitting the difference
Tools & Resources
- The tactical playbook for getting 20-40% more comp (without sounding greedy) | Jacob Warwick — a detailed resource on discovery-led interviewing, live negotiation, and scope-based comp movement .
- How are you tracking AI-assisted work on your project boards? — a useful thread if your team is already experimenting with MCP, coding agents, or agent-to-board automation .
- TrajanCloud — one commenter described it as automated team progress tracking based on commits, LOC submitted, and activity windows across the last 24, 48 hours and 7 or 14 days .
- Why startup meetings sometimes feel chaotic even when the team agrees — a short post worth saving for facilitation and meeting design .
20VC with Harry Stebbings
andrew chen
Elena Verna
Big Ideas
1) Tab count is a fast AI opportunity filter
Andrew Chen's heuristic is simple: the number of browser tabs or alt-tabs in a workflow is a proxy for how much AI can compress that work into a single experience . His example is person/company research, which used to require LinkedIn, X, Google, notes, and Slack, but can now be collapsed into one prompt in about 10 seconds . He says the biggest opportunities sit in workflows where users alt-tab 20+ times per task, especially in sales, recruiting, research, compliance, and procurement .
Why it matters: it gives PMs a concrete way to prioritize AI work around workflow compression rather than novelty .
How to apply: audit a few high-frequency jobs your users perform, count tabs and copy-paste loops, and prioritize the flows with the most context-switching first .
"AI doesn’t need to be superintelligent to be wildly useful. it just needs to be good enough to close the tabs"
2) AI monetization needs flexibility, not pricing dogma
Elena Verna argues current monetization models are not right for every AI company because many teams are still passing through expensive LLM costs to users . She expects LLM costs to fall and says monetization will need to move toward outcomes as models commoditize . She is also explicit that subscription-only monetization is a poor fit for bursty usage; at Lovable, adding top-ups on top of subscription increased monetization capture and improved retention .
Why it matters: if usage is uneven and model costs are moving, pricing becomes part of product strategy, not a one-time packaging decision .
How to apply: test ad hoc purchases alongside subscription for bursty use cases, and make pricing changes operationally easy instead of treating them as annual events .
3) For productivity tools, meaningful frequency beats intensity
Verna frames activation around product engagement: define the aha moment, the steps to reach it, and the early habit loops that bring users back . She argues intensity can be an anti-metric for simple productivity tools, because more time may mean users are stuck, while daily or weekly usage sits in the habitual zone and monthly usage drifts into the forgettable zone . She also warns against login-based metrics and prefers value-creating actions instead .
Why it matters: teams often mistake activity for value .
How to apply: choose one or two actions that clearly represent user value, then track repeat frequency on a daily or weekly basis rather than visits or logins .
4) "Minimum lovable" is part of the product bar
Verna argues teams should aim for a minimum lovable product in every feature, because software is increasingly judged by the emotion, trust, and connection it creates, not just by basic functionality . In her framing, the progression is: it works, users trust it, then users connect with it .
Why it matters: she argues personality and emotional connection are becoming a minimum bar to kickstart growth .
How to apply: during reviews, evaluate not just whether a feature works, but whether it creates trust and a recognizable product feel .
Tactical Playbook
1) Run a tab-count audit before you scope an AI feature
Use this sequence:
- List the tabs, docs, and tools a user opens to finish one job; Chen's core idea is that tab count signals compressibility .
- Mark every copy-paste handoff; Chen says eliminating 6+ tabs and a copy-paste loop is immediately useful to users .
- Prioritize jobs with extreme context switching; he highlights workflows with 20+ alt-tabs per task .
- Prototype the whole flow as one AI-native experience; his example collapses LinkedIn, X, Google, notes, and Slack into a single prompt-driven workflow .
Why it matters: this turns abstract AI brainstorming into a concrete prioritization method .
2) Redefine activation around value, not logins
A practical setup from Verna's framework:
- Write down the user's aha moment and the steps required to get there .
- Decide which action proves value; at Lovable, examples include building an app or receiving traffic on a published app .
- Track whether that action repeats daily or weekly, because that is the habitual zone Verna wants to see .
- Treat raw logins as a vanity metric and be careful with time-spent metrics if your product is supposed to feel simple .
Why it matters: it aligns your core metric with value creation instead of mere presence .
3) Use a two-speed launch system
Lovable's operating rhythm suggests a clear playbook:
- Ship customer-facing improvements daily, not just bug fixes .
- Let the people closest to the work share releases; Lovable encourages engineers to post launches socially and then "beeswarms" those posts for amplification .
- Reserve major narrative effort for bundled launches every 1-2 months, when multiple capabilities add up to a story and a step-function change .
- Treat ongoing visibility as part of retention and resurrection, not just acquisition; Verna says the constant noise brings people back because the product feels alive and evolving .
Why it matters: it separates release velocity from storytelling cadence without losing either .
4) Treat freemium as a marketing channel with its own metric
Verna's framing is unusually direct: a free user has value if they get delighted and then market the product on your behalf . Lovable tracks this with a "lovable score" that measures how often users refer the product to someone else .
How to apply:
- Define what a successful free experience looks like before conversion .
- Track referral behavior explicitly, not just free-to-paid conversion .
- Protect the parts of the free experience most likely to create delight and sharing .
Why it matters: it gives PMs a clearer way to value free usage in products where word of mouth matters .
Case Studies & Lessons
1) Lovable turned shipping cadence into retention infrastructure
At Lovable, engineering releases improvements every day, employees post about those releases on social, the company amplifies them internally, and marketing concentrates on bigger tier-one launches every 1-2 months . Verna says that constant noise is part of retention and resurrection because users feel the product is "living, breathing" and worth revisiting .
Key takeaway: if your category is moving quickly, consistent visible improvement can be part of the product experience, not just a marketing layer .
2) A Romanian accountant SaaS validated the workflow before polishing the brand
One founder started with a very specific problem: accountants were spending 3-5 hours each month chasing invoices, bank statements, and receipts over WhatsApp . Validation was lightweight and direct: they messaged about 50 Romanian accountants on WhatsApp, got repeated confirmation, and built the MVP in 2 weeks . The product itself stayed close to the workflow: each client gets a personal upload link with no account or onboarding, and the accountant sees a dashboard showing who sent documents and who did not . On day one, the product saw 172 visitors, 18 signup reaches, 2 registered accounts, 2 Stripe checkout visits, and a 59% bounce rate .
A commenter highlighted the strongest decision: the product started from a real workflow rather than "cool tech," and recommended 5-10 Zoom walkthroughs of actual month-end work to surface edge cases before chasing more traffic . The founder's own lesson was that niche, non-English B2B can be slow, but each signup is more likely to be a real customer than a curiosity click .
Key takeaway: tight workflow validation plus narrow positioning can produce higher-signal early learning than broad top-of-funnel traffic .
3) Lovable used top-ups to fit bursty AI usage
Verna says Lovable introduced ad hoc top-ups on top of subscription and the response was "absolutely wild" . Her claim is that this kind of purchase adds incrementally rather than cannibalizing recurring revenue, and that retention improves when users get this flexibility .
Key takeaway: when usage comes in bursts, a hybrid pricing model can capture more value than subscription alone .
Career Corner
1) Compare roles by daily work loop, not just by title
In one Product Management community thread, the choice was between an IT Requirements Engineer role in IAM and a Product Owner role in another area . The IT Requirements Engineer description centered on gathering requirements for identity and access management systems and translating business needs into technical specifications , while the Product Owner role centered on stakeholder work, product requirements, backlog prioritization, and guiding development teams .
Why it matters: the titles sound adjacent, but the day-to-day work is different .
How to apply: evaluate career options against growth, compensation, job security, work-life balance, domain interest, longevity, and pay—not title prestige alone .
2) Use community signals carefully when assessing AI exposure
In the same thread, one commenter said an IT Requirements Engineer sounds closer to a Business Analyst role . Another suggested IAM may be more repetitive, but also less likely to be handed over to AI than a Product Owner role .
Why it matters: job security discussions are already being filtered through assumptions about which work AI will and will not absorb .
How to apply: treat this as community signal, not settled fact, and stress-test any role by asking which parts of the job are domain-heavy, stakeholder-heavy, or easy to standardize .
3) Pricing and engagement design are becoming stronger PM differentiators in AI products
Across Verna's interview, two recurring responsibilities stand out: defining meaningful engagement signals instead of vanity metrics and building the infrastructure to test monetization model changes quickly as AI costs and economics shift .
Why it matters: these are product problems that cannot be solved by feature delivery alone .
How to apply: if you want to broaden your scope, volunteer for activation metric design or pricing and packaging experiments rather than limiting yourself to backlog management .
Tools & Resources
- Andrew Chen's tab-count post — a compact framework for identifying AI opportunities by counting tabs, alt-tabs, and copy-paste loops in a workflow .
- Tab-count worksheet — create a simple table with columns for job-to-be-done, tabs opened, copy-paste handoffs, and whether the flow could be collapsed into one AI-native experience .
- Elena Verna: How Lovable Launches Product & Hacks Social to Go Viral — useful for PMs working on launch cadence, activation metrics, freemium, and AI monetization design .
- Meaningful action scorecard — document the aha moment, the action that proves value, the target frequency, and the anti-metric you want to avoid, such as logins or excessive time spent .
- Romanian accountant workflow-first case study — a useful teardown of direct problem validation, narrow MVP scope, simple pricing, and day-one funnel metrics in a niche B2B market .
scott belsky
Tony Fadell
John Cutler
Big Ideas
1) Taste at speed is becoming a real PM advantage
Aakash Gupta’s framing is that the emerging skill is taste at speed: the ability to evaluate working software quickly, kill most of it, and ship the survivors . In that model, AI does not just speed building; it changes the bottleneck from can we build it to should we ship it. The workflow contrast is sharp: the older path runs from idea to PRD to design to engineering to QA to ship in 8-12 weeks, while the AI-era loop described here is idea → 5 prototypes → evaluate → kill 4 → spec the survivor → ship in 1-2 weeks.
“The spec didn’t disappear. It moved from step 2 to step 6.”
- Why it matters: the leverage comes from filtering, not from shipping everything faster; the cited 80% kill rate is the point .
- How to apply: for ambiguous problems, require multiple working prototypes, review them against empathy, simulation, strategy, taste, and creative execution, then spec only the winner .
2) Context and rituals are becoming the real operating leverage
John Cutler argues that the first place to inspect a product organization is its rituals: the daily and weekly interactions people have through meetings, dashboards, Slack, and other tools . The weak spot is often the layer between front lines and leadership, where information has to move fluidly across the organization . He also warns that AI makes it easy to generate documentation, but more documentation does not create intentionality . Leah Tharin makes the complementary product point: context is the real value, not the model, and a jobs lens like “I want to listen to music on the go” opens a much broader solution space than a demographic profile .
“Frameworks are models and all models are useful but wrong.”
- Why it matters: without shared context and deliberate rituals, faster output just creates faster drift. Cutler also points to a collective memory problem where teams keep re-documenting old issues because context is co-created over time .
- How to apply: build living context around recurring customer challenges, not just one-off deliverables, and deliberately design how information moves up, down, and across the org .
3) AI speed makes alignment and work-shape clarity non-negotiable
Scott Belsky argues AI creates a stronger case for talent density and far more alignment than usual because teams can now move very quickly in the wrong direction . Tony Fadell’s reminder is simpler: knowing the destination helps people self-prioritize and decide what and how to build . Cutler adds that most organizations have parallel motions at once—some work is large, coordinated, and governance-heavy, while other work is highly iterative—and pretending everything should run through one process is damaging .
- Why it matters: more prototyping capacity increases the cost of fuzzy goals and one-size-fits-all process.
- How to apply: make the destination explicit, separate high-coordination work from iterative work, and align your operating rhythm to each motion rather than forcing one template across both .
Tactical Playbook
1) Run a prototype-first decision loop
- Start with multiple working options. The model here is five fast prototypes, not one polished plan .
- Evaluate against five lenses. Check empathy, simulation, strategy, taste, and creative execution while looking at working software .
- Kill aggressively. An 80% kill rate is framed here as a feature, not a failure .
- Write the spec after you have a winner. In this flow, the spec follows the prototype, not the reverse .
- Keep a human gate before production. Anthropic still requires an engineer to approve changes before anything goes live .
- Why it matters: it compresses false certainty early and moves discussion onto concrete artifacts.
- How to apply: pilot this on one fuzzy feature before greenlighting a full PRD or roadmap commitment.
2) Use challenge memory in discovery
- State the customer challenge in job terms. Prefer “listen to music on the go” over a demographic profile .
- Capture the surrounding context. Tharin’s argument is to build memory around the challenge, not just a single job statement .
- Feed better context into the system. More correct context improves the odds of better decisions and of spotting bad data early .
- Revisit old knowledge before reopening old problems. Cutler points to teams repeatedly documenting the same issue because collective memory is weak .
- Why it matters: it broadens solution space and reduces rediscovery waste.
- How to apply: keep one artifact per problem area with the job to be done, prior evidence, edge cases, and what the team already learned.
3) Repair the operating system through rituals
- Map the current rituals. Start with the daily and weekly interactions people actually have, not the official process deck .
- Design information cadence intentionally. Cutler’s advice is to get information moving up, down, and across the org deliberately; documentation alone is not enough .
- Name parallel motions. Separate large, coordinated efforts from fast iterative streams so each gets the right governance .
- Label relationships honestly. Do not force work into fake linear hierarchies when frontline teams can sometimes move business metrics directly .
- Treat new habits as repeated experiments. One kickoff meeting or spreadsheet rarely survives without sustained reps .
- Why it matters: many execution problems are really information-flow and habit-formation problems.
- How to apply: redesign one recurring meeting and one update channel before adding another template or framework.
Case Studies & Lessons
1) Anthropic’s Claude Code workflow pairs extreme speed with explicit review gates
Boris Cherny is described as shipping 20-30 PRs a day using five parallel Claude instances, with a third of his code potentially started from the iOS app and 100% of his own code written with Claude Code . This sits inside an unusually technical culture where everyone shares the title Member of Technical Staff and PMs, designers, data scientists, and even finance code . Company-wide, Claude Code writes about 80% of code, and productivity per engineer is cited as up 200% since launch even as Anthropic tripled headcount . The process is not review-free: every PR is first reviewed by Claude Code, which catches about 80% of bugs, and a human engineer still does the second pass and approves anything before production .
- Lesson: the interesting move is not just AI-assisted coding; it is AI-assisted coding plus automated review plus human approval.
- Boundary condition: Gupta notes this setup fits a small, senior team with deep shared context, where the product is the AI tool itself, though the prototype-first discipline can translate beyond Anthropic .
2) Anthropic’s product work uses volume to improve judgment, not just output
The reported iteration counts are unusually high: agent teams went through “probably hundreds of versions” before shipping; condensed file view saw about 30 prototypes followed by a month of internal dogfooding; the terminal spinner had roughly 50-100 iterations, with about 80% not shipping. One example is plugins: Daisy reportedly used a swarm of a couple hundred agents over a weekend, producing about 100 tasks and an implementation that became “pretty much the version of plugins that we shipped” .
“And it’s a filtering function, not an acceleration function. The 80% kill rate is the whole point.”
- Lesson: faster tools matter most when the team is willing to discard most versions.
3) PM hiring loops are being redesigned because old homework signals got cheaper to fake
Andrew Chen says homework has become a common interview step for PMs and other knowledge roles because it can surface real work output . His update is that, in recent weeks, these responses have been flooded with “AI slop”—long, meandering documents instead of a short, high-signal point of view . His proposed fixes are a recorded presentation, where candidates sign off on what they wrote and can be questioned later, and a true work trial reserved for the end of the funnel . He also cautions that overly structured formats can alienate top-end talent .
- Lesson: when a signal becomes easy to mass-produce, move evaluation closer to live reasoning or real work.
Career Corner
1) Build taste reps now, before the gap compounds
The clearest career signal in this set is volume of evaluation. A PM who reviews 15 prototypes a week builds judgment faster than one reviewing one spec a month; over six months, Gupta argues that becomes a widening taste gap and then a career gap . He also says PMs who start building these reps now will have a massive head start .
- Why it matters: the compounding advantage comes from pattern-matching on working software.
- How to apply: keep a log of prototypes you killed, what you learned, and which evaluation lens changed your mind.
2) Nontraditional backgrounds still map well to core PM work
In community discussion, PMs pointed to several transferable skills from psychology, behavior analysis, and research backgrounds: observing people use a product, noticing nonverbal cues, asking better questions, and using surveys and statistics to understand how broad a problem is . One experienced PM described these as among the most important things PMs do , while a former cognitive neuroscience researcher said they successfully switched into product and enjoy the work . Another poster said the technical side of a specific product or industry looked like the stimulating challenge, not a blocker .
- Why it matters: PM hiring still rewards human observation and research judgment, not just AI fluency.
- How to apply: translate prior work into PM language: user observation, hypothesis formation, research design, and statistical interpretation.
3) Candidates and hiring managers both need an AI-era interview upgrade
For candidates, the implication of Chen’s post is straightforward: concise thinking and live defense now matter more than polished take-homes alone . For hiring managers, recorded walkthroughs can scale better than full work trials, while the most realistic evaluation should still happen late in the funnel . Chen’s warning against overly structured formats is also a reminder not to optimize the process so tightly that you filter out strong candidates .
- Why it matters: interview loops are becoming tests of reasoning ownership, not document generation.
- How to apply: if you’re a candidate, practice explaining trade-offs live; if you’re hiring, keep one live defense step in the loop.
Tools & Resources
- There’s a New PM Skill. It’s Called Taste at Speed — the clearest source here on prototype-first PM work, spec-after-prototype sequencing, and Anthropic’s reported metrics
- Claude Code / Cowork — examples of AI-assisted building workflows to watch; Cowork is described as a full product for non-engineers built in about 10 days and part of Anthropic’s push to bring this style to non-engineers
- The missing layer in government tech? A real operating system. — John Cutler on rituals, information flow, and why a real operating system is more than a framework
- 10 print Hello World — Leah Tharin on context as value, jobs framing, and building memory around customer challenges
- Second Axis — a community example of tools targeting the “messy middle” between idea and execution by generating docs, tickets, and edge cases from a feature idea; treat this as a category to watch rather than a vetted recommendation
Aakash Gupta
Product Management
Lenny Rachitsky
Big Ideas
1) The scarce skill shifted from building to deciding
Creation is easier; judgment is now the differentiator. Hiten Shah frames the new bottleneck as deciding what should exist, who it is for, what problem it solves, and whether it should be built at all . Rekhi makes the operational PM version: AI has sped up delivery enough that discovery and design can become the constraint, increasing the risk of feature factories if teams stop validating with customers .
The bottleneck moved.
- Why it matters: faster output does not automatically create better products.
- How to apply: raise the bar on problem selection before prompting or prototyping starts, and require a clear statement of what is being built and why .
2) AI should speed discovery, not replace product intuition
Rekhi's core caution is to stay in the loop. PMs still need to read customer feedback, sample interview recordings, and digest research themselves to build intuition and empathy . In his experience, end-to-end workflows that ask AI to synthesize feedback, invent the feature, and write the spec produce poor results; the better pattern is to let AI surface pain points and summaries, then apply human judgment to the solution .
- Why it matters: many PM decisions still happen without full information, so intuition remains a real operating asset .
- How to apply: ask for verbatims alongside summaries, and keep solutioning as a human step .
3) Discovery is becoming a continuous operating system
The notable change in Rekhi's workflow is not one tool but a connected system: continuous surveys, feedback rivers, AI-generated interview guides, interview synthesis, AI-moderated interviews, functional prototypes with analytics, and natural-language data analysis . He describes moving from quarterly NPS work with a marketing team to continuous collection and weekly automated reporting .
- Why it matters: PMs can keep pace with faster engineering without waiting on long research or analytics queues.
- How to apply: replace one-off studies with a recurring loop: collect, synthesize, validate, instrument, and revisit.
4) Some products may need to design for agent-to-agent use
Aakash Gupta argues that many teams still optimize for a single interaction model - human opens app, human types, AI responds - while the next surface may be agent-to-agent, where a user's assistant contacts the product directly . His suggested PM questions shift to what the product exposes to other agents, what permissions it grants, and where it escalates when an AI cannot answer .
- Why it matters: onboarding flows, nudges, and empty states assume a human is watching .
- How to apply: inventory which actions, permissions, and escalation paths would still work if another agent were the caller .
Tactical Playbook
1) Build a continuous signal stack
- Run surveys continuously. In Rekhi's example, Claude calculated an overall NPS of 41 from 2,600 responses, showed 58% promoters, visualized monthly trends, and supported deeper segmentation work .
- Automate the reporting layer. His agent checks the CSV, runs numerical analysis and verbatim themes, then generates both an HTML report and a Gamma presentation .
- Add a feedback river. Tools in this category pull from sources such as App Store, Google Play, G2, Zendesk, cancellation surveys, and NPS, then group feedback into themes with trend lines and counts . In his example, one theme showed 138 complaints about Evernote import issues .
- Keep the customer voice attached. Review exact verbatims and use the linked identities or contact records when you need follow-up conversations .
Why it matters: this turns customer signal from a quarterly project into a weekly operating rhythm.
Start here: automate one recurring survey report, then add one support or review source next.
2) Upgrade interview work end to end
- Seed AI with strong interviewing principles. Rekhi uses a summary of actionable best practices from The Mom Test plus a research brief to generate better interview guides .
- Transcribe and summarize recordings against a template. His workflow extracts takeaways, pain points, workflow/tools, feature requests, and direct quotes from each interview .
- Ask for cross-interview patterns. The same workflow can cluster themes across a batch and count how often pain points appear across 10 interviews .
- Use AI-moderated interviews when speed matters. Rekhi says these tools can run async concept interviews, probe with follow-up questions, and return summarized results by the next morning .
- Still watch some interviews yourself. He explicitly keeps sampling recordings to build customer and product intuition .
Why it matters: scripting, transcription, and synthesis compress from hours or days into an overnight workflow.
Start here: standardize one interview template before automating anything else.
3) Use prototypes as research instruments
- Build a functional prototype, not just a mockup. Rekhi used Bolt to create a working Ask AI concept .
- Instrument it like a real product. He adds in-product surveys, retention tracking, session replay, and heatmaps through analytics tooling such as PostHog .
- Use synthetic users for fast usability feedback. In his example, synthetic feedback surfaced a hidden entry point, 5-20 second waits, missing example prompts, and the lack of multi-note querying .
- Then let real behavior settle design choices. Heatmaps showed users clicking the Ask AI button in the top right, not the bottom-right floating button .
- Do not confuse synthetic feedback with product-market fit. Rekhi says it is useful for usability-style feedback, not for determining product-market fit .
Why it matters: this gives teams pre-launch signal on usability and return behavior that mockups cannot provide.
Start here: add one survey question and one heatmap before you send a prototype out.
4) Make analytics self-serve - carefully
- Connect AI to your data safely. Rekhi uses MCP or a database dump with read-only access so the model can inspect schema and query data .
- Ask questions in plain English. His workflow reads tables and columns, writes SQL, groups results, and returns charts or recurring dashboards from natural-language prompts .
- Teach the system with examples. His advice is to add real question-and-SQL pairs because the model learns patterns from examples well .
- Document schema quirks explicitly. A simple instruction such as using
canonical_sourceinstead ofsourceimproved accuracy in his example . - Audit early, then share. Rekhi says he audited a few dozen queries to build confidence, and recommends sharing the Claude Project so the whole team benefits from the learned context .
Why it matters: PMs can answer more of their own product questions without waiting on a data queue.
Start here: teach one high-value dataset first; do not trust a zero-context setup .
Case Studies & Lessons
1) Tesla optimized for the moment of doubt
Tesla's Supercharger spacing is described here as a product decision, not a simple infrastructure rule. The cited explanation is that chargers were placed around where drivers typically hit 15-20% battery - when range anxiety begins and people start doing mental math - rather than at uniform intervals . The deeper move was optimizing for the user's emotional state at a key moment, even though that is harder to put in a dashboard than coverage or utilization .
- Why it worked: it reduced the moment where users begin to doubt the journey .
- How to apply: identify the point in your journey where users start calculating, hesitating, or seeking reassurance, and design around that moment first .
2) Notejoy's Ask AI concept shows layered discovery in practice
Synthetic users first surfaced usability issues in Rekhi's Ask AI prototype: the entry point was not obvious before a note was selected, response time felt too slow, example prompts were missing, and cross-note querying was absent . After that, prototype instrumentation showed that users preferred the top-right Ask AI button over the bottom-right alternative .
- Why it worked: different methods answered different questions - synthetic users for early usability friction, real usage for design preference.
- How to apply: use simulated feedback to narrow what to test, then use real telemetry to settle decisions .
3) LocalMind was interesting, but not frequent enough
Lenny Rachitsky describes LocalMind as an app on top of Foursquare that let people ask questions of users checked into places around the world, such as whether there was a long line at a location . His conclusion was that it solved a real problem, but only occasionally, which made it hard to support as a standalone business .
- Why it matters: novelty and utility do not guarantee repeated use.
- How to apply: pressure-test frequency of need early, not just whether the concept is clever or helpful .
Career Corner
1) Strong PM narratives still center on impact, judgment, and alignment
Lenny's shorthand for the role is impact, collaboration, judgment, and alignment, with coordination close behind . He also frames the PM job as delivering business impact by prioritizing and solving the most impactful business problems, while thinking the way a CEO would think about the product's success .
My words are impact, collaboration, judgment, alignment.
How to use it: in interviews and promotion narratives, explain the business problem, the decision you drove, the alignment work, and the outcome.
2) AI hiring signals are moving from prompts to systems
Aakash Gupta says PM interviews now commonly ask how candidates use AI in their workflow, and the answer interviewers want is a system, not a generic tool mention . His examples of strong setups include custom GPTs for PRD drafts, Claude Projects loaded with company design principles, Gemini Gems for competitive analysis, and Claude Code with context; he estimates roughly two hours of setup for 5+ hours of weekly return .
How to use it: build one reusable system for writing, one for analysis, and one for recurring tasks - and be ready to explain inputs, outputs, and guardrails .
3) Protect the human skills AI can erode
Rekhi's guidance is consistent across research and analytics: use AI to speed production, but keep empathy, pattern recognition, and solution judgment with the PM . He continues to watch interviews and read research himself even after automating much of the process .
How to use it: keep a weekly habit of reviewing raw customer material, not just summaries.
4) Classic PM strengths still compound
Lenny points to organization, a high bar for quality, and succinct communication as the traits that carried over from PM work into creator work . He also describes iterating heavily rather than shipping first drafts, sometimes revising a newsletter post dozens of times .
How to use it: treat memos, specs, and stakeholder updates the way you treat product flows - refine them until the point is easy to grasp.
Tools & Resources
- Feedback rivers: Reforge Insights, Interpret, Craftful, Birdie, Miro Insights, Unwrap, and Productboard for aggregating reviews, support, and survey data into themes and trends .
- Interview synthesis: NotebookLM and Claude for transcribing recordings, summarizing interviews, and finding patterns across batches .
- AI-moderated interviews: Reforge, Listen, Outset, and Maze for asynchronous concept testing with dynamic follow-up questions .
- Synthetic user testing: Reforge and Simile for persona-based usability feedback on prototypes .
- Prototype stack: Bolt for fast functional prototypes; PostHog for surveys, retention, heatmaps, and session replay .
- Analytics stack: Claude Projects or Claude Code plus MCP for natural-language SQL and dashboards, especially when paired with example queries and shared instructions .
- Reporting output: Gamma for auto-generated presentations from recurring analysis workflows .
- Interview prompt seed:The Mom Test as a best-practices source for AI-generated customer interview guides .
Hiten Shah
Melissa Perri
Big Ideas
1) PM workflow is moving from sequential handoffs to parallel, compounding systems
Aakash Gupta argues that the fastest PMs are discarding sequential work: they plan one feature while another builds, iterate on UI while the data layer assembles, and run multiple workstreams in parallel because the tooling now supports it . Dave Killeen's Claude Code setup shows the operational version of that shift: a single /dailyplan command pulls calendar, CRM, meeting notes, LinkedIn messages, YouTube transcripts, newsletters, and quarterly goals into one page , while hooks inject priorities, preferences, and past mistakes every time a new session starts .
“Skills are what you do. MCP is how you connect. Hooks are how you compound.”
- Why it matters: less tab switching, faster context loading, and more room for judgment and experimentation .
- How to apply: start with one repeatable command, one connected system, and one persistent file structure; treat AI as infrastructure, not just a faster chat window .
2) Strong problem framing means moving across layers, not polishing one sentence
The Beautiful Mess argues there is no perfect articulation of a problem. Teams need to move between multiple layers: what question to ask, what the term means, the surrounding environment, what is happening now, why it happens, why it matters, what better futures exist, and how success should be evaluated . Product leadership, in this framing, is not just defining the problem and handing it off; it is creating conditions for people to engage it from several elevations at once .
“The trick is to dance between layers.”
- Why it matters: vague statements like “too slow” or “too heavy” have little diagnostic value .
- How to apply: force teams to articulate behavior, causes, stakes, and measurement separately before jumping to solutions .
3) The best growth teams hunt friction, then measure whether relationships deepen
Brian Hale's contrast is consistent across seven dimensions: excellent growth teams own the full user journey, remove blockages, maximize learning rate rather than experiment count, practice product growth instead of growth hacking, optimize relationships rather than raw activity, compound wins, and hire people who diagnose readiness before joining .
“Excellent growth teams relentlessly do the most important thing, even when it’s unglamorous, non-obvious, or uncomfortable.”
Robby Stein adds a measurement lens: early PMF looks like flat or J-curve retention through day 30/60/90, followed by organic week-over-week user and usage growth and rising intensity of use over time .
- Why it matters: activity can rise because power users do more, while new or hesitant users stay stuck .
- How to apply: look for where users hesitate, then track people-count metrics, cohort retention, and deeper usage instead of celebrating experiment volume alone .
4) Large product orgs are getting clearer about alignment: fund capacity, not just projects
Across Vanguard, Chase, and Affirm, the operating model is similar: align teams to business outcomes through OKRs and portfolio reviews , fund product/design/data/engineering capacity at the domain level and let teams prioritize within that capacity , rebalance selectively without constant budget swings , use weekly forums for problem validation, sequencing, and dependencies , and embed legal/compliance early when needed .
“Empowerment without alignment is chaos.”
- Why it matters: this creates room for local judgment without losing strategic coherence .
- How to apply: review business, product, and quality/engineering metrics separately, run weekly decision forums, and make capacity allocation a portfolio choice rather than a feature-by-feature fight .
5) PMs need a money language, not just a product language
Rich Mironov's “money stories” framework flips user stories for executive audiences: the question is not how a feature works, but roughly how much revenue or retained revenue a set of work could return . His advice is to use order-of-magnitude ranges, sort ideas by digit count, label roadmap swim lanes with value ranges, and avoid mythical feature-level ROI claims .
- Why it matters: executives often decide at the level of business impact, not implementation detail .
- How to apply: tell simple upsell or churn stories, pressure-test them with sales, marketing, or finance, and use team-level “earning your keep” logic instead of pretending every ticket has a precise ROI .
Tactical Playbook
1) Build a minimal PM operating system
- Create one command for your morning workflow. Dave's version checks whether digests already ran, pulls structured data through MCP, and outputs priorities, account context, and suggested Slack messages .
- Connect one tool first. The guidance here is to start with calendar, find the API/MCP/CLI docs, give Claude the docs plus an API key, and let it build the server .
- Separate skills, MCP, and hooks correctly: use skills for flexible judgment, MCP for deterministic integrations, and hooks for session-start context .
- Keep the knowledge base alive. Use stakeholder/project/company pages, a mistakes file, working preferences, and a short Claude.MD map that points to deeper files .
- For product execution, let the system move from backlog to PRD to Kanban, but keep Dave's own caveat in mind: AI PRDs are strong first drafts, yet work use still needs tighter commercial context and metrics .
- Why it matters: it turns scattered PM admin into a reusable operating loop .
- Apply it this week: clone DEX and run
/setup, or create your own first command and one persistent project file .
2) Diagnose a fuzzy problem before you prioritize it
- Start with the question you are actually trying to answer .
- Define key terms .
- Describe what users are doing today and the surrounding environment .
- State the most plausible causes .
- Explain why the problem matters and what improving it would make possible .
- Generate alternatives and define how you will know it is better .
- Why it matters: the initiative-creation example shows that “too heavy” only becomes useful once you separate ambiguity, missing information, and definition confusion from the downstream planning impact .
- Apply it: require teams to bring at least one sentence per layer before solution reviews.
3) Run prioritization with three lenses: customer, business, and build reality
- Anchor on customer evidence, even if discovery is fast .
- Make the business unlock explicit: revenue, complaints, expense, or another OKR .
- Expose the speed-versus-complexity trade-off instead of hiding it .
- Bring decisions into weekly cross-functional forums for sequencing, dependency, and partnership questions .
- In regulated or high-risk contexts, pull legal/compliance into ideation rather than near launch .
- Monitor three dashboard buckets: business, product, and quality/engineering .
- Why it matters: this keeps prioritization grounded without collapsing into either feature prescription or vague empowerment .
- Apply it: make every roadmap discussion show the customer signal, the business unlock, and the delivery trade-off on one page.
4) Translate roadmap bets into money stories
- Use a range, not false precision, because executive trade-offs are usually order-of-magnitude decisions .
- Sort ideas by digit count so you can ignore three- and four-digit requests early .
- For upsell work, multiply target customers × price delta × expected upgrade rate .
- For retention work, multiply customers at risk × annual value × expected churn reduction .
- Put the value range on the roadmap swim lane, not on every ticket, and require an equivalent revenue case before disrupting the lane .
- Sanity-check the numbers with sales, marketing, or a finance partner .
- Why it matters: it converts stakeholder fights into business trade-offs instead of backlog theater .
5) Verify PMF before you scale
- Start with 10 trusted testers; if 10 people do not like it, scale will not fix that .
- Move to a small opt-in launch without pushing marketing hard .
- Look for J-curve or flat retention across day 0, 30, 60, and 90 .
- Then ask whether users and usage are growing week over week without unnatural intervention .
- Watch engagement depth and follow-up behavior, not just shallow reach .
- Why it matters: it separates genuine product pull from novelty or forced distribution.
Case Studies & Lessons
1) DoorDash: the problem was communication, not the offer
DoorDash had users hesitating because they feared delivery fees even when a zero-delivery-fee offer existed . The growth win came from making the existing value impossible to miss, not inventing a new incentive.
- Why it matters: some conversion problems are explanation problems.
- Apply it: audit places where users hesitate because they do not understand an existing benefit .
2) Instagram Close Friends: fix the feedback loop, then simplify everything around it
Close Friends initially failed because the experience was confusing, poorly translated in some regions, and lacked a feedback loop . It worked only after Instagram made it a Stories behavior, changed the name, added the green ring indicator, and made it easy to create a 20-30 person list so replies would actually happen .
- Why it matters: the smallest viable experience is not always the one with the fewest UI elements; it is the one whose behavior makes sense to users.
- Apply it: map the desired social or functional loop first, then strip away everything that does not reinforce it.
3) Instagram Reels: similar surfaces can hide incompatible incentives
Instagram first launched Reels as a Stories-adjacent experience in Brazil because the surface looked similar: full-screen video . That failed because creators wanted persistence and the possibility of going viral, not content that disappeared in a day .
- Why it matters: product teams often overfit to interface similarity and underweight creator or user incentives.
- Apply it: test whether the surface reinforces the creator or user payoff, not just whether the format looks familiar.
4) Google AI search: strong product fit can still create ecosystem trade-offs
Robby Stein says AI Mode worked for Google because users already came to Google with hard informational tasks and showed latent demand by explicitly adding “AI” to queries after AI Overviews launched . In a separate analysis, Hiten Shah argues Google also chose to own zero-click behavior, citing organic CTR down 61%, paid CTR down 68%, and zero-click news searches up from 56% to 69% over 18 months . He also cites search ad spend up 9% year over year versus 4% click growth .
- Why it matters: the same move can look like better user alignment from inside the product and ecosystem compression from outside it.
- Apply it: when you add AI to a core surface, evaluate both native task fit and the second-order effects on partners, monetization, and traffic flows.
5) Vanguard: outcome framing outperformed feature prescription
Vanguard describes giving teams outcome goals tied to helping investors take the next best action, rather than prescribing features . In one financial wellness experience, the opening survey reached nearly 80% completion versus a cited 10-20% benchmark for comparable retirement-industry tools .
- Why it matters: teams get smarter when they own the problem, not just the brief.
- Apply it: define the customer outcome, then let the team design the experience that gets there.
Career Corner
1) The PM job market is splitting between manual operators and system builders
Aakash Gupta frames two groups: PMs still writing PRDs and updates manually, and PMs with tuned Claude.MD files, custom skills, and PRD writers that generate most of a shipping-ready doc in minutes . The second group reinvests time into users, engineering relationships, and strategy .
- Why it matters: the compounding advantage comes from systems, not one-off prompts.
- How to apply: automate one recurring artifact per week and spend the recovered time on discovery or stakeholder work.
2) Build your promotion case continuously
Dave's career MCP listens for evidence of skills, feedback, and outcomes, maps gaps against goals, and calculates promotion readiness before review season arrives .
- Why it matters: most PMs track the backlog better than their own growth .
- How to apply: keep an evidence log by skill and outcome, review gaps weekly, and enter performance reviews with assembled proof rather than memory.
3) Financial fluency is becoming part of PM credibility
Rich Mironov argues PMs are not being trained to talk about money, even though the basics are often just P&L and cost accounting 101 . His suggestion: learn how the company makes money, what one more unit contributes, and what the product team costs .
- Why it matters: you cannot defend priorities or funding well if you cannot tell a money story.
- How to apply: find the finance partner in your company, ask basic questions, and learn the economics of your product line .
4) Owner mindset still matters, but leaders should spend detail-time on the few bets that matter most
Robby Stein describes successful builders as people with a strong internal locus of control who fully own outcomes . He also says leaders should pick a small number of projects where the upside is five-to-ten-year value and where their direct intervention is uniquely useful, then co-create intensely until the work is on track . Chase describes a similar expectation: product teams should think end to end and understand the P&L with their business partners .
- Why it matters: high agency without focus becomes thrash.
- How to apply: own results end to end, but reserve deep involvement for the few bets where leadership detail changes the outcome.
Tools & Resources
- DEX — open-source PM OS;
/setupscaffolds the system around your role and goals in minutes . - This CPO Uses Claude Code to Run his Entire Work Life | Dave Killeen, Field CPO @ Pendo — practical walkthrough of daily planning, backlog-to-PRD flow, and career evidence capture .
- Claude Code setup — useful if you want a Claude.MD structure built around progressive disclosure .
- TBM 410: Dancing With Problems — compact framework for turning fuzzy problem statements into decision-quality framing .
- What Excellent Growth Teams See That Others Miss — Brian Hale's seven contrasts between okay and excellent growth teams .
- How to communicate the value of your product work — Rich Mironov on money stories, ROI ranges, and executive communication .
- Google VP of Product on The Future of Search and AI Mode — Robby Stein on PMF metrics, testing progression, and AI Mode decisions .
- Episode 264: Product at Scale Inside the World’s Largest Financial Institutions — operating model examples for outcome alignment, capacity funding, and metrics design .