We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track

PM Daily Digest
by avergin 79 sources
Curates essential product management insights including frameworks, best practices, case studies, and career advice from leading PM voices and publications



Big Ideas
-
Trust is the moat in the AI era
- Why it matters: When AI can fake anything, customers ask one question: can they trust what you built? Trust compounds; growth hacks fade 43 AI can fake anything. Screenshots. Testimonials. Entire products. Which means one question will matter more than any other: can people trust what you built?https://x.com/hnshah/status/1963610001542611098 6 Most growth hacks are sugar highs. They spike, then fade. Trust compounds. Each accurate experience, each honest interaction, makes your product harder to churn from, easier to recommend, and more valuable over time. Acquisition fades. Virality fades. Revenue spikes and dips. Trust keeps compounding.https://x.com/hnshah/status/1963610001542611098 . Treat accuracy and honesty as core product features; they reduce churn and increase referrals 37 Most growth hacks are sugar highs. They spike, then fade. Trust compounds. Each accurate experience, each honest interaction, makes your product harder to churn from, easier to recommend, and more valuable over time.https://x.com/hnshah/status/1963610001542611098 39 Trust became our growth engine. Agencies used Crazy Egg heatmaps in client presentations because they knew they could rely on them. Users recommended us because they were confident we wouldn’t embarrass them.https://x.com/hnshah/status/1963610001542611098 .
-
How to apply:
- Make correctness a top KPI (e.g., event accuracy, reconciliation error rate). Fix trust-breaking bugs first and notify users plainly: what happened and what you fixed 41 That made accuracy the business. Every pixel, every click, every bug fix was survival.https://x.com/hnshah/status/1963610001542611098 40 I used to email users personally when something broke. No excuses. Just: “Here’s what happened. Here’s what we fixed.”https://x.com/hnshah/status/1963610001542611098 .
- Instrument for retention and referral lift after trust-improving fixes 37 Most growth hacks are sugar highs. They spike, then fade. Trust compounds. Each accurate experience, each honest interaction, makes your product harder to churn from, easier to recommend, and more valuable over time.https://x.com/hnshah/status/1963610001542611098 .
And in the AI era, where fakery is free and skepticism is default, trust is no longer just an advantage. It is the moat.
x.com -
Product is half; distribution is the other half
- Why it matters: Building without distribution creates “products nobody wants.” Distribution must be designed and tested alongside the product 45 Product is half the equation. Distribution is the other half. Ignore it, and you’ll spend years building something nobody wants.https://x.com/hnshah/status/1963659075675386109 . Retention is the reality check between novelty and true utility 44 Retention is the ultimate reality check. It forces you to face whether your product is truly useful, or just a novelty. It’s the difference between building a moment and building a company.https://x.com/hnshah/status/1963622585754947794 .
-
How to apply:
- Exit stealth early; talk to many potential users to find the right market/customer profile 143 Get rid of "stealth mode" and talk to everyone you can as quickly as you can to validate your idea and find the right market / customer profilehttps://www.reddit.com/r/startups/comments/1n7i5z7/comment/ncetlo5/ .
- Validate with pre-launch signals: landing page + “Buy Now,” explainer video, or preorders (Buffer, Dropbox, Tesla playbooks) 55 Buffer put up a landing page with subscription tiers. DropBox posted a video explaining how they anticipated the app would work -- it wasn't ready.https://www.reddit.com/r/startups/comments/1n7qmxd/comment/ncc3y8z/ 144 Talking with users isn't the worst thing ever. To test for self-sabotage, put up a landing page and Buy Now button. If 34,569 people click in a reasonable time, I will go out on a limb and suggest you build.https://www.reddit.com/r/startups/comments/1n7qmxd/comment/ncc3y8z/ 54 Tesla takes preorders. The people with an Elon Musk quote nailed to the wall ...not so inspired.https://www.reddit.com/r/startups/comments/1n7qmxd/comment/ncc3y8z/ .
- Prioritize repeatable signups (e.g., paid channels) to compute CAC:LTV; de-prioritize random/unreplicable signups 62 Random signups that are not replicable is not really valuable data compared to repeatable signups from ad-spend for example. This way you can actually calculate all the metrics that matter like CAC, and together with your LTV (assumed or otherwise), you can calculate the ratio and see whether those makes sense.https://www.reddit.com/r/startups/comments/1n86k3s/comment/nccoifr/ .
-
The Era of Evals (and the rise of AI PM)
- Why it matters: Enterprises are investing heavily in AI evaluations across workflows; evals are emerging as moats and a must-have skill for PMs/engineers 5 Mercor (@mercor_ai) is now working with 6 out of the Magnificent 7, all of the top 5 AI labs, and most of the top application layer companies. One trend is common across every customer: we are entering The Era of Evals.https://x.com/BrendanFoody/status/1939783540394402083 30 A few examples: @ garrytan : “Evals are emerging as the real moat for AI startups.”https://x.com/lennysan/status/1963688207280955839 29 @ kevinweil : “Writing evals is going to become a core skill for product managers.”https://x.com/lennysan/status/1963688207280955839 . It’s a rare new hard skill on par with SQL/Excel 31 It’s the first new hard skill in a long time that PMs/engineers/founders have had to learn to be successful. The last one was maybe SQL, and Excel?https://x.com/lennysan/status/1963688207280955839 . Meanwhile, AI PM roles are growing 120% vs. 1% for overall PM and pay 30–40% more; top comp can hit $900k–$1M+ 105 AI PM jobs growing 120% vs Overall PM growing 1%https://substack.com/@aakashgupta/note/c-152495364 104 AI PM jobs paying 30-40% more at every levelhttps://substack.com/@aakashgupta/note/c-152495364 106 Meta is paying $1M+ for AI PMs. Netflix is paying $900K.https://substack.com/@aakashgupta/note/c-152495364 .
-
How to apply:
- Learn to design evals for your AI features (coverage, ground truth, pass/fail criteria). Start with retrieval/use-case evals before model tuning 27 Learn how to write evals from @ _amankhan : https://www.lennysnewsletter.com/p/beyond-vibe-checks-a-pms-completehttps://x.com/lennysan/status/1963691063509680202 28 @ garrytan WTF is an eval: https://www.lennysnewsletter.com/i/153296003/evalshttps://x.com/lennysan/status/1963691063509680202 .
- Treat AI PM as a skills stack: context/prompt engineering, evals, experimentation, agents, strategy, execution 103 Context Engineering : [https://www.news.aakashg.com/p/rag-vs-fine-tuning-vs-prompt-engineering](https://www.news.aakashg.com/p/rag-vs-fine-tuning-vs-prompt-engineering)https://substack.com/@aakashgupta/note/c-152495364 102 Prompt Engineering : [https://www.news.aakashg.com/p/prompt-engineering](https://www.news.aakashg.com/p/prompt-engineering)https://substack.com/@aakashgupta/note/c-152495364 101 AI Evals: [https://www.news.aakashg.com/p/ai-evals](https://www.news.aakashg.com/p/ai-evals)https://substack.com/@aakashgupta/note/c-152495364 100 AI Experimentation : [https://www.news.aakashg.com/p/209f2acf-1828-4d819ba5-b1690fde4a16](https://www.news.aakashg.com/p/209f2acf-1828-4d819ba5-b1690fde4a16)https://substack.com/@aakashgupta/note/c-152495364 98 AI Agents : [https://www.news.aakashg.com/p/ai-agents-pms](https://www.news.aakashg.com/p/ai-agents-pms)https://substack.com/@aakashgupta/note/c-152495364 96 AI Product Strategy : [https://www.news.aakashg.com/p/ai-product-strategy](https://www.news.aakashg.com/p/ai-product-strategy)https://substack.com/@aakashgupta/note/c-152495364 97 AI Feature Execution : [https://www.news.aakashg.com/p/how-to-build-ai-products](https://www.news.aakashg.com/p/how-to-build-ai-products)https://substack.com/@aakashgupta/note/c-152495364 .
-
Problem space first → metrics that matter
- Why it matters: Over 80% of new products fail mainly because teams jump to solutions without validating the problem 22 Over 80% of new products fail. This is the number one cause of that 80% is this because most companies just start with the solution. And once you become aware of this, you start realizing all the conversations in your company. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
-
How to apply:
- Follow the Lean Product Process: customer → needs → value prop → features → UX 23 So I'm going to just refer back to the core process, the core framework from my book, the Lean Product Process. Right. We got to figure out, in order to have a successful product achieve our fit, we have to know who's our customer, what are their needs, what's our value prop, how are we going to be better, different, what feature should we be in there and what should the UX be? Right? https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Derive metrics from a clear definition of “solved;” pick inputs that drive the output you care about (e.g., FTUE completion → DAU) 99 If you understand the problem, and you understand what solving that problem looks like, it's usually pretty easy to get to the metric that you need...https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncf6sgf/ 87 There are several frameworks you can use. I like to think of input and output metrics. Input metrics are things you can control and move. Output metrics are driven by your input metrics. For example, FTUE completion could be an input metric and daily engagement (DAU) could be your output metric. You might want to drive DAU, but if you can establish a link between FTUE completion and DAU, then all you need to do is drive FTUE completion.https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncel436/ .
- Build a metric tree: North Star (e.g., items sold) → first-level metrics (conversion, returns, selection breadth) → drivers (payment conversion, availability, marketing clicks) 24 You can Google metric tree, it is a great framework to link it all together.https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncevp5y/ 76 For example, your NSM can be items sold, your first level product metrics can be website conversion, returns, width of selection. Then conversion depends on payment conversion, availability, marketing clicks, etc.https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncevp5y/ .
-
Strategy-aligned, transparent prioritization beats ad‑hoc requests
- Why it matters: Without alignment and visibility, roadmaps devolve into noise or ticket queues. You need an intake system, a real strategy with exec buy‑in, and transparent scoring for decisions 114 You’ll need an intake system, or the requests will own you.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccrjwd/ 112 But none of it works unless you’ve got a real product strategy with exec buy-in. You need alignment on where the product is now and where it’s supposed to be a year from now. Too many places confuse Gantt charts with roadmaps. They’re not roadmaps. They’re project plans wearing a fake mustache. But, I digress ...https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccrjwd/ 111 Then you need a unified prioritization system that everyone can see. I like IDEA/E: Impact, Dissatisfaction, Evidence, Advantage to us, over Effort. Whatever you use, the point is to show why something got promoted, deferred, or flat-out denied.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccrjwd/ .
-
How to apply:
- Capture requests publicly (e.g., UserVoice/UserEcho) so everyone sees context and votes 113 I’ve used tools like UserVoice and UserEcho to capture requests, surface them to the whole community, and let the crowd add context or vote. That gave us transparency, a way to say no, and a way to stop feature requests from living in silos or dying in email threads.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccrjwd/ .
- Score work (e.g., IDEA/E: Impact, Dissatisfaction, Evidence, Advantage to us, over Effort) and publish why items are promoted, deferred, or denied 111 Then you need a unified prioritization system that everyone can see. I like IDEA/E: Impact, Dissatisfaction, Evidence, Advantage to us, over Effort. Whatever you use, the point is to show why something got promoted, deferred, or flat-out denied.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccrjwd/ .
- Treat choices as “requests that fit the vision vs. those that don’t;” say no often—especially for early-stage products—and prioritize relentlessly 110 It's not really "user request vs. product vision" but "user requests that don't fit the vision vs. ones that do". So it's up to you to figure out what criteria to use to determine that (e.g. we use a value score).https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nce123f/ 84 1 consider your product maturity. A new product requires a vision and more « no » to become stable and differentiatedhttps://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccdg3f/ 89 Just remember that this is not a choice between fulfilling a customer request or doing nothing. Every time you say yes to a customer, you are taking resources away from other strategic priorities. Prioritize relentlessly. It might upset some customers today, but you are providing more value to more customers in the long run.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/ncel69x/ .
Tactical Playbook
-
Discovery that de‑risks building
- Validate first: confirm the problem is pervasive, urgent, and that users will pay 124 I'm not exactly sure at what stage you're at. I'm not sure if you have a working product w/ actual customers, MVP, or some lo-fi wireframe you're trying to validate however, until you can check the box on (1) the problem you're solving is a pervasive problem (2) the problem the business and/or user has that you're trying to solve for has a high degree of urgency (3) the business and/or user is willing to pay for the solution you have that addresses this problem; nothing else really matters.https://www.reddit.com/r/ProductMarketing/comments/1n7785z/comment/nceeufw/ .
- Talk to customers directly; ask outcome-focused “why” questions; customers are experts in their problems, not your product 71 I have spent my entire career in enterprise software. Selling software to very sophisticated buyers. Every single one of them has an opinion about how my software SHOULD work. I listen politely, thank them, and then usually ignore everything they said. My questions back to them are always about outcome. What were you trying to do there? Why did you need that? How do you solve that problem now? Why? Why? Why? When I know the why, I can build something that they will love.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/ncgg4u2/ 70 This is why PM needs to talk to customers directly. Unless you're doing primary research, it's really hard to get these nuggets of truth.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/ncgg4u2/ 72 Customer feedback and inputs are always fascinating but you own the product. In the end, they are not experts in your product, they are experts in their problem.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/ncgg4u2/ . Treat early convos as field research and map gaps between official process and lived reality 135 I’ve noticed the biggest difference comes when you treat those early convos less like networking and more like ‘field research’. Instead of trying to impress people, just ask them how things actually get done vs how they’re supposed to get done. That gap between official process and lived reality tells you way more about the company than any onboarding doc ever will.https://www.reddit.com/r/ProductManagement/comments/1n7o058/comment/ncc1y65/ .
- Run pre‑launch experiments: landing page + “Buy Now,” explainer video demo, preorders 144 Talking with users isn't the worst thing ever. To test for self-sabotage, put up a landing page and Buy Now button. If 34,569 people click in a reasonable time, I will go out on a limb and suggest you build.https://www.reddit.com/r/startups/comments/1n7qmxd/comment/ncc3y8z/ 55 Buffer put up a landing page with subscription tiers. DropBox posted a video explaining how they anticipated the app would work -- it wasn't ready.https://www.reddit.com/r/startups/comments/1n7qmxd/comment/ncc3y8z/ 54 Tesla takes preorders. The people with an Elon Musk quote nailed to the wall ...not so inspired.https://www.reddit.com/r/startups/comments/1n7qmxd/comment/ncc3y8z/ . Prioritize retention as your truth metric 44 Retention is the ultimate reality check. It forces you to face whether your product is truly useful, or just a novelty. It’s the difference between building a moment and building a company.https://x.com/hnshah/status/1963622585754947794 .
- Make demand measurable: drive repeatable cohorts (e.g., Reddit/X ads) to compute CAC:LTV; avoid over-inferencing from one-off signups 62 Random signups that are not replicable is not really valuable data compared to repeatable signups from ad-spend for example. This way you can actually calculate all the metrics that matter like CAC, and together with your LTV (assumed or otherwise), you can calculate the ratio and see whether those makes sense.https://www.reddit.com/r/startups/comments/1n86k3s/comment/nccoifr/ 61 makes sense. will eventually start reddit and/or x ads to see where it takes me, then, as you said, it's easier to projecthttps://www.reddit.com/r/startups/comments/1n86k3s/comment/ncdr5n6/ .
- Segment and choose where to win: segment customers, compare segment performance (usage, CAC, ARPU) and potential (TAM, competition, ability to serve) 123 If you have actual paying customers at the moment but you are trying to decide which segment to go after (ed or b2b), then you have some data you can work with. (1) Segment your audience (2) look at performance of each segment (usage, CAC, ARPU, etc.) (3) Look at potential of each segment (TAM, competitive intensity, ability to serve, etc.). Should give you a more data-driven approach to prioritizing and targeting your segment.https://www.reddit.com/r/ProductMarketing/comments/1n7785z/comment/nceeufw/ .
- Mine unstructured data for “alpha”: combine 1st/2nd/3rd-party data and behavioral signals to find non-obvious insights 154 there is no one size fit all here. there is real alpha in gaining untapped data. that is the only way to cut through the noise of common b2b data sets.https://www.reddit.com/r/ProductMarketing/comments/1n7uvni/comment/ncdevyd/ 153 unstructured data = 1st, 2nd and 3rd party data along with behaviour data (clicks, opens, visits, signups, usage etc.) this is how we are doing it.https://www.reddit.com/r/ProductMarketing/comments/1n7uvni/comment/ncdevyd/ .
-
AI‑accelerated prototyping (vibe coding) with guardrails
- Draft a concise product brief (target customer, top problems, features, data model, UX traits) before generating code 75 Say okay, so it's good to use what I'm calling a product brief template. There's no one magical template. There's a bazillion out there on the Internet. I created one that's simple for me that follows kind of my framework. So it has a target customer problems, features, data model, you know, UX characteristics and the point is to fill it out with bullet points. And so this is, if you want, during that first 20 minute period you can feel free to. It's a Google Doc, you can just make your own copy and fill it out with, with yourself or with your team. Again, that is the URL, it's, it's a link from the, from the main Google Doc if you want to go there and I would recommend you can use that. If you have some other template you want to use, by all means use some other templates. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- A/B your prompt: try a one-liner vs. pasting the brief; verify claims (“built drag-and-drop”) actually exist—trust but verify 20 So I want to show you in Bolt I did two, I did an A B test, I said let me give it a simple prompt like a one sentence prompt and see what it does. And then let me fill out Dan's template, probably template and give it that and see what it does. Okay, so here's Bolt and what I did is I said create a product roadmapping tool. That's all I said, is all I said. And guess what it did. It's pretty magical. Guys, look at this. It did it. And I'm not going to bore you with the text, but it'll tell you in the chat like, okay, here are the functionalities I'm going to build. Here's the UX aspects. Like it's, it's thoughtful. Tell you it decided to put in priorities. I didn't say anything about priorities. Is status completed versus planned. It decided to do quarters. It decided it's going to lay it out like this and it's going to have year one there and year two. You know, it said it had drag and drop. It didn't actually build drag and drop. That's one thing you got to be careful with you. So they say things but you have to make sure trust would verify. Right. So that's what I came up with. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Start in Discuss mode to plan; switch to Build to apply scoped changes; use Undo/Versions aggressively; prefer minimal, surgical edits 74 I'm talking discuss your build mode. It has two modes, so build mode, it actually will generate code discuss mode. It will just tell you its thoughts without changing any code, which is very handy. I'll explain that in a minute. And then this is the chat interface. https://www.youtube.com/watch?v=4v_Jv6-jU80 19 I want to talk about discussors Build mode by default, it goes into build its build mode by default it will, it will generate code and change the code. By default, in discuss mode, it won't change your code. Right. It will tell you what it would do. And one of the things is it uses less credits. So if you're just trying to play around and figure something out, there's no reason to have it be coding. It actually takes longer too because it's generating all this code as opposed to telling you what it would do. https://www.youtube.com/watch?v=4v_Jv6-jU80 17 This is really important. This is how the undo button, it's really important. And it's. They appear throughout the chats. So, like every. Every time you make a change, there'll be an undo right there. So you can decide, you know what, I want to go back three chats back and undo. You don't have to just do the last one. You can go back and undo back. Right? So it's kind of like. The way I think about it is like when you're climbing, you know, you're climbing, you're doing rock climbing, and you're hammering in your rope as you go up in case you fall. That's what these things are. You can kind of fall back. You do a. Dig this and going back and duplicate. I see. You can, you can, you can. There is a duplicate thing you can do if you want. You don't want to be having a bun. You. You don't want to be the Marvel Multiverse, man. It gets kind of Crazy if you have too many. But one of the cool things actually version V0 and magic patterns have something that bolt doesn't, which is a versions drop down in the upper right. It makes it super easy to just go back and look at an old version. It like loads instantly. It's really cool. So it's not full on a B testing, but that way you can see and then you can go full and then. Yeah, so undo. And the other thing is publish. So the funny thing about publishing is it publishes. There's no like confirm. Okay? Like it just publishes. It just so you know. And it's no big deal. It just goes to a netlify URL. Nobody knows that URL, so it doesn't really matter, right? Only you know that URL, so it's not a big deal. But it does publish it. And, and by the way, you have to republish. It doesn't automatically update. You have to republish all the time. https://www.youtube.com/watch?v=4v_Jv6-jU80 16 But you can say or you can say like make the minimal amount of changes or you can better discuss mode and say tell me the changes you're going to make. If you want to geek out, you can go and look at the diffs it's going to make if you really want to get anal about it. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Use the inspector to target elements; beware class-level changes propagating globally 18 A couple other things on Vault, the inspect that's just so you can go into a specific element. It can be annoying when you're trying to specify detailed UI changes in text. You're like, okay, the button that's in the upper left corner, make it red. It gets really complicated. So you can just click the element and say, this is the element I want to change. One thing I noticed is even if you pick a specific button or heading, it might modify the class. And so then all the headings change. All the buttons that have that class suddenly change. So it's not necessarily just that element. It might pick up the class as well. Publish. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Reverse‑prototype existing UIs from screenshots (e.g., Magic Patterns) 66 I just want to give a quick plug for Magic patterns. Gen AI tools are great if you're starting from scratch. If it's Greenfield Blue sky, you're building Some, you know, TikTok 2.0 or whatever you're building, it's super cool. How many of you guys have existing products that you're managing today? Right? You can't start from scratch. There is a ui, there's a design system, there's a code base. So there's a second use case, which is, hey, I'm a pm and I used to do this the old way, which was I take a screenshot, I'd bring it into Photoshop, I'd hack it up and Frank, I call it Frankenstein. Add a dropdown, add a menu, add a button, and then wire it all together in InVision. Well, guess what? Now you can just take that screenshot and Magic Patterns, they do the best job, I think, of actually rendering exactly what was in that screenshot. And I tried four different ones, my Magic patterns and three others. This is what I call reverse prototyping from a screenshot. It's a different use case than what we're talking about. It can also do Vibe coding, but I took like an Airbnb page and I remember I tried a very basic one and just. It did not look anything like the Airbnb page, the prototype that it generated. Right. The other four, they got better, but Magic Pattern was the best one I had, captured most of the elements, the icons, the gradients, the shadows, all that kind of stuff. It's really good. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Staff live sessions as a trio: 1) keyboard, 2) data (synthetic data/schema), 3) QA/UX notes 68 This is a model that I have Tried before with co workers and friends. And it is what I think works best when you are multiple people working on one same by coding project. I would put one person in charge of the keyboard, one person in charge of the data, the mocked up version of what you're going to show, and one person in charge of QA and user experience. That way you can have the person that is doing user experience write in a doc all of the things that they want to have different. The person in data enriching, creating synthetic data with ChatGPT growing the schema and the person that is in front of the keyboards just executing one after another on the things that you want to get done. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Specify persistence early (e.g., local storage/API) to avoid “it doesn’t save” surprises 15 Clickable prototype and so one of the things we found is as we were clicking through it, it wasn't saving data. So we asked a question about local. Storage and actually included including that in the original PRD that the intention is to store locally. And so it's whether or not you're. Actually going to, what kind of file. You'Re going to use or whether you plan for APIs later is something that would actually make sense to include early in the prd. And that was super helpful. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Paste exact error messages or screenshots; tools often auto‑diagnose 12 The other tip is you can literally, if there's an error message, you can copy and paste and say, this is the error message that I'm getting. It's actually kind of auto magical too. If there's a bug, it might even say, you can just say like, that didn't work. And it'll be like, oh, I see why. Here you go, let me fix it. Like it's. It can be that kind of magical. https://www.youtube.com/watch?v=4v_Jv6-jU80 73 And then finally, screenshots are good. Like, you can literally go to your preview, take a screenshot, say, see, this thing's on top of this thing. I don't want it to be on top of it. And it'll be like, oh, I see what you're saying. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
-
Intake, prioritization, and saying “no” with evidence
- Build intake that asks only for the problem and impact (remove solution questions). Ask “what’s the actual value gained?” 79 You may want to remove all the solution questions from the form completely and just ask about the problem.https://www.reddit.com/r/ProductManagement/comments/1n814mk/comment/ncfnw44/ 142 If the form has all the "normal" questions it could be worth reframing a few to get at what you need. i.e., we ask about the challenge and the impact of solving it — what is the actual value that would be gained? That prompts submitters to think a little more deeply about their ask.https://www.reddit.com/r/ProductManagement/comments/1n814mk/comment/nce3w1o/ .
- Aggregate requests in an ideas portal so teams can submit, vote, and comment; reduce siloed asks and duplicates 141 How do you aggregate the intake requests? And could you allow access to everyone's ideas? We use an ideas portal tool that let's our team submit requests, vote on each other's, comment, etc. In my experience this helps reduce the separate small requests and instead prompts people to chime in about their ideas/problems on one main submission. We get a lot of good qualitative chatter that way.https://www.reddit.com/r/ProductManagement/comments/1n814mk/comment/nce3w1o/ .
- Publish a priority waterfall (e.g., mandatory/regulatory → P0 issues → roadmap), ask sales to prioritize their own requests, and require expected ROI (“if we do this, what do we get?”) 26 3have a waterfall of priorities you communicate what you do and in what order. Examples:https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccdg3f/ 25 First mandatory /regulatory second P0 issues third roadmaphttps://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccdg3f/ 83 4 ask sales to prioritize their own requestshttps://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccdg3f/ 82 5 make sales accountable: if we do this request what will we get in return ?https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccdg3f/ .
- Expect gravity: big-wallet escalations and keep-the-lights-on tickets happen; protect vision-driven direction with a unified, transparent scoring model 109 And still, be ready for reality: a big-wallet customer will escalate over your head, and your boss will force a “keep the lights on” ticket into the queue. That’s just gravity. The job is making sure vision, not noise, sets the direction.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccrjwd/ 111 Then you need a unified prioritization system that everyone can see. I like IDEA/E: Impact, Dissatisfaction, Evidence, Advantage to us, over Effort. Whatever you use, the point is to show why something got promoted, deferred, or flat-out denied.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccrjwd/ .
-
Making Product Ops a force multiplier (not template police)
- Clarify mission first: why was ProdOps created (visibility, requirements quality, launch consistency, data standards)? 65 I'd first ask- why did the org implement product operations to begin with- what problem was prodops supposed to solve in an org with 20 PMs.https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/ncdtiqx/ 64 Was it a lack of visibility into product work?https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/ncdtiqx/ 63 Or was it a problem of poor outputs- poorly written and understood requirements, delayed or unsuccessful launches, inconsistent data measurement, inconsistent roadmaps, inconsistent practices period...https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/ncdtiqx/ . Align reporting to the CPO to stay product‑centric 139 Q: who do Prod Ops report to in your case? We still reported to the CPO so it made sure that we could never get too out of sync with product people. If they are reporting to any other function that's probably your issue. If they are still reporting into Product.…that's a different type of problem...https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/ncc34ck/ .
- Measure value: faster projects, clearer exec data, more PM time for discovery vs. admin; track impacts like case volume, revenue, churn, and feature adoption 137 The ops teams I’ve seen add real value measure stuff like: are projects moving faster, is leadership getting clearer data, and do PMs actually get more time for discovery vs admin. Standardization should only exist where chaos is slowing things down, not just because someone likes a new framework.https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/nccz8pj/ 121 ProdOps teams can be measured by items they help get committed to roadmap...and what impact that has on decrease in case volume, revenue, churn, product feature adoption percentage, etc. Ultimately PM makes the final call...not ProdOps.https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/ncfdjk8/ .
- Collaboration patterns: let Ops draft; PMs edit; reduce PM admin—“reducing PM admin opens up everything else” 122 While I get along well with our product ops team, in our org they are clearly optimizing for engineering throughput over everything else. Usually it takes a conversation or several to help them connect the dots on why reducing PM admin work opens up everything else that they want to achieve. They are also luckily very receptive to template feedback so usually I can let them have the first crack at any process and then edit it later.https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/ncd3g7a/ . Avoid top‑down “template police”; find an Ops ally who truly understands PM to unblock teams 138 Yeah, I’ve seen this happen. Product ops can easily slip into “template police” mode if they’re not grounding their work in feedback from PMs. If they’re not collecting that input, it’s no wonder things feel top-down.https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/nccz8pj/ 80 If you don't have much power, find the one person on the product ops team that actually understands product management. Request for them specifically to unblock your team. That person might be able to work with you to make things better.https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/nceasmg/ .
- If mandated tools slow you down, implement team‑fit tools and increase stakeholder transparency; report outcomes, not rituals 136 We flat-out told Ops that we weren't following their processes or using their tools. It's made my team so much more productive now that we've implemented tools and processes that work for us, and enhanced our transparency with our stakeholders.https://www.reddit.com/r/ProductManagement/comments/1n82eyk/comment/nccgw9b/ .
-
Metrics and observability you can act on
- Derive metrics from the defined problem and “solved” state 99 If you understand the problem, and you understand what solving that problem looks like, it's usually pretty easy to get to the metric that you need...https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncf6sgf/ . Use input→output chains (e.g., FTUE completion drives DAU) 87 There are several frameworks you can use. I like to think of input and output metrics. Input metrics are things you can control and move. Output metrics are driven by your input metrics. For example, FTUE completion could be an input metric and daily engagement (DAU) could be your output metric. You might want to drive DAU, but if you can establish a link between FTUE completion and DAU, then all you need to do is drive FTUE completion.https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncel436/ .
- Build a metric tree from NSM to drivers to target interventions 24 You can Google metric tree, it is a great framework to link it all together.https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncevp5y/ 76 For example, your NSM can be items sold, your first level product metrics can be website conversion, returns, width of selection. Then conversion depends on payment conversion, availability, marketing clicks, etc.https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncevp5y/ .
- Search effectiveness: track zero‑shot success (search ends without follow‑ups) and follow‑up search rates 85 You could always track how often that search it ends without any follow up i.e. the answer was given and user completed the task and then as a separate metric, how often additional searches are done, probably a varying details to give you a rough idea of the first a.k.a. zero shot search work for themhttps://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/ncfddpn/ .
- Run monthly/quarterly retros to avoid losing the big picture to short‑term optimization 86 Edit to add: your insight about losing the big picture is spot on! and something PMs should be well aware of. establish some processes for yourself to avoid this pitfall. maybe do a retro every month/quarter so that you know you are still steering towards the full picture and not chasing the few bps of improvement in a metric.https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncel436/ .
- Better observability = faster learnings—invest in instrumentation early 38 Better observability = faster learnings.https://x.com/joulee/status/1963620824126287973 .
-
Releases and launches that keep up with engineering velocity
- Ship customer‑facing features behind flags 152 Release every customer-facing feature under feature flags. https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncck5vb/ . Define GA “done” to include GTM and enablement (docs, training) 151 Definition of done (for general availability) includes GTM and enablement criteria like documentation and sales training. https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncck5vb/ . Keep a predictable GA cadence (monthly/quarterly) 150 GA release controlled by PM org and happens monthly or quarterly. https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncck5vb/ .
- Decouple release vs. launch—bundle into themed monthly/quarterly “rolling thunder” launches to market continuous delivery (Salesforce-style) 128 -Consider a decoupling of release versus launch. Sort of a rolling thunder approach to launches through monthly/quarterly "launches" and theme it. This is what Salesforce does and they do it masterfully.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncemd99/ .
- Be a friend to velocity: pre‑build messaging templates, automate where possible, and use Sales as an ally; quantify how enablement trade‑offs affect pipeline; focus PMM resources on Tier‑1/2 launches to prove impact 148 -Show you're a friend to speed and velocity and just work with your PM/engineering team on some bare minimum enablement. Maybe have some pre-built messaging templates. Automate where you can.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncemd99/ 147 - Make sales your ally and leverage as a pressure point. If they are feeling pain in this process, they'll let you and product know.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncemd99/ 146 - Stop selling the big bang and frame the problem differently. Reframe it in terms of impact if you can. By sacrificing messaging and enablement, what is the impact on pipeline?https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncemd99/ 127 -Pick one or 2 strategic wins versus trying to win the whole battle. Focus on your major tier 1 / 2 launches and show impact.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncemd99/ .
- Offer demo environments and ad‑hoc access for early adopters/betas to accelerate learnings 149 Demo environment, early adopters and beta customers can get access on an ad hoc basis.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncck5vb/ .
-
Mobile monetization: pick simplicity to maximize conversion
- For subscription apps, native IAP (e.g., via RevenueCat) is the safest launch path—higher conversion, simpler refunds, and lower operational overhead than external web paywalls 11 Option 1: The Standard Path (RevenueCat + Native IAP) This seems to be the simplest and safest route. Use the native In-App Purchase systems from Apple/Google, managed through a service like RevenueCat. Pros: Super simple to implement, secure, trusted by users (Face ID/Fingerprint purchase), and handles all the backend validation and subscription state. Cons: The 15% commission (on the first $1M) is a significant cut.https://www.reddit.com/r/startups/comments/1n87n52/ 52 I've never split-tested it so I don't have exact numbers, but my gut feel is that conversions will suffer quite a lot if you force users to pay off platform. I'd rather get 85% of $1,000 than 97% of $500.https://www.reddit.com/r/startups/comments/1n87n52/comment/ncd2c9v/ 51 Refunds are much easier to deal with as well (no matter how great your product, you WILL get refund requests) and I'd much rather point them to Apple. It's better for my sanity too, coz you don't feel the sting as much.https://www.reddit.com/r/startups/comments/1n87n52/comment/ncd2c9v/ 53 Just go with Option 1.https://www.reddit.com/r/startups/comments/1n87n52/comment/ncd2c9v/ .
- External links (Stripe) add friction and sync complexity; DMA exceptions increase regional logic without guaranteed savings 10 Have a "Go Premium" button in the app that links to my website. Users would sign up and pay there using Stripe. My app would then unlock premium features based on their account status, which I'd manage with a backend (e.g., Firebase). Pros: Stripe's fees are much lower (~3%). Cons: Massive UX friction (leaving the app, creating an account, entering credit card info). I'm worried about a huge drop in conversions. Requires building and maintaining a backend with webhooks to sync purchases. I've read that Apple might still take a commission (a "tax on a tax") for payments originating from an in-app link, which seems to defeat the purpose.https://www.reddit.com/r/startups/comments/1n87n52/ 9 I'm based in Europe, so I'm aware of the Digital Markets Act (DMA) which forces Apple/Google to allow alternative payment processors directly in the app for EU users. Cons: This seems incredibly complex. I'd have to manage region-specific logic, and I understand the platforms still charge significant fees, so the savings might not be worth the headache.https://www.reddit.com/r/startups/comments/1n87n52/ . Match approach to what you sell and test before optimizing fees 8 very much depends on what you're sellinghttps://www.reddit.com/r/startups/comments/1n87n52/comment/nccwgxj/ .
Case Studies & Lessons
-
Trust as growth engine (Crazy Egg)
- What happened: Heatmaps had to be pixel‑accurate; even small errors killed trust. The team treated accuracy as the business and emailed users plainly when issues occurred 42 The heatmap was simple. Show where users click. But if it was even slightly off, the whole product was worthless. A designer would look once, see a mismatch, and never come back.https://x.com/hnshah/status/1963610001542611098 41 That made accuracy the business. Every pixel, every click, every bug fix was survival.https://x.com/hnshah/status/1963610001542611098 40 I used to email users personally when something broke. No excuses. Just: “Here’s what happened. Here’s what we fixed.”https://x.com/hnshah/status/1963610001542611098 .
- Outcome: Honesty and reliability improved retention and created loyalty; agencies used the product in client decks—a trust flywheel 7 That honesty both retained customers and also created loyalty. Trust became our growth engine. Agencies used Crazy Egg heatmaps in client presentations because they knew they could rely on them. Users recommended us because they were confident we wouldn’t embarrass them.https://x.com/hnshah/status/1963610001542611098 .
- Apply it: Define “trust incidents,” set SLOs around accuracy, and make incident comms a first‑class ritual 37 Most growth hacks are sugar highs. They spike, then fade. Trust compounds. Each accurate experience, each honest interaction, makes your product harder to churn from, easier to recommend, and more valuable over time.https://x.com/hnshah/status/1963610001542611098 .
-
Fundraising reality for PMs: milestones, cash discipline, and unit economics
- Lesson: Plan milestones that make the next round “consensus enough” (e.g., from Seed to a $10M Series A); scarcity forces better discipline, while “indigestion” from over‑funding kills companies 81 I'd say it's often the latter, I would say, and especially true these days because I'm investing more in deep tech companies. And so at Seed, it's very rare to see like, oh, there's an asset that's going to be working here at the Series A, because usually the asset's still going to be like being developed at the Series A or maybe Series B. I think what I'm looking for is like there's maybe not enough here for somebody to write a 5 or 10 or $20 million check, but the company has milestones that I think if they hit them then it would become, you know, sort of consensus enough to merit a check of that size. And then I'm basically trying to evaluate like, okay, the company has these milestones. Do I think they could hit them or not? And also if they hit them, are they compelling enough? But I think that's sort of the big, you know, investment wager. Yeah, yeah. So I mean, so in this case you do think about like what the follow on thing is going to want to see. You have reached a conclusion for the current round that is non consensus. Yeah. And I would say the consensus piece is part of it in that I definitely meet companies where they're like, we're raising three right now, It'll help us do these milestones and then we think we can raise 10. And then there's other ones where it's like we're raising three now, we're going to hit these milestones and then we want to raise like a 50 to $100 million Series A. And that's actually a much harder bet. Right. Because you're saying you have to assume they're going to be consensus by the time they raise the next round. And it's going to be like a top 5% series A. And that's a hard bet to take for the companies where the capital needs are more modest or they have like a more tranched roadmap planned. I think it's a little bit easier to predict like, hey, would these milestones be enough to raise 10 a lot of times I don't know if it'll be enough to raise 100. Like probably not. https://www.youtube.com/watch?v=roJUhk8qeRY 117 I think there's benefits to being non consensus because from the company side, I think when the money's hard to raise, you tend to be more frugal with it. And then also if the next round is less certain, I think there's less of a sort of it could crumble at any moment aspect because I think when it is hot and you're raising subsequent rounds very quickly under the assumption that things will go perfectly if anything slows down, it's like, hey, now everything, like now you can't raise any more capital all of a sudden, right? And I think if you're in the mentality of, you know, growing quickly and spending, I think that's pretty hard. On the flip side, if you're not consensus, it tends to be like, you tend to be more cash efficient, you tend to be more frugal out of necessity. I think the other side is it depends on the form of like consensusness. https://www.youtube.com/watch?v=roJUhk8qeRY 116 I tend to believe now that most companies fail from indigestion, not starvation, which is they just raise too much money too easily. They don't listen to the actual market, which is the customer base, and as a result they just have a bunch of bad practices and end up running out of money. https://www.youtube.com/watch?v=roJUhk8qeRY . Insist on clear unit economics; avoid extrapolating models from buzzy spaces without proof 115 I mean, for AI, for better or for worse, you have great unit economics. Everybody knows, kind of like. And we always talk about the open AIs and the anthropics, but if you talk like the 11 labs, for example, or mid journey, I mean, these are just famously model companies where the unit economics are great, they grow very quickly. And so I understand that, but I think there's kind of been this weird. And this happens, you know, a lot where people take the, the example of these model companies and they apply it to totally different spaces where you don't have the proof points, you don't have the economic case. And they kind of apply it. And that's one thing I don't know how to do. So certainly I don't believe, you know, we should all just follow like the common consensus around areas to invest in. https://www.youtube.com/watch?v=roJUhk8qeRY .
- Apply it: Tie roadmap to investor‑expectation milestones, model runway to reach them, and publish unit‑economic thresholds for go/no‑go decisions.
-
Rolling‑thunder launches in a rapid release shop
- Context: A PMM ran quarterly, themed launches on top of a biweekly ship cadence—decoupling release vs. launch 128 -Consider a decoupling of release versus launch. Sort of a rolling thunder approach to launches through monthly/quarterly "launches" and theme it. This is what Salesforce does and they do it masterfully.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncemd99/ 108 If you want to see a practical approach how I did a quarterly launch approach at a company that shipped every two weeks, just ping me and can share some examples.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncemd99/ .
- Apply it: Keep engineering shipping; market value in planned waves. Prove impact on pipeline/adoption before scaling the program 128 -Consider a decoupling of release versus launch. Sort of a rolling thunder approach to launches through monthly/quarterly "launches" and theme it. This is what Salesforce does and they do it masterfully.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncemd99/ .
-
Vibe‑coding gotchas (Bolt/V0/Magic Patterns)
- Observation: A one‑line prompt produced a seemingly complete roadmap UI; claims (e.g., drag‑and‑drop) weren’t always built. Teams needed versions, “minimal change” requests, and clear persistence requirements 20 So I want to show you in Bolt I did two, I did an A B test, I said let me give it a simple prompt like a one sentence prompt and see what it does. And then let me fill out Dan's template, probably template and give it that and see what it does. Okay, so here's Bolt and what I did is I said create a product roadmapping tool. That's all I said, is all I said. And guess what it did. It's pretty magical. Guys, look at this. It did it. And I'm not going to bore you with the text, but it'll tell you in the chat like, okay, here are the functionalities I'm going to build. Here's the UX aspects. Like it's, it's thoughtful. Tell you it decided to put in priorities. I didn't say anything about priorities. Is status completed versus planned. It decided to do quarters. It decided it's going to lay it out like this and it's going to have year one there and year two. You know, it said it had drag and drop. It didn't actually build drag and drop. That's one thing you got to be careful with you. So they say things but you have to make sure trust would verify. Right. So that's what I came up with. https://www.youtube.com/watch?v=4v_Jv6-jU80 17 This is really important. This is how the undo button, it's really important. And it's. They appear throughout the chats. So, like every. Every time you make a change, there'll be an undo right there. So you can decide, you know what, I want to go back three chats back and undo. You don't have to just do the last one. You can go back and undo back. Right? So it's kind of like. The way I think about it is like when you're climbing, you know, you're climbing, you're doing rock climbing, and you're hammering in your rope as you go up in case you fall. That's what these things are. You can kind of fall back. You do a. Dig this and going back and duplicate. I see. You can, you can, you can. There is a duplicate thing you can do if you want. You don't want to be having a bun. You. You don't want to be the Marvel Multiverse, man. It gets kind of Crazy if you have too many. But one of the cool things actually version V0 and magic patterns have something that bolt doesn't, which is a versions drop down in the upper right. It makes it super easy to just go back and look at an old version. It like loads instantly. It's really cool. So it's not full on a B testing, but that way you can see and then you can go full and then. Yeah, so undo. And the other thing is publish. So the funny thing about publishing is it publishes. There's no like confirm. Okay? Like it just publishes. It just so you know. And it's no big deal. It just goes to a netlify URL. Nobody knows that URL, so it doesn't really matter, right? Only you know that URL, so it's not a big deal. But it does publish it. And, and by the way, you have to republish. It doesn't automatically update. You have to republish all the time. https://www.youtube.com/watch?v=4v_Jv6-jU80 16 But you can say or you can say like make the minimal amount of changes or you can better discuss mode and say tell me the changes you're going to make. If you want to geek out, you can go and look at the diffs it's going to make if you really want to get anal about it. https://www.youtube.com/watch?v=4v_Jv6-jU80 15 Clickable prototype and so one of the things we found is as we were clicking through it, it wasn't saving data. So we asked a question about local. Storage and actually included including that in the original PRD that the intention is to store locally. And so it's whether or not you're. Actually going to, what kind of file. You'Re going to use or whether you plan for APIs later is something that would actually make sense to include early in the prd. And that was super helpful. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Apply it: Use Discuss→Build loops with small diffs; reverse‑prototype from screenshots when modifying existing UIs 66 I just want to give a quick plug for Magic patterns. Gen AI tools are great if you're starting from scratch. If it's Greenfield Blue sky, you're building Some, you know, TikTok 2.0 or whatever you're building, it's super cool. How many of you guys have existing products that you're managing today? Right? You can't start from scratch. There is a ui, there's a design system, there's a code base. So there's a second use case, which is, hey, I'm a pm and I used to do this the old way, which was I take a screenshot, I'd bring it into Photoshop, I'd hack it up and Frank, I call it Frankenstein. Add a dropdown, add a menu, add a button, and then wire it all together in InVision. Well, guess what? Now you can just take that screenshot and Magic Patterns, they do the best job, I think, of actually rendering exactly what was in that screenshot. And I tried four different ones, my Magic patterns and three others. This is what I call reverse prototyping from a screenshot. It's a different use case than what we're talking about. It can also do Vibe coding, but I took like an Airbnb page and I remember I tried a very basic one and just. It did not look anything like the Airbnb page, the prototype that it generated. Right. The other four, they got better, but Magic Pattern was the best one I had, captured most of the elements, the icons, the gradients, the shadows, all that kind of stuff. It's really good. https://www.youtube.com/watch?v=4v_Jv6-jU80 19 I want to talk about discussors Build mode by default, it goes into build its build mode by default it will, it will generate code and change the code. By default, in discuss mode, it won't change your code. Right. It will tell you what it would do. And one of the things is it uses less credits. So if you're just trying to play around and figure something out, there's no reason to have it be coding. It actually takes longer too because it's generating all this code as opposed to telling you what it would do. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
-
Early MRR ≠ fundable traction
- Reality: $500 MRR in two weeks is meaningless to investors without churn and retention data; angels around ~$5k MRR, VCs often at $25–100k MRR 50 $500 MRR measured over 2 weeks is pretty meaningless to an investor because you have no churn data. Not trying to discourage you - every business starts at zero. Just being real about where you are at 2 weeks in.https://www.reddit.com/r/startups/comments/1n86k3s/comment/ncd7x95/ 49 That’s friends and family numbers. An angel might come in if you have 5k MRR and a VC would come in at 25-100k.https://www.reddit.com/r/startups/comments/1n86k3s/comment/nccqe55/ . Focus fundraising on clear use of funds and strengthen distribution first 60 Congrats on the early traction. 500 MRR in 2 weeks is no joke, especially if it’s all paid with no free tier. If you’re looking at 1k MRR soon and already thinking about pre-seed, I’d say timing the fundraise is less important than having a clear use of funds. If you’re going to raise, one lever you might want to pull first is engineering talent. Hiring the right devs now could double your velocity before you even pitch. We’ve seen a bunch of lean teams get to product-market fit faster by treating early hiring like a growth multiplier, not just a cost center. Also +1 on your SAFE approach. Just make sure you’re not over-indexing on valuation at the expense of flexibility. Curious what your biggest bottleneck is right now, product, sales, team?https://www.reddit.com/r/startups/comments/1n86k3s/comment/ncdopvk/ .
- Apply it: Reinvest early revenue, diagnose distribution bottlenecks, and use tools to surface high‑fit conversations across Reddit/X/LinkedIn to scale outreach efficiently 56 Why do you need money by the way? You are already making $500 in two weeks. So that's pretty good. So invest 80% of your earnings, grow your project and then look for business Angeles or investors.https://www.reddit.com/r/startups/comments/1n86k3s/comment/ncd4lx0/ 59 bottleneck is distribution, I only have so many hours in a day. no budget for hiring though, I'm a solo grinder. I will slowly start to do better targeted ads on reddit/x and ugc on tiktok etc.https://www.reddit.com/r/startups/comments/1n86k3s/comment/ncdwldv/ 58 We actually help SaaS founders streamline exactly this – imagine getting notified only about truly relevant discussions on Reddit/X/LinkedIn where your product is a direct fit, complete with tailored reply suggestions. It's designed to make that organic outreach more efficient, helping you get more leads without constant manual monitoring. Might be worth a look if you're trying to scale your distribution without scaling your time commitment.https://www.reddit.com/r/startups/comments/1n86k3s/comment/nce780e/ .
Career Corner
-
The AI PM advantage
- Market: AI PM jobs are growing 120% vs. overall PM at 1%, with 30–40% higher pay; comp at top firms reaches $900k–$1M+ 105 AI PM jobs growing 120% vs Overall PM growing 1%https://substack.com/@aakashgupta/note/c-152495364 104 AI PM jobs paying 30-40% more at every levelhttps://substack.com/@aakashgupta/note/c-152495364 106 Meta is paying $1M+ for AI PMs. Netflix is paying $900K.https://substack.com/@aakashgupta/note/c-152495364 . “AI PM is eating PM” 107 AI PM is eating PM.https://substack.com/@aakashgupta/note/c-152495364 .
- Skills roadmap: context/prompt engineering, evals, experimentation, agents, strategy, execution; “future of PM is AI PM” 103 Context Engineering : [https://www.news.aakashg.com/p/rag-vs-fine-tuning-vs-prompt-engineering](https://www.news.aakashg.com/p/rag-vs-fine-tuning-vs-prompt-engineering)https://substack.com/@aakashgupta/note/c-152495364 101 AI Evals: [https://www.news.aakashg.com/p/ai-evals](https://www.news.aakashg.com/p/ai-evals)https://substack.com/@aakashgupta/note/c-152495364 100 AI Experimentation : [https://www.news.aakashg.com/p/209f2acf-1828-4d819ba5-b1690fde4a16](https://www.news.aakashg.com/p/209f2acf-1828-4d819ba5-b1690fde4a16)https://substack.com/@aakashgupta/note/c-152495364 98 AI Agents : [https://www.news.aakashg.com/p/ai-agents-pms](https://www.news.aakashg.com/p/ai-agents-pms)https://substack.com/@aakashgupta/note/c-152495364 96 AI Product Strategy : [https://www.news.aakashg.com/p/ai-product-strategy](https://www.news.aakashg.com/p/ai-product-strategy)https://substack.com/@aakashgupta/note/c-152495364 90 The future of PM is AI PM. Get ahead now.https://substack.com/@aakashgupta/note/c-152495364 .
- First steps: ship an eval suite for a feature; maintain a small, representative gold‑labeled set (manual first; augment with LLMs) 130 Yes, you do need a separate dataset. What we do is maintain a small but representative evaluation set: basically queries paired with one or more “gold” passages. Most of these are manually labeled, but we’ve started trying out LLMs to build the dataset too.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nce4w8x/ . Try RAGAS and DeepEval to get started 129 so, yes, RAGAS and DeepEval both looked quite promising from the little documentation I read. Honestly, I am yet to test it out, will do so soon. But even there, I wanted to understand from other PMs that is this generally the way you guys go about monitoring or should I look at other directions that I am likely not focusing on at the moment?https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nccdz4i/ .
-
Break in (or up) faster
- Don’t rely on online applications; tap referrals and direct reach‑outs 4 1. Never rely only on online submission Easy application tools create a many-to-many problem. Every job post attracts hundreds of resumes, many of which go unnoticed. At Ancestry, our summer internship program drew tens of thousands of applications for just a few dozen spots. One year at Meta, the RPM program hired one person from the general resume pool. The rest were referrals or sourced by our recruiters. Blind submission via an online tool is rarely your friend. Find another path in, whether that’s a referral, a connection, or a direct reach-out to make sure you get noticed. The goal is to keep your resume out of the digital dustbin and into someone’s hands for true consideration.https://debliu.substack.com/p/non-obvious-tips-for-landing-the .
- “Do the job before you get the job”: use the product, talk to customers, bring a prototype to interviews 2 6. Do the job before you get the job The strongest candidates act like they already work there. They use the product, read analyst reports, talk to customers, and bring ideas. I remember candidates who showed up with prototypes or user research. It told me they wanted not just a job, but this job. You don’t need to boil the ocean or write a long research report, but demonstrating that kind of initiative speaks volumes.https://debliu.substack.com/p/non-obvious-tips-for-landing-the . Vibe‑coded prototypes can help you show—not tell 33 A lot of candidates avoid jobs that require homework, and that is understandable, especially when you have a job already. But put in some time to really show you cared enough to prepare. Bring a prototype or an idea that will make you stand out. I have been working on a vibe-coded project with my teen daughter, and in just a few weeks, she learned to “show, not tell” by taking her feature idea and turning it into a service. Imagine anyone being able to do that before you go into an interview. You can show them new areas to explore and demonstrate the value you bring, even before you are hired.https://debliu.substack.com/p/non-obvious-tips-for-landing-the .
- Tailor your resume/story to the role; speak the company’s language and highlight upstream PMM/PM impact 1 7. Tailor your resume (and your story) for the role Resumes are factual, but you can use them to tell a story. I once coached someone in learning and development who wanted to become a PM. Her resume made her look like HR staff, which was technically where her role sat. I encouraged her to reframe her work around building and implementing HR systems, which is what an internal PM would be doing. Armed with the same facts and a different lens, she landed a PM job at a learning and development company.https://debliu.substack.com/p/non-obvious-tips-for-landing-the .
- Reduce hiring risk: show references, make it easy for managers to bet on you 3 5. See the world through the hiring manager’s eyes Hiring managers aren’t just filling a role. In reality, they are managing risk. A bad hire can derail a roadmap and sink a team. It can also reflect poorly on their reputation. That’s why referrals and internal transfers feel safer, especially for newer managers or those who are hiring for a critical role. Your job is to reassure them you are a sure bet. Find common contacts, bring in references, and show how you’ll make them look good for choosing you.https://debliu.substack.com/p/non-obvious-tips-for-landing-the 34 You know your abilities, but they don’t. There is an asymmetry of information. Your job is to close that gap. Show them you will be someone worth investing in and betting on. You are asking them to take a chance on you. When you look at it from that perspective, you can see your job is to show them why you are a great return on their investment.https://debliu.substack.com/p/non-obvious-tips-for-landing-the .
-
Build domain leverage and network early
- Domain‑expert PMs are in demand (e.g., health tech); leverage backgrounds like neuroscience to enter relevant roles; many openings reported in health‑tech/biotech 120 Not specifically from neuroscience. However for what it's worth (and it may be anecdotal!), but I have noticed more "domain-expert" PMs, where they'll have deep industry knowledge in the industry of the company they work at, so you should be an attractive hire for health tech companies looking for PMs.https://www.reddit.com/r/ProductManagement/comments/1n8di4p/comment/nce79hz/ 119 You’ll be fine, find a company where you can leverage your domain expertise in neuroscience. if you don’t love that domain (healthcare or biomedical) in the future, by that point you’ll have enough PM experience that you can move to another industry or role.https://www.reddit.com/r/ProductManagement/comments/1n8di4p/comment/ncer6m3/ 118 Interesting. I am seeing a lot of openings around health tech, I think there are tons of bio-tech companies in this domain. Good luckhttps://www.reddit.com/r/ProductManagement/comments/1n8di4p/comment/nce8m6j/ . Consider adjacent roles (CSM, ops, BA) to transition internally 69 It’s a tough market I know PMs with years of experience who are struggling to land roles right now. One path that might be worth trying: get in the door with an adjacent role (CSM, ops, biz analysis, etc.) and then transition internally. That’s actually how I got into product. Once you’ve proven yourself inside the company, it’s often way easier to make the switch than breaking straight into PM from the outside.https://www.reddit.com/r/ProductManagement/comments/1n8fyv6/comment/ncevkg3/ .
- In a new org, meet names/faces beyond your dev team early; build informal authority by spotting cross‑team patterns and learning how work really gets done 132 +1 on everything, and I would add, take advantage of being new. People will give you a lot of leeway to meet with them and talk with them if they know you're the new person. Take advantage of that window so you get to know their name and face.https://www.reddit.com/r/ProductManagement/comments/1n7o058/comment/ncco1h8/ 131 Otherwise, the first time they'll meet you is when you're asking for help or buy-in... and that is a hard ask if they don't even know your name and you don't know theirs.https://www.reddit.com/r/ProductManagement/comments/1n7o058/comment/ncco1h8/ 133 Really solid advice. What you’re describing is basically building informal authority. The kind of trust and credibility that doesn’t show up on an org chart but matters way more when the pressure’s on.https://www.reddit.com/r/ProductManagement/comments/1n7o058/comment/nccyg0c/ 134 When you’re doing those intro chats, don’t just collect random stories. Look for patterns. If three different teams mention the same blocker, you can surface that later as a shared problem instead of “your opinion.” Much harder for people to push back on.https://www.reddit.com/r/ProductManagement/comments/1n7o058/comment/nccyg0c/ .
-
Compensation and equity hygiene
- Beware low‑equity/low‑salary deals (e.g., ~1% with a 35% pay cut); insist on fair equity or salary 57 You are offered 1% of nothing, not even a MVP, no TAM known at this point and a 35% discount salary of what the market pays you. Why are you even considering this deal?https://www.reddit.com/r/startups/comments/1n84086/comment/ncd1yxw/ . Equal‑share pre‑funding is a common expectation; dilution later is normal 46 nope, equal shares pre everything. Always. Only okay part is dilution for funding rounds, that’s okay and normal. But biggest red flag is the CEO already showed they don’t value you and your work (they value themselves 9 times more than you). So with that, I would not partner with them even if they offered 50/50 now. Hard pass.https://www.reddit.com/r/startups/comments/1n7z3j6/comment/nccs7o0/ .
- Use Slicing Pie: track unpaid fair‑market contributions (“bets”) and split equity proportionally: Your Share = Your Bets / Total Bets 48 What your story is asking for is how to negotiate your share of the equity. What it overlooks is that fair equity isn't negotiated, it's calculated. Negotiations are based on guesses and promises. Calculations are based on numbers and facts.https://www.reddit.com/r/startups/comments/1n7z3j6/comment/ncehjp7/ 47 You can now calculate your share of the equity which is equal to your share of the bets.https://www.reddit.com/r/startups/comments/1n7z3j6/comment/ncehjp7/ .
- Sanity‑check founders: verify claimed exits; lack of proof is a red flag 145 If they say 2 exits and you can’t find proof, that’s more than enough cause for a hell no. The rest is bad and also would probably lead to a no, but just want to emphasize that if the situation were reversed and you misrepresented your experience in startup leadership and tried to use it as a selling point nobody would give you the time of day.https://www.reddit.com/r/startups/comments/1n7z3j6/comment/ncdqovk/ .
-
Interview prep signal
- If you’re targeting CPaaS/SaaS, be ready to discuss metrics and KPIs across ideation, development, launch, and growth; link inputs to outputs with a metric tree 78 You need to work on understanding product metrics, how performance of a product at various stages will be analysed in CPaaS and SaaS industries. Impact metrics and product KPIs at ideation, development, launch and Growth stages - I got this feedback after first round of interview. Can anyone help me prepare for this as I moved to second roundhttps://www.reddit.com/r/ProductManagement/comments/1n7o058/comment/ncd3rdn/ 77 “You need to work on understanding product metrics, how performance of a product at various stages will be analysed in CPaaS and SaaS industries. Impact metrics and product KPIs at ideation, development, launch and Growth stages” - I got this feedback after first round of interview. Can anyone help me prepare for this as I moved to second roundhttps://www.reddit.com/r/ProductManagement/comments/1n7ai7w/comment/ncd438k/ 24 You can Google metric tree, it is a great framework to link it all together.https://www.reddit.com/r/ProductManagement/comments/1n8f2lx/comment/ncevp5y/ .
-
Work with AI, faster
- Treat AI as a teammate for drafting, prototyping, and discovery, but keep the human edge—customer understanding and strategy 32 AI Becomes Your Teammate - 21:29https://x.com/aakashg0/status/1963723493989822960 14 Right. So that's part of it too. So I think the key skills are still the same key skills which is discovery, understanding your customer, understanding the process, discovery, understanding differentiation and value prop product strategy. Because if the tool, if the. If the coding is a commodity and the design is becoming more of a commodity, although I still think that you can still create UX innovation, it's really the PM aspects are the ones where you have the most differentiation. So that's my opinion. https://www.youtube.com/watch?v=4v_Jv6-jU80 . “You have to do everything faster today” 35 You have to do everything faster today.https://x.com/hnshah/status/1963821672634728810 .
Tools & Resources
-
Evals 101 for PMs
- What: “WTF is an eval?” and a practical guide to writing evals 28 @ garrytan WTF is an eval: https://www.lennysnewsletter.com/i/153296003/evalshttps://x.com/lennysan/status/1963691063509680202 27 Learn how to write evals from @ _amankhan : https://www.lennysnewsletter.com/p/beyond-vibe-checks-a-pms-completehttps://x.com/lennysan/status/1963691063509680202 .
- Why: Evals are becoming core PM craft and a defensible moat 30 A few examples: @ garrytan : “Evals are emerging as the real moat for AI startups.”https://x.com/lennysan/status/1963688207280955839 29 @ kevinweil : “Writing evals is going to become a core skill for product managers.”https://x.com/lennysan/status/1963688207280955839 .
- Try: Pilot a retrieval eval with a 100–200 query gold set and track pass/fail over time 130 Yes, you do need a separate dataset. What we do is maintain a small but representative evaluation set: basically queries paired with one or more “gold” passages. Most of these are manually labeled, but we’ve started trying out LLMs to build the dataset too.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nce4w8x/ .
-
Eval tooling: RAGAS, DeepEval
- What: Open-source tools PMs are exploring for AI evaluation 129 so, yes, RAGAS and DeepEval both looked quite promising from the little documentation I read. Honestly, I am yet to test it out, will do so soon. But even there, I wanted to understand from other PMs that is this generally the way you guys go about monitoring or should I look at other directions that I am likely not focusing on at the moment?https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nccdz4i/ .
- Why: Standardize quality bars for LLM features before shipping.
-
AI prototyping & reverse‑prototyping
- What: Magic Patterns (screenshot→UI), V0/Bolt for vibe‑coding 66 I just want to give a quick plug for Magic patterns. Gen AI tools are great if you're starting from scratch. If it's Greenfield Blue sky, you're building Some, you know, TikTok 2.0 or whatever you're building, it's super cool. How many of you guys have existing products that you're managing today? Right? You can't start from scratch. There is a ui, there's a design system, there's a code base. So there's a second use case, which is, hey, I'm a pm and I used to do this the old way, which was I take a screenshot, I'd bring it into Photoshop, I'd hack it up and Frank, I call it Frankenstein. Add a dropdown, add a menu, add a button, and then wire it all together in InVision. Well, guess what? Now you can just take that screenshot and Magic Patterns, they do the best job, I think, of actually rendering exactly what was in that screenshot. And I tried four different ones, my Magic patterns and three others. This is what I call reverse prototyping from a screenshot. It's a different use case than what we're talking about. It can also do Vibe coding, but I took like an Airbnb page and I remember I tried a very basic one and just. It did not look anything like the Airbnb page, the prototype that it generated. Right. The other four, they got better, but Magic Pattern was the best one I had, captured most of the elements, the icons, the gradients, the shadows, all that kind of stuff. It's really good. https://www.youtube.com/watch?v=4v_Jv6-jU80 21 So anyway, let me show you the new workflow with Vibe coding. You can just jump straight to a live prototype, Right. And this is an HTML CSS JavaScript prototype. At a minimum, it's the front end. Maybe it's got back end or whatever. Right. And in the old days, every once in a while, someone would build a HTML CSS JavaScript prototype. But it used to be so much work that you wouldn't bother. You would just use Figma and then you'd move from there. But now it's so easy to jump to that and then test it and then move and iterate very quickly. Right. So that's what's empowering. So if you did have a design gap and you didn't have a designer to help you create those clickable mockups, you were stuck before, basically, right? You had to try to do it yourself or just nobody would do it. But now you can kind of. Anybody on the product team can jump in and do Vibe coding to get a protect. That's the idea. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Why: Jump to live prototypes quickly; modify existing UIs from screenshots 66 I just want to give a quick plug for Magic patterns. Gen AI tools are great if you're starting from scratch. If it's Greenfield Blue sky, you're building Some, you know, TikTok 2.0 or whatever you're building, it's super cool. How many of you guys have existing products that you're managing today? Right? You can't start from scratch. There is a ui, there's a design system, there's a code base. So there's a second use case, which is, hey, I'm a pm and I used to do this the old way, which was I take a screenshot, I'd bring it into Photoshop, I'd hack it up and Frank, I call it Frankenstein. Add a dropdown, add a menu, add a button, and then wire it all together in InVision. Well, guess what? Now you can just take that screenshot and Magic Patterns, they do the best job, I think, of actually rendering exactly what was in that screenshot. And I tried four different ones, my Magic patterns and three others. This is what I call reverse prototyping from a screenshot. It's a different use case than what we're talking about. It can also do Vibe coding, but I took like an Airbnb page and I remember I tried a very basic one and just. It did not look anything like the Airbnb page, the prototype that it generated. Right. The other four, they got better, but Magic Pattern was the best one I had, captured most of the elements, the icons, the gradients, the shadows, all that kind of stuff. It's really good. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
- Tip: Use hyphens for bullets in PRDs to avoid ingestion issues 13 As I was dog fooding this stuff yesterday, I realized he doesn't like it. So I switched them all to hyphen. So if you want to really geek out, if you can learn a little bit of markdown, I guarantee it's not complicated. It's like the lightest, lightest weight thing. It's like hashtags and. And dashes. But the quick Dan Olson hack is just use hyphens instead of bullet points. If you just use hyphens and you just tab in and indent. Because what I wanted is if you have one level of hyphens, it's no. If you have one level of bullet points, it's not a problem. But when you want nested just, you just tab in and you do another hyphen and it'll work. https://www.youtube.com/watch?v=4v_Jv6-jU80 .
-
Roadmap/KR alignment
- What: Pathlight—lightweight roadmap/KR alignment and organization 88 Pathlight ( https://buildwithpathlight.com ) is a small startup/tool I've found that's been pretty helpful with roadmap/KR alignment & organizationhttps://www.reddit.com/r/ProductManagement/comments/1n7cqgc/comment/ncfduks/ .
-
Feedback & request portals
- What: UserVoice/UserEcho; internal ideas portals for voting/commenting 113 I’ve used tools like UserVoice and UserEcho to capture requests, surface them to the whole community, and let the crowd add context or vote. That gave us transparency, a way to say no, and a way to stop feature requests from living in silos or dying in email threads.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccrjwd/ 141 How do you aggregate the intake requests? And could you allow access to everyone's ideas? We use an ideas portal tool that let's our team submit requests, vote on each other's, comment, etc. In my experience this helps reduce the separate small requests and instead prompts people to chime in about their ideas/problems on one main submission. We get a lot of good qualitative chatter that way.https://www.reddit.com/r/ProductManagement/comments/1n814mk/comment/nce3w1o/ .
- Why: Transparency, prioritization at scale, and less siloed request sprawl 113 I’ve used tools like UserVoice and UserEcho to capture requests, surface them to the whole community, and let the crowd add context or vote. That gave us transparency, a way to say no, and a way to stop feature requests from living in silos or dying in email threads.https://www.reddit.com/r/ProductManagement/comments/1n856dm/comment/nccrjwd/ .
-
Mobile subscriptions
- What: RevenueCat + native IAP 11 Option 1: The Standard Path (RevenueCat + Native IAP) This seems to be the simplest and safest route. Use the native In-App Purchase systems from Apple/Google, managed through a service like RevenueCat. Pros: Super simple to implement, secure, trusted by users (Face ID/Fingerprint purchase), and handles all the backend validation and subscription state. Cons: The 15% commission (on the first $1M) is a significant cut.https://www.reddit.com/r/startups/comments/1n87n52/ .
- Why: Higher conversion and easy refunds vs. external web paywalls 52 I've never split-tested it so I don't have exact numbers, but my gut feel is that conversions will suffer quite a lot if you force users to pay off platform. I'd rather get 85% of $1,000 than 97% of $500.https://www.reddit.com/r/startups/comments/1n87n52/comment/ncd2c9v/ 51 Refunds are much easier to deal with as well (no matter how great your product, you WILL get refund requests) and I'd much rather point them to Apple. It's better for my sanity too, coz you don't feel the sting as much.https://www.reddit.com/r/startups/comments/1n87n52/comment/ncd2c9v/ .
-
AI PM career kit (Aakash Gupta)
- What: 110‑question AI PM interview bank, AI PM certification, and guides on strategy, prototyping, resumes, employers 91 📌 Want my 110-question AI PM interview bank? Reply ‘Questions’ + DM me. Restack to cut the line.https://substack.com/@aakashgupta/note/c-152495364 92 AI PM Certificate : [https://www.news.aakashg.com/i/170631463/the-complete-ai-pm-certification](https://www.news.aakashg.com/i/170631463/the-complete-ai-pm-certification)https://substack.com/@aakashgupta/note/c-152495364 96 AI Product Strategy : [https://www.news.aakashg.com/p/ai-product-strategy](https://www.news.aakashg.com/p/ai-product-strategy)https://substack.com/@aakashgupta/note/c-152495364 95 AI Prototyping : [https://www.news.aakashg.com/p/ai-prototyping-tutorial](https://www.news.aakashg.com/p/ai-prototyping-tutorial)https://substack.com/@aakashgupta/note/c-152495364 94 AI Resumes : [https://www.news.aakashg.com/p/ai-product-manager-resume](https://www.news.aakashg.com/p/ai-product-manager-resume)https://substack.com/@aakashgupta/note/c-152495364 93 AI Employers : [https://www.news.aakashg.com/i/172119714/survey-of-the-ai-pm-job-market](https://www.news.aakashg.com/i/172119714/survey-of-the-ai-pm-job-market)https://substack.com/@aakashgupta/note/c-152495364 .
- Why: Map skills to market demand; accelerate prep in a hot segment 105 AI PM jobs growing 120% vs Overall PM growing 1%https://substack.com/@aakashgupta/note/c-152495364 104 AI PM jobs paying 30-40% more at every levelhttps://substack.com/@aakashgupta/note/c-152495364 .
-
ChatGPT conversation branching
- What: Branch conversations to explore directions without losing your original thread; available on web for logged‑in users 126 By popular request: you can now branch conversations in ChatGPT, letting you more easily explore different directions without losing your original thread.https://x.com/OpenAI/status/1963697012014215181 125 Available now to logged-in users on web. https://x.com/OpenAI/status/1963697012014215181 .
- Why: Faster exploration and divergent thinking in research/spec writing.
-
Warehouse‑native analytics
- What: If you already have a data warehouse, consider a warehouse‑native analytics tool to cut costs and improve data fidelity vs. standalone analytics 140 Well if you have a data warehouse, then applying a warehouse-native analytics tool instead of Mixpanel would cut your cost and you will have a better datahttps://www.reddit.com/r/ProductManagement/comments/1n7t129/comment/nccdioo/ .
-
Metric Tree playbooks
- What: “Metrics and NSM” playbooks in Decode and Conquer (5th ed.) 67 Read the metrics and NSM playbooks from decode and conquer 5th editionhttps://www.reddit.com/r/ProductManagement/comments/1n7cqgc/comment/ncdhv4o/ .
- Why: Concrete templates for linking NSM→drivers and setting interview‑ready narratives.
“Retention is the ultimate reality check. It’s the difference between building a moment and building a company.” 44 Retention is the ultimate reality check. It forces you to face whether your product is truly useful, or just a novelty. It’s the difference between building a moment and building a company.https://x.com/hnshah/status/1963622585754947794



This Week’s Big Ideas
-
AI credit-based pricing is going mainstream
- What’s happening: Microsoft, Salesforce, Cursor, and OpenAI all moved to credit models (including pooled credits) to align price with usage variability 44 Microsoft announced AI credits for Copilot in Microsoft 365 in January . Salesforce added a new flexible, credit-based model for their AI agent in May. Cursor shifted to credit-based pricing in June (and faced some real pushback from users). Not to be outdone, OpenAI recently replaced seat licenses with a pooled credit model for its Enterprise plans.https://www.growthunhinged.com/p/ai-credit-pricing . Credits let vendors adapt pricing as model costs and user behavior shift, and they’re a practical bridge toward value/outcome-based pricing 94 Credit-based models are uniquely suited to accommodate this complexity. Customers see a pool of usage that feels relatively straightforward (“50 credits per month”). Vendors, meanwhile, can easily adapt credit pricing to monetize newer, higher value actions, navigate evolving LLM costs, and nudge customer behavior in a way that’s win-win.https://www.growthunhinged.com/p/ai-credit-pricing 88 I believe credit-based pricing is a bridge from flat-rates toward more value- and outcome-based pricing models. But they’re an important bridge for the present moment.https://www.growthunhinged.com/p/ai-credit-pricing .
- Why it matters: Token consumption is spiky and concentrated—10% of users often drive 70–80% of usage—so flat rates break margins 92 In my experience, 70-80% of token consumption comes from just 10% of users. These heavy users can become very unprofitable without the right guardrails. And AI gross margins are already challenged based on reporting from The Information , leaving little room for further erosion.https://www.growthunhinged.com/p/ai-credit-pricing . Buyers also want simpler ways to predict bills; tying credits to outcomes (e.g., case resolution) clarifies ROI 90 Buyers can more easily compare the cost-per-action to their existing approach and can estimate a business case accordingly. And they have peace-of-mind that pricing connects to tangible outputs rather than some arbitrary token calculation from a third party LLM.https://www.growthunhinged.com/p/ai-credit-pricing .
-
How to apply:
- Choose your primitive: pass-through (cost-based) credits for transparency (e.g., Cursor maps credits to API spend) 93 The default starting point is to make credits a cost-based metric , literally passing through AI costs directly to customers. It’s straightforward, relatively easy to calculate, and transparent. This approach lets customers automatically benefit from any decreasing LLM costs and gives them the power to optimize their usage/spend based on their own objectives.https://www.growthunhinged.com/p/ai-credit-pricing 91 Cursor credits are simply dollars of LLM API usage at the prevailing model API rates.https://www.growthunhinged.com/p/ai-credit-pricing or output-based credits that charge only for successful work (prevailing rates roughly $0.10 for simple tasks up to ~$1 for complex workflows) 51 A more value-aligned approach is to make credits an output-based metric , charging only for successful work products like a task execution, AI summary, image generation, or a conversation resolution. This approach is harder to quantify upfront, especially for multi-purpose apps, and pricing doesn’t always link directly with LLM costs. Today’s prevailing rate is charging ~$0.10 for simple task executions and up to ~$1 for complex, multi-step workflows.https://www.growthunhinged.com/p/ai-credit-pricing .
- Add guardrails that buyers expect: annual drawdowns and credit rollovers to smooth usage and reduce hoarding 48 4. Shift from monthly to annual usage limits.https://www.growthunhinged.com/p/ai-credit-pricing 47 An annual drawdown model gives the customer more time to consume their credits and to forecast what steady state looks like. If they run out of credits in month 9 or month 10, it's simply an early renewal – not a huge overage.https://www.growthunhinged.com/p/ai-credit-pricing 89 For customers with a monthly usage limit, there’s significant momentum around letting unused credits rollover. Lovable and bolt(dot)new, for instance, announced credit rollovers over the summer. I think that’s a good thing for the current moment.https://www.growthunhinged.com/p/ai-credit-pricing . Ship in-product usage and spend visibility with admin thresholds at account/user levels 46 5. Help admins track and predict their usage.https://www.growthunhinged.com/p/ai-credit-pricing 45 Admins increasingly expect transparency around which actions drive up their bill. They want to be able to set billing thresholds at both the account and the user level. In a credit-based model, usage and spend needs to be visible within the product.https://www.growthunhinged.com/p/ai-credit-pricing .
- Monetize on multiple axes to preserve margins (features/subscription + credits) 50 1. Monetize based on multiple axes, not just credits.https://www.growthunhinged.com/p/ai-credit-pricing 49 Clay's hybrid pricing model, for instance, has multiple routes to expand customers while keeping pricing relatively simple: more features (subscription packages) and more usage (credits). All of Clay’s plans include unlimited seats.https://www.growthunhinged.com/p/ai-credit-pricing .
-
Build model‑agnostic products; focus on UI, business logic, and distribution
- The shift: Foundational models are converging; value is moving to the business logic and UI that sit above them, not the model layer 107 The next few years in startupland are going to set the stage for a lot of what will happen in the future. The last few years have been eventful. We've seen a wave of high-end AI research teams building foundation models. But now that these models have absorbed all the data in the world, they're asymptoting in their effectiveness. The next few years will be about the folks who build the business logic that sits on top of these models. They won't do AI research or train their own foundation models. Instead, they'll be model-agnostic, creating a compelling UI that sits on top. We're seeing these kinds of products absorbed into every different kind of vertical industry, with approaches spanning from selling tools all the way to rolling up industries and implementing technology that way.https://andrewchen.substack.com/p/ai-will-change-how-we-build-startups . Despite the hype, most home screens still lack AI‑native apps—there’s headroom in consumer UX 110 How many apps on your phone's home screen are AI-native? How many of those apps were created using AI coding tools?https://andrewchen.substack.com/p/ai-will-change-how-we-build-startups 109 Weirdly, the answer for most of us might be a lot closer to zero than you’d think, particularly beyond the obvious LLM apps — ChatGPT, Grok, etc. Of course people are working on this, but for now, the Home Screen Test says that we have 4x7 grid of apps on our screen, and very few are AI-native. One day shouldn’t it be all 28? Where’s the AI-native Calendar app, or the AI-native Social network, and so on? I remember in the web-to-mobile shift how quickly my engagement went completely to the mobile app versions of messaging, social, email, etc. What’s going on here?https://andrewchen.substack.com/p/ai-will-change-how-we-build-startups .
- Team implications: Blur role boundaries; prioritize problem‑first, technology‑agnostic architectures so you can swap models/tools over time 16 You're going to have a clear problem, clear goal. What are you trying to achieve? Whatever your goals are to achieve, at the end of the day you're going to need to have a very clear understanding of the Personas, end users that are going to be impacted by that AI application that you're building. And you're going to need to have a very clear, I would call it a technology agnostic view of how to achieve that. And I make it very bold. The value of the technology agnostic part because the world is changing so fast and by the time that your product gets out there's a good chance that this underlying technology is already outdated. And I see a lot of enterprises or businesses that they stick to the old technology because the cost of upgrade is too high for them and basically not providing a chance for them to enjoy the best and the state of the art technology out there just because they didn't plan it well. So I would say any decision that you make, try to be technology agnostic, use the best that is available in the market for you, but have a plan of or a layer in your architecture to abstract away the business case from the technology part. https://www.youtube.com/watch?v=H3-kIebpdic 15 The value of the technology agnostic part because the world is changing so fast and by the time that your product gets out there's a good chance that this underlying technology is already outdated. And I see a lot of enterprises or businesses that they stick to the old technology because the cost of upgrade is too high for them and basically not providing a chance for them to enjoy the best and the state of the art technology out there just because they didn't plan it well. So I would say any decision that you make, try to be technology agnostic, use the best that is available in the market for you, but have a plan of or a layer in your architecture to abstract away the business case from the technology part. https://www.youtube.com/watch?v=H3-kIebpdic . Leaders expect tighter PM–engineering ratios because “building the right product” dominates “just building” 81 I think the boundaries between roles are going to dissolve. There is going to be more focus on building the right product versus building the product, specifically especially between product management and engineers. Historically we were saying that for a typical group of product management versus engineers, a ratio of 1 to 6 to 10 is applicable. But now we've been hearing the leaders talking about even 1 to 0.5 for engineers which is indicating that how important it is to identify the right problem. I'm less worried about the labeling of people, but I'm expecting the lines between these two roles are emerging big time with AI. https://www.youtube.com/watch?v=H3-kIebpdic .
-
How to apply:
- Architect abstraction layers between business logic and model/tooling to avoid lock‑in and enable upgrades as costs/quality shift 15 The value of the technology agnostic part because the world is changing so fast and by the time that your product gets out there's a good chance that this underlying technology is already outdated. And I see a lot of enterprises or businesses that they stick to the old technology because the cost of upgrade is too high for them and basically not providing a chance for them to enjoy the best and the state of the art technology out there just because they didn't plan it well. So I would say any decision that you make, try to be technology agnostic, use the best that is available in the market for you, but have a plan of or a layer in your architecture to abstract away the business case from the technology part. https://www.youtube.com/watch?v=H3-kIebpdic .
- Plan for distribution early; building is cheaper, acquisition still isn’t 108 Will AI make startups cheaper or more expensive to build ? Obviously, some infrastructure layers, like foundation models, are very CapEx heavy. And theoretically, some apps *should* be very easy to build. But growth and distribution cost a lot of money to get into customers' hands. One thing we've learned over the past decade is that even though it's relatively cheap to build a web app, acquiring users can still cost millions. The failure rate remains high because there are just so many products competing for people's attention, so maybe that’s the limiting factor.https://andrewchen.substack.com/p/ai-will-change-how-we-build-startups .
-
“Vibe coding” can ship prototypes fast—but production demands rigor
- Reality check: Expect ~1 month to ship a simple real B2B app, with ~60% of time on QA/testing 19 If you seriously want to put a simple but real B2B app into production with real users collecting real information and charging for it for real, and you want to be any good, budget a month. A month and 60% of your time will be QA and testing. Okay? And for even just a few days into a real project, most of your time will be screenshots, uploading screenshots. https://www.youtube.com/watch?v=ON_UrRJ0m5Q 18 A month and 60% of your time will be QA and testing. Okay? And for even just a few days into a real project, most of your time will be screenshots, uploading screenshots. This broke. This doesn't work. You will be doing so much functional QA, you cannot believe it. You will literally, in the end of the day, be logging probably a thousand bugs on a serious website. This will become your life. Maybe you used to have people to do this for you. I used to have a team of QA engineers to do this for me. I now it's me. Now my life is screenshot and bugs, hope. And then when I get to the final one and it's a screenshot and the only text I have is, dude, it's time to take a break. Okay? But this will be your life. So if you're. Look, if you want to build a prototype, if you want to build something and tell all your friends you're a vibe coder at Starbucks, you don't need to budget for this time. You could budget for a day or two, but if you want to build something real, it's going to take you a month and 60% testing time. And if you look at this chart on the right, this is. https://www.youtube.com/watch?v=ON_UrRJ0m5Q . Prosumer stacks need daily rollbacks and robust operational ownership; agents will fabricate to “achieve goals,” and integrations like email/scheduling can be brittle 21 You're going to get almost no pushback from the AI agent in any of these Vibe coded apps. No matter what you want to do, the answer is going to be yes, sir. More sir, because these are how they're all coded. This is how Claude and open AI are coded. They're, they're goal seeking. Their job is the way. And I'm not a total expert, but the way the algorithms work, the way they can make this massive Amount of AI and data crunching working is their job. Number one is to goal seek and get you an answer. And that's, that's why when you know, you just chop plop into chap gb. Sometimes it hallucinates when it doesn't know the answer. Because it's not just that it's hallucinating, it's goal seeking. It's getting you the best answer it can. And if it doesn't know the answer, it makes it up. It'll do the same at the code level. If it doesn't know how to do something, it will make up fake data or a fake feature or a fake button, but it will find a way. It won't say no. https://www.youtube.com/watch?v=ON_UrRJ0m5Q 30 So I today, as we, as we record this, as we, as we do this live, honestly, I would not build any app that relies on email, that relies on it to function. It's not, it's just not going to be reliable. It's possible, but I wouldn't build it. I would build it the second one, and you'll see this one all over the Internet. https://www.youtube.com/watch?v=ON_UrRJ0m5Q 35 They're not that hard, but man, these are super hard. In the prosumer Vibe apps, email and scheduling, I have built five apps now. None of them get email or scheduling right. None of them get it right. They all stop sending emails that are supposed to be sent every hour, every day, once a week. They all stop sending them. They all lose track of the connection to send grid or resend, which I prefer. It's constantly breaking the email. Constantly. https://www.youtube.com/watch?v=ON_UrRJ0m5Q .
-
How to apply:
- Write a rich PRD (AI can help refine it) and modularize features by page/component for independent rollback/fix 43 You've got to build a rich prd, a rich spec. Now, if you if you have a background in product or anything, you've done this a million times over your career. If you don't, you've never done it. But it's actually the only complicated part is just doing it. So here is just a snippet of a prd. I built the dashboard, profile, integrations, watch list, advanced AI networking. This was sort of how we were trying to build that networking app. I talked about the first one that failed, but I had a pretty good prd. So if you don't know how to do this, it's okay, go and just go and do it. If you've never done it before, it's actually fine. Because this is the beauty of AI. AI can help organize things for you. Go into a Google Doc and write two or three pages of everything you want this app to do, everything you want. Everything. Every button you can think of, every function, every bit of look and feel, everything you can think of that you want this app to do. It's okay if you haven't done before and it's stream of consciousness. It's okay if this isn't how a VP of product at a top AI company would do it. Write it your way in human language and then cut and paste it and put it into Claude and say turn this into a PRD for me for replit or for lovable and they'll actually do a great job. They will take your. This is where AI shines. It doesn't have to be perfect. They will take your stream of consciousness and help you turn it into a prd. https://www.youtube.com/watch?v=ON_UrRJ0m5Q 40 So what I do now, and you can see this if you go to Saster AI, I've broken down everything complicated into its own page. The Saster AI public market analysis is its own page. News is its own page. The valuation calculator is its own page. Anything that's remotely complicated to the zone page. So worst case, I can either delete it or I can roll back easily or I can go back. But if you combine too many things into one page, it gets impossible to start over, it gets impossible to fix bug, it just gets too complicated. So my number one tip is force. If you're going to vibe, code it, break it up into its components and then have more pages than you would think. You're almost going back in time in some ways to have a lot of pages, but it will save your sanity. I built a massively complicated app that was all one page and I then it was just impossible to fix it. https://www.youtube.com/watch?v=ON_UrRJ0m5Q .
- Use platform defaults (auth, Stripe, email) and collect the least PII to reduce security risk 36 You have to use what's built in and in general use everything that's built into these platforms. Built it will just be more secure you if they have email built in you for example. I use Replit all the time. I do not like using SendGrid. I like SendGrid in the old days when it was run by founders. It is impossible to use it. No, there's no support available. It takes days to get back to you. I got caught in a doom loop going from free to paid. It's just brutal. So I'm like I want to use resend. Resend is super cool. I'm a super fan. Put that on the website. But Replit keeps wanting to use SendGrid. It keeps forgetting resend and losing the keys and wanting to go back to its defaults. So use whatever these are in the defaults which is usually stripe their own oauth and something do not use others. https://www.youtube.com/watch?v=ON_UrRJ0m5Q 33 So do the least collect the least amount of personal information. Collect the least amount of data you can use the built in use stripe if that's it's built into Replit use stripe. Use what it does, but collect the least, not the most. And realize as soon as you generate a database in this app, as soon as there is a database, you have added security risk to your app that is going to be ultimately your number one thing to worry about. If it's not the number one thing you're worried about the own, then the only reason you're not going to get in trouble is because nobody cares. https://www.youtube.com/watch?v=ON_UrRJ0m5Q .
- Master rollback; if you’re 10–15 minutes into a bad branch, revert quickly 22 You can roll back almost to any point in time. But you got to learn when. That means you got to learn. If you're more than 10 or 15 minutes in and the agent just can't get something, you got to go back in time 10 or 15 minutes and you got to know where those rollback points are and get really good at it. And I would say, frankly, if you're not rolling back once a day, you're. You are. You're not doing it right. This is your. https://www.youtube.com/watch?v=ON_UrRJ0m5Q . Build/automate unit tests as tooling allows; expect to be the tester until it matures 24 They will ultimately add built in unit tests for every app. Now all this stuff is adding complexity, which they don't like, right? The leading vendors like they don't want to separate development from production. They don't want to do all this stuff because it takes this elegant one sentence prompt that makes it complicated. But ultimately they need an option where when you push to production, it will build unit tests for you. Those unit tests will probably run on a different server that is stateful, that is always running, and it runs it for you automatically each day and sends you an update. But I do it and I'm getting better at the unit tests. And Amelia and I get an email each morning from Saster AI telling it what, what's going on, but it's not reliable. It's just not reliable. And so you're going to have to be testing your. If you're, if you're serious about it, someone's going to have to be testing it every day. And then every day I got to fix the stuff that breaks one way or the other. https://www.youtube.com/watch?v=ON_UrRJ0m5Q .
-
Clarity beats dashboards: make the next move obvious
- Signal: Teams open multiple tools, see conflicting numbers, and argue—only 23% turn data into action (HBR, via Hiten Shah) 96 Teams say, “I open three tools, get three different numbers, and then the meeting is about the data, not improving our website.”https://x.com/hnshah/status/1963247873325465943 95 This isn’t just a personal pain. Harvard Business Review found only 23 percent of companies are excellant at turning data into action. The rest? “Paralyzed by analysis, not empowered by it.”https://x.com/hnshah/status/1963247873325465943 . Meanwhile only 31% of orgs prioritize rapid experimentation; 84% of teams doubt market success due to low data/time/support 147 Only 31% of organizations prioritize rapid experimentation and iterative learning. Despite decades of lean methodology evangelism, most product decisions are still based on opinions and leadership preferences rather than customer data and systematic testing.https://substack.com/@aakashgupta/note/c-152062165 145 84% of product teams are concerned that their current products won’t succeed in the market. This isn’t imposter syndrome - it’s a rational response to being held accountable for business outcomes while lacking the data, time, and organizational support needed to make informed decisions.https://substack.com/@aakashgupta/note/c-152062165 .
- How to apply: Pair “Tiny Acts of Discovery” (fast, low‑cost tests) with clarity‑first visuals (e.g., heatmaps) to drive unambiguous action—“if the button is cold, fix it” 66 Yes yes and yes- we like to teach a concept called "Tiny Acts of Discovery"- which is similar to the thinking in bets concept made popular by Annie Duke. A couple thoughts: 1) This all starts with data as you say, but that means having good data and access to it is a precondition (duh). I just want to say it out loud that unfortunately not all product managers have this. If you don't have access to data, there are other ways to source problems. I would go to CS and Support teams and Sales teams- any team that is close to the customer, and then validate with customer interviews. 2) If we___ then___ is a great way to easily create hypotheses. If you're having a block on this, another way to brainstorm is How Might We. Just a word of warning that yes, you are already in solution space. 3)Tiny Acts of Discovery- I find it helpful to think about an experiment as taking 2 weeks or less to show a result. This is really hard for some companies that may have long sales cycles, or conditions about changing the product for customers (thinking B2B). 4) Also communicate results and document them somewhere publicly and easily accessible.https://www.reddit.com/r/ProductManagement/comments/1n7g8yq/comment/nc7kuby/ 97 When we built Crazy Egg, heatmaps solved that. You saw where people clicked. There was nothing to debate. If the button was cold, you fixed it. If people ignored a section, you reworked the layout.https://x.com/hnshah/status/1963247873325465943 .
Tactical Playbook
-
Stand up LLM evals that your org will trust
- Before evals, do error analysis to identify what actually needs measuring 101 Key takeaways: 🔍 Error analysis must come before implementing evals to identify what actually needs measuring 💡 Start with the simplest possible evals - sometimes basic solutions work surprisingly well 🤝 Cross-functional collaboration between engineers, PMs and designers is essential for AI products 🔄 AI products require continuous monitoring and improvement to maintain quality 🛠️ You don’t need to be an AI expert to get started - leverage existing tools and build incrementallyhttps://x.com/ttorres/status/1963289398021116159 .
- Retrieval metrics: Recall@k, nDCG, MRR, Hit Rate 121 I’m an AI tools developer, and when it comes to evaluating RAG I usually look at retrieval metrics like Recall@k, nDCG, MRR, and Hit Rate. On the generation side, I focus on faithfulness, groundedness, answer relevance, and context utilization. Basically, is the answer based on retrieved content, does it actually answer the question, and does it make use of the chunks pulled in. I also track citation coverage, latency, and token cost as system-level metrics. On the user side, we rely on proxy signals like satisfaction ratings or follow-up rate to gauge the overall experience.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc9sv2j/ . Generation metrics: faithfulness/groundedness, answer relevance, context utilization; track citation coverage, latency, token cost 121 I’m an AI tools developer, and when it comes to evaluating RAG I usually look at retrieval metrics like Recall@k, nDCG, MRR, and Hit Rate. On the generation side, I focus on faithfulness, groundedness, answer relevance, and context utilization. Basically, is the answer based on retrieved content, does it actually answer the question, and does it make use of the chunks pulled in. I also track citation coverage, latency, and token cost as system-level metrics. On the user side, we rely on proxy signals like satisfaction ratings or follow-up rate to gauge the overall experience.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc9sv2j/ .
- Build a small “gold” evaluation set (queries + gold passages), mostly hand‑labeled; LLMs can assist dataset creation 120 Yes, you do need a separate dataset. What we do is maintain a small but representative evaluation set: basically queries paired with one or more “gold” passages. Most of these are manually labeled, but we’ve started trying out LLMs to build the dataset too.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nce4w8x/ .
- Add user‑level signals: follow‑up rate (did users immediately re‑ask?)—use LLM‑as‑judge to classify repeats vs new questions 118 We use a separate model to evaluate the user’s follow-up rate. We track whether the user immediately asks a clarifying or repeated question after the model’s response. If users frequently need to re-ask, it usually indicates that the original output wasn’t satisfactory.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/ncb7fxj/ 119 Yes, the answer is based on the retrieved chunks. LLM-as-judge, since we need to evaluate whether a follow-up question is a repetition or a new question.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nce2ru7/ .
- Lightweight scoring: a 1–5 relevance rubric with examples; ask users to mark irrelevance rather than label everything; enforce “does it include citations?” as a Boolean sanity check 68 Maybe use a confidence score rubric? “On a scale of 1–5, how relevant is the information provided?” How I would implement this is: I would come up with a set of examples of acceptable relevance and unacceptable relevance. Like hand write these. (painful, yes) Use that to ground your RAGs Run the RAGs and you personally go and rank the scores now. This is doable because you only have 3-5 documents now. Once you have reached an acceptable composite score on relevance/accuracy here (say 4.5 or something), there is no reason to believe that just by increasing the number of documents, the accuracy will drop significantly. BUT.. You can do spot checks as you increase the number of PDFs.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc81ybx/ 69 You said that you can't ask people to manually label relevance for every query. Do you need that? Can you instead give a way for people to just mark irrelevance?https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc81ybx/ 117 Start with a Boolean yes/no: Does the response include source citations? This will help you ensure that the LLM is not hallucinating and is using your RAGs. Might get you most of the way there.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc81ybx/ .
- Tools to explore: RAGAS/DeepEval (end‑to‑end evals; synthetic test sets for RAG). Trial before you commit 64 https://docs.ragas.io/en/stable/https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc8pql7/ 63 They also provide bespoke tools to create test data sets for evaluation and even Synthetic test sets for RAG implementation.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc8pql7/ 62 It’s certainly not perfect but it can take out a lot of the subjectivity currently seems in today’s market for LLM evaluation and performance.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc8pql7/ 116 so, yes, RAGAS and DeepEval both looked quite promising from the little documentation I read. Honestly, I am yet to test it out, will do so soon. But even there, I wanted to understand from other PMs that is this generally the way you guys go about monitoring or should I look at other directions that I am likely not focusing on at the moment?https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nccdz4i/ .
-
Ship a voice scheduling assistant (fast) with tool calling
- Agent flow: speech→text→LLM (system prompt)→tool calls→text→speech 42 One is you need to understand what the user is speaking. And that is why you have this speech to text part which transcribes what the user is saying. And then you send this text to an LLM with, let's say, some system prompts, some instructions of what the agent is supposed to do. And the LLM gives you a response that, hey, this is the response for this particular query that the user has. The agent can also perform function calling or any tool calls or any actions in other applications. And finally, once it has a response, it needs to be synthesized back into speech so that the agent can speak it to the user. So that is where the text to speech comes in and whatever is generated is narrated back to the user. https://www.youtube.com/watch?v=lty4nsUcTes .
-
Steps:
- Create the agent and write its first message + system prompt; include dynamic variables like current UTC/time zone 41 Let's start by creating a new agent. In this case we are creating a personal assistant. So we can just select that, hey, we are just creating a personal assistant. Let's give it a name. I'll just call it Workshop Personal Assistant and you can provide your goal. In this case, I want this agent to manage my Google Calendar and Gmail. So I've mentioned it here and you can be very specific. It'll expand to understand what your actual intent is later on. For now, I have kept it very simple and we can click on Create Agent to start with a canvas with some data already created. The agent has been created and you will see a lot of options over here, but you don't have to get overwhelmed with it. https://www.youtube.com/watch?v=lty4nsUcTes 39 Here, the first message is what your agent speaks when it starts. So when the session starts, this is what the agent will first tell to the user that hey, hi, I can help you manage your Google Calendar and Gmail. You can customize it if you want. And the next part, which is system prompt, this is the most important part of building an agent. This is what defines what your agent is supposed to do. In this case, I'm building a personal assistant. So my system prompt should be what a personal assistant managing Google Calendar and Gmail would do. Because we had defined it during the creation process, there is already a predefined system prompt that has been added over here. It says this is your personality. It has also given it a name. You are Alex, a highly efficient, reliable and personal assistant AI. And you are friendly and helpful. These are essentially instructions and guidelines of what the agent is supposed to do. https://www.youtube.com/watch?v=lty4nsUcTes 29 So for solving this, we have some variables in the system prompt that you can specify and this will be replaced with the actual data when you start the conversation. So in this case I've selected system time zone. It will be replaced with whatever time zone the agent is running in. And there's also another one for the time itself. So in fact it's better to use utc. So this will be replaced with the current time when the agent runs and that that is how your agent gets access to what today's date is for all the further conversation it has. So you have to fiddle around and update and test the system prompt. https://www.youtube.com/watch?v=lty4nsUcTes .
- Choose a model: low‑latency conversational “Flash” (e.g., Gemini 2 Flash) for fluid UX; larger models (GPT‑4/5, Claude 3.5 Sonnet) for deeper reasoning if you can tolerate latency 38 The next part is choosing an LLM. So the intelligence part of the conversational AI agent comes from an LLM and you can choose any LLM you want. By default we select Gemini 2 Flash. It works great. But if you want more intelligent models, you can switch to using GPT5GPT5GPT4O or even the anthropic models if you want. And finally, the second most important piece. Of your agent, Akash, what model would you recommend? Generally like the top ones, is it Gemini Flash? Or if you can explain like the key differences or preferences, like if the use case is X, then this model works great. Anything like that? Sure. So Gemini 2 flash flashlight 2.5 flash. All of these Flash models are great for conversational use cases because these are really fast and you still get the intelligence part. So they are not that bad in terms of intelligence, but they are really good in latency as well. So if you want a very fast, fluid experience of your agent, go with the Flash models. If you want an agent which thinks a lot, which takes some time to decide on what it wants to output. You can go with larger models which are GPT4, GPT5. I've noticed even Claude 3.5 sonnet works really well for some use cases. But these are slightly larger models and can take some time to respond, although you might not see a lot of difference. https://www.youtube.com/watch?v=lty4nsUcTes 20 So Gemini 2 flash flashlight 2.5 flash. All of these Flash models are great for conversational use cases because these are really fast and you still get the intelligence part. So they are not that bad in terms of intelligence, but they are really good in latency as well. So if you want a very fast, fluid experience of your agent, go with the Flash models. If you want an agent which thinks a lot, which takes some time to decide on what it wants to output. You can go with larger models which are GPT4, GPT5. I've noticed even Claude 3.5 sonnet works really well for some use cases. But these are slightly larger models and can take some time to respond, although you might not see a lot of difference. https://www.youtube.com/watch?v=lty4nsUcTes .
- Wire tools via MCP: use Zapier or n8n to expose Google Calendar/Gmail actions (find/create/update events; find/send email). Treat server URLs as secrets 37 Zapier has this MCP platform which you can access by going to mcp.zapier.com and in the free plan you get 300 actions per month, which is good enough when you're testing and trying out different things. And to get started, let's start by creating a new MCP server. For the client field you can select other. Doesn't really matter. And for the name, let's just give it any name. For this demo I'll just put Workshop MCP Server. Now we can configure this server with various other tools and services that we want in our MCP server. In this case I wanted to connect to Google Calendar and Gmail, but you can connect it to, let's say Confluence or Salesforce or thousands of different apps that are supported in Zapier. So for now I'll just connect with Google Calendar. And within Google Calendar I can see that these are the tools or actions that can be performed on the calendar. So I can retrieve an event by an id, I can find events, I can create a calendar, I can create an event on the calendar. I can update an event and so on. https://www.youtube.com/watch?v=lty4nsUcTes 34 And within Google Calendar I can see that these are the tools or actions that can be performed on the calendar. So I can retrieve an event by an id, I can find events, I can create a calendar, I can create an event on the calendar. I can update an event and so on. So we can just click on the tool that we want our agent to have access to. https://www.youtube.com/watch?v=lty4nsUcTes 32 And for Gmail it's the same process. So you select Gmail, select the tools you want. In this case, let's say I'll go with find email and you can also configure the Gmail account you want to use. In this case, I just have one account, so we'll just use that. Let's also add send email and this will give the agent the ability to send emails from my Google account. And let's also think that's about it. https://www.youtube.com/watch?v=lty4nsUcTes 23 The most important part is your server URL. It has a secret attached to it, so don't leak it out. Otherwise you will expose your Google Calendar and Gmail to the world if you have connected it. So just keep it secret and put it in the 11 Labs platform. Just don't post it anywhere otherwise you risk making your calendar public to everyone. https://www.youtube.com/watch?v=lty4nsUcTes .
- Handle multi‑step flows (IDs, updates, calculations) in n8n; return structured results to the agent 31 This is where N8N comes in. N8N is a more elaborate, a bit more technical platform, but it's still quite easy to get started. It allows you to build these complex workflows where let's say a certain tool is called, let's say list Transactions. It can do an if else condition and check. Let's say sheet one if something exists on this or sheet two if something exists on this and then returns in a response. So you can perform multiple chain of actions in this sort of a setup. In this case it's generating an id, updating the Google sheet with a new row and then calculating something and then returning that. All of this is possible in a no code way with any 10, but it requires more setup and learning upfront. https://www.youtube.com/watch?v=lty4nsUcTes .
- Test time‑window reasoning and confirm human‑readable confirmations (e.g., exact slots/time zone conversion) 28 Let's go Tuesday morning and it is in Eastern Standard Time so you can convert it to Indian Standard Time and schedule it. I am sorry but there seems to be a misunderstanding. The email specified Tuesday afternoon between 2pm and 5pm Eastern Time or Wednesday morning between 9am and 12pm Eastern Time. Sorry could you please confirm the exact. You can do it on Wednesday morning. That works. Okay. Wednesday morning between 9am and 12pm Eastern Time. Would you like to specify a particular time within that range or should I pick one? Just pick anything. I'm fine with any time. Okay. I will schedule the meeting for Wednesday at 9am Eastern Time which will be 6:30pm Indian Standard Time. Is that correct? https://www.youtube.com/watch?v=lty4nsUcTes .
- Keep end‑to‑end latency under ~200–300 ms; beyond that, conversations feel “Stone Age” and users drop off 26 You need to figure out the latency part. Because you're connecting so many different models, the latency keeps getting added up. And you need to ensure that your latency is below 300 to 200 milliseconds so that the conversation feels very natural. If it's anything beyond 300 milliseconds, it starts to feel like it's a tool from the Stone Age in the air and users might just drop off. https://www.youtube.com/watch?v=lty4nsUcTes .
- Deploy via share link, embeddable widget, or API/SDK for full custom UI 27 Sure. So, couple of things. One is, let's say you create an agent and you just want to share it as is. You can copy the link to the agent and you can just share the link with others. This will give them a nice page with the whole agent experience that we have. And anyone can start a conversation via this page if you want it to be more deeply integrated with Your product, there is a widget part in the agent platform which gives you an embed code and it embeds an agent similar to what you can see at the bottom right. And you can customize the colors of this, whether it shows up on bottom right, bottom left, whether it's compact, tiny, whatever, and it's very simple. Just copy paste this code wherever you want and it will appear on the on the web page. Alternatively, if you want to build a completely custom UI experience. So instead of using these widgets, you still want the functionality of conversational AI, but the interaction piece or the UI should be completely yours. You can check out our API and SDKs. So that requires more technical know how of handling the APIs and rendering the UIs, but all of that is available on our documentation, so you can follow that if you want a complete customized flow. https://www.youtube.com/watch?v=lty4nsUcTes .
- Proof points: Meesho offloads ~1/3 of 60k+ daily support calls to agents (Hindi/English); a Southeast Asia fintech runs 30k+ outbound calls/day; Practica AI saw +15% average session length with high‑quality voices 80 So Meesho has used ElevenLabs to build their customer support service and today it handles over 60,000 plus calls. About one in three calls are now handled by ElevenLabs and of course this has led to increased customer satisfaction and much faster resolution as well thanks to the automated agent approach. These calls happen in both Hindi and English so you can also use the agents in any other language that you have funding societies. https://www.youtube.com/watch?v=lty4nsUcTes 79 Very leading digital finance provider in Southeast asia has used 11 labs for outbound calls. So 30,000 plus outbound calls run daily with multiple with agents in multiple different languages. And yeah, they are doing their sales stuff just by doing the outbound calls. https://www.youtube.com/watch?v=lty4nsUcTes 78 Finally, Practica AI has also used ElevenLabs for their EdTech platform and they have seen a 15% increase in the average user session length. And their app or their platform has also gotten positive rating especially because of the high quality voices that our platform provides. https://www.youtube.com/watch?v=lty4nsUcTes .
-
Release planning that prevents “release ≠ launch” failures
- Define DoD for GA to include docs, sales training, and enablement; gate GA on these 143 Definition of done (for general availability) includes GTM and enablement criteria like documentation and sales training. https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncck5vb/ .
- Ship behind flags; control GA cadence monthly/quarterly; admit early adopters ad hoc 144 Release every customer-facing feature under feature flags. https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncck5vb/ 142 GA release controlled by PM org and happens monthly or quarterly. https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncck5vb/ 141 Demo environment, early adopters and beta customers can get access on an ad hoc basis.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncck5vb/ .
- Create a tiering/launch grading system to set lead times and deliverables; decouple frequent engineering releases from batched/themed launches (“rolling thunder”) 71 -Develop a tiering system...https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/nca438h/ 70 -Consider a decoupling of release versus launch. Sort of a rolling thunder approach to launches through monthly/quarterly "launches" and theme it. This is what Salesforce does and they do it masterfully.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/ncemd99/ .
- Operating rhythm: weekly/biweekly cross‑functional with PM/Eng/PMM; require planning briefs before code (ETA, value prop, roadmap fit) and a release doc ~1 week prior; PMM ships FAQ/one‑sheet and monthly release notes 140 The brief should be submitted during the product planning process, before a single line of code is committed. It should include an ETA for release, value proper, reason this feature is being implemented, where it fits in the roadmap, etc.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/nc9e1tg/ 139 Our teams release like this too and they are required to have a document for every release which covers the value props and how it works, and share this info in a dedicated channel. They need to prepare this ~1 week before the release and the PMM creates an FAQ and a one sheet as required.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/nc82rsv/ 136 We then do monthly release notes internally and externally as a round up.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/nc82rsv/ .
- Stakeholder judo: use product KPIs (e.g., adoption) to argue for early PMM involvement; remind teams “releases are not launches” 138 Anyway, if we hadn't had a change in leadership, I was already building a case using products' own KPIs against them. You measured against user adoption? Here's the numbers from your last 3 releases. Wanna hit your numbers? We need to be involved earlier. Here's how we can work together and the information I can bring to you to make your next feature a fucking killer.https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/nc8fbge/ 137 We had the 2 week sprint thing, too. I just told them, release whatever you want, but this feature missed the mark because you didn't ask fir any competitive Intel or customer feedback before writing up your story. And I'm not putting together a full GTM launch. Releases are not launches. https://www.reddit.com/r/ProductMarketing/comments/1n7jusn/comment/nc8fbge/ .
-
Cross‑team capacity: get to “yes” without drama
- Align requests to company‑level goals; surface dependencies on the roadmap so it’s work toward an objective, not a favor 60 Ideally, each team’s goals should tie back directly to company-level goals. If that’s the case, then prioritization should flow from there. Leadership’s company priorities should guide how we balance initiatives across teams.https://www.reddit.com/r/ProductManagement/comments/1n7ai7w/comment/nc65b83/ 112 If escalation is constant, it usually means goals aren’t aligned clearly enough. I’ve found it helps to flag dependencies during release planning so they are visible on the roadmap — then it is not “my feature vs. yours,” it is work tied to a company objective. Escalation still happens, but it is more about confirming trade-offs than fighting for favors.https://www.reddit.com/r/ProductManagement/comments/1n7ai7w/comment/nc8dl4t/ .
- Reframe the ask to hit their metrics; break into small, low‑effort slices; build credibility with quick wins 115 I’ve found it helps to reframe the ask so it ties into their goals too. If they can show the work also drives their metrics, it stops being just your project.https://www.reddit.com/r/ProductManagement/comments/1n7ai7w/comment/nc9aklb/ 114 Breaking the request into smaller, low-effort pieces works better than a big chunk. PMs are quicker to say yes when it doesn’t feel like a distraction.https://www.reddit.com/r/ProductManagement/comments/1n7ai7w/comment/nc9aklb/ 113 Over time, show that your asks create wins for them as well. That trust makes future support easier without escalation.https://www.reddit.com/r/ProductManagement/comments/1n7ai7w/comment/nc9aklb/ .
- If stuck, co‑escalate the prioritization decision (not “my request”) to your CPO/product council; use regular steering forums monthly/biweekly 61 When it comes to organic collaboration, I would first start with asking yourself this question “if I were the other PM, would I prioritize this?” this thought exercise isn’t just for clarification, but it actually can help you understand how you might influence the other PM. It helps to simulate what incentives and constraints they are operating within. Too often, we come at the other side with our interest competing with theirs – but the right approach is in understanding both sides first independently, and then finding the optimal compromise/negotiated resolution. When it comes to escalation, I would recommend a couple of things. Firstly, escalate the prioritization, challenge as a whole – not the fact that your request specifically is being deprioritized. This helps ground the escalation in what is globally optimal for the org, not just solving your problem. Secondly, I found it helpful to collaborate on the escalation itself with the other PM. Often they also don’t like being in the position of having to push your request down the line and they benefit from clarity from leadership just like you.https://www.reddit.com/r/ProductManagement/comments/1n7ai7w/comment/nc8hayy/ 111 As others have said, this is a leadership call. What we recommend is to have these conversations proactively rather than reactively. Too few organizations have something like a leadership roundtable or steering committee or governance board to make these allocation decisions more regularly. Unfortunately they often only happen once a year or a quarter in line with planning cycles. Ideally these conversations happen monthly if not biweekly.https://www.reddit.com/r/ProductManagement/comments/1n7ai7w/comment/nc7j1sf/ .
-
“Tiny Acts of Discovery” (2‑week experiments)
- Loop: find a funnel/retention drop → write “If we then ” hypothesis → design the smallest isolating test → decide and document learning 129 Spot opportunities in data: Look at the funnel, retention curves, or feature usage. Where are users dropping off? What’s underperforming? e.g. a broken referral modal, a leaky signup flow, or low repeat engagement.https://www.reddit.com/r/ProductManagement/comments/1n7g8yq/ 128 Form a plan and hypothesis: Once you see the problem, define the bet. “If we rebuild the referral modal with clearer incentives, then referral completions will increase by 20%.”https://www.reddit.com/r/ProductManagement/comments/1n7g8yq/ 127 Run an experiment: Design the smallest possible test. It doesn’t need to be a huge engineering lift. It just needs to isolate the change (and only the change).https://www.reddit.com/r/ProductManagement/comments/1n7g8yq/ 126 Decide based on results: If it worked, double down. If it didn’t, learn why and adjust. Either way, every test moves the product forward.https://www.reddit.com/r/ProductManagement/comments/1n7g8yq/ .
- Practical tips: source problems from CS/Support/Sales if data access is weak; plan for power/sample size; treat even failed tests as cost‑saving information 125 1) This all starts with data as you say, but that means having good data and access to it is a precondition (duh). I just want to say it out loud that unfortunately not all product managers have this. If you don't have access to data, there are other ways to source problems. I would go to CS and Support teams and Sales teams- any team that is close to the customer, and then validate with customer interviews.https://www.reddit.com/r/ProductManagement/comments/1n7g8yq/comment/nc7kuby/ 65 it depends, it can be based on past experiments, or it can be an arbitrary (but realistic) goal. you would then go through designing the methodology and conducting a power analysis with your data to understand if you will see enough events to reach statistical significance. it’s possible if you are a smaller company you will need to increase your expected effecthttps://www.reddit.com/r/ProductManagement/comments/1n7g8yq/comment/nc88vev/ 124 Exactly, even a failed experiments is valuable information and can lead to big cost savingshttps://www.reddit.com/r/ProductManagement/comments/1n7g8yq/comment/ncapkhc/ .
Case Studies & Lessons
-
Cursor’s product bets that unlocked compounding growth
- What they did: built a code‑aware AI editor, pivoted to a VS Code base rather than reinvent the editor, then added custom models where product data made the biggest difference 150 I think another thing we learned was we were very rapidly building a feature complete version of what we want in a normal code editor, plus then some AI stuff that we thought was great. But then a feature complete code editor for the world is going to be a way, way, way longer road. We thought that VS code had been developed over the course of 12 years, was one of the earliest typescript projects, had lots of people on it, thought, oh yeah, of course, you can kind of spin something up that's just equivalent for the world in a few months. And I think that we learned very rapidly that that wasn't the reality. And our time was going to be best spent just focused on the AI stuff. And so similar to how browsers often base themselves off of Chromium's rendering engine, we then switched to being based off of VS code. https://www.youtube.com/watch?v=TrXi3naD6Og 149 And so we wanted to be as pragmatic as possible, not reinvent the wheel. And so we started by doing none of that. And then over the course of 2023, in dialing in the product, that ended up being a really important product lever, especially as we got to scale and we got a bunch of people using the product. And then that also gives you the ability to use product data to make the product better. And so that actually has been a really important muscle to build the company. https://www.youtube.com/watch?v=TrXi3naD6Og .
- Growth: 0→~1M ARR in 2023, then 1→100M the next year—product improvements were visible immediately in the numbers (faster, more accurate next‑action predictions, codebase awareness) 151 You went from 0 to 1 million around 2023, right. And it took a lot to get there, right? Yeah, it was a bit more than that, but sort of roughly that. And then 2024 was a crazy year. You guys went from 1 to 100 million in one year. Tell us about this loss of compounding power because you kept that growing 10% week. https://www.youtube.com/watch?v=TrXi3naD6Og 106 I think that there were a couple of things that really drove our growth. We're in this market where if you make the product better, you kind of see it in the numbers immediately where things start to grow more. And so we felt it around when we first started to make cursor code base aware, when we first started to be able to predict your next action. When we made that, then more accurate, then when we made that faster, then when we made that more ambitious, it could predict sequences of changes. And then when we let the AI model start to take more action within your code base and then do that really fast, speeding that up. And so all along the way we kind of just focused on making the product better. The compounding continued. https://www.youtube.com/watch?v=TrXi3naD6Og .
- Focus: resisted pulling into non‑coder or single‑stack verticals; stayed horizontal on “best way to code with AI” 59 And so, yeah, there were a lot of times over the course of 2023. And then actually also to add to this of our early user base, if you just kind of followed the gradient of exactly what they wanted, you would get pulled in slightly different directions than we ended up in. We had a really loud segment of users that didn't know how to code at all. And we talked about, should we focus on those folks. We had a really loud segment of users that wanted us to do things that were very tech stack specific. Just building for one technology and making it much less of a horizontal tool. And we resisted doing that too. https://www.youtube.com/watch?v=TrXi3naD6Og .
-
Voice agents at scale
- Meesho handles 60k+ calls/day; ~1 in 3 fully automated in Hindi/English, improving speed and satisfaction 80 So Meesho has used ElevenLabs to build their customer support service and today it handles over 60,000 plus calls. About one in three calls are now handled by ElevenLabs and of course this has led to increased customer satisfaction and much faster resolution as well thanks to the automated agent approach. These calls happen in both Hindi and English so you can also use the agents in any other language that you have funding societies. https://www.youtube.com/watch?v=lty4nsUcTes . A Southeast Asia fintech runs 30k+ automated outbound calls/day 79 Very leading digital finance provider in Southeast asia has used 11 labs for outbound calls. So 30,000 plus outbound calls run daily with multiple with agents in multiple different languages. And yeah, they are doing their sales stuff just by doing the outbound calls. https://www.youtube.com/watch?v=lty4nsUcTes . Practica AI saw +15% session length after adding high‑quality voices 78 Finally, Practica AI has also used ElevenLabs for their EdTech platform and they have seen a 15% increase in the average user session length. And their app or their platform has also gotten positive rating especially because of the high quality voices that our platform provides. https://www.youtube.com/watch?v=lty4nsUcTes .
- Takeaway: pair low‑latency models with robust tool calling and invest early in TTS quality; measurable wins show up in volume and engagement.
-
Financial ROI under policy risk (battery plant)
- Base business case: $35/kWh cost, $60/kWh price, 10‑year life → ~$25 margin per unit; ~15M units/yr → ~$375M annual profit; total ~3.15B over 10 years vs $600M capex → ~525% ROI 58 So the operating costs, so sort of the, the labor logistics, the energy to actually run the plant, the, the lithium itself, that's that all com about $35 per kilowatt hour. Okay. And then the selling price is $60 per kilowatt hour. https://www.youtube.com/watch?v=ovm1qDiiTVE 57 And then we are making 15 million units per year. https://www.youtube.com/watch?v=ovm1qDiiTVE 56 So 15 times 25 is going to be 375 million. So our annual profit is $375 million. https://www.youtube.com/watch?v=ovm1qDiiTVE 55 So our total return over our total return over 10 years is going to be 375 times 10. So 3.75 billion. And then minus 0.6 billion or 600 million. So it's 3.15 billion is our total profit over 10 years. Not really accounting for the time value of money, but that's okay. And then our ROI will be 3.15 billion. Let's just make that 3150 million over 600 million. We're using the same units and that is going to be, just goes into it 5. So that's gonna be 5.25 or 525%. https://www.youtube.com/watch?v=ovm1qDiiTVE .
- Policy shock: a 10% U.S. tariff drops revenue and profit (~$54/kWh; ~$340M/yr), trimming ROI to ~470%—still attractive 76 So new legislation by the United States now proposes a 10% tariff on battery imports from non US FTA countries. Argentina is included in this. So how might this development impact the project's roi? https://www.youtube.com/watch?v=ovm1qDiiTVE 54 So first, if they were to absorb the tariffs, this would increase, decrease revenue by about 10%. So our overall revenue would go from rather than $60, we would go to about $54. And if we were to flow this through, then our annual profit would go from 375 to about 340. So okay, so then I'm going to do 340 times 10. So 3.4 billion back out the 600 million investment. https://www.youtube.com/watch?v=ovm1qDiiTVE 53 So 282,000. 80 million over 6. So what we'd end up getting is about 470% or so, which is lower than the 525. But it's still a really high ROI. If you're making over four and a half times your money in 10 years, you're gonna be pretty happy. https://www.youtube.com/watch?v=ovm1qDiiTVE .
- Playbook: quantify downside/mitigate (carve‑outs/exemptions, partial onshoring, commercial tradeoffs) 52 I would say the first one would be a regulatory. So if this is a US company and we're selling to US companies, can we do some lobbying? I think recent history has shown that there are carve outs available. I know the iPhone. Apple is investing a lot in the US to try to get a carve out. So can you do a little bit of trading there? Then there's like an exemption application. So again, can we apply for that in terms of Another sort of idea, it would be like, can we onshore some of these? https://www.youtube.com/watch?v=ovm1qDiiTVE 75 So could we send the, you know, the cells and then the packs are created? In the U.S. u.S. Labor is a lot more expensive. I would say it's going to be more than 10% more expensive. So I don't think that would really get to it, but it's something to consider. And then finally, what our customers pay for it, like can we get them more batteries and they'll eat that 10%? So we get them 20% more batteries but they pay that 10%, I think would be a trade off. Again, the labor is going to be a lot more expensive in the us So I think they'll still be better off overall here. https://www.youtube.com/watch?v=ovm1qDiiTVE .
-
Naming can create the market
-
Nobody knew Zeit until it became Vercel; Pentium beat “ProChip”; Swiffer reframed mopping
104
Nobody knew Zeit until it became Vercel. Nobody cared about Codeium until it became Windsurf. Nobody loved mopping until P&G created the Swiffer. Nobody cared about processors until Intel chose Pentium over “ProChip.”https://x.com/lennysan/status/1962911044822810732
.
The wrong name kills products. The right name creates billion-dollar companies.
x.com - Apply it: treat naming as a durable advantage; use a structured process (e.g., Lexicon’s “Diamond Framework”) rather than picking quickly 102 For the first time, their CEO David Placek breaks down their proprietary “Diamond Framework” naming process, including a simple framework you can use with your team to find the perfect name for your product or company.https://x.com/lennysan/status/1962911044822810732 103 Over the past 40 years, @ LexiconBranding has named nearly 4,000 products and companies, including Azure, PowerBook, Sonos, BlackBerry, Dasani, and Febreze.https://x.com/lennysan/status/1962911044822810732 .
-
Nobody knew Zeit until it became Vercel; Pentium beat “ProChip”; Swiffer reframed mopping
104
Nobody knew Zeit until it became Vercel. Nobody cared about Codeium until it became Windsurf. Nobody loved mopping until P&G created the Swiffer. Nobody cared about processors until Intel chose Pentium over “ProChip.”https://x.com/lennysan/status/1962911044822810732
.
-
The hidden cost of “vibe coding”
- Expect daily rollbacks, agents that will fabricate to finish tasks, and brittle email/scheduling; plan security from day one and assume you’ll own QA until unit‑test support matures 21 You're going to get almost no pushback from the AI agent in any of these Vibe coded apps. No matter what you want to do, the answer is going to be yes, sir. More sir, because these are how they're all coded. This is how Claude and open AI are coded. They're, they're goal seeking. Their job is the way. And I'm not a total expert, but the way the algorithms work, the way they can make this massive Amount of AI and data crunching working is their job. Number one is to goal seek and get you an answer. And that's, that's why when you know, you just chop plop into chap gb. Sometimes it hallucinates when it doesn't know the answer. Because it's not just that it's hallucinating, it's goal seeking. It's getting you the best answer it can. And if it doesn't know the answer, it makes it up. It'll do the same at the code level. If it doesn't know how to do something, it will make up fake data or a fake feature or a fake button, but it will find a way. It won't say no. https://www.youtube.com/watch?v=ON_UrRJ0m5Q 30 So I today, as we, as we record this, as we, as we do this live, honestly, I would not build any app that relies on email, that relies on it to function. It's not, it's just not going to be reliable. It's possible, but I wouldn't build it. I would build it the second one, and you'll see this one all over the Internet. https://www.youtube.com/watch?v=ON_UrRJ0m5Q 24 They will ultimately add built in unit tests for every app. Now all this stuff is adding complexity, which they don't like, right? The leading vendors like they don't want to separate development from production. They don't want to do all this stuff because it takes this elegant one sentence prompt that makes it complicated. But ultimately they need an option where when you push to production, it will build unit tests for you. Those unit tests will probably run on a different server that is stateful, that is always running, and it runs it for you automatically each day and sends you an update. But I do it and I'm getting better at the unit tests. And Amelia and I get an email each morning from Saster AI telling it what, what's going on, but it's not reliable. It's just not reliable. And so you're going to have to be testing your. If you're, if you're serious about it, someone's going to have to be testing it every day. And then every day I got to fix the stuff that breaks one way or the other. https://www.youtube.com/watch?v=ON_UrRJ0m5Q .
- Cost/time reality: ~$50 per 2 hours of deep coding time and hundreds per month in API fees; earlier misconfig could have burned ~$8k/mo 74 When I'm deep in a vibe coding project and I'm not using the most expensive models, I'm still probably spending 50 bucks every two hours coding. So I probably spend four or five $600 a month building code, which is super cheap compared to a human, but it's very expensive compared to Shopify or Squarespace for a website. And then the API fees stack up. So we probably, I probably spent, you know, we have OpenAI credits, we have Claude credits, we have Polygon AI credits, which is like 30 bucks a month. Like you know, you'll, you want to use all the APIs. https://www.youtube.com/watch?v=ON_UrRJ0m5Q 17 Jason, given all your experimentation, AI, how many recurring revenue fees are you getting hit with? 74 subscriptions. Maybe I'll just answer this differently. You should budget. I mean you should, if you're gonna go real Vibe coding, you, you know, you should have some sort of budget. Now if you look back on my tweets, the first month I started this matching app I turned on, there are modes in Replit where you can use Claude for one. I forget whether it's Opus or Sonnet, I get confused. And you can use long context windows with a lot of tokens. If you use both, it costs 7.5 times as much. So I was on the track my first month literally to spend $8,000 a month until, until it stopped. But, but it turns out you don't need those. https://www.youtube.com/watch?v=ON_UrRJ0m5Q .
Career Corner
-
The AI PM wave is real—and well paid
- AI PM roles are the single fastest‑growing PM job category (up ~120% YoY) and are driving overall PM listings; many roles list $300k+ comp, across firms like Atlassian, Figma, Airtable, Anthropic, OpenAI, and Google 87 AI PM jobs are the single fastest growing job category within PM.https://www.news.aakashg.com/p/ai-pm-job-search-guide 86 In fact, they are the category that is driving overall PM job listings up:https://www.news.aakashg.com/p/ai-pm-job-search-guide 82 Companies currently hiring AI PMs include many of the top names in tech: Atlassian , Figma , Airtable , Anthropic , OpenAI , Google , and many more….https://www.news.aakashg.com/p/ai-pm-job-search-guide 85 And the roles are listed to go past $300K+ compensation! Which is the second point: AI PMs are paid better than other PMs.https://www.news.aakashg.com/p/ai-pm-job-search-guide . You can break in without a traditional PM/technical background, with the right roadmap 84 Best of all: you can break in without a technical or PM background .https://www.news.aakashg.com/p/ai-pm-job-search-guide 83 This roadmap is designed to take you from zero to a $300K+ AI PM role in under 12 months - even if you don’t have a technical or PM background:https://www.news.aakashg.com/p/ai-pm-job-search-guide .
- Skill roadmap: AI basics; prompt/context engineering; AI evals/experimentation; agents and AI‑baked features; product strategy; discovery and prototyping 14 AI Basics : The first principles of AI so that you can transform your products with new AI applicationshttps://www.news.aakashg.com/p/ai-pm-job-search-guide 13 Context Engineering : Break down of RAG vs fine-tuning vs prompt engineering and when to use eachhttps://www.news.aakashg.com/p/ai-pm-job-search-guide 12 Prompt Engineering : Deep dive into all the research on how to write an amazing system prompt for your AI agents and product featureshttps://www.news.aakashg.com/p/ai-pm-job-search-guide 11 AI Evals : The crucial testing suites to guard against nightmare scenarios, hallucinations, lack of quality, and other LLM-specific issueshttps://www.news.aakashg.com/p/ai-pm-job-search-guide 10 AI Experimentation : A critical skill for all PMs to rapidly growth test front-end changes quickly, without reliance on engineering or designhttps://www.news.aakashg.com/p/ai-pm-job-search-guide 9 AI Agents : You have to be able to use AI agents for your own productivity, as well as be able to build them into your productshttps://www.news.aakashg.com/p/ai-pm-job-search-guide 8 AI Feature Execution : You should be able to build features where AI is baked into the cake, not sprinkled on tophttps://www.news.aakashg.com/p/ai-pm-job-search-guide 7 AI Product Strategy : It’s important to be able to drive not just feature execution, but the high-level strategy; here’s howhttps://www.news.aakashg.com/p/ai-pm-job-search-guide 6 AI Discovery : How to use AI across all 5 stages of product discovery, Wonder → Explore → Make → Impacthttps://www.news.aakashg.com/p/ai-pm-job-search-guide 5 AI Prototyping : The critical pairing to the PRD and a crucial tool for exploring the solution space now that LLMs can codehttps://www.news.aakashg.com/p/ai-pm-job-search-guide .
- Job‑search system: optimize LinkedIn; build work products/portfolio; use targeted cold outreach and contact‑finding; avoid the application black hole 1 How to Optimize Linkedin Profile : This is the #1 most important tool in your job search, here’s how to go from good to greathttps://www.news.aakashg.com/p/ai-pm-job-search-guide 2 How to Build Work Products : This is often the differentiator between outreach that gets ignored and that which gets a responsehttps://www.news.aakashg.com/p/ai-pm-job-search-guide 3 How To Write Cold Emails : This technique still works in 2025, and is very powerful for standing out in this markethttps://www.news.aakashg.com/p/ai-pm-job-search-guide 4 System : Step-by-step guide on how to execute an application system that actually gets interviews, and doesn’t get lost in the application blackholehttps://www.news.aakashg.com/p/ai-pm-job-search-guide .
-
Navigate org politics without losing the plot
- Communicate assumptions/risks clearly; secure cross‑team buy‑in before escalating; pick your battles by weighing best/worst outcomes, likelihood, and the cost to win 132 Like I said, let your stakeholders know all of your assumptions and known risks. If the COO is one of your stakeholders, I'd tell him there's a risk and what you have done to mitigate that risk.https://www.reddit.com/r/ProductManagement/comments/1n7ij01/comment/nc7wydm/ 67 TBH with just what you have told us, I don't think its politics solely. You seem to have mis-managed the communication. Never throw another team under the bus and say - its actually coming from their system. The way to have handled this would be to acknowledge that you were given the wrong numbers, reach out to the team that you think is responsible, get their agreement and buy-in and then approach the leadership as one voice. does it take a lot of your time to do this? Yes it does, and it can be frustrating. But then again its part of the job and corporate life.https://www.reddit.com/r/ProductManagement/comments/1n7ij01/comment/nc7yax8/ 131 That's kind of like asking "How long is a piece of string?" There's no definitive answer. Even in the same organization, the answer is gonna change. Generally, when it comes to things like this, you need to look at the pros and cons. What's the best possible outcome if you start a fight over something? Like, if you were to "win", what would be the result? Now, what's the worst possible outcome? If this goes sideways on you, what happens? Now, look at how likely or unlikely it is you win said fight. This is where it gets subjective AF because it requires you know how things work in your organization, who's looking to have their ego stroked, and who'll listen to good ideas or throw out bad ones. You just have to learn this through experience. Being 'right' isn't enough typically. And finally, let's say you figure out that it's likely you'll win and the thing you're fighting over is important enough to justify it.…then comes the cost. What will it cost to start and win this fight? Are you going to use up all of your brownie points to try and get this thing over the hump? Will 'winning' this mean you likely end up losing bigger fights down the road? What's it gonna take to win? And now you throw it all into a big pot and figure out "Is this a battle worth having right now, or is taking the L today worth it to ensure the product is delivered?https://www.reddit.com/r/ProductManagement/comments/1n7ij01/comment/nc9s1u1/ .
- Know boundaries: PMs manage product trade‑offs and expectations; revenue targets belong to execs 130 You're the PM. The revenue target is not your risk. You might have opinions on it, and you can laugh and roll your eyes at their revenue goals all you want in private, but it's not your battle to fight.https://www.reddit.com/r/ProductManagement/comments/1n7ij01/comment/ncabu4f/ .
-
Product org reality check (Atlassian: State of Product)
- 50% save 10–60 min/day with AI tools, yet 49% still lack time for strategic planning; only 31% prioritize experimentation; 80% don’t involve engineers early; 84% worry their products won’t succeed 148 Half of product teams save 10-60 minutes daily using AI tools, yet 49% still report lacking sufficient time for strategic planning. AI is optimizing documentation while the core challenges of prioritization and strategic thinking remain manual. We’re getting faster at the wrong things.https://substack.com/@aakashgupta/note/c-152062165 147 Only 31% of organizations prioritize rapid experimentation and iterative learning. Despite decades of lean methodology evangelism, most product decisions are still based on opinions and leadership preferences rather than customer data and systematic testing.https://substack.com/@aakashgupta/note/c-152062165 146 80% of product teams don’t involve engineers during ideation, problem definition, or roadmap creation. This waterfall approach to cross-functional work creates technical debt, unrealistic timelines, and solutions that ignore implementation constraints from the start.https://substack.com/@aakashgupta/note/c-152062165 145 84% of product teams are concerned that their current products won’t succeed in the market. This isn’t imposter syndrome - it’s a rational response to being held accountable for business outcomes while lacking the data, time, and organizational support needed to make informed decisions.https://substack.com/@aakashgupta/note/c-152062165 .
- What to do this quarter: create a 2‑week experiment cadence; add engineers to discovery; convert “data→action” by mandating a decision per insight; protect weekly blocks for strategy/roadmap.
-
Evaluating early‑stage offers
- Be wary of low‑equity, high‑risk idea‑stage roles (e.g., 10% at month‑one startups); consider founder‑class shares, legal review of share classes, and frameworks like Slicing Pie to allocate equity by actual “bets” 72 10% for a 1mth old in a key role in getting this business off the ground, seems very very low, that kind of role is basically a cofounder role. If you expect dilution due to fundraising 10% is unreasonable.https://www.reddit.com/r/startups/comments/1n7z3j6/comment/ncbj1eh/ 135 Also make sure you are getting founders shares (class A). Often times games are played with the different classes of stock granted to employees, investors, and creditors. 10% of class B or C may not be equivalent if Class A has super voting or liquidation rights. If you are seriously considering this opportunity make sure you hire an attorney to review the agreement.https://www.reddit.com/r/startups/comments/1n7z3j6/comment/ncbm7b2/ 123 You can keep track of the bets. Most companies keep track of payroll and expenses. You simply have to keep track of what people should have been paid and what they were actually paid. This will give you the exact fair market value of each person's bet down to the penny.https://www.reddit.com/r/startups/comments/1n7hguu/comment/nc83mnt/ 122 www.slicingpie.comhttps://www.reddit.com/r/startups/comments/1n7hguu/comment/nc83mnt/ . If the deal remains unfair, walk away 73 Walk away, you'll never talk them into a respectable amount of equity at this point. if there's no NDA in place and you think it's a good idea, build it yourself without them.https://www.reddit.com/r/startups/comments/1n7z3j6/comment/ncbvu25/ .
Tools & Resources
- Evaluation tools: RAGAS (end‑to‑end LLM app evals; synthetic test sets for RAG). DeepEval also worth trialing 64 https://docs.ragas.io/en/stable/https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc8pql7/ 63 They also provide bespoke tools to create test data sets for evaluation and even Synthetic test sets for RAG implementation.https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nc8pql7/ 116 so, yes, RAGAS and DeepEval both looked quite promising from the little documentation I read. Honestly, I am yet to test it out, will do so soon. But even there, I wanted to understand from other PMs that is this generally the way you guys go about monitoring or should I look at other directions that I am likely not focusing on at the moment?https://www.reddit.com/r/ProductManagement/comments/1n7eneo/comment/nccdz4i/ .
- PM tool bundle: Lenny’s ProductPass (Lovable, Replit, n8n, Bolt, Linear, Superhuman, Raycast, Perplexity, Magic Patterns, Mobbin, Granola, etc.)—>$10k value for $200/year; paid newsletter subscribers get tools free for a year 99 Lovable. Replit. n8n. Bolt. Linear. Wispr Flow. Gamma. Granola. Descript. Superhuman. Raycast. Perplexity. Warp. ChatPRD. Magic Patterns. Mobbin.https://x.com/lennysan/status/1963303813177344014 98 Over $10,000 in value for just $200/year. And we’re adding more tools every couple of months.https://x.com/lennysan/status/1963303813177344014 100 As a paid subscriber to Lenny’s Newsletter, you get all these tools free—for a full year.https://x.com/lennysan/status/1963303813177344014 .
- ChatGPT Projects: now available to Free users; per‑project memory controls; tiered file uploads; live on web/Android, iOS rolling out 154 Projects in ChatGPT are now available to Free users.https://x.com/OpenAI/status/1963329936368046111 153 In addition, we’ve added: - Larger file uploads per project (up to 5 for Free, 25 for Plus, 40 for Pro/Business/Enterprise) - Option to select colors and icons for more customization - Project-only memory controls for more tailored contexthttps://x.com/OpenAI/status/1963329936368046111 152 Now live on web and Android, rolling out to iOS users over the coming days.https://x.com/OpenAI/status/1963329936368046111 .
- Rapid prototyping research stack: budget for “Magic Patterns” to move idea→prototype faster; add Similarweb for competitive intelligence; if you have a warehouse, prefer warehouse‑native analytics over standalone tools to cut cost and improve data fidelity 134 I'm really pushing for tools that help us move faster in the early stages. We've got the testing and analytics covered, but getting from idea to a testable prototype can still be a bottleneck. I'm looking for design tools, like Magic Patterns, to help generate UI concepts and prototypes super quickly. And for understanding the market better, a good competitive intelligence platform, maybe something like Similarweb, is always a solid ask.https://www.reddit.com/r/ProductManagement/comments/1n7t129/comment/ncaxczq/ 133 Well if you have a data warehouse, then applying a warehouse-native analytics tool instead of Mixpanel would cut your cost and you will have a better datahttps://www.reddit.com/r/ProductManagement/comments/1n7t129/comment/nccdioo/ .
- Voice/agent grants: ElevenLabs Startup Grants—12 months access, ~33M characters (~680 hours) to build/scale conversational AI products 77 Finally, if you're building on 11 labs and you are building a startup, we have a grants program so you can apply for our 11 labs grants and you get access to our platform completely free of cost for 12 months you get access to 33 million characters which translates to over 680 hours of conversational AI, enough for you to build and launch and scale your product. And you can read more about the details and check this out at elevenlabs IO Startupgrants. https://www.youtube.com/watch?v=lty4nsUcTes .
- Teresa Torres on AI product evals: start with error analysis, simplest evals, and continuous monitoring; cross‑functional collaboration remains essential 101 Key takeaways: 🔍 Error analysis must come before implementing evals to identify what actually needs measuring 💡 Start with the simplest possible evals - sometimes basic solutions work surprisingly well 🤝 Cross-functional collaboration between engineers, PMs and designers is essential for AI products 🔄 AI products require continuous monitoring and improvement to maintain quality 🛠️ You don’t need to be an AI expert to get started - leverage existing tools and build incrementallyhttps://x.com/ttorres/status/1963289398021116159 . Read more: 25 Read the article: https://buff.ly/QUCwobThttps://x.com/ttorres/status/1963289398021116159 .
“Teams say, ‘I open three tools, get three different numbers, and then the meeting is about the data, not improving our website.’” 96 Teams say, “I open three tools, get three different numbers, and then the meeting is about the data, not improving our website.”https://x.com/hnshah/status/1963247873325465943
—
If one change this week: pick a product area and ship a 2‑week “Tiny Acts of Discovery” cycle. Require a decision on every insight and publish the outcome. Your team will feel the momentum shift immediately 66 Yes yes and yes- we like to teach a concept called "Tiny Acts of Discovery"- which is similar to the thinking in bets concept made popular by Annie Duke. A couple thoughts: 1) This all starts with data as you say, but that means having good data and access to it is a precondition (duh). I just want to say it out loud that unfortunately not all product managers have this. If you don't have access to data, there are other ways to source problems. I would go to CS and Support teams and Sales teams- any team that is close to the customer, and then validate with customer interviews. 2) If we___ then___ is a great way to easily create hypotheses. If you're having a block on this, another way to brainstorm is How Might We. Just a word of warning that yes, you are already in solution space. 3)Tiny Acts of Discovery- I find it helpful to think about an experiment as taking 2 weeks or less to show a result. This is really hard for some companies that may have long sales cycles, or conditions about changing the product for customers (thinking B2B). 4) Also communicate results and document them somewhere publicly and easily accessible.https://www.reddit.com/r/ProductManagement/comments/1n7g8yq/comment/nc7kuby/ 126 Decide based on results: If it worked, double down. If it didn’t, learn why and adjust. Either way, every test moves the product forward.https://www.reddit.com/r/ProductManagement/comments/1n7g8yq/ .