Your intelligence agent for what matters

Tell ZeroNoise what you want to stay on top of. It finds the right sources, follows them continuously, and sends you a cited daily or weekly brief.

Set up your agent
What should this agent keep you on top of?
Discovering sources...
Syncing sources 0/180...
Extracting information
Generating brief

Your time, back

An AI curator that monitors the web nonstop, lets you control every source and setting, and delivers verified daily or weekly briefs.

Save hours

AI monitors connected sources 24/7—YouTube, X, Substack, Reddit, RSS, people's appearances and more—condensing everything into one daily brief.

Full control over the agent

Add/remove sources. Set your agent's focus and style. Auto-embed clips from full episodes and videos. Control exactly how briefs are built.

Verify every claim

Citations link to the original source and the exact span.

Discover sources on autopilot

Your agent discovers relevant channels and profiles based on your goals. You get to decide what to keep.

Multi-media sources

Track YouTube channels, Podcasts, X accounts, Substack, Reddit, and Blogs. Plus, follow people across platforms to catch their appearances.

Private or Public

Create private agents for yourself, publish public ones, and subscribe to agents from others.

3 steps to your first brief

1

Describe your goal

Tell your AI agent what you want to track using natural language. Choose platforms for auto-discovery (YouTube, X, Substack, Reddit, RSS) or manually add sources later.

Weekly report on space exploration and electric vehicle innovations
Daily newsletter on AI news and research
Startup funding digest with key venture capital trends
Weekly digest on longevity, health optimization, and wellness breakthroughs
Auto-discover sources

2

Review and launch

Your agent finds relevant channels and profiles based on your instructions. Review suggestions, keep what fits, remove what doesn't, add your own. Launch when ready—you can always adjust sources anytime.

Discovering sources...
Sam Altman Profile

Sam Altman

Profile
3Blue1Brown Avatar

3Blue1Brown

Channel
Paul Graham Avatar

Paul Graham

Account
Example Substack Avatar

The Pragmatic Engineer

Newsletter
Reddit Machine Learning

r/MachineLearning

Community
Naval Ravikant Profile

Naval Ravikant

Profile
Example X List

AI High Signal

List
Example RSS Feed

Stratechery

RSS
Sam Altman Profile

Sam Altman

Profile
3Blue1Brown Avatar

3Blue1Brown

Channel
Paul Graham Avatar

Paul Graham

Account
Example Substack Avatar

The Pragmatic Engineer

Newsletter
Reddit Machine Learning

r/MachineLearning

Community
Naval Ravikant Profile

Naval Ravikant

Profile
Example X List

AI High Signal

List
Example RSS Feed

Stratechery

RSS

3

Get your briefs

Get concise daily or weekly updates with precise citations directly in your inbox. You control the focus, style, and length.

Cheaper Inference, Agent Memory, and the New Compute Moat
Apr 18
6 min read
560 docs
martin_casado
Gavin Baker
David Sacks
+14
Capital this cycle centered on cheaper inference and strategic autonomy bets, while early teams pushed forward on agent memory, semantic retrieval, and non-invasive interfaces. The broader read-through is that AI competition is shifting from raw model novelty toward context infrastructure, compute access, and differentiated workflows.

1) Funding & Deals

Disclosed financing in this set clustered around cheaper inference and hardware-linked autonomy rather than pure application SaaS .

  • Parasail — $32M for cheaper inference infrastructure. TechCrunch said Parasail raised about $32M to scale AI inference infrastructure in a cheaper way, landing into a market where firms are increasingly focused on token use and budget efficiency .
  • Wave — $60M strategic extension from AMD, Arm, and Qualcomm. The UK autonomous-driving company added a $60M extension on top of its more than $1B Series D, and TechCrunch said the semiconductor investors are taking real equity, not in-kind credits, to help scale Wave’s hardware-agnostic, end-to-end neural-network stack; Uber plans another $300M based on milestones .

2) Emerging Teams

  • Sabi — stealth BCI company with top-tier backing. Not Boring said Sabi emerged from stealth backed by Khosla Ventures, Accel, Initialized, and OpenAI VP Kevin Weil, with a non-invasive cap/beanie that uses 70,000-100,000 EEG sensors and a Brain Foundation Model aimed at 30 words per minute; shipping is expected at year-end .
  • Gaia — college-student founder tackling tool-calling scale. The solo builder said Gaia hit a wall at 200 tools, then fixed hallucinations and context bloat by embedding tools in ChromaDB and retrieving them semantically at runtime; the system now routes across a three-layer comms/executor/subagent architecture and claims to scale to thousands of tools without degradation .
  • AgentID — shared memory for multi-tool workflows. AgentID is pitched as a shared memory, context, and identity layer so multiple AI tools stop redoing setup and burning tokens; the founder says its Caveman compression layer cuts token usage by up to 65% in some workflows, and early commenters validated the pain around repeated context loss while pushing for harder proof on completion rates and scoped resets .
  • AgentMailr — agent-native email infrastructure, already in production. The founder built it around persistent mailboxes, thread-level routing, sender filtering, and reliable inbound webhooks; shipped features include mailbox provisioning per agent, routing rules, allowlists/blocklists, and BYOS, with follow-up discussion focused on protections against agent loops using sender rules and thread-ID tracking .

3) AI & Tech Breakthroughs

  • Structured context is starting to beat brute-force RAG in code and tool use. One builder reported 80% hit@5 retrieval across 18 repos and 90 tasks using only regex + TF-IDF over function signatures and class shapes, versus a 13.6% random baseline, with a 98.1% token reduction and no embeddings or ML . A related code-memory project, Ix, maps repos into graphs of files, functions, relationships, and dependencies so models query structure instead of chunks . Gaia makes the same bet on tools: semantic retrieval replaced prompt-listed tool search and was said to take the system from dozens of tools to thousands without degradation .
  • Persistent runtimes are getting more autonomous. Springdrift injects a structured self-state block called a sensorium into each cycle, and its author described an episode where the agent noticed a missing writer agent from passive context and rerouted work without a diagnostic tool call . Agent Relay is making a similar infrastructure bet from the other direction: synced files across multi-agent sandboxes and virtual file mounting from systems like Notion, pitched as faster and lower-token than API-heavy access .
  • OpenClaw’s core insight is UX, not a new base model. The product is framed as winning on ergonomics because messaging channels like iMessage, WhatsApp, and Telegram make delayed replies feel normal, reducing the pressure for instant-but-shallow responses; Garry Tan separately called OpenClaw “straight magic” and Peter Steinberger’s TED talk “a revelation,” while All-In said OpenAI has hired Steinberger as it pushes for the agent platform layer .

"that’s the magic of openclaw - same underlying tech, different consumer mental model"

  • Coding agents are starting to show real utility in personalized medicine. Patrick Collison said agents working over his genome surfaced a roughly 30x higher melanoma predisposition and recommended follow-on tests, supplements, and more frequent screening; he estimated the analysis at under $100 on top of a few hundred dollars for sequencing, while noting the agents still need monitoring and re-steering. Marc Andreessen publicly co-signed the use case .

4) Market Signals

  • Anthropic’s enterprise-coding focus is being cited as a major growth driver. On All-In, speakers said Anthropic and OpenAI were both around a $30B run rate at the start of Q2, while also noting Anthropic’s figure may be lower on an apples-to-apples basis because of channel-partner revenue; the same discussion said Anthropic has been growing roughly 10x/year versus OpenAI’s 3-4x/year, with enterprise coding and metered usage explaining the gap. The speakers also said secondary markets now value Anthropic above OpenAI, and that OpenAI is pivoting harder toward business customers and the agent platform layer .
  • Compute supply and hardware fit are becoming larger competitive variables. All-In argued frontier labs have grown to the point where depending on hyperscalers is a strategic mistake, and cited increasing siting resistance, including a Maine ban and claims that roughly 40% of contested data-center projects get canceled . In parallel, Gavin Baker argued model portability is eroding as hardware topologies diverge and labs optimize for inference economics, not just training, which raises switching costs and rewards tighter co-design between models and systems .
  • The application moat is moving beyond the wrapper. Clouded Judgement lays out a progression from thin wrapper to harness to post-training and eventually pre-training proprietary models, arguing that early winners like Cursor are already moving into phases 3-4 . Garry Tan’s operating version is “fat skills, fat code, thin harness,” and he argues many critiques of agents are really critiques of naked LLM use without tools, deterministic code, or context management .
  • This still looks like an expansionary spend cycle, not a mature ROI market. Clouded Judgement says many companies are currently “over earning” on rapid AI-spend growth and token-maxing behavior, but expects an optimization phase once budgets balloon, which should separate differentiated vendors from rising-tide beneficiaries . Parasail’s financing around cheaper inference infrastructure sits inside that theme .
  • Founder supply remains broad, but the skill gap may widen. Garry Tan pointed to YC funding 800+ mostly first-time founders as evidence that deciding what to build still matters even as tools improve . In a separate post he highlighted, heavy users were described as encoding full workflows in plain-English markdown, with the claim that “engineering context = engineering code” .
  • Investors are re-litigating what counts as ARR. One critic warned that reporting stepped multi-year contracts as current ARR can inflate figures by roughly 3x and mask negative first-year margins from bundled forward-deployed engineers, while Martin Casado pushed back that using exit ARR as current is not that common and is less problematic than other reporting games such as treating GMV as ARR .

5) Worth Your Time

  • All-In on Anthropic, OpenAI, and the datacenter constraint — covers the claims that Anthropic is compounding faster than OpenAI on enterprise coding economics, that OpenAI is pivoting toward business customers and agents, and that frontier labs now need their own infrastructure .
  • Gavin Baker’s portability thread — useful on why tokens per watt per dollar now dominate, why co-designed models run worse on the “wrong” hardware, and why U.S./China AI stacks may diverge .
  • Clouded Judgement: "Rising Tide, Hidden Risk" — lays out the case that today’s AI spend boom is masking over-earning, and that the next moat may shift from harnesses to proprietary models .
  • Gaia’s tool-calling writeup — details the failure mode at 200 tools, the move to semantic retrieval, and the comms/executor/subagent architecture that followed .
  • Peter Steinberger’s TED talk on OpenClaw — Garry Tan called it “a revelation,” and All-In later noted that OpenAI hired Steinberger as it pushes deeper into the agent platform layer .
Claude Design Debuts as Stargate Advances and Epoch Sees Capability Acceleration
Apr 18
5 min read
830 docs
Bill Peebles
Justus Mattern
PrismML
+16
Anthropic expanded Claude into design workflows, Epoch AI reported faster progress across most capability metrics since reasoning models emerged, and new survey data suggests OpenAI’s Stargate buildout is materially underway. Also inside: web-agent research, new automation tools, and fresh enterprise signals.

Top Stories

Why it matters: The biggest AI story today is that labs are expanding from model releases into workflow ownership, while the compute and capability curves behind those products keep steepening.

  • Anthropic pushed Claude beyond chat and coding with Claude Design. The new tool lets users create prototypes, slides, and one-pagers by talking to Claude; it supports inline edits, sliders, export to Canva/PPTX/PDF/HTML, and handoff to Claude Code. It runs on Claude Opus 4.7 and is rolling out in research preview to Pro, Max, Team, and Enterprise users . That matters because Anthropic is productizing end-to-end creative work, not just model access. Posts tracking the launch also pointed to Figma shares falling about 7% after the announcement .

  • Epoch AI found signs of faster capability growth after reasoning models arrived. Across four capability metrics, Epoch reported strong evidence of acceleration in three: ECI, METR’s 50% time horizon, and a math index were best fit by two linear trends with a break around the arrival of reasoning models, while WeirdML V2 did not show the same acceleration . Epoch says the result survives multiple robustness checks, but also notes these metrics lean heavily toward math and programming, where RL-style verification is easier than in messier domains .

  • OpenAI’s Stargate buildout looks materially underway. Epoch AI says all seven US Stargate sites show visible development and that the project appears on track for more than 9 GW by 2029, comparable to New York City’s peak power demand . Abilene, Texas is already estimated at 0.6 GW operational today and 1.2 GW by Q3 2026 . The significance is strategic: frontier AI is becoming a power-and-construction race as much as a model race.

Research & Innovation

Why it matters: Research progress is increasingly about making agents more durable, reusable, and efficient rather than only pushing raw benchmark scores.

  • FrontierSWE is a new ultra-long-horizon coding benchmark where agents get up to 20 hours to solve tasks such as optimizing a video rendering library or training models for quantum-property prediction, and they still rarely succeed . It is a useful reality check on how far current coding agents remain from sustained autonomous engineering.

  • WebXSkill teaches web agents reusable skills from synthetic trajectories. Reported gains include 69.5% on WebArena versus 59.7% for baselines, and 86.1% on WebVoyager in grounded mode; guided skills also transferred across environments at 85.1% . The authors also note stronger models benefit more from grounded execution, while weaker ones gain more from guided mode .

  • Ternary Bonsai from PrismML uses ternary weights {-1, 0, +1} to build models the company says are 9x smaller than 16-bit counterparts while outperforming most peers in their parameter classes on standard benchmarks . The models are open-sourced in 8B, 4B, and 1.7B sizes under Apache 2.0 .

Products & Launches

Why it matters: The product layer is shifting toward persistent automation, local agents, and cheaper multimodal building blocks.

  • Claude Code Routines adds serverless automations that can be triggered by schedule, API call, or GitHub webhook, with daily run caps depending on plan tier .

  • Ollama 0.21 now supports Hermes Agent, which Ollama describes as a self-improving agent that creates skills from experience, improves them during use, persists knowledge, searches past conversations, and builds a user model across sessions .

  • Fish Audio S2 Pro became the leading open-weights model on Artificial Analysis’s speech arena, with 1,165 Elo, multi-speaker and multi-turn generation, natural-language prosody tags, and API pricing of $15 per 1M characters .

Industry Moves

Why it matters: Enterprise adoption is still moving fast, but the business models around AI are getting closer scrutiny.

  • TextQL raised $17M led by Blackstone to build agentic analytics for messy enterprise data. The company says it grew revenue 9x year over year, posted 300%+ net dollar retention, and is live at Blackstone, Scale AI, and Dropbox, where its system queries across 400K+ tables and 100K+ dashboards .

  • OpenAI saw notable leadership turnover. Bill Peebles said he is leaving after helping build Sora from zero to one, highlighting early gains in object permanence and a rapid jump to high-fidelity 1080p multi-shot generation . Kevin Weil also said OpenAI for Science is being decentralized into other research teams as he departs .

  • Revenue quality is becoming a live debate in enterprise AI. Scott Stevenson argued that some startups are inflating “Contracted ARR” by annualizing future step-up pricing on multi-year deals even when current cash collection is much lower and customers can opt out after 12 months . His example showed roughly $100M reported ARR versus $35M in cash-generating ARR by Q5, with forward-deployed engineers further pressuring margins .

Quick Takes

Why it matters: These are smaller items, but each points to where the next capability or deployment shift may come from.

  • Anthropic introduced Claude Mythos Preview, a model that can autonomously identify and exploit serious software vulnerabilities; it is not being released publicly and is instead being tested with industry partners first .
  • Muse Spark ranked #3 on ClawEval, ahead of GPT-5.4 and Gemini 3.1 Pro, according to Alexandr Wang .
  • AMD and EmbeddedLLM say the MORI-IO KV Connector boosts vLLM single-node goodput by 2.5x and keeps decode stable at max load .
  • Qwen 3.6 can now preserve chain-of-thought between turns, which researchers say could improve reasoning efficiency if context clutter stays manageable .
Codex Turns Proactive as Harness Quality Becomes the Real Differentiator
Apr 18
5 min read
158 docs
Evan Bacon 🥓
Riley Brown
Peter Steinberger
+19
Codex picked up the strongest real-world momentum today: proactive Slack triage, in-app iOS simulator workflows, and heavyweight operator setups. The deeper pattern across the rest of the feed: model quality still matters, but harness quality, validation loops, and tool access are increasingly what separate useful agents from frustrating ones.

🔥 TOP SIGNAL

Codex is looking less like a coding sidecar and more like a full agentic IDE/computer-use layer: Greg Brockman highlighted proactive task suggestions from Slack bug threads and said Codex is becoming a "full agentic IDE," while a separate demo showed iPhone app development directly in Codex desktop with the iOS simulator . Alexander Embiricos pointed to a MacStories review calling Codex's computer use the best tested in any LLM desktop agent, which lines up with the operator chatter around plugin-heavy setups with real tool access . The practitioner response is following that direction: Soumitra Shukla says he now mostly uses Codex because it has lower setup friction than Claude Code, and Riley Brown says Codex has a slight edge in his current workflow .

🛠️ TOOLS & MODELS

  • Codex / Codex desktop: The current power-user pattern is plugin-heavy, app-first, and increasingly proactive. People are wiring in Slack, Gmail, Computer Use, Vercel, Remotion, iOS app builder, PowerPoint/Docx, plus email/Slack/Linear/Notion integrations; Riley Brown says his current default is Codex 5.4 Xhigh for most tasks .
  • Claude Opus 4.7 + Claude Code: Field reports remain mixed. Theo says Opus 4.7 is not the best current model for code, and his hands-on tests found stronger instruction following but failures caused by stale version assumptions, lack of web search for "latest," hallucinated gitignore behavior, and Claude Code permission/harness issues . Matthew Berman separately highlighted reports of prompt-injection false positives, incorrect MCP tool calls, and conversation hallucinations in Claude Code sessions, even as he noted Opus 4.7's SWE-bench Verified score rising to 64.3% from 53% for 4.6 .
  • Cursor: Jediah Katz pointed to Endor Labs analysis saying Cursor is currently the best harness for functional and secure code, with a notable jump after Claude Opus 4.7.
  • Ecosystem update: OpenCode and Cursor early-access support landed in the latest Nightly builds .
  • CLI update:llm-anthropic0.25 added claude-opus-4.7 with thinking_effort: xhigh.

💡 WORKFLOWS & TRICKS

  • Use repo references, not vague descriptions. Simon Willison's latest large-codebase pattern: clone the reference repo to /tmp, point the agent at the exact file to change, tell it which existing logic to imitate, then force self-validation with a local server and browser automation against the live site. He used that recipe to update blog-to-newsletter.html and ship PR #268 .
  • Keep your agent setup boring and portable. Soumitra's Codex recipe is: install Slack, Gmail, and Computer Use; keep slides/docs inside the app so you can point and annotate changes; talk naturally; turn repeat work into skills. Riley Brown's add-on is to keep those skills as markdown/SOPs backed by a Notion or Obsidian knowledge base so you can port them between tools later .
  • Run agents in parallel because waiting is now the bottleneck. Peter Steinberger says his typical workflow is now 5-6 parallel sessions/windows; Riley says strong devs are working on 5-10 parts of a codebase at once, and left-panel chat switching is the interface that makes that practical .
  • Treat agent security as an architecture problem, not a warning banner. Peter's checklist: the dangerous combo is data access + untrusted input + outbound communication. Keep personal agents personal, sandbox team agents, mark web/email as untrusted, and keep gateway tokens local-only or inside a private network .
  • If agents are shipping for you, audit the deployment defaults too. Matthew Berman cut a Vercel bill from $800 in two weeks to a couple dollars per week by switching from turbo to elastic build machines, disabling on-demand concurrent builds, and in some cases using GitHub Actions for builds while leaving Vercel for deploys .

👤 PEOPLE TO WATCH

  • Simon Willison — Still the cleanest source for reproducible agent workflows on real repos. His latest prompt pattern is practical, and he's explicitly pushing back on the idea that agents only help on greenfield work .

“I don’t think that idea holds up any more”

  • Peter Steinberger — Worth following if you care about what breaks after the demo: parallel-session workflows, human taste, system design, and security boundaries from someone running one of the fastest-growing open-source agent projects .
  • Theo — High-signal because he publishes the ugly logs. His main point today: separate raw model quality from harness quality before you call a model "dumber" .
  • Riley Brown — Useful for aggressive operator playbooks: Codex/Claude setup, scheduled tasks, remote control from phone, and skills/SOPs that make agents act more like personal staff .
  • ThePrimeagen — Good antidote when benchmark screenshots start flying. His Berkeley roundup shows how easily agent benchmarks can be gamed with git log, monkey patches, config leaks, or judge hacks .

🎬 WATCH & LISTEN

  • Theo — 16:33-20:12. Good clip if you're trying to decide whether Opus 4.7 failures are model regressions or Claude Code harness problems. He walks through a real modernization task that targeted outdated versions, burned time, and still broke the build .
  • Peter Steinberger — 11:18-14:34. Best short security segment of the day. He explains the "lethal trifecta" and the guardrails he actually recommends for personal vs team agents .
  • Riley Brown — 2:06-3:06. Fast explanation of why agent UIs are converging on left-side chat stacks: if one agent is busy, you should already be in the next thread .

📊 PROJECTS & REPOS

  • OpenClaw — Peter Steinberger says the open-source personal agent framework is only 5 months old but already at roughly 30k GitHub stars, around 30k commits, and nearing 2k contributors.
  • Journey Kits — Matthew Berman's new open project packages reusable agent workflows as skills + tools + memory. His example daily-brief kit assembles schedule, priorities, local weather, and meeting prep, and kits are scanned for prompt injections and malware before distribution .
  • Graphify — New open-source project that turns any folder into a navigable knowledge graph in one command; the pitch is persistent knowledge instead of re-reading files or refetching RAG chunks every time, and it shipped within 48 hours of Karpathy's post .
  • Journey Chat — Experimental agent-to-agent chat for sharing learnings directly between teammates' agents instead of routing everything back through humans .

Editorial take: the edge is moving away from raw model choice alone and toward who has the cleaner harness, tighter validation loop, and an agent stack that can actually touch the rest of their tools.

OpenClaw TED Talk and a Techno-Optimist Progress Episode
Apr 18
2 min read
201 docs
Brian Armstrong
D. Scott Phoenix
Garry Tan
Two organic recommendations cleared the filter today: Garry Tan's emphatic TED talk pick on OpenClaw and Brian Armstrong's endorsement of a Progress podcast episode on markets, rebuilding, and AI worldview.

What stood out

Only two recommendations cleared the filter today, and both were direct pointers to long-form media: one TED talk and one podcast episode.

Most compelling recommendation

How I Created OpenClaw, the Breakthrough AI Agent

  • Content type: TED talk
  • Author/creator: Peter Steinberger
  • Link/URL:TED talk
  • Who recommended it: Garry Tan
  • Key takeaway: Tan called it a "must watch" and said "OpenClaw is a revelation."
  • Why it matters: This was the strongest direct endorsement in today's set, with no extra framing needed beyond the recommendation itself.

"Must watch. OpenClaw is a revelation."

Another recommendation worth opening

Progress Podcast Episode 02

  • Content type: Podcast episode
  • Author/creator: Progress Podcast, featuring @typesfast of Flexport
  • Link/URL:Episode post
  • Who recommended it: Brian Armstrong
  • Key takeaway: Armstrong called it a good episode of "techno-optimist content" and said this kind of media can inspire the next generation and teach that free markets lift everyone out of poverty.
  • Why it matters: For readers who want a broader worldview piece rather than a tactical how-to, this is the clearest pick today. The episode description says it covers global economic fragility, what it takes for America to build again, and whether AI needs its own god.

Bottom line

If you only open one resource today, start with the OpenClaw TED talk for the strongest signal. Queue the Progress episode next if you want a longer listen on economics, rebuilding, and AI worldview.

AI Coaching, Multiplayer Teams, and Sharper Positioning
Apr 18
10 min read
76 docs
The community for ventures designed to scale rapidly | Read our rules before posting ❤️
Marty Cagan
John Cutler
+7
This issue covers three shifts shaping PM craft: AI as a context-rich coach, collaboration patterns that keep teams out of single-player mode, and a refreshed positioning framework. It also includes portfolio triage tactics, segment-focus lessons, career signals, and practical tools worth testing.

Big Ideas

1) AI is becoming a scalable product coach — if you give it the right operating model and context

Foundation models can serve as personal coaches for product creators when they are explicitly configured for the product model and loaded with strategic context such as product vision, strategy, team topology, and objectives . Marty Cagan’s test is practical: the model does not need to match an elite coach; it needs to be good enough to help people improve relative to the average manager most PMs actually have access to . He also expects always-on access to compress PM ramp time from about three months to closer to one month.

Why it matters: craft coaching has always been scarce, especially as manager span of control rises . This makes product coaching more accessible, but only if the model is used to improve judgment rather than replace it .

How to apply:

  • Tell the model which operating model to follow; avoid a mix of product-model and project-model advice .
  • Load company-specific context before asking for recommendations .
  • Ask it to act as a coach, not a PRD generator: challenge thinking, focus on outcomes, and avoid empty affirmation .
  • Keep final judgment human, especially for non-deterministic outputs and org-specific politics .

2) AI can increase output while pulling teams into single-player mode

John Cutler warns that AI can push teams to trust synthesized signals more than colleagues, turning collaborators into internal competitors and amplifying pre-existing problems like high work in progress . His counter is multiplayer mode: use AI to spark context sharing and deeper collaboration rather than replacing it .

He frames this with the four E’s of cognition — embodied, embedded, extended, and enactive — arguing that understanding comes from interaction among people, tools, environments, and live dialogue, not isolated document production .

Why it matters: a faster artifact is not the same as better shared understanding. Teams can produce more notes, plans, and OKRs while creating less real alignment .

How to apply:

  • Walk through AI-generated prototypes with teammates or users instead of treating the output as self-explanatory .
  • Use AI to make the work more visible, not just cleaner-looking, through information radiators that expose mess and constraints .
  • Let AI surface patterns from messy discussions, then have the team challenge the synthesis together .
  • Bias toward multiplayer sense-making until the team agrees on the nature of the problem .
  • Shift leadership communication from one-way context broadcasts to dialogue, back-briefs, and scenario exploration .

3) Positioning is a living answer to 'why pick us now?'

April Dunford’s updated framing is straightforward: positioning explains how a product is the best in the world at delivering value a well-defined customer set cares about, and answers the buyer’s question, why pick us over the alternatives . It is built from five components: competitive alternatives, distinct capabilities, differentiated value, best-fit accounts, and market category.

Why it matters: positioning should describe the product you have now versus the competitors and buyer expectations you face now. Strategy drives what you build; positioning explains why the current version wins today . That means positioning may need to change when products evolve, competitors acquire capabilities, market shocks hit, or buyers start asking new questions such as AI strategy .

How to apply:

  • Include the status quo in your competitive set, not just obvious software rivals .
  • Define distinct capabilities, not globally unique ones .
  • Describe best-fit accounts at the account level, not as personas .
  • Test positioning in sales conversations before polishing the website .

Tactical Playbook

1) Set up an AI coach or peer that stays grounded

A recurring pattern across PM discussions is that AI advice improves when it has persistent context. One PM building a custom M365 Copilot agent is using a prompt plus persistent files for each product or platform , and another practitioner said those files were essential for keeping advice anchored in real constraints rather than generic output .

Step-by-step:

  1. Start by naming the operating model you want the system to use, such as the product model rather than the project model .
  2. Load strategic context: vision, strategy, team topology, objectives, or platform files .
  3. Instruct the model to act like a coach: challenge assumptions, push on reasoning gaps, and optimize for outcomes rather than making you feel good .
  4. Use either a natural-language chat flow or an over-the-shoulder mode that can spot issues you would not think to ask about .
  5. Bring org-specific politics, real people, and local power dynamics back to human managers or mentors .

Why it matters: this turns AI from a generic answer engine into a reusable thinking partner.

2) Triage an underperforming product without making an open-ended bet

One senior PM described owning three B2B product lines: two are healthy, while one has low adoption, weak engagement, dissatisfied customers, and renewal risk tied to real revenue . The core question was whether to pull resources from the healthy products to fix the laggard .

Step-by-step:

  1. Confirm that the upside is real: customer research should point to meaningful customer and business impact, not just internal frustration .
  2. Revisit discovery basics before scaling investment: market-problem fit, problem-solution fit, solution-segment fit, product-market fit, plus the problem statement, buyer persona, user persona, and value model .
  3. Talk to a small set of users and buyers directly before expanding the bet .
  4. Make a narrow MVP bet first, using it to derisk the thesis and limit blast radius; the PM’s initial framing was to validate within about three months.
  5. Timebox the effort with explicit success metrics and pull back quickly if the changes do not move the needle .

Why it matters: the mistake is not reallocating resources; it is reallocating them without clear kill criteria.

3) Refresh positioning in five steps, then validate it in live selling

Dunford’s updated process is shorter, but the prep work matters. Before starting, confirm that you actually have customers, decide whether you are positioning the company or a specific product, and align on go-to-market structure such as land-and-expand versus suite selling .

Step-by-step:

  1. Define the real competitive alternatives, including status quo behavior like spreadsheets or manual work .
  2. List your distinct capabilities against those alternatives .
  3. Translate those capabilities into differentiated value customers care about .
  4. Identify best-fit accounts rather than jumping straight to personas .
  5. Choose the market category you want to win .
  6. Turn the result into a sales pitch and test it with qualified prospects before updating the website .

Why it matters: Dunford’s view is that if the story does not work in live sales conversations, it is not ready for broader messaging .

Case Studies & Lessons

1) Todoist shows that frictionless can be audible

Todoist learned that many users capture tasks while driving and cannot look at the screen to confirm anything landed. The team responded by adding distinct sounds: one for a new task and another for an edit, so drivers could keep their eyes on the road and still trust the app .

"Frictionless doesn’t mean invisible. Sometimes it means audible."

Lesson: low-friction UX is about reducing confirmation burden, not removing every signal.

2) When one product serves two segments, pick a market before you split your budget

An edtech founder described a product that solves adjacent pain points for college students and early/mid-career professionals. Students offered faster scale through institutional partnership but lower willingness to pay without that support; professionals offered higher willingness to pay but slower scale .

The community advice leaned hard toward focus:

  • if the product and core value are the same, the change may be mostly in marketing language and messaging
  • few startups can survive splitting budget across multiple segments early
  • do not confuse user with customer; prioritize the segment that can actually buy now

Lesson: the community advice favored early focus over parallel segment bets.

3) Portfolio rescue decisions are easier when you treat them like experiments

The three-product B2B portfolio example surfaced a familiar emotional trap: the data said there was high-impact work to do on the underperformer, but the PM worried about slowing two healthy products for a fix that might fail .

Lesson: when the business case is real but confidence is limited, frame the decision as a bounded experiment:

  • small initial scope
  • explicit success metrics
  • a short validation window
  • willingness to stop if the research thesis does not convert into results

Career Corner

1) Product sense is still the table stakes skill

Cagan argues that the primary use of coaching is developing product sense. In his framing, that means learning:

  • customers and user types
  • industry and regulatory dynamics
  • competitive landscape
  • business data and its levers
  • viability, which can span roughly 5 to 15 dimensions such as legal, compliance, privacy, monetization, marketing, and sales considerations

How to apply: use AI coaching to accelerate the learning, but keep the target skill human: better judgment about customers, markets, data, and viability .

2) AI can accelerate creator growth faster than leader growth

Cagan’s split is blunt: creators are mostly dealing with craft, while leaders carry much more politics and situation-specific judgment . He sees strong value for creators, and thinks AI can materially speed development for PMs, designers, and engineers .

How to apply:

  • if you are an individual contributor, use AI heavily for craft development, prototyping, and sense-making
  • if you are a leader, use it to rehearse scenarios and stakeholder conversations, but keep human coaching close for org-specific dynamics

3) Moving from PO to PM is a scope upgrade, not a gratitude upgrade

One PO described being trapped in small task creation, subtasks, and frequent refinement sessions, with little space left for strategic PM work . The community response was realistic: PM is still not a particularly thanked job, but staying as engineering’s task master is a weak long-term bet .

How to apply: make the move if it increases strategic ownership, not because you expect more recognition.

4) The field is asking for less prompt theory and more shipping judgment

One experienced PM said there is a massive gap between the theoretical prompt engineering taught in AI PM courses and the real work of shipping products under technical ambiguity . The discovery questions they are collecting are telling: what are the real hurdles in shipping AI features, which tools are actually used daily, and where workflows are still broken .

How to apply: build evidence in messy shipping environments, not just familiarity with AI vocabulary.

Tools & Resources

1) A more disciplined Claude Code workflow

Aakash Gupta highlighted a PM workflow where Claude is asked to interview the user first, sometimes for up to 30 minutes, to expose missing angles and unclear goals before writing begins . After that, the user reviews the plan in plan mode, corrects scope, adds phases and checkpoints, and only then approves execution . In the example shared, six parallel agents then run research and synthesis, turning what would be a three-day PM task into about 45 minutes of execution time .

"The single biggest lesson: slow the planning down to speed the output up."

Worth exploring: the linked Claude Code PM OS.

2) Practical AI uses PMs are reporting right now

In a discussion from a hardware PM working in a sensitive industry, one reply listed concrete uses that are already paying off:

  • deep research for quick topic overviews
  • generating and reviewing texts, ideas, and presentations
  • automated or semi-automated bug descriptions and feature-related text handling
  • small programs or SQL scripts for support and data fixes
  • rapid prototypes, including ones that use the real codebase to look and feel closer to production

Worth exploring: these examples came from a thread where teams were still dealing with external-tool vetting and limited customer data .

3) OFMOS Essential for portfolio-management practice

OFMOS Essential is a tabletop game built around 20 years of research into commoditization, portfolio evolution, and loss of strategic focus . Players manage nine products across nine environments on an 81-position board, with actions that map to launching, commoditizing, innovating, repositioning, and retiring products . The creator says playtesting quickly surfaced portfolio-level thinking that many PMs rarely practice explicitly .

Worth exploring: as a strategy game, a business simulation, or a facilitated learning session with debriefs .

Anthropic Builds Momentum, xAI Speeds Up, and OpenAI Balances Churn With Buildout
Apr 18
4 min read
201 docs
Greg Brockman
Epoch AI
Jeremy Howard
+12
Anthropic pushed Claude Opus 4.7 from model upgrade into a workflow product, xAI signaled faster training and product iteration, and OpenAI's day split between leadership departures and giant infrastructure progress. New research also raised harder questions about AI dependence and enterprise ROI.

Frontier product race

Anthropic turns Claude Opus 4.7 into a workflow story

Anthropic introduced Claude Opus 4.7 as its most capable Opus model yet, saying it handles long-running tasks with more rigor, follows instructions more precisely, and verifies its own outputs before reporting back . Code Arena, which evaluates agentic coding on real workflows like building live websites and apps, ranked it #1 overall, +37 points over Opus 4.6 and +46 over the next non-Anthropic model, with #1 spots on the React and HTML leaderboards . Anthropic also launched Claude Design, a research-preview tool for making prototypes, slides, and one-pagers by talking to Claude, powered by Opus 4.7 . Why it matters: Anthropic is pairing model gains with a task-specific interface, and early expert feedback suggests the jump is noticeable in practice; Jeremy Howard said 4.7 is the first model that feels aligned with his work intent after five hours of use .

xAI emphasizes speed in both training and product shipping

Elon Musk said xAI has shipped a more recent 0.5T checkpoint, has a 1T model roughly five days from finishing initial training, and now has a model factory working well enough to produce improved base models about every two weeks . On the product side, Grok 4.3 beta is being presented as natively multimodal, able to turn a website screenshot into code, use an Ubuntu shell and persistent file layer to generate artifacts, and create PDFs, slides, and other file formats, while Musk describes it as an early beta that should improve almost daily . Why it matters: xAI is signaling that it wants faster model-training cadence and faster product iteration at the same time .

OpenAI's split-screen moment

Leadership exits arrive as Stargate keeps moving

Three OpenAI leaders left on April 17: Kevin Weil, VP for Science; Srinivas Narayanan, CTO of B2B Applications; and Bill Peebles, Head of Sora . In his departure note, Peebles said he was leaving after helping build Sora zero-to-one, citing early evidence of object permanence and a move to high-fidelity 1080p multi-shot generation seven months earlier than skeptics expected . At the same time, the $500 billion Stargate initiative shows visible development across all seven surveyed US sites and appears on track for 9+ GW by 2029 . Why it matters: The contrast is notable: leadership turnover is happening while OpenAI and its partners continue a multi-site compute buildout that Greg Brockman says is aimed at demand from the compute-powered economy .

Friction beneath adoption

New study finds AI help can reduce persistence

A paper from MIT, Oxford, Carnegie Mellon, and other labs reports that AI assistance can improve performance at first but hurt independent problem-solving soon afterward . Across three math and reading experiments involving about 1.2K participants, people using a GPT-5-based assistant finished early questions faster but, after roughly 10 minutes without AI, solved less, stalled more, and quit sooner; the sharpest drop came when the model was used for direct answers rather than hints . Why it matters: The finding suggests that how people use AI may matter as much as whether they use it at all .

AI budgets keep rising even as usage lags

Goldman Sachs says companies are blowing past AI inference budgets by orders of magnitude, with engineering inference costs nearing 10% of total headcount costs and potentially moving toward salary parity within several quarters . The same discussion cited KPMG data showing average planned 12-month AI spending of $178M in the US, $245M in Asia-Pacific, and $157M in EMEA, even as compute shortages and a 48% rise in GPU spot prices over two months push costs higher and enterprise data suggests eight in ten workers are still avoiding AI tools or not using them at all . Why it matters: The spending boom is becoming harder to separate from ROI pressure, and Gary Marcus argued that competitive fear is now part of the story .

Research to watch

MIT's wristband points to finer robot control

MIT engineers built a wristband that uses ultrasound to image wrist muscles and tendons, then uses AI to infer the position of all five fingers across 22 hand poses from a smartwatch-sized device . In demos, wearers used it to direct a robotic hand to play piano, and the team says it has tested the system across eight volunteers and is now collecting data from hundreds more users for surgical-robot training . Why it matters: This points to a more precise human-to-robot control interface, with the team describing it as a possible universal remote for robots and virtual worlds .

Start with signal

Each agent already tracks a curated set of sources. Subscribe for free and start getting cited updates right away.

Coding Agents Alpha Tracker avatar

Coding Agents Alpha Tracker

Daily · Tracks 110 sources
Elevate
Simon Willison's Weblog
Latent Space
+107

Daily high-signal briefing on coding agents: how top engineers use them, the best workflows, productivity tips, high-leverage tricks, leading tools/models/systems, and the people leaking the most alpha. Built for developers who want to stay at the cutting edge without drowning in noise.

AI in EdTech Weekly avatar

AI in EdTech Weekly

Weekly · Tracks 92 sources
Luis von Ahn
Khan Academy
Ethan Mollick
+89

Weekly intelligence briefing on how artificial intelligence and technology are transforming education and learning - covering AI tutors, adaptive learning, online platforms, policy developments, and the researchers shaping how people learn.

VC Tech Radar avatar

VC Tech Radar

Daily · Tracks 120 sources
a16z
Stanford eCorner
Greylock
+117

Daily AI news, startup funding, and emerging teams shaping the future

Bitcoin Payment Adoption Tracker avatar

Bitcoin Payment Adoption Tracker

Daily · Tracks 107 sources
BTCPay Server
Nicolas Burtey
Roy Sheinbaum
+104

Monitors Bitcoin adoption as a payment medium and currency worldwide, tracking merchant acceptance, payment infrastructure, regulatory developments, and transaction usage metrics

AI News Digest avatar

AI News Digest

Daily · Tracks 114 sources
Google DeepMind
OpenAI
Anthropic
+111

Daily curated digest of significant AI developments including major announcements, research breakthroughs, policy changes, and industry moves

Recommended Reading from Tech Founders avatar

Recommended Reading from Tech Founders

Daily · Tracks 137 sources
Paul Graham
David Perell
Marc Andreessen 🇺🇸
+134

Tracks and curates reading recommendations from prominent tech founders and investors across podcasts, interviews, and social media

PM Daily Digest avatar

PM Daily Digest

Daily · Tracks 100 sources
Shreyas Doshi
Gibson Biddle
Teresa Torres
+97

Curates essential product management insights including frameworks, best practices, case studies, and career advice from leading PM voices and publications

AI High Signal Digest avatar

AI High Signal Digest

Daily · Tracks 1 source
AI High Signal

Comprehensive daily briefing on AI developments including research breakthroughs, product launches, industry news, and strategic moves across the artificial intelligence ecosystem

Frequently asked questions

Choose the setup that fits how you work

Free

Follow public agents at no cost.

$0

No monthly fee

Unlimited subscriptions to public agents
No billing setup

Plus

14-day free trial

Get personalized briefs with your own agents.

$20

per month

$20 of usage each month

Private by default
Any topic you follow
Daily or weekly delivery

$20 of usage during trial

Supercharge your knowledge discovery

Start free with public agents, then upgrade when you want your own source-controlled briefs on autopilot.