Your intelligence agent for what matters

Tell ZeroNoise what you want to stay on top of. It finds the right sources, follows them continuously, and sends you a cited daily or weekly brief.

Set up your agent
What should this agent keep you on top of?
Discovering sources...
Syncing sources 0/180...
Extracting information
Generating brief

Your time, back

An AI curator that monitors the web nonstop, lets you control every source and setting, and delivers verified daily or weekly briefs.

Save hours

AI monitors connected sources 24/7—YouTube, X, Substack, Reddit, RSS, people's appearances and more—condensing everything into one daily brief.

Full control over the agent

Add/remove sources. Set your agent's focus and style. Auto-embed clips from full episodes and videos. Control exactly how briefs are built.

Verify every claim

Citations link to the original source and the exact span.

Discover sources on autopilot

Your agent discovers relevant channels and profiles based on your goals. You get to decide what to keep.

Multi-media sources

Track YouTube channels, Podcasts, X accounts, Substack, Reddit, and Blogs. Plus, follow people across platforms to catch their appearances.

Private or Public

Create private agents for yourself, publish public ones, and subscribe to agents from others.

3 steps to your first brief

1

Describe your goal

Tell your AI agent what you want to track using natural language. Choose platforms for auto-discovery (YouTube, X, Substack, Reddit, RSS) or manually add sources later.

Weekly report on space exploration and electric vehicle innovations
Daily newsletter on AI news and research
Startup funding digest with key venture capital trends
Weekly digest on longevity, health optimization, and wellness breakthroughs
Auto-discover sources

2

Review and launch

Your agent finds relevant channels and profiles based on your instructions. Review suggestions, keep what fits, remove what doesn't, add your own. Launch when ready—you can always adjust sources anytime.

Discovering sources...
Sam Altman Profile

Sam Altman

Profile
3Blue1Brown Avatar

3Blue1Brown

Channel
Paul Graham Avatar

Paul Graham

Account
Example Substack Avatar

The Pragmatic Engineer

Newsletter
Reddit Machine Learning

r/MachineLearning

Community
Naval Ravikant Profile

Naval Ravikant

Profile
Example X List

AI High Signal

List
Example RSS Feed

Stratechery

RSS
Sam Altman Profile

Sam Altman

Profile
3Blue1Brown Avatar

3Blue1Brown

Channel
Paul Graham Avatar

Paul Graham

Account
Example Substack Avatar

The Pragmatic Engineer

Newsletter
Reddit Machine Learning

r/MachineLearning

Community
Naval Ravikant Profile

Naval Ravikant

Profile
Example X List

AI High Signal

List
Example RSS Feed

Stratechery

RSS

3

Get your briefs

Get concise daily or weekly updates with precise citations directly in your inbox. You control the focus, style, and length.

Carmack’s Law Resurfaces as Keith Rabois Shares Stress, Standards, and History Picks
Apr 13
4 min read
190 docs
Lenny's Podcast
Tim Sweeney
Keith Rabois
After filtering out self-promotional material, today’s authentic picks split between Tim Sweeney’s resurfacing of Carmack’s Law source material and Keith Rabois’s recommendations on stress, performance, human behavior, history, and hiring.

What made the cut

After filtering out self-promotional material, today’s useful signal came in two clusters: Tim Sweeney resurfacing the primary sources behind Carmack’s Law, and Keith Rabois recommending books and videos on stress, elite standards, human behavior, history, and hiring .

Most compelling recommendation

Doom 3 interviews

  • Content type: Webpage / interview archive
  • Author/creator: Fabien Sanglard
  • Who recommended it: Tim Sweeney
  • Key takeaway: Sweeney points readers to this page as the original source for Carmack’s Law, then argues that the prediction proved prescient in computer graphics and is now clearly so in AI .
  • Why it matters: This is the strongest pick because Sweeney is not sharing it as trivia; he explicitly places the idea alongside Moore’s Law in importance .

Companion read: Gordon Moore’s 1965 paper

  • Content type: Paper
  • Author/creator: Gordon Moore
  • Who recommended it: Tim Sweeney
  • Key takeaway: He shares Moore’s paper as the comparison point for the scale of Carmack’s Law .
  • Why it matters: It gives readers the historical benchmark Sweeney invokes when making his case .

"Ultimately the prediction, which proved prescient in computer graphics and is now clearly so of AI, is as important as Moore’s..."

Keith Rabois’s operating recommendations

Rabois’s picks are less a conventional startup reading list than a compact set of resources on how ambitious people handle pressure, standards, judgment, and hiring .

The Upside of Stress

  • Content type: Book
  • Author/creator: Kelly McGonigal
  • Link/URL: Not provided in source material
  • Who recommended it: Keith Rabois
  • Key takeaway: He says the book argues that if you want to be happy, healthy, or wealthy, you need more stress in your life, not less, and he calls the evidence compelling and the book transformative .
  • Why it matters: It is his clearest recommendation for reinterpreting a condition many operators try to avoid rather than use .

The Jordan Rules

  • Content type: Book
  • Author/creator: Not specified in source material
  • Link/URL: Not provided in source material
  • Who recommended it: Keith Rabois
  • Key takeaway: He recommends it as the better guide to elite performance, with the blunt lesson that if you want to be Michael Jordan, you have to act like Michael Jordan .
  • Why it matters: It is the sharpest performance-culture recommendation in today’s batch .

Shakespeare

  • Content type: Plays / collected works
  • Author/creator: Shakespeare
  • Link/URL: Not provided in source material
  • Who recommended it: Keith Rabois
  • Key takeaway: Rabois says that everything important to learn about humans was written by Shakespeare, and that reading Shakespeare is better than customer research .
  • Why it matters: It shows he treats human observation as a practical advantage, not just a cultural interest .

Nuremberg Trial

  • Content type: Film
  • Author/creator: Not specified in source material
  • Link/URL: Not provided in source material
  • Who recommended it: Keith Rabois
  • Key takeaway: He says it contains lessons applicable to the modern world, taught him several things he did not know, and is useful for understanding historical travesties and how to prevent them .
  • Why it matters: This is the clearest recommendation today for sharpening present-day judgment through history .

Eric Glyman’s hiring speech

  • Content type: Talk / online video
  • Author/creator: Eric Glyman
  • Link/URL: Not provided in source material
  • Who recommended it: Keith Rabois
  • Key takeaway: Rabois says the speech covers hiring in detail and is fairly similar to his own views on the topic .
  • Why it matters: It is the only recommendation in today’s mix aimed directly at a core startup function: hiring .

Bottom line

The highest-signal item today is Sweeney’s Carmack’s Law source material because it comes with a precise claim about lasting importance for AI. The rest of the list is most useful as a pattern: Rabois’s recommendations repeatedly point readers toward better handling of stress, higher standards, deeper models of human behavior, and stronger historical judgment .

Distillation Gets Faster, Open Agents Self-Improve, and AI Supply Risks Turn Geopolitical
Apr 13
9 min read
480 docs
Perplexity
sarah guo
Together AI
+32
This brief covers a major distillation advance in TRL, the emergence of self-improving open agents around Hermes, a medical-monitoring use case from Penn, and new signals on distribution, compliance, and compute risk across the AI industry.

Top Stories

Why it matters: This cycle’s biggest developments were about leverage and constraints: better ways to compress frontier models, open agents that iterate on themselves, a concrete medical-monitoring use case, and external limits on compute supply.

1) Frontier distillation got materially more practical

TRL’s rebuilt on-policy distillation trainer now supports teacher models above 100B parameters and is reported to run more than 40x faster than naive implementations through buffer and payload optimizations . In the cited example, the team distilled Qwen3-235B into a 4B student and gained 39+ points on AIME25 . The same setup is presented as usable across Llama, Qwen, and Gemma model families . Blog: TRL distillation trainer.

Impact: This is a concrete post-training efficiency story: very large teacher models can now be used to produce much smaller students with less engineering overhead than before .

2) The open-agent stack is moving toward self-improvement

NousResearch’s hermes-agent-self-evolution repo automates a loop where the system reviews past task histories, identifies errors and root causes, proposes prompt or code changes, tests variants, and keeps the best-performing version . Today it focuses on optimizing SKILL.md, with a roadmap to extend the same loop to tool descriptions, system prompts, and underlying code . The repo is API-only for mutation and evaluation, includes test gates, file-size and scope limits, and still requires human-reviewed PRs before merge . Repo: hermes-agent-self-evolution.

Related activity around Hermes points the same direction: one contributor said Hermes was evolved into an autonomous ML/RL research lab with 94 Orchestra primitives for zero-shot GRPO and distillation , while Teknium said more is coming . Separately, a quality-of-life update made Hermes about 20% more likely to load the right skill for a task, and Teknium said the agent performed the prompt improvement and benchmarking itself .

Impact: Open agents are being pushed beyond static tool wrappers toward systems that can revise parts of their own operating instructions under guardrails .

3) A Nature paper showed a strong real-world medical use case for LLMs

University of Pennsylvania researchers analyzed more than five years of Reddit discussion—about 400,000 posts from 70,000 users—about GLP-1 drugs including Ozempic and Mounjaro, using LLMs to map patient language to medical terminology . The study surfaced side effects not fully reflected in current labels, including menstrual irregularities, hot flashes, chills, and fatigue. The authors frame this as computational social listening and argue that mainstream drugs may need real-time social-media monitoring as an early warning system beyond voluntary FDA-style reporting . Paper: Nature.

Impact: This is one of the clearest examples in this cycle of LLMs being used to expand surveillance capacity in a high-stakes domain rather than just improve a chatbot .

4) AI infrastructure risk is becoming geopolitical, not just technical

One analysis tied Middle East tensions to four AI supply-chain bottlenecks: helium, energy, shipping, and advanced chip concentration . It said the strike on Qatar’s Ras Laffan facility removed about a third of global helium supply, that helium is essential for sub-3nm fabrication, and that South Korea and Taiwan are highly exposed through suppliers such as SK Hynix and TSMC . The same thread argued that Strait closures could disrupt 25% of world energy supply, threaten Taiwan’s power availability, and block about 30% of global container shipping for semiconductor equipment, chemicals, and finished chips . It tied those risks directly to the roughly $500B AI data-center buildout planned for 2026 .

Impact: This frames compute risk as a question of energy, logistics, and industrial gases—not only model quality and GPU counts .

Research & Innovation

Why it matters: The research pipeline this cycle focused on new runtimes, cheaper arithmetic, and better long-context mechanics—not just bigger models.

  • Neural Computers: Meta AI and KAUST proposed Neural Computers, a learned system where computation, memory, and I/O move into the model’s runtime state . Early prototypes use video-model-style prediction to simulate terminal and GUI behavior directly from instructions, pixels, and user actions rather than executing code in a conventional OS . The authors and commentators are explicit that long-horizon reasoning, stable symbolic computation, and durable reuse are not solved yet, framing this as a first step toward a Completely Neural Computer. Paper: arXiv 2604.06425.

  • Huawei HiFloat4 FP4: Huawei proposed HiFloat4 (HiF4) for Ascend NPUs, reporting BF16-comparable performance with large compute-efficiency gains . A follow-on note summarized the key claim more precisely: about 90% of training computation can run in FP4 while keeping the loss gap within roughly 1.5% of a full-precision baseline . The same note said chips designed around this are still more than 1.5 years away. Paper: arXiv 2604.08826.

  • cuLA for linear attention:cuLA packages handwritten CUDA kernels for linear attention variants including GLA, KDA, GDN, and Lightning, built with CuTe DSL and CUTLASS for Hopper and Blackwell GPUs . The pitch is straightforward: linear attention gives O(N) scaling instead of O(N²), enabling million-token settings, while cuLA reports speedups from 1.32x–1.45x for KDA on Blackwell to 2.19x for Lightning Attention against an FLA Triton baseline . The migration path is also small—a one-line import swap from fla.ops.kda to cula.kda.

  • Retrieval models are back under debate: A fresh thread argued that ColBERTv2—a 100M late-interaction retriever—still beats Qwen3-Embed-8B on BrowseComp+ and LIMIT, despite the dense retriever being about 80x larger. Supporters took that as evidence that single-vector dense models generalize poorly out of domain . Critics replied that LIMIT is artificial, that synthetic datasets can mislead, and that quantization often trades only 1–5% recall for roughly 100x memory reduction and major speed gains . A separate observation from the same discussion is that retrieval papers are increasingly built on decoder-only LLMs rather than encoder-only models like BERT .

Products & Launches

Why it matters: New releases were less about generic chat and more about workflow control, evaluation, search, and document manipulation.

  • OpenClaw v2026.4.11 added ChatGPT import, a new Memory Palace for exploring chats as structured memory, guided plugin setup, richer chat rendering, better video generation handling, and stronger integrations with Teams, Feishu, WhatsApp, and Telegram . Changelog: openclaw v2026.4.11.

  • Opik was highlighted as an open-source tool for debugging, evaluating, and monitoring LLM apps, RAG systems, and agent workflows with tracing, automated evals, and dashboards . Repo: comet-ml/opik.

  • Qdrant Query Language (QQL) introduced a SQL-like interface for vector search, combining hybrid retrieval, filtering, and semantics in a more structured query flow . Article: QQL overview.

  • Verso AI launched with an engine that converts PowerPoint’s OOXML format into a representation that is easier for LLMs to edit; the team said its system beats competing tools on its internal benchmark .

  • Hermes Telegram Dashboard was announced for Hermes users, bringing terminal access, system-resource monitoring, and cron management into Telegram .

Industry Moves

Why it matters: Distribution, platform reach, and infrastructure economics continue to matter as much as model releases.

  • MiniMax M2.7 widened its distribution footprint. Together AI said the model is live on its platform, with day-0 availability on both serverless and dedicated infrastructure . Fireworks separately announced day-0 hosted availability, describing the model as production-ready with 200K+ context, strong reasoning, and native multi-agent support . At the same time, separate discussion around the open-weight release focused on the non-commercial license, with one user saying the restriction prevents at-home use .

  • Perplexity launched a company-building competition around Perplexity Computer. The Billion Dollar Build is an eight-week program for teams building a company with a path to a $1B valuation, with finalists eligible for up to $1M from the Perplexity Fund and up to $1M in Computer credits .

  • Infrastructure remains a constrained market. Notes from a HumanX fireside chat said the AI infra market is still early, that AI-native firms are using post-training to reach frontier quality more cheaply and quickly, and that inference remains constrained by both supply and talent .

  • Claude’s consumer reach kept rising. Similarweb reported 341.26% YoY growth in Claude website traffic in Q1 . One commentator attributed some of the appeal to Claude’s less chatty, more concise responses .

Policy & Regulation

Why it matters: This cycle had fewer formal government actions, but access control, compliance, and labor-policy debates continued to shape how AI systems are deployed.

  • Capability control is tightening around agent ecosystems. A Zhihu Frontier roundup said Anthropic removed OpenClaw from the Claude whitelist, describing it as a sign of tighter control over agent ecosystems . The same roundup said Anthropic Mythos was released only to select partners, framing that as a capability-control issue rather than a broad public rollout .

  • Compliance remains a live cross-border constraint. The same roundup said Slack’s shutdown in Greater China was presented as a compliance matter rather than an abrupt exit .

  • AI-risk activism remained under scrutiny after the attack on Sam Altman’s home. PauseAI said it unequivocally condemned the attack, said the suspect had only minimal participation in its public Discord, and stated he had no role in the group’s campaigns or events . The organization also argued that concern about advanced AI risk is shared by major researchers, legislators, and institutions, and positioned its own work around protests, petitions, and policy advocacy as a peaceful outlet .

  • The labor-policy conversation is widening. Sam Altman was cited as proposing a 4-day, 32-hour workweek and a new social contract in response to AI and robotics . At the same time, Alex Karp argued AI will destroy humanities jobs and increase the importance of vocational training and hands-on skills , while Aaron Levie argued the opposite dynamic may hold in law, predicting more lawyers because AI will generate more legal questions and new areas of compliance work .

Quick Takes

Why it matters: Smaller items still help map where capability, demand, and uncertainty are moving next.*

  • Unitree R1 is now available for global pre-order on AliExpress starting at $6,806, with June 30 deliveries for the base R1 AIR trim; the EDU version adds custom software support, full SDK access, and optional Jetson Orin .

  • Muse Spark kept attracting favorable external reactions. Meta said it rebuilt its AI stack from scratch and that Muse Spark now powers Meta AI . François Fleuret said the model passed his tests, including image generation , and Alexandr Wang said it is particularly good at finding open-source data and analyzing it .

  • Claude Opus 4.6 sparked another benchmark dispute. BridgeBench claimed its hallucination accuracy fell from 83.3% to 68.3% and described the model as nerfed . Critics replied that the newer run used 30 tasks instead of 6, and said shared-task scores changed only from 87.6% to 85.4%, which they argued could be statistical noise . Separate commentary noted that inference at scale can legitimately create performance variability across services .

  • LangChain clarified its agent-SDK stack. Harrison Chase described create-agent as minimal, deepagents as more batteries-included, and middleware as the advanced customization layer across both . In a related exchange, he pointed to deepagents when asked for a strong open-source alternative to the Claude Agent SDK .

  • Open-source AI security tooling got a fresh checklist. A roundup highlighted NeMo Guardrails, Promptfoo, LLM Guard, garak, DeepTeam, Llama Prompt Guard 2-86M, ShieldGemma 2, OpenGuardrails, Cupcake, and CyberSecEval 3 – Visual Prompt Injection as lightweight alternatives while Mythos remains closed .

Perplexity’s Builder Bet, New Agent Memory Layers, and the Inference Margin Fight
Apr 13
5 min read
486 docs
Perplexity
sarah guo
Keith Rabois
+11
Perplexity is using capital and credits to source AI-native founders, while early teams in agent memory, accessibility, and vertical apps are showing traction. The broader backdrop is an inference-led market shift: infrastructure bottlenecks, stealth churn in horizontal SaaS, and a growing expectation that investors build with the tools they back.

1) Funding & Deals

  • Perplexity's Billion Dollar Build is the clearest capital signal in this batch. The company launched an 8-week competition for teams using Perplexity Computer to build a company with a path to $1B; finalists can receive up to $1M in investment from the Perplexity Fund and up to $1M in Computer credits . Aravind Srinivas framed it as a bet that a new generation of builders can create very large businesses with these tools .

  • Keith Rabois restated a founder-first seed/Series A thesis. He says early investing starts with whether a founder has a non-zero chance of changing an industry or the world. In the interview he pointed to Fair—founded by former Square colleagues Max Rich and Jeff Collison—as a case where execution tempo stood out early , and said Ramp's progress toward shipping cards in 3 months instead of the usual 9–12 months was one reason he preempted the Series A .

2) Emerging Teams

  • GrayMatter is a lightweight persistent memory layer for agents that drops in with three lines of code, stores observations, checkpoints, and a knowledge graph in a local file, and pulls back only relevant context with hybrid retrieval . The solo builder reports up to 90% token savings after 100+ sessions, with native MCP support for Claude Code and Cursor, offline operation, and an open-source repo at github.com/angelnicolasc/graymatter.

  • slashlast30days v3 has notable open-source traction at 20,000+ GitHub stars and a differentiated search stack . The new version makes Reddit, X, and YouTube transcripts free without API keys, adds a Python pre-research engine that resolves the right handles, subreddits, hashtags, and channels before search, and ships cross-source cluster merging, comparisons, person-mode, and ELI5 mode . Garry Tan described the broader direction as "personal open source software" .

  • Hearica is an accessibility-driven Windows app that captions or translates all PC audio through a floating overlay and adds speaker separation, session saving, and custom context profiles . The founder says it was motivated by personal hearing loss, launched last month, already has paid subscribers, and builds on an earlier OSS system-captioning project using Whisper models .

  • AdmitOdds comes from a founder with direct user empathy: an 18-year-old high school senior built the product from his own college admissions pain point . Reported traction is 415 user accounts, 18 paying subscribers at $19.99/month, and $0 ad spend, with growth coming from Reddit, TikTok, and word of mouth . The stack is Next.js, Supabase, Stripe, and Claude/GPT .

3) AI & Tech Breakthroughs

  • Hermes shows a more explicit self-improvement loop for agents. In the cited architecture discussion, Hermes does not rely on offline trajectory mining or a separate skill-extraction model; the runtime notices reusable workflows, writes them into durable artifacts through the skill interface, and reuses them in the same act-notice-write-reuse loop . Garry Tan highlighted it as evidence that app-level applied research is increasingly happening in the open through open source .

  • BDH fast weights offers a concrete memory mechanism outside the normal context-window tradeoff. The setup keeps the transformer backbone frozen, updates only an isolated memory buffer for 300 gradient steps, and retained 20/20 unrelated facts after a cold reload with median probability 0.997 and cross-contamination below 0.03. The demo used a 15M-parameter model trained on 250M tokens on a single consumer GPU, and the code is open source .

  • Coding-agent differentiation may be thinner than branding suggests. A technical analysis of Cursor 3.0 claims that Cursor Agent is effectively Claude Code behind a local proxy with prompt find-and-replace, bundled Anthropic claude-agent-sdk and claude-code packages, and a custom fine-tuned Claude 3.7 Sonnet model . Garry Tan cited the report as further confirmation of the "thin harness, fat skill" pattern .

4) Market Signals

  • Horizontal SaaS may be masking AI-driven churn. SaaStr argues that specialized AI tools are quietly replacing specific workflows for power users even while overall accounts still look healthy . In the example, Canva work had already been split across Reve for images, Opus Pro for clips, and Higgsfield for short-form video . The practical takeaway is to track power-user engagement separately and talk to users who are still paying but have stopped showing up .

  • Inference is where more investors now see the bottleneck—and the margin fight. One analysis projects inference spend approaching a 10:1 ratio to training by mid-2026, with competition moving to the lowest marginal cost and highest reliability . In parallel, HumanX takeaways summarized by Sarah Guo say the AI infra market is still early, AI-native companies are using post-training to reach frontier quality cheaper and faster, and inference remains supply- and talent-constrained . The moats highlighted in the essay are upstream: custom silicon, power and cooling access, proprietary data loops, and system-level integration .

  • The investor bar is moving toward practitioners, not observers. Harry Stebbings argues investors who are not building with AI will miss the real opportunities, bottlenecks, and failure modes . Keith Rabois's version of the same pressure is operational: in fast-moving AI companies, long roadmaps make less sense, business-minded engineers gain leverage, and accumulating advantages matter more because foundation-model progress is compressing time horizons .

"If you are not building with AI today you simply should not be investing."

  • AI-generated code is creating repeatable diligence risks. One security scanner builder says a missing Stripe webhook signature check appears in almost every AI-built SaaS they review, enabling fake payment succeeded events if an endpoint accepts unsigned payloads . The claimed fix is four lines, and the same builder now offers xploitscan.com to scan for this and roughly 150 other AI-generated code patterns .

5) Worth Your Time

  • Jason Kneen's Cursor 3.0 report — Useful teardown if you are tracking coding-agent defensibility and how much value sits in the harness versus the provider layer .

  • The 3 Year Inference Landscape — A compact framework for where margins may accrue across chips, models, hosting, power, and system integration as inference scales .

  • SaaStr on Canva and stealth AI churn — A practical retention lens for boards: watch the power user who quietly stopped showing up before renewal pressure hits the numbers .

  • slashlast30days v3 thread — Worth scanning if you care about agent-led search, especially the pre-research layer that decides where to search before the report is assembled .

Cursor Details Its Async-Agent Playbook; DevTools MCP Adds Agent QA Skills
Apr 13
5 min read
41 docs
Kent C. Dodds ⚡
Lenny Rachitsky
Aman Sanger
+5
Aman Sanger shared rare production metrics and concrete operating patterns for cloud coding agents, from artifact-first review to planner/subplanner trees. Also in today’s signal: DevTools MCP adds built-in web quality checks, Kent C. Dodds shows an agent-to-agent repair loop, and Addy Osmani puts guardrails around parallel-agent overload.

🔥 TOP SIGNAL

Cursor founder Aman Sanger shared rare production datapoints: async coding agents running in cloud VMs are already generating 30% of merged PRs internally, and by the end of 2025 agent requests exceeded tab accepts even though tab can fire on every keystroke . This is not toy output: he cites an eight-hour React-to-Rust video-rendering refactor that made the system 25x faster, plus a 10,000-line PR for network policy controls . The actionable part is the workflow: give long-running agents a full computer, let them test their own work, and review video/report artifacts before you spend time reading code .

🛠️ TOOLS & MODELS

  • Cursor Cloud Agents: model routing is task routing. Cursor uses OpenAI models for higher-level planning/orchestration, Gemini and Anthropic for computer use and multimodal work, Anthropic for stronger UI generation, and faster models for simpler subagent tasks where they match performance at lower latency .
  • Chrome DevTools MCP just got more useful for agent QA. Addy Osmani says it now includes Lighthouse performance checks, a memory leak detection skill, an accessibility debugging skill, an LCP optimization skill, and an experimental CLI .
  • Codex UX signal from a power user. Riley Brown likes Codex because it stays bare-bones, keeps project work in left-sidebar folders with multiple threads, and should auto-spin new threads when a request splits into multiple tasks instead of adding extra interface layers like a Codex Cowork middle step .
  • Local model reality check. Armin Ronacher says local models are not universally there yet, but Qwen3-Coder-Next on a beefy Mac is useful for its speed and decent tool calling; his broader point is not to dismiss local models outright, especially for tasks like STT .

💡 WORKFLOWS & TRICKS

  • Artifact-first review loop for long-running agents.
    1. Run the agent in a cloud VM with the same kind of tools you'd use locally.
    2. Let it implement and test the change itself.
    3. Review its artifacts first, especially a demo video or research report.
    4. If something looks off, reprompt from the artifact instead of diving straight into the diff.
    5. Only review code once the behavior looks right .
  • Break long jobs into a planner tree. Cursor's working pattern is planner -> subplanners -> workers. Keep each leaf task simple enough to stay within training distribution, cap the outer agent to hundreds of thousands of tokens rather than tens of millions, and use faster models for simpler subagent tasks when they achieve the same result .
  • Use event-driven agents, not just ad hoc prompts. Cursor's automations fire on issues, pages, training runs, PRs, and security events. Their examples: an agent investigates a middle-of-the-night page and proposes a single-click fix, watches training metrics and logs to catch failures early, and finds PR vulnerabilities before shipping .
  • Let one agent escalate to another agent. Kent C. Dodds describes a clean loop with Kody: when Kody hits a problem, he tells it to kick off a Cursor Cloud Agent, and Kody already knows the problem context well enough to compose the repair prompt itself .
  • Parallel agents are not free throughput. Addy Osmani's advice: time-box long agentic sessions like deep work, tighten the scope for each agent, and treat finding your personal ceiling as a skill. Addy's longer note: Your parallel Agent limit.

I can fire up four agents in parallel and have them work on four different problems, and by 11am I am wiped out for the day.

  • On unfamiliar platforms, shrink the blast radius. Trash Dev says he shipped his first iPhone app with 100% AI-generated code despite no prior Swift or iOS experience, but kept it local-only, manually checked local storage and external calls, and used a separate design tool when the UI was the bottleneck .

👤 PEOPLE TO WATCH

  • Aman Sanger. Shared production metrics and concrete architecture details: PR share, artifact review, multi-agent decomposition, model routing, and event-driven automations .
  • Addy Osmani. Posted on two useful fronts today: human-in-the-loop constraints for parallel agents and practical agent QA tooling via DevTools MCP .
  • Kent C. Dodds. Showed an agent-to-agent repair loop where the agent using the software can hand off to a cloud repair agent without re-explaining the bug .
  • Armin Ronacher. Shared a grounded local-model take: one concrete setup that works on strong Mac hardware today instead of blanket yes or no claims about local coding models .

🎬 WATCH & LISTEN

  • 3:48-4:48 — Why async agents need their own computers. Sanger lays out why long-running agents need cloud VMs with full desktop access and testing, not just editor-side autocomplete .
  • 6:40-8:14 — Review artifacts before code. Sanger explains why videos and research reports are faster review surfaces than raw diffs for early agent iterations .
  • 8:18-9:05 — Zero-Swift iOS app, but still audit the risky path. Trash Dev explains the 100% AI-coded experiment, then grounds it by keeping the app local-only and manually checking storage and external calls before shipping .

📊 PROJECTS & REPOS

  • Chrome DevTools MCP. New quality-check surface for agents: Lighthouse performance checks, memory leak detection, accessibility debugging, LCP optimization, plus an experimental CLI. Repo: chrome-devtools-mcp.
  • Cursor's browser prototype. Sanger says Cursor ran a one-week multi-agent build that spent billions of tokens and tens of thousands of dollars of compute to produce a working, but still far-from-production, browser .

Editorial take: today's serious workflows looked less like autocomplete and more like ops — cloud VMs, event triggers, review artifacts, and explicit limits on human attention.

AI Reshapes PM Roles as Pricing and Team Design Move Upstream
Apr 13
11 min read
89 docs
scott belsky
Eric Glyman
Keith Rabois
+5
This brief covers four major PM signals: AI is shifting the role toward business judgment and demos, pricing is moving upstream into product design, team throughput depends more on initiative owners than headcount, and internal AI enablement is becoming a company advantage. It also includes practical playbooks, pricing case studies, career signals, and lightweight resources to use this week.

Big Ideas

1) AI is moving PM leverage from long roadmaps to business judgment

Keith Rabois argued that conventional PM work—collecting customer inputs and translating them into year-long sequential roadmaps—makes less sense when AI capabilities change week to week . In his framing, the enduring skill is business acumen: deciding what to build, why it matters, and how it moves the company’s equation, with communication shifting from slideware to working demos . Separately, an experienced PM on Reddit described feeling like a beginner again and struggling to find advanced material on model-first products, non-deterministic outcomes, and responsible scaling .

“Every presentation on product has to be a workable demo”

Why it matters: The role bar is moving away from artifact production and toward judgment, speed, and comfort with ambiguous AI behavior .

How to apply: Shorten planning cadence, ask teams to demo working product instead of presenting static decks, and spend learning time on model behavior, system constraints, and business impact—not just prompt tactics .

2) More headcount does not create more parallel progress

Rabois’s “barrels and ammunition” framework argues that initiative capacity is limited by the number of people who can independently take something from inception to success, not by total employee count . He describes ammunition as the supporting roles around them, and notes that post-funding teams often raise burn without increasing velocity because they add people without increasing barrel count . His examples: PayPal had roughly 12-17 barrels among 254 people, while a strong company might only have 2 .

Why it matters: If barrel count stays flat, adding people can increase coordination tax instead of throughput .

How to apply: Count true initiative owners first, cap the number of major parallel bets to that number, then add the right ammunition per initiative rather than assuming blanket hiring will fix execution .

3) Pricing and packaging are upstream product decisions—especially in AI

The Mind the Product discussion framed pricing as a cross-functional system spanning product, sales, marketing, finance, and rev ops, with product either leading or deeply involved because pricing connects how value is created and how it is captured . A recurring warning was to avoid building the product first and only then asking what to charge, because architecture, billing, GTM, and packaging may already be misaligned . For AI products, speakers argued the old fundamentals still apply—price to value—but hybrid models, clear plan tiers, and constrained usage patterns may be safer than open-ended chat experiences that create unpredictable token costs .

“Pricing’s never done. It’s not a one and done.”

Why it matters: Pricing errors can come from org design and product design, not just bad math . In AI, poor packaging can also weaken margin discipline .

How to apply: Move pricing decisions earlier, tie them to ICPs and value metrics, and test packaging before launch rather than after revenue stalls .

4) Internal AI enablement is becoming a company-level product advantage

Ramp’s Glass case study is a concrete example: 99% of employees were already using AI daily, but many were stuck because setup was painful and fragmented . Ramp responded by building a day-one AI workspace with SSO integrations, a marketplace of 350+ reusable skills built by colleagues, persistent memory, and scheduled automations so one person’s improved workflow could propagate across a team . Scott Belsky summarized the broader implication succinctly .

“don’t just build a differentiated product, build a differentiated company”

Why it matters: The advantage may come less from model access itself and more from how fast an organization turns AI usage into reusable workflows .

How to apply: Remove setup friction, standardize access, and create a way for employee-discovered workflows to become shared assets instead of personal hacks .

Tactical Playbook

1) Run a pricing and packaging sprint before you ship

  1. Align execs on the business objective. Start with strategy and commercial intent; if leaders are not aligned on why the product exists, the rest of the pricing work will drift .
  2. Segment the customer base. Break customers down by ARR, ACV, revenue, or deal size, then identify which ICPs actually get the most value .
  3. Write the value proposition by segment. Tie packaging and plan language to positioning, not just features .
  4. Quantify derived economic value. Ask whether you help customers make money, save money, or reduce risk, and use concrete proxies where needed; one example cited small business time at about $30 per hour .
  5. Test willingness to pay. Use methods like conjoint, Van Westendorp, and Gabor Granger; Van Westendorp specifically tests the “too cheap,” “bargain,” “expensive but acceptable,” and “too expensive” range .
  6. Package deliberately. Use plan tiers, a scaling value metric, add-ons, and a leader/filler/killer pass so core value is obvious .
  7. Revisit continuously. Pricing is iterative, not a one-time launch choice .

Why this matters: It prevents the common failure mode of shipping a product whose architecture, billing, and GTM do not support monetization .

2) Match discovery method to market shape

  1. Use direct customer conversations when there are identifiable decision-makers. Rabois explicitly said enterprise customer development is useful when you can talk to the actual buyer or a small set of must-win accounts .
  2. Ask value questions in those conversations. The pricing discussion recommends derived economic value questions as a standard part of customer interviews .
  3. Be more skeptical of stated preferences in consumer and SMB. Rabois argued that for consumer, SMB, and micro-merchant products, customer interviews can be directionally wrong because users struggle to explain subconscious purchase behavior .
  4. Lean more on observed behavior and economics in those markets. His suggested fallback was instincts plus outcomes like ticket sales, CAC, and LTV rather than over-weighting a handful of interviews .
  5. Do not universalize one method. Community responses on healthy PM orgs still emphasized consistent customer engagement overall, so the practical lesson is to adapt the discovery loop to the market rather than follow one doctrine everywhere .

Why this matters: Discovery quality drops when teams apply enterprise-style interview logic to markets where buyer motivation is hard to verbalize—or ignore customers entirely where decision-makers are knowable .

3) Audit initiative capacity before you add headcount

  1. Count barrels honestly. A barrel is someone who can take the company “over the hill” from idea to result, including motivating people, gathering resources, and measuring outcomes .
  2. Map each major initiative to a barrel. That defines how many important bets you can truly run in parallel .
  3. Pause blanket hiring if barrel count has not changed. Otherwise you risk more burn with the same or less output .
  4. Add ammunition per problem, not by org chart default. Some initiatives need design, engineering, PM, or data support; others need very little .
  5. Hire or promote barrels first when growth demands more throughput. That is the lever Rabois says increases parallel initiative capacity .

Why this matters: It reframes scaling from “more people” to “more independently completable bets” .

4) Design PM work so signal and autonomy can coexist

  1. Set vision and strategy at the top. One community description of a healthy org started with clear product vision and strategy from leadership .
  2. Give PMs room to steer within that frame. The same comment described full freedom to run a product line as long as it aligns with strategy, with accountability for wins and losses .
  3. Protect PM focus. Strong POs or similar roles can absorb dev communication, tickets, and documentation so PMs can stay focused on customers, markets, stakeholders, and roadmap quality .
  4. Keep feedback capture lightweight. A shared spreadsheet can be enough to record customer observations and requests .
  5. Treat manager quality as a force multiplier. Multiple community responses stressed that a good boss often determines whether the environment is workable day to day .

Why this matters: Healthy product teams are defined less by perfection than by clear direction, usable signal, and operating conditions that amplify PM strengths .

Case Studies & Lessons

1) File & Fight: packaging alone drove a 3x deal-size increase

One speaker said their packaging launch produced a 3x increase in price or deal size.

Why it matters: Monetization upside can come from packaging and plan structure even when the core product is unchanged .

How to apply: Before adding new features, ask whether clearer tiers, a better value metric, or better add-ons would unlock higher willingness to pay .

2) Aura: prove value first, then monetize with a stronger story

The Aura example emphasized getting the product in, proving value with NZ Police, and then using that proof as both monetization support and a marketing story for other police forces .

Why it matters: In some B2B settings, sequencing matters more than the first quoted price .

How to apply: If adoption friction is high, consider a GTM motion that lets customers experience value before you optimize pricing—and turn the best proof points into referenceable case studies .

3) Tracksuit: use price as a discovery input, not just an output

Tracksuit reportedly started with a blunt question: “What would you pay $10,000 for?” and used the answers to shape the initial feature set .

Why it matters: Pricing questions can reveal which problems are valuable enough to deserve a roadmap slot .

How to apply: In early discovery, ask customers what outcome or feature set would justify a meaningful budget line, then build around that signal instead of brainstorming features in isolation .

4) Lifetime pricing: great for validation, risky for scaling

A startup founder said lifetime access priced at about 3x the monthly plan helped generate first revenue and validate the product when there were no users yet . The downside came later: every new feature and cost increased customer value but not revenue from those lifetime users, pushing the founder toward recurring pricing instead . Community replies distilled the trade-off: lifetime reduces friction early, but the complexity reappears later; one suggested keeping core access while charging separately for future costly features such as AI .

Why it matters: Early validation pricing can quietly determine whether a SaaS business remains monetizable as costs rise .

How to apply: If you use lifetime offers for early traction, define clear boundaries up front around future features, usage, and premium add-ons .

5) Ramp Glass: internal AI adoption becomes reusable product infrastructure

Ramp saw heavy AI usage already in place, but setup friction blocked broader leverage . Glass addressed that with a standardized workspace and a marketplace of 350+ reusable skills, letting one person’s better workflow spread across the team .

Why it matters: Internal productivity gains compound when workflow knowledge becomes shareable infrastructure rather than isolated experimentation .

How to apply: Treat internal AI workflows like product surfaces: standardize onboarding and make the best patterns easy to copy .

Career Corner

1) Senior PMs are feeling the AI skill reset in real time

A PM with 12 years in industry and 9 years in product described feeling like “a beginner again” because the field is moving from traditional strategy and user empathy toward model-primary products, non-deterministic behavior, and responsible scaling . Another commenter replied simply: “I’m in the same boat” .

Why it matters: This is not just a junior-skills problem; experienced PMs are openly describing a missing bridge between classic PM strength and AI-native product depth .

How to apply: Move from passive reading to hands-on building with engineers and designers, even on small MVPs, and target specific gaps like model-centric architecture and managing probabilistic behavior .

2) A healthy PM org is usually easier to describe operationally than culturally

Community answers converged on a few concrete signs: high autonomy inside a clear top-level strategy, the ability to focus on core PM work, consistent customer engagement, a strong boss, and simple systems for capturing feedback .

Why it matters: These are interviewable signals. “Healthy” is less abstract when you can test for direction, decision rights, manager quality, and how customer signal actually flows .

How to apply: In interviews, ask who owns strategy, how PMs interact with customers, who handles dev-groundwork tasks, where feedback lives, and how success or failure is assigned .

3) The PM market is still tight, so adjacent roles matter

One experienced PM described months of unemployment and repeated late-stage rejections despite having design and engineering chops . Replies called the market “brutal” and pointed to adjacent roles such as solutions architecture, technical sales, sales engineering, forward deployment, Product Owner, Agile practitioner, or even coding as fallback paths . Another community note suggested that candidates without strong pedigree signals may need to target early-stage startups and build something with real metrics first .

Why it matters: Career resilience may depend on how well you can reframe PM experience into nearby commercial or technical roles .

How to apply: Tailor applications toward adjacent roles when needed and, if you are trying to break back into PM, keep a concrete project with proven metrics ready to show .

Tools & Resources

1) Simon Kutcher’s value-based packaging framework

Use it to structure tiers around good/better/best, a scaling value metric, and add-ons like implementation or support, then sort features into leaders, fillers, and killers .

Best use: When you need a shared language for pricing conversations across product, sales, CS, and marketing .

2) The pricing research toolkit: conjoint, Van Westendorp, Gabor Granger

These were presented as the pricing equivalent of user research techniques, with Van Westendorp offering a straightforward acceptable-price-range test .

Best use: When stakeholder debate is outrunning actual willingness-to-pay evidence .

3) The penetration × willingness-to-pay feature matrix

Map features by high or low penetration and high or low willingness to pay to separate core leaders from niche add-ons or low-value distractions .

Best use: When executives want to pack too much into the base plan or when roadmap debates need a monetization lens .

4) A lightweight customer feedback repository

One community answer reminded PMs that the system can be simple: a shared spreadsheet of customer observations and requests .

Best use: When the team needs a fast, low-friction place to accumulate signal before investing in heavier tooling .

5) Two sessions worth watching this week

Apple’s Siri Platform Bet, Hassabis’s AGI Safety Push, and Pressure on Open Models
Apr 13
4 min read
200 docs
Demis Hassabis
Ben Thompson
Sebastian Raschka
+7
Apple appears set to widen Siri’s role as an AI access layer while defending its hardware talent base against OpenAI. Demis Hassabis paired a near-term AGI estimate with calls for minimum safety standards, while MiniMax’s license shift and new research signals pointed to tightening AI economics and evolving technical directions.

Platform control is the clearest strategic theme

Today’s clearest thread is control of the interface: Apple is widening Siri’s model options while the hardware talent battle with OpenAI is already visible .

Apple is preparing to make Siri a gateway for outside AI

Apple is preparing an iOS 27 Siri overhaul that would let installed apps such as Google Gemini and Anthropic Claude handle queries inside Siri, extending the current ChatGPT integration . In Ben Thompson’s analysis, that would let Apple aggregate outside AI through the device and the App Store, rather than matching hyperscaler spending on model infrastructure .

Why it matters: If Apple keeps the user interface and subscription relationship while model providers compete underneath, it strengthens Apple’s position as the point of integration .

Apple is already responding to OpenAI’s hardware recruiting

Apple awarded rare out-of-cycle bonuses to iPhone hardware designers amid concern about departures to AI startups, with OpenAI described as a particular threat . OpenAI’s hardware effort is run in part by former Apple executive Tang Tan, advised by Jony Ive, and has hired several dozen Apple engineers across iPhone, iPad, Apple Watch, and Vision Pro teams .

Why it matters: The competition is no longer only about models or apps; it is also about who controls the next AI-centric device and the teams that can build it .

Hassabis pairs a near-term AGI estimate with safety coordination

Demis Hassabis says AGI may be ‘maybe five years away’

Google DeepMind CEO Demis Hassabis said AGI may be ‘maybe five years away’ and framed AI as a scientific tool for understanding the universe and tackling medicine, energy, and environmental challenges . He also said there is a non-zero chance things go badly if the technology is built the wrong way, arguing for cautious optimism, minimum standards among leading labs, and more international cooperation as companies and nations race toward the technology .

"There’s definitely ... a non zero chance that things could go quite badly wrong if the technology is not built in the right way."

Why it matters: A leading frontier-lab CEO is coupling a relatively near AGI timeline with explicit coordination asks, not just capability forecasts .

Open models are running into harder economics

MiniMax moves to a non-commercial license

MiniMax has shifted to a non-commercial license, with attribution requirements for users above $30M in revenue or 100M users, alongside an acceptable use policy . Nathan Lambert said the move looks like what happens when open-model companies ‘start to worry about money’ and reiterated his view that MiniMax, Moonshot AI, and Zhipu AI could face financial strain if their strategies hold . He separately said usage of 30-200B open models appears to be surging, though attribution is still hard to pin down .

Why it matters: Demand for open models may be rising, but this license change is a concrete sign that free-use frontier releases are getting harder to fund .

Research signals worth watching

‘Neural Computers’ push world models into the interface layer

A new ‘Neural Computers’ paper proposes learning a computer interface directly as a world model: the system takes keystrokes, mouse clicks, and previous screen pixels, then generates the next frames, readable text, and cursor movement without a traditional operating system . The authors describe it as a first step toward a ‘Completely Neural Computer’ where computation, memory, and I/O move into a learned runtime state; the paper is here.

Why it matters: Instead of putting an AI agent on top of software, this line of work asks whether part of the software runtime itself can move inside the model .

Raschka’s 2026 LLM readout centers on efficiency and agent use

Sebastian Raschka used recent releases including Nemotron 3 Super and Qwen 3.5 as anchor points for reading where LLM design is going . His shortlist of standout trends includes hybrid transformer/SSM designs such as Qwen 3.5’s gated delta-net layers and Nemotron 3’s hybrid attention, multi-head latent attention to compress KV cache, learned sparse attention for long contexts, and RLVR-driven reasoning that spreads through distillation . He also said Qwen 3.5 is being built with agentic use, tool calling, and more affordable long contexts in mind .

Why it matters: The emphasis is moving toward architectures that make long-context use, reasoning, and agent behavior cheaper and more practical .

Start with signal

Each agent already tracks a curated set of sources. Subscribe for free and start getting cited updates right away.

Coding Agents Alpha Tracker avatar

Coding Agents Alpha Tracker

Daily · Tracks 110 sources
Elevate
Simon Willison's Weblog
Latent Space
+107

Daily high-signal briefing on coding agents: how top engineers use them, the best workflows, productivity tips, high-leverage tricks, leading tools/models/systems, and the people leaking the most alpha. Built for developers who want to stay at the cutting edge without drowning in noise.

AI in EdTech Weekly avatar

AI in EdTech Weekly

Weekly · Tracks 92 sources
Luis von Ahn
Khan Academy
Ethan Mollick
+89

Weekly intelligence briefing on how artificial intelligence and technology are transforming education and learning - covering AI tutors, adaptive learning, online platforms, policy developments, and the researchers shaping how people learn.

VC Tech Radar avatar

VC Tech Radar

Daily · Tracks 120 sources
a16z
Stanford eCorner
Greylock
+117

Daily AI news, startup funding, and emerging teams shaping the future

Bitcoin Payment Adoption Tracker avatar

Bitcoin Payment Adoption Tracker

Daily · Tracks 107 sources
BTCPay Server
Nicolas Burtey
Roy Sheinbaum
+104

Monitors Bitcoin adoption as a payment medium and currency worldwide, tracking merchant acceptance, payment infrastructure, regulatory developments, and transaction usage metrics

AI News Digest avatar

AI News Digest

Daily · Tracks 114 sources
Google DeepMind
OpenAI
Anthropic
+111

Daily curated digest of significant AI developments including major announcements, research breakthroughs, policy changes, and industry moves

Recommended Reading from Tech Founders avatar

Recommended Reading from Tech Founders

Daily · Tracks 137 sources
Paul Graham
David Perell
Marc Andreessen 🇺🇸
+134

Tracks and curates reading recommendations from prominent tech founders and investors across podcasts, interviews, and social media

PM Daily Digest avatar

PM Daily Digest

Daily · Tracks 100 sources
Shreyas Doshi
Gibson Biddle
Teresa Torres
+97

Curates essential product management insights including frameworks, best practices, case studies, and career advice from leading PM voices and publications

AI High Signal Digest avatar

AI High Signal Digest

Daily · Tracks 1 source
AI High Signal

Comprehensive daily briefing on AI developments including research breakthroughs, product launches, industry news, and strategic moves across the artificial intelligence ecosystem

Frequently asked questions

Choose the setup that fits how you work

Free

Follow public agents at no cost.

$0

No monthly fee

Unlimited subscriptions to public agents
No billing setup

Plus

14-day free trial

Get personalized briefs with your own agents.

$20

per month

Includes $20 of usage each month

Private by default
Any topic you follow
Daily or weekly delivery

Includes $20 of usage during trial

Supercharge your knowledge discovery

Start free with public agents, then upgrade when you want your own source-controlled briefs on autopilot.