We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track
Your intelligence agent for what matters
Tell ZeroNoise what you want to stay on top of. It finds the right sources, follows them continuously, and sends you a cited daily or weekly brief.
Your time, back
An AI curator that monitors the web nonstop, lets you control every source and setting, and delivers verified daily or weekly briefs.
Save hours
AI monitors connected sources 24/7—YouTube, X, Substack, Reddit, RSS, people's appearances and more—condensing everything into one daily brief.
Full control over the agent
Add/remove sources. Set your agent's focus and style. Auto-embed clips from full episodes and videos. Control exactly how briefs are built.
Verify every claim
Citations link to the original source and the exact span.
Discover sources on autopilot
Your agent discovers relevant channels and profiles based on your goals. You get to decide what to keep.
Multi-media sources
Track YouTube channels, Podcasts, X accounts, Substack, Reddit, and Blogs. Plus, follow people across platforms to catch their appearances.
Private or Public
Create private agents for yourself, publish public ones, and subscribe to agents from others.
3 steps to your first brief
Describe your goal
Tell your AI agent what you want to track using natural language. Choose platforms for auto-discovery (YouTube, X, Substack, Reddit, RSS) or manually add sources later.
Review and launch
Your agent finds relevant channels and profiles based on your instructions. Review suggestions, keep what fits, remove what doesn't, add your own. Launch when ready—you can always adjust sources anytime.
Sam Altman
3Blue1Brown
Paul Graham
The Pragmatic Engineer
r/MachineLearning
Naval Ravikant
AI High Signal
Stratechery
Sam Altman
3Blue1Brown
Paul Graham
The Pragmatic Engineer
r/MachineLearning
Naval Ravikant
AI High Signal
Stratechery
Get your briefs
Get concise daily or weekly updates with precise citations directly in your inbox. You control the focus, style, and length.
Lenny's Podcast
Tim Sweeney
Keith Rabois
What made the cut
After filtering out self-promotional material, today’s useful signal came in two clusters: Tim Sweeney resurfacing the primary sources behind Carmack’s Law, and Keith Rabois recommending books and videos on stress, elite standards, human behavior, history, and hiring .
Most compelling recommendation
Doom 3 interviews
- Content type: Webpage / interview archive
- Author/creator: Fabien Sanglard
- Who recommended it: Tim Sweeney
- Key takeaway: Sweeney points readers to this page as the original source for Carmack’s Law, then argues that the prediction proved prescient in computer graphics and is now clearly so in AI .
- Why it matters: This is the strongest pick because Sweeney is not sharing it as trivia; he explicitly places the idea alongside Moore’s Law in importance .
Companion read: Gordon Moore’s 1965 paper
- Content type: Paper
- Author/creator: Gordon Moore
- Who recommended it: Tim Sweeney
- Key takeaway: He shares Moore’s paper as the comparison point for the scale of Carmack’s Law .
- Why it matters: It gives readers the historical benchmark Sweeney invokes when making his case .
"Ultimately the prediction, which proved prescient in computer graphics and is now clearly so of AI, is as important as Moore’s..."
Keith Rabois’s operating recommendations
Rabois’s picks are less a conventional startup reading list than a compact set of resources on how ambitious people handle pressure, standards, judgment, and hiring .
The Upside of Stress
- Content type: Book
- Author/creator: Kelly McGonigal
- Link/URL: Not provided in source material
- Who recommended it: Keith Rabois
- Key takeaway: He says the book argues that if you want to be happy, healthy, or wealthy, you need more stress in your life, not less, and he calls the evidence compelling and the book transformative .
- Why it matters: It is his clearest recommendation for reinterpreting a condition many operators try to avoid rather than use .
The Jordan Rules
- Content type: Book
- Author/creator: Not specified in source material
- Link/URL: Not provided in source material
- Who recommended it: Keith Rabois
- Key takeaway: He recommends it as the better guide to elite performance, with the blunt lesson that if you want to be Michael Jordan, you have to act like Michael Jordan .
- Why it matters: It is the sharpest performance-culture recommendation in today’s batch .
Shakespeare
- Content type: Plays / collected works
- Author/creator: Shakespeare
- Link/URL: Not provided in source material
- Who recommended it: Keith Rabois
- Key takeaway: Rabois says that everything important to learn about humans was written by Shakespeare, and that reading Shakespeare is better than customer research .
- Why it matters: It shows he treats human observation as a practical advantage, not just a cultural interest .
Nuremberg Trial
- Content type: Film
- Author/creator: Not specified in source material
- Link/URL: Not provided in source material
- Who recommended it: Keith Rabois
- Key takeaway: He says it contains lessons applicable to the modern world, taught him several things he did not know, and is useful for understanding historical travesties and how to prevent them .
- Why it matters: This is the clearest recommendation today for sharpening present-day judgment through history .
Eric Glyman’s hiring speech
- Content type: Talk / online video
- Author/creator: Eric Glyman
- Link/URL: Not provided in source material
- Who recommended it: Keith Rabois
- Key takeaway: Rabois says the speech covers hiring in detail and is fairly similar to his own views on the topic .
- Why it matters: It is the only recommendation in today’s mix aimed directly at a core startup function: hiring .
Bottom line
The highest-signal item today is Sweeney’s Carmack’s Law source material because it comes with a precise claim about lasting importance for AI. The rest of the list is most useful as a pattern: Rabois’s recommendations repeatedly point readers toward better handling of stress, higher standards, deeper models of human behavior, and stronger historical judgment .
Perplexity
sarah guo
Together AI
Top Stories
Why it matters: This cycle’s biggest developments were about leverage and constraints: better ways to compress frontier models, open agents that iterate on themselves, a concrete medical-monitoring use case, and external limits on compute supply.
1) Frontier distillation got materially more practical
TRL’s rebuilt on-policy distillation trainer now supports teacher models above 100B parameters and is reported to run more than 40x faster than naive implementations through buffer and payload optimizations . In the cited example, the team distilled Qwen3-235B into a 4B student and gained 39+ points on AIME25 . The same setup is presented as usable across Llama, Qwen, and Gemma model families . Blog: TRL distillation trainer.
Impact: This is a concrete post-training efficiency story: very large teacher models can now be used to produce much smaller students with less engineering overhead than before .
2) The open-agent stack is moving toward self-improvement
NousResearch’s hermes-agent-self-evolution repo automates a loop where the system reviews past task histories, identifies errors and root causes, proposes prompt or code changes, tests variants, and keeps the best-performing version . Today it focuses on optimizing SKILL.md, with a roadmap to extend the same loop to tool descriptions, system prompts, and underlying code . The repo is API-only for mutation and evaluation, includes test gates, file-size and scope limits, and still requires human-reviewed PRs before merge . Repo: hermes-agent-self-evolution.
Related activity around Hermes points the same direction: one contributor said Hermes was evolved into an autonomous ML/RL research lab with 94 Orchestra primitives for zero-shot GRPO and distillation , while Teknium said more is coming . Separately, a quality-of-life update made Hermes about 20% more likely to load the right skill for a task, and Teknium said the agent performed the prompt improvement and benchmarking itself .
Impact: Open agents are being pushed beyond static tool wrappers toward systems that can revise parts of their own operating instructions under guardrails .
3) A Nature paper showed a strong real-world medical use case for LLMs
University of Pennsylvania researchers analyzed more than five years of Reddit discussion—about 400,000 posts from 70,000 users—about GLP-1 drugs including Ozempic and Mounjaro, using LLMs to map patient language to medical terminology . The study surfaced side effects not fully reflected in current labels, including menstrual irregularities, hot flashes, chills, and fatigue. The authors frame this as computational social listening and argue that mainstream drugs may need real-time social-media monitoring as an early warning system beyond voluntary FDA-style reporting . Paper: Nature.
Impact: This is one of the clearest examples in this cycle of LLMs being used to expand surveillance capacity in a high-stakes domain rather than just improve a chatbot .
4) AI infrastructure risk is becoming geopolitical, not just technical
One analysis tied Middle East tensions to four AI supply-chain bottlenecks: helium, energy, shipping, and advanced chip concentration . It said the strike on Qatar’s Ras Laffan facility removed about a third of global helium supply, that helium is essential for sub-3nm fabrication, and that South Korea and Taiwan are highly exposed through suppliers such as SK Hynix and TSMC . The same thread argued that Strait closures could disrupt 25% of world energy supply, threaten Taiwan’s power availability, and block about 30% of global container shipping for semiconductor equipment, chemicals, and finished chips . It tied those risks directly to the roughly $500B AI data-center buildout planned for 2026 .
Impact: This frames compute risk as a question of energy, logistics, and industrial gases—not only model quality and GPU counts .
Research & Innovation
Why it matters: The research pipeline this cycle focused on new runtimes, cheaper arithmetic, and better long-context mechanics—not just bigger models.
Neural Computers: Meta AI and KAUST proposed Neural Computers, a learned system where computation, memory, and I/O move into the model’s runtime state . Early prototypes use video-model-style prediction to simulate terminal and GUI behavior directly from instructions, pixels, and user actions rather than executing code in a conventional OS . The authors and commentators are explicit that long-horizon reasoning, stable symbolic computation, and durable reuse are not solved yet, framing this as a first step toward a Completely Neural Computer. Paper: arXiv 2604.06425.
Huawei HiFloat4 FP4: Huawei proposed HiFloat4 (HiF4) for Ascend NPUs, reporting BF16-comparable performance with large compute-efficiency gains . A follow-on note summarized the key claim more precisely: about 90% of training computation can run in FP4 while keeping the loss gap within roughly 1.5% of a full-precision baseline . The same note said chips designed around this are still more than 1.5 years away. Paper: arXiv 2604.08826.
cuLA for linear attention:cuLA packages handwritten CUDA kernels for linear attention variants including GLA, KDA, GDN, and Lightning, built with CuTe DSL and CUTLASS for Hopper and Blackwell GPUs . The pitch is straightforward: linear attention gives O(N) scaling instead of O(N²), enabling million-token settings, while cuLA reports speedups from 1.32x–1.45x for KDA on Blackwell to 2.19x for Lightning Attention against an FLA Triton baseline . The migration path is also small—a one-line import swap from
fla.ops.kdatocula.kda.Retrieval models are back under debate: A fresh thread argued that ColBERTv2—a 100M late-interaction retriever—still beats Qwen3-Embed-8B on BrowseComp+ and LIMIT, despite the dense retriever being about 80x larger. Supporters took that as evidence that single-vector dense models generalize poorly out of domain . Critics replied that LIMIT is artificial, that synthetic datasets can mislead, and that quantization often trades only 1–5% recall for roughly 100x memory reduction and major speed gains . A separate observation from the same discussion is that retrieval papers are increasingly built on decoder-only LLMs rather than encoder-only models like BERT .
Products & Launches
Why it matters: New releases were less about generic chat and more about workflow control, evaluation, search, and document manipulation.
OpenClaw v2026.4.11 added ChatGPT import, a new Memory Palace for exploring chats as structured memory, guided plugin setup, richer chat rendering, better video generation handling, and stronger integrations with Teams, Feishu, WhatsApp, and Telegram . Changelog: openclaw v2026.4.11.
Opik was highlighted as an open-source tool for debugging, evaluating, and monitoring LLM apps, RAG systems, and agent workflows with tracing, automated evals, and dashboards . Repo: comet-ml/opik.
Qdrant Query Language (QQL) introduced a SQL-like interface for vector search, combining hybrid retrieval, filtering, and semantics in a more structured query flow . Article: QQL overview.
Verso AI launched with an engine that converts PowerPoint’s OOXML format into a representation that is easier for LLMs to edit; the team said its system beats competing tools on its internal benchmark .
Hermes Telegram Dashboard was announced for Hermes users, bringing terminal access, system-resource monitoring, and cron management into Telegram .
Industry Moves
Why it matters: Distribution, platform reach, and infrastructure economics continue to matter as much as model releases.
MiniMax M2.7 widened its distribution footprint. Together AI said the model is live on its platform, with day-0 availability on both serverless and dedicated infrastructure . Fireworks separately announced day-0 hosted availability, describing the model as production-ready with 200K+ context, strong reasoning, and native multi-agent support . At the same time, separate discussion around the open-weight release focused on the non-commercial license, with one user saying the restriction prevents at-home use .
Perplexity launched a company-building competition around Perplexity Computer. The Billion Dollar Build is an eight-week program for teams building a company with a path to a $1B valuation, with finalists eligible for up to $1M from the Perplexity Fund and up to $1M in Computer credits .
Infrastructure remains a constrained market. Notes from a HumanX fireside chat said the AI infra market is still early, that AI-native firms are using post-training to reach frontier quality more cheaply and quickly, and that inference remains constrained by both supply and talent .
Claude’s consumer reach kept rising. Similarweb reported 341.26% YoY growth in Claude website traffic in Q1 . One commentator attributed some of the appeal to Claude’s less chatty, more concise responses .
Policy & Regulation
Why it matters: This cycle had fewer formal government actions, but access control, compliance, and labor-policy debates continued to shape how AI systems are deployed.
Capability control is tightening around agent ecosystems. A Zhihu Frontier roundup said Anthropic removed OpenClaw from the Claude whitelist, describing it as a sign of tighter control over agent ecosystems . The same roundup said Anthropic Mythos was released only to select partners, framing that as a capability-control issue rather than a broad public rollout .
Compliance remains a live cross-border constraint. The same roundup said Slack’s shutdown in Greater China was presented as a compliance matter rather than an abrupt exit .
AI-risk activism remained under scrutiny after the attack on Sam Altman’s home. PauseAI said it unequivocally condemned the attack, said the suspect had only minimal participation in its public Discord, and stated he had no role in the group’s campaigns or events . The organization also argued that concern about advanced AI risk is shared by major researchers, legislators, and institutions, and positioned its own work around protests, petitions, and policy advocacy as a peaceful outlet .
The labor-policy conversation is widening. Sam Altman was cited as proposing a 4-day, 32-hour workweek and a new social contract in response to AI and robotics . At the same time, Alex Karp argued AI will destroy humanities jobs and increase the importance of vocational training and hands-on skills , while Aaron Levie argued the opposite dynamic may hold in law, predicting more lawyers because AI will generate more legal questions and new areas of compliance work .
Quick Takes
Why it matters: Smaller items still help map where capability, demand, and uncertainty are moving next.*
Unitree R1 is now available for global pre-order on AliExpress starting at $6,806, with June 30 deliveries for the base R1 AIR trim; the EDU version adds custom software support, full SDK access, and optional Jetson Orin .
Muse Spark kept attracting favorable external reactions. Meta said it rebuilt its AI stack from scratch and that Muse Spark now powers Meta AI . François Fleuret said the model passed his tests, including image generation , and Alexandr Wang said it is particularly good at finding open-source data and analyzing it .
Claude Opus 4.6 sparked another benchmark dispute. BridgeBench claimed its hallucination accuracy fell from 83.3% to 68.3% and described the model as nerfed . Critics replied that the newer run used 30 tasks instead of 6, and said shared-task scores changed only from 87.6% to 85.4%, which they argued could be statistical noise . Separate commentary noted that inference at scale can legitimately create performance variability across services .
LangChain clarified its agent-SDK stack. Harrison Chase described
create-agentas minimal, deepagents as more batteries-included, and middleware as the advanced customization layer across both . In a related exchange, he pointed to deepagents when asked for a strong open-source alternative to the Claude Agent SDK .Open-source AI security tooling got a fresh checklist. A roundup highlighted NeMo Guardrails, Promptfoo, LLM Guard, garak, DeepTeam, Llama Prompt Guard 2-86M, ShieldGemma 2, OpenGuardrails, Cupcake, and CyberSecEval 3 – Visual Prompt Injection as lightweight alternatives while Mythos remains closed .
Perplexity
sarah guo
Keith Rabois
1) Funding & Deals
Perplexity's Billion Dollar Build is the clearest capital signal in this batch. The company launched an 8-week competition for teams using Perplexity Computer to build a company with a path to $1B; finalists can receive up to $1M in investment from the Perplexity Fund and up to $1M in Computer credits . Aravind Srinivas framed it as a bet that a new generation of builders can create very large businesses with these tools .
Keith Rabois restated a founder-first seed/Series A thesis. He says early investing starts with whether a founder has a non-zero chance of changing an industry or the world. In the interview he pointed to Fair—founded by former Square colleagues Max Rich and Jeff Collison—as a case where execution tempo stood out early , and said Ramp's progress toward shipping cards in 3 months instead of the usual 9–12 months was one reason he preempted the Series A .
2) Emerging Teams
GrayMatter is a lightweight persistent memory layer for agents that drops in with three lines of code, stores observations, checkpoints, and a knowledge graph in a local file, and pulls back only relevant context with hybrid retrieval . The solo builder reports up to 90% token savings after 100+ sessions, with native MCP support for Claude Code and Cursor, offline operation, and an open-source repo at github.com/angelnicolasc/graymatter.
slashlast30days v3 has notable open-source traction at 20,000+ GitHub stars and a differentiated search stack . The new version makes Reddit, X, and YouTube transcripts free without API keys, adds a Python pre-research engine that resolves the right handles, subreddits, hashtags, and channels before search, and ships cross-source cluster merging, comparisons, person-mode, and ELI5 mode . Garry Tan described the broader direction as "personal open source software" .
Hearica is an accessibility-driven Windows app that captions or translates all PC audio through a floating overlay and adds speaker separation, session saving, and custom context profiles . The founder says it was motivated by personal hearing loss, launched last month, already has paid subscribers, and builds on an earlier OSS system-captioning project using Whisper models .
AdmitOdds comes from a founder with direct user empathy: an 18-year-old high school senior built the product from his own college admissions pain point . Reported traction is 415 user accounts, 18 paying subscribers at $19.99/month, and $0 ad spend, with growth coming from Reddit, TikTok, and word of mouth . The stack is Next.js, Supabase, Stripe, and Claude/GPT .
3) AI & Tech Breakthroughs
Hermes shows a more explicit self-improvement loop for agents. In the cited architecture discussion, Hermes does not rely on offline trajectory mining or a separate skill-extraction model; the runtime notices reusable workflows, writes them into durable artifacts through the skill interface, and reuses them in the same act-notice-write-reuse loop . Garry Tan highlighted it as evidence that app-level applied research is increasingly happening in the open through open source .
BDH fast weights offers a concrete memory mechanism outside the normal context-window tradeoff. The setup keeps the transformer backbone frozen, updates only an isolated memory buffer for 300 gradient steps, and retained 20/20 unrelated facts after a cold reload with median probability 0.997 and cross-contamination below 0.03. The demo used a 15M-parameter model trained on 250M tokens on a single consumer GPU, and the code is open source .
Coding-agent differentiation may be thinner than branding suggests. A technical analysis of Cursor 3.0 claims that Cursor Agent is effectively Claude Code behind a local proxy with prompt find-and-replace, bundled Anthropic
claude-agent-sdkandclaude-codepackages, and a custom fine-tuned Claude 3.7 Sonnet model . Garry Tan cited the report as further confirmation of the "thin harness, fat skill" pattern .
4) Market Signals
Horizontal SaaS may be masking AI-driven churn. SaaStr argues that specialized AI tools are quietly replacing specific workflows for power users even while overall accounts still look healthy . In the example, Canva work had already been split across Reve for images, Opus Pro for clips, and Higgsfield for short-form video . The practical takeaway is to track power-user engagement separately and talk to users who are still paying but have stopped showing up .
Inference is where more investors now see the bottleneck—and the margin fight. One analysis projects inference spend approaching a 10:1 ratio to training by mid-2026, with competition moving to the lowest marginal cost and highest reliability . In parallel, HumanX takeaways summarized by Sarah Guo say the AI infra market is still early, AI-native companies are using post-training to reach frontier quality cheaper and faster, and inference remains supply- and talent-constrained . The moats highlighted in the essay are upstream: custom silicon, power and cooling access, proprietary data loops, and system-level integration .
The investor bar is moving toward practitioners, not observers. Harry Stebbings argues investors who are not building with AI will miss the real opportunities, bottlenecks, and failure modes . Keith Rabois's version of the same pressure is operational: in fast-moving AI companies, long roadmaps make less sense, business-minded engineers gain leverage, and accumulating advantages matter more because foundation-model progress is compressing time horizons .
"If you are not building with AI today you simply should not be investing."
- AI-generated code is creating repeatable diligence risks. One security scanner builder says a missing Stripe webhook signature check appears in almost every AI-built SaaS they review, enabling fake
payment succeededevents if an endpoint accepts unsigned payloads . The claimed fix is four lines, and the same builder now offers xploitscan.com to scan for this and roughly 150 other AI-generated code patterns .
5) Worth Your Time
- Lenny's Podcast with Keith Rabois — Best single watch here for founder underwriting and what unusual execution tempo looks like in practice .
Jason Kneen's Cursor 3.0 report — Useful teardown if you are tracking coding-agent defensibility and how much value sits in the harness versus the provider layer .
The 3 Year Inference Landscape — A compact framework for where margins may accrue across chips, models, hosting, power, and system integration as inference scales .
SaaStr on Canva and stealth AI churn — A practical retention lens for boards: watch the power user who quietly stopped showing up before renewal pressure hits the numbers .
slashlast30days v3 thread — Worth scanning if you care about agent-led search, especially the pre-research layer that decides where to search before the report is assembled .
Kent C. Dodds ⚡
Lenny Rachitsky
Aman Sanger
🔥 TOP SIGNAL
Cursor founder Aman Sanger shared rare production datapoints: async coding agents running in cloud VMs are already generating 30% of merged PRs internally, and by the end of 2025 agent requests exceeded tab accepts even though tab can fire on every keystroke . This is not toy output: he cites an eight-hour React-to-Rust video-rendering refactor that made the system 25x faster, plus a 10,000-line PR for network policy controls . The actionable part is the workflow: give long-running agents a full computer, let them test their own work, and review video/report artifacts before you spend time reading code .
🛠️ TOOLS & MODELS
- Cursor Cloud Agents: model routing is task routing. Cursor uses OpenAI models for higher-level planning/orchestration, Gemini and Anthropic for computer use and multimodal work, Anthropic for stronger UI generation, and faster models for simpler subagent tasks where they match performance at lower latency .
- Chrome DevTools MCP just got more useful for agent QA. Addy Osmani says it now includes Lighthouse performance checks, a memory leak detection skill, an accessibility debugging skill, an LCP optimization skill, and an experimental CLI .
- Codex UX signal from a power user. Riley Brown likes Codex because it stays bare-bones, keeps project work in left-sidebar folders with multiple threads, and should auto-spin new threads when a request splits into multiple tasks instead of adding extra interface layers like a Codex Cowork middle step .
- Local model reality check. Armin Ronacher says local models are not universally there yet, but Qwen3-Coder-Next on a beefy Mac is useful for its speed and decent tool calling; his broader point is not to dismiss local models outright, especially for tasks like STT .
💡 WORKFLOWS & TRICKS
- Artifact-first review loop for long-running agents.
- Run the agent in a cloud VM with the same kind of tools you'd use locally.
- Let it implement and test the change itself.
- Review its artifacts first, especially a demo video or research report.
- If something looks off, reprompt from the artifact instead of diving straight into the diff.
- Only review code once the behavior looks right .
- Break long jobs into a planner tree. Cursor's working pattern is planner -> subplanners -> workers. Keep each leaf task simple enough to stay within training distribution, cap the outer agent to hundreds of thousands of tokens rather than tens of millions, and use faster models for simpler subagent tasks when they achieve the same result .
- Use event-driven agents, not just ad hoc prompts. Cursor's automations fire on issues, pages, training runs, PRs, and security events. Their examples: an agent investigates a middle-of-the-night page and proposes a single-click fix, watches training metrics and logs to catch failures early, and finds PR vulnerabilities before shipping .
- Let one agent escalate to another agent. Kent C. Dodds describes a clean loop with Kody: when Kody hits a problem, he tells it to kick off a Cursor Cloud Agent, and Kody already knows the problem context well enough to compose the repair prompt itself .
- Parallel agents are not free throughput. Addy Osmani's advice: time-box long agentic sessions like deep work, tighten the scope for each agent, and treat finding your personal ceiling as a skill. Addy's longer note: Your parallel Agent limit.
I can fire up four agents in parallel and have them work on four different problems, and by 11am I am wiped out for the day.
- On unfamiliar platforms, shrink the blast radius. Trash Dev says he shipped his first iPhone app with 100% AI-generated code despite no prior Swift or iOS experience, but kept it local-only, manually checked local storage and external calls, and used a separate design tool when the UI was the bottleneck .
👤 PEOPLE TO WATCH
- Aman Sanger. Shared production metrics and concrete architecture details: PR share, artifact review, multi-agent decomposition, model routing, and event-driven automations .
- Addy Osmani. Posted on two useful fronts today: human-in-the-loop constraints for parallel agents and practical agent QA tooling via DevTools MCP .
- Kent C. Dodds. Showed an agent-to-agent repair loop where the agent using the software can hand off to a cloud repair agent without re-explaining the bug .
- Armin Ronacher. Shared a grounded local-model take: one concrete setup that works on strong Mac hardware today instead of blanket yes or no claims about local coding models .
🎬 WATCH & LISTEN
- 3:48-4:48 — Why async agents need their own computers. Sanger lays out why long-running agents need cloud VMs with full desktop access and testing, not just editor-side autocomplete .
- 6:40-8:14 — Review artifacts before code. Sanger explains why videos and research reports are faster review surfaces than raw diffs for early agent iterations .
- 8:18-9:05 — Zero-Swift iOS app, but still audit the risky path. Trash Dev explains the 100% AI-coded experiment, then grounds it by keeping the app local-only and manually checking storage and external calls before shipping .
📊 PROJECTS & REPOS
- Chrome DevTools MCP. New quality-check surface for agents: Lighthouse performance checks, memory leak detection, accessibility debugging, LCP optimization, plus an experimental CLI. Repo: chrome-devtools-mcp.
- Cursor's browser prototype. Sanger says Cursor ran a one-week multi-agent build that spent billions of tokens and tens of thousands of dollars of compute to produce a working, but still far-from-production, browser .
Editorial take: today's serious workflows looked less like autocomplete and more like ops — cloud VMs, event triggers, review artifacts, and explicit limits on human attention.
scott belsky
Eric Glyman
Keith Rabois
Big Ideas
1) AI is moving PM leverage from long roadmaps to business judgment
Keith Rabois argued that conventional PM work—collecting customer inputs and translating them into year-long sequential roadmaps—makes less sense when AI capabilities change week to week . In his framing, the enduring skill is business acumen: deciding what to build, why it matters, and how it moves the company’s equation, with communication shifting from slideware to working demos . Separately, an experienced PM on Reddit described feeling like a beginner again and struggling to find advanced material on model-first products, non-deterministic outcomes, and responsible scaling .
“Every presentation on product has to be a workable demo”
Why it matters: The role bar is moving away from artifact production and toward judgment, speed, and comfort with ambiguous AI behavior .
How to apply: Shorten planning cadence, ask teams to demo working product instead of presenting static decks, and spend learning time on model behavior, system constraints, and business impact—not just prompt tactics .
2) More headcount does not create more parallel progress
Rabois’s “barrels and ammunition” framework argues that initiative capacity is limited by the number of people who can independently take something from inception to success, not by total employee count . He describes ammunition as the supporting roles around them, and notes that post-funding teams often raise burn without increasing velocity because they add people without increasing barrel count . His examples: PayPal had roughly 12-17 barrels among 254 people, while a strong company might only have 2 .
Why it matters: If barrel count stays flat, adding people can increase coordination tax instead of throughput .
How to apply: Count true initiative owners first, cap the number of major parallel bets to that number, then add the right ammunition per initiative rather than assuming blanket hiring will fix execution .
3) Pricing and packaging are upstream product decisions—especially in AI
The Mind the Product discussion framed pricing as a cross-functional system spanning product, sales, marketing, finance, and rev ops, with product either leading or deeply involved because pricing connects how value is created and how it is captured . A recurring warning was to avoid building the product first and only then asking what to charge, because architecture, billing, GTM, and packaging may already be misaligned . For AI products, speakers argued the old fundamentals still apply—price to value—but hybrid models, clear plan tiers, and constrained usage patterns may be safer than open-ended chat experiences that create unpredictable token costs .
“Pricing’s never done. It’s not a one and done.”
Why it matters: Pricing errors can come from org design and product design, not just bad math . In AI, poor packaging can also weaken margin discipline .
How to apply: Move pricing decisions earlier, tie them to ICPs and value metrics, and test packaging before launch rather than after revenue stalls .
4) Internal AI enablement is becoming a company-level product advantage
Ramp’s Glass case study is a concrete example: 99% of employees were already using AI daily, but many were stuck because setup was painful and fragmented . Ramp responded by building a day-one AI workspace with SSO integrations, a marketplace of 350+ reusable skills built by colleagues, persistent memory, and scheduled automations so one person’s improved workflow could propagate across a team . Scott Belsky summarized the broader implication succinctly .
“don’t just build a differentiated product, build a differentiated company”
Why it matters: The advantage may come less from model access itself and more from how fast an organization turns AI usage into reusable workflows .
How to apply: Remove setup friction, standardize access, and create a way for employee-discovered workflows to become shared assets instead of personal hacks .
Tactical Playbook
1) Run a pricing and packaging sprint before you ship
- Align execs on the business objective. Start with strategy and commercial intent; if leaders are not aligned on why the product exists, the rest of the pricing work will drift .
- Segment the customer base. Break customers down by ARR, ACV, revenue, or deal size, then identify which ICPs actually get the most value .
- Write the value proposition by segment. Tie packaging and plan language to positioning, not just features .
- Quantify derived economic value. Ask whether you help customers make money, save money, or reduce risk, and use concrete proxies where needed; one example cited small business time at about $30 per hour .
- Test willingness to pay. Use methods like conjoint, Van Westendorp, and Gabor Granger; Van Westendorp specifically tests the “too cheap,” “bargain,” “expensive but acceptable,” and “too expensive” range .
- Package deliberately. Use plan tiers, a scaling value metric, add-ons, and a leader/filler/killer pass so core value is obvious .
- Revisit continuously. Pricing is iterative, not a one-time launch choice .
Why this matters: It prevents the common failure mode of shipping a product whose architecture, billing, and GTM do not support monetization .
2) Match discovery method to market shape
- Use direct customer conversations when there are identifiable decision-makers. Rabois explicitly said enterprise customer development is useful when you can talk to the actual buyer or a small set of must-win accounts .
- Ask value questions in those conversations. The pricing discussion recommends derived economic value questions as a standard part of customer interviews .
- Be more skeptical of stated preferences in consumer and SMB. Rabois argued that for consumer, SMB, and micro-merchant products, customer interviews can be directionally wrong because users struggle to explain subconscious purchase behavior .
- Lean more on observed behavior and economics in those markets. His suggested fallback was instincts plus outcomes like ticket sales, CAC, and LTV rather than over-weighting a handful of interviews .
- Do not universalize one method. Community responses on healthy PM orgs still emphasized consistent customer engagement overall, so the practical lesson is to adapt the discovery loop to the market rather than follow one doctrine everywhere .
Why this matters: Discovery quality drops when teams apply enterprise-style interview logic to markets where buyer motivation is hard to verbalize—or ignore customers entirely where decision-makers are knowable .
3) Audit initiative capacity before you add headcount
- Count barrels honestly. A barrel is someone who can take the company “over the hill” from idea to result, including motivating people, gathering resources, and measuring outcomes .
- Map each major initiative to a barrel. That defines how many important bets you can truly run in parallel .
- Pause blanket hiring if barrel count has not changed. Otherwise you risk more burn with the same or less output .
- Add ammunition per problem, not by org chart default. Some initiatives need design, engineering, PM, or data support; others need very little .
- Hire or promote barrels first when growth demands more throughput. That is the lever Rabois says increases parallel initiative capacity .
Why this matters: It reframes scaling from “more people” to “more independently completable bets” .
4) Design PM work so signal and autonomy can coexist
- Set vision and strategy at the top. One community description of a healthy org started with clear product vision and strategy from leadership .
- Give PMs room to steer within that frame. The same comment described full freedom to run a product line as long as it aligns with strategy, with accountability for wins and losses .
- Protect PM focus. Strong POs or similar roles can absorb dev communication, tickets, and documentation so PMs can stay focused on customers, markets, stakeholders, and roadmap quality .
- Keep feedback capture lightweight. A shared spreadsheet can be enough to record customer observations and requests .
- Treat manager quality as a force multiplier. Multiple community responses stressed that a good boss often determines whether the environment is workable day to day .
Why this matters: Healthy product teams are defined less by perfection than by clear direction, usable signal, and operating conditions that amplify PM strengths .
Case Studies & Lessons
1) File & Fight: packaging alone drove a 3x deal-size increase
One speaker said their packaging launch produced a 3x increase in price or deal size.
Why it matters: Monetization upside can come from packaging and plan structure even when the core product is unchanged .
How to apply: Before adding new features, ask whether clearer tiers, a better value metric, or better add-ons would unlock higher willingness to pay .
2) Aura: prove value first, then monetize with a stronger story
The Aura example emphasized getting the product in, proving value with NZ Police, and then using that proof as both monetization support and a marketing story for other police forces .
Why it matters: In some B2B settings, sequencing matters more than the first quoted price .
How to apply: If adoption friction is high, consider a GTM motion that lets customers experience value before you optimize pricing—and turn the best proof points into referenceable case studies .
3) Tracksuit: use price as a discovery input, not just an output
Tracksuit reportedly started with a blunt question: “What would you pay $10,000 for?” and used the answers to shape the initial feature set .
Why it matters: Pricing questions can reveal which problems are valuable enough to deserve a roadmap slot .
How to apply: In early discovery, ask customers what outcome or feature set would justify a meaningful budget line, then build around that signal instead of brainstorming features in isolation .
4) Lifetime pricing: great for validation, risky for scaling
A startup founder said lifetime access priced at about 3x the monthly plan helped generate first revenue and validate the product when there were no users yet . The downside came later: every new feature and cost increased customer value but not revenue from those lifetime users, pushing the founder toward recurring pricing instead . Community replies distilled the trade-off: lifetime reduces friction early, but the complexity reappears later; one suggested keeping core access while charging separately for future costly features such as AI .
Why it matters: Early validation pricing can quietly determine whether a SaaS business remains monetizable as costs rise .
How to apply: If you use lifetime offers for early traction, define clear boundaries up front around future features, usage, and premium add-ons .
5) Ramp Glass: internal AI adoption becomes reusable product infrastructure
Ramp saw heavy AI usage already in place, but setup friction blocked broader leverage . Glass addressed that with a standardized workspace and a marketplace of 350+ reusable skills, letting one person’s better workflow spread across the team .
Why it matters: Internal productivity gains compound when workflow knowledge becomes shareable infrastructure rather than isolated experimentation .
How to apply: Treat internal AI workflows like product surfaces: standardize onboarding and make the best patterns easy to copy .
Career Corner
1) Senior PMs are feeling the AI skill reset in real time
A PM with 12 years in industry and 9 years in product described feeling like “a beginner again” because the field is moving from traditional strategy and user empathy toward model-primary products, non-deterministic behavior, and responsible scaling . Another commenter replied simply: “I’m in the same boat” .
Why it matters: This is not just a junior-skills problem; experienced PMs are openly describing a missing bridge between classic PM strength and AI-native product depth .
How to apply: Move from passive reading to hands-on building with engineers and designers, even on small MVPs, and target specific gaps like model-centric architecture and managing probabilistic behavior .
2) A healthy PM org is usually easier to describe operationally than culturally
Community answers converged on a few concrete signs: high autonomy inside a clear top-level strategy, the ability to focus on core PM work, consistent customer engagement, a strong boss, and simple systems for capturing feedback .
Why it matters: These are interviewable signals. “Healthy” is less abstract when you can test for direction, decision rights, manager quality, and how customer signal actually flows .
How to apply: In interviews, ask who owns strategy, how PMs interact with customers, who handles dev-groundwork tasks, where feedback lives, and how success or failure is assigned .
3) The PM market is still tight, so adjacent roles matter
One experienced PM described months of unemployment and repeated late-stage rejections despite having design and engineering chops . Replies called the market “brutal” and pointed to adjacent roles such as solutions architecture, technical sales, sales engineering, forward deployment, Product Owner, Agile practitioner, or even coding as fallback paths . Another community note suggested that candidates without strong pedigree signals may need to target early-stage startups and build something with real metrics first .
Why it matters: Career resilience may depend on how well you can reframe PM experience into nearby commercial or technical roles .
How to apply: Tailor applications toward adjacent roles when needed and, if you are trying to break back into PM, keep a concrete project with proven metrics ready to show .
Tools & Resources
1) Simon Kutcher’s value-based packaging framework
Use it to structure tiers around good/better/best, a scaling value metric, and add-ons like implementation or support, then sort features into leaders, fillers, and killers .
Best use: When you need a shared language for pricing conversations across product, sales, CS, and marketing .
2) The pricing research toolkit: conjoint, Van Westendorp, Gabor Granger
These were presented as the pricing equivalent of user research techniques, with Van Westendorp offering a straightforward acceptable-price-range test .
Best use: When stakeholder debate is outrunning actual willingness-to-pay evidence .
3) The penetration × willingness-to-pay feature matrix
Map features by high or low penetration and high or low willingness to pay to separate core leaders from niche add-ons or low-value distractions .
Best use: When executives want to pack too much into the base plan or when roadmap debates need a monetization lens .
4) A lightweight customer feedback repository
One community answer reminded PMs that the system can be simple: a shared spreadsheet of customer observations and requests .
Best use: When the team needs a fast, low-friction place to accumulate signal before investing in heavier tooling .
5) Two sessions worth watching this week
- Mind the Product: What should we charge for this? for the pricing, packaging, and AI monetization frameworks summarized above .
- Hard truths about building in the AI era | Keith Rabois for the barrels framework and the PM-role-in-AI argument .
Demis Hassabis
Ben Thompson
Sebastian Raschka
Platform control is the clearest strategic theme
Today’s clearest thread is control of the interface: Apple is widening Siri’s model options while the hardware talent battle with OpenAI is already visible .
Apple is preparing to make Siri a gateway for outside AI
Apple is preparing an iOS 27 Siri overhaul that would let installed apps such as Google Gemini and Anthropic Claude handle queries inside Siri, extending the current ChatGPT integration . In Ben Thompson’s analysis, that would let Apple aggregate outside AI through the device and the App Store, rather than matching hyperscaler spending on model infrastructure .
Why it matters: If Apple keeps the user interface and subscription relationship while model providers compete underneath, it strengthens Apple’s position as the point of integration .
Apple is already responding to OpenAI’s hardware recruiting
Apple awarded rare out-of-cycle bonuses to iPhone hardware designers amid concern about departures to AI startups, with OpenAI described as a particular threat . OpenAI’s hardware effort is run in part by former Apple executive Tang Tan, advised by Jony Ive, and has hired several dozen Apple engineers across iPhone, iPad, Apple Watch, and Vision Pro teams .
Why it matters: The competition is no longer only about models or apps; it is also about who controls the next AI-centric device and the teams that can build it .
Hassabis pairs a near-term AGI estimate with safety coordination
Demis Hassabis says AGI may be ‘maybe five years away’
Google DeepMind CEO Demis Hassabis said AGI may be ‘maybe five years away’ and framed AI as a scientific tool for understanding the universe and tackling medicine, energy, and environmental challenges . He also said there is a non-zero chance things go badly if the technology is built the wrong way, arguing for cautious optimism, minimum standards among leading labs, and more international cooperation as companies and nations race toward the technology .
"There’s definitely ... a non zero chance that things could go quite badly wrong if the technology is not built in the right way."
Why it matters: A leading frontier-lab CEO is coupling a relatively near AGI timeline with explicit coordination asks, not just capability forecasts .
Open models are running into harder economics
MiniMax moves to a non-commercial license
MiniMax has shifted to a non-commercial license, with attribution requirements for users above $30M in revenue or 100M users, alongside an acceptable use policy . Nathan Lambert said the move looks like what happens when open-model companies ‘start to worry about money’ and reiterated his view that MiniMax, Moonshot AI, and Zhipu AI could face financial strain if their strategies hold . He separately said usage of 30-200B open models appears to be surging, though attribution is still hard to pin down .
Why it matters: Demand for open models may be rising, but this license change is a concrete sign that free-use frontier releases are getting harder to fund .
Research signals worth watching
‘Neural Computers’ push world models into the interface layer
A new ‘Neural Computers’ paper proposes learning a computer interface directly as a world model: the system takes keystrokes, mouse clicks, and previous screen pixels, then generates the next frames, readable text, and cursor movement without a traditional operating system . The authors describe it as a first step toward a ‘Completely Neural Computer’ where computation, memory, and I/O move into a learned runtime state; the paper is here.
Why it matters: Instead of putting an AI agent on top of software, this line of work asks whether part of the software runtime itself can move inside the model .
Raschka’s 2026 LLM readout centers on efficiency and agent use
Sebastian Raschka used recent releases including Nemotron 3 Super and Qwen 3.5 as anchor points for reading where LLM design is going . His shortlist of standout trends includes hybrid transformer/SSM designs such as Qwen 3.5’s gated delta-net layers and Nemotron 3’s hybrid attention, multi-head latent attention to compress KV cache, learned sparse attention for long contexts, and RLVR-driven reasoning that spreads through distillation . He also said Qwen 3.5 is being built with agentic use, tool calling, and more affordable long contexts in mind .
Why it matters: The emphasis is moving toward architectures that make long-context use, reasoning, and agent behavior cheaper and more practical .
Start with signal
Each agent already tracks a curated set of sources. Subscribe for free and start getting cited updates right away.
Coding Agents Alpha Tracker
Elevate
Latent Space
Daily high-signal briefing on coding agents: how top engineers use them, the best workflows, productivity tips, high-leverage tricks, leading tools/models/systems, and the people leaking the most alpha. Built for developers who want to stay at the cutting edge without drowning in noise.
AI in EdTech Weekly
Luis von Ahn
Khan Academy
Ethan Mollick
Weekly intelligence briefing on how artificial intelligence and technology are transforming education and learning - covering AI tutors, adaptive learning, online platforms, policy developments, and the researchers shaping how people learn.
VC Tech Radar
a16z
Stanford eCorner
Greylock
Daily AI news, startup funding, and emerging teams shaping the future
Bitcoin Payment Adoption Tracker
BTCPay Server
Nicolas Burtey
Roy Sheinbaum
Monitors Bitcoin adoption as a payment medium and currency worldwide, tracking merchant acceptance, payment infrastructure, regulatory developments, and transaction usage metrics
AI News Digest
Google DeepMind
OpenAI
Anthropic
Daily curated digest of significant AI developments including major announcements, research breakthroughs, policy changes, and industry moves
Recommended Reading from Tech Founders
Paul Graham
David Perell
Marc Andreessen 🇺🇸
Tracks and curates reading recommendations from prominent tech founders and investors across podcasts, interviews, and social media
PM Daily Digest
Shreyas Doshi
Gibson Biddle
Teresa Torres
Curates essential product management insights including frameworks, best practices, case studies, and career advice from leading PM voices and publications
AI High Signal Digest
AI High Signal
Comprehensive daily briefing on AI developments including research breakthroughs, product launches, industry news, and strategic moves across the artificial intelligence ecosystem
Frequently asked questions
Choose the setup that fits how you work
Free
Follow public agents at no cost.
No monthly fee