Your intelligence agent for what matters

Tell ZeroNoise what you want to stay on top of. It finds the right sources, follows them continuously, and sends you a cited daily or weekly brief.

Set up your agent
What should this agent keep you on top of?
Discovering sources...
Syncing sources 0/180...
Extracting information
Generating brief

Your time, back

An AI curator that monitors the web nonstop, lets you control every source and setting, and delivers verified daily or weekly briefs.

Save hours

AI monitors connected sources 24/7—YouTube, X, Substack, Reddit, RSS, people's appearances and more—condensing everything into one daily brief.

Full control over the agent

Add/remove sources. Set your agent's focus and style. Auto-embed clips from full episodes and videos. Control exactly how briefs are built.

Verify every claim

Citations link to the original source and the exact span.

Discover sources on autopilot

Your agent discovers relevant channels and profiles based on your goals. You get to decide what to keep.

Multi-media sources

Track YouTube channels, Podcasts, X accounts, Substack, Reddit, and Blogs. Plus, follow people across platforms to catch their appearances.

Private or Public

Create private agents for yourself, publish public ones, and subscribe to agents from others.

3 steps to your first brief

1

Describe your goal

Tell your AI agent what you want to track using natural language. Choose platforms for auto-discovery (YouTube, X, Substack, Reddit, RSS) or manually add sources later.

Weekly report on space exploration and electric vehicle innovations
Daily newsletter on AI news and research
Startup funding digest with key venture capital trends
Weekly digest on longevity, health optimization, and wellness breakthroughs
Auto-discover sources

2

Review and launch

Your agent finds relevant channels and profiles based on your instructions. Review suggestions, keep what fits, remove what doesn't, add your own. Launch when ready—you can always adjust sources anytime.

Discovering sources...
Sam Altman Profile

Sam Altman

Profile
3Blue1Brown Avatar

3Blue1Brown

Channel
Paul Graham Avatar

Paul Graham

Account
Example Substack Avatar

The Pragmatic Engineer

Newsletter
Reddit Machine Learning

r/MachineLearning

Community
Naval Ravikant Profile

Naval Ravikant

Profile
Example X List

AI High Signal

List
Example RSS Feed

Stratechery

RSS
Sam Altman Profile

Sam Altman

Profile
3Blue1Brown Avatar

3Blue1Brown

Channel
Paul Graham Avatar

Paul Graham

Account
Example Substack Avatar

The Pragmatic Engineer

Newsletter
Reddit Machine Learning

r/MachineLearning

Community
Naval Ravikant Profile

Naval Ravikant

Profile
Example X List

AI High Signal

List
Example RSS Feed

Stratechery

RSS

3

Get your briefs

Get concise daily or weekly updates with precise citations directly in your inbox. You control the focus, style, and length.

Seed Conviction Rises as Agent Infrastructure and Deep-Tech Teams Emerge
May 13
6 min read
773 docs
Aravind Srinivas
Perplexity
Andrew Ng
+15
A* III and a16z speedrun sharpen the early-stage funding picture while YC and indie founders surface in data-center cooling, agent software, biotech, and on-device AI. Enterprise buyers are moving multi-model, inference costs keep falling, and workflow connectivity is emerging as the next control layer.

1) Funding & Deals

  • A* III announced a $450M early-stage fund. Kevin Hartz’s thesis is straightforward: be a founder’s first believer, invest before consensus, traction, or even a product, then concentrate time and capital from day zero. A* said that approach has already produced early positions in Ramp, Decagon, Whop, Cape, Simile, Paraform, Watney Robotics, and Mercor, and that the firm now manages $1B+ AUM less than five years after launch.
  • a16z speedrun 007 is another strong signal that elite capital keeps moving earlier. The program says it will invest up to $1M in brand-new startups, including some that are pre-launch, pre-traction, or even pre-idea, and pair that with $5M in tool credits plus operator support across recruiting, GTM, marketing, HR, and more. Andrew Chen’s rationale is that AI has compressed team size, timelines, distribution, and iteration speed, making it unusually good timing for small teams to start sooner. Applications close May 17 for the July 27–October 11 San Francisco cohort.
  • Isomorphic Labs raised $2.1B to accelerate AI drug discovery. The company tied the raise to its AlphaFold lineage and its mission to reimagine drug discovery. For early-stage investors, this is less a seed datapoint than a category-level validation signal for AI-native therapeutics.

2) Emerging Teams

YC launches

  • InstaAgent has the clearest traction signal in the YC batch: YC says the company helps B2C brands scale social media marketing across hundreds of personas and has already reached $1M ARR in 10 months. Founders: @klwongkyle and @tseungcolin.
  • Madrone is a hard-tech team worth screening against the AI infrastructure buildout. YC says its dew-point cooling systems can cut power and water use by 30% at Texas data-center sites. Founders: @akshaytree and @ErikMeike.
  • Superlog is a sharp devtools wedge: a wizard configures logs, traces, alerts, and dashboards daily, while an agent investigates incidents and posts one mergeable PR per issue into Slack. Founders: @nicolomagnante and @arseniycodes.
  • FinalDose is one of the more ambitious biotech launches: YC describes it as a programmable drug platform built around a smart molecule that finds diseased cells by DNA and destroys them, starting with cancers. Founders: @Jeffliu6068Liu, @sklin_lite, and @liyaohuang2.

Indie watchlist

  • Alt pairs a clear privacy wedge with serious technical execution. The KAIST student team says the note-taking app runs fully on-device, offers unlimited free transcription, uses a quantized 1.6GB voice model for Apple Silicon, hits 12ms per audio chunk versus a 46ms benchmark, and runs local Pyannote diarization. It works offline on M-series Macs, iPhones, and iPads, with an optional $4/month tier for cloud summaries and translations.
  • firsteyes AI shows an early lean-distribution signal: six weeks after launch, the solo founder reports 900+ visitors, 100+ signups, 250+ audits, and small revenue with zero paid ads, though the landing page still creates confusion for some visitors about the core value proposition.

3) AI & Tech Breakthroughs

  • Perplexity’s GB200 work is the clearest infrastructure update in the batch. The team published how it serves post-trained Qwen3 235B MoE models on NVIDIA GB200 NVL72 Blackwell racks, arguing GB200 is a major step up over Hopper for high-throughput inference on large MoEs and that it changes how prefill/decode disaggregation should be done.
  • LlamaIndex’s liteparse-server turns document parsing into self-hostable infrastructure: an open-source HTTP server that parses PDFs, Office files, and images and generates screenshots while staying 100% self-hosted, private by default, and built for production. The underlying LiteParse parser is model-free, handles 50+ document types, parses complex layouts and tables in seconds, and includes lightweight OCR.
  • The deployment surface around liteparse-server is mature enough to matter. It ships as a Docker container or serverless Express API and integrates with Redis, OpenTelemetry, Jaeger, Prometheus, and Grafana, which makes it easier to drop into production document workflows.
  • Hugging Face crossing 1,000,000 public datasets is a quieter but important platform shift. The dataset base doubled in the last 8 months after taking 4 years to reach the first 500,000, and Clement Delangue says better data is becoming the next bottleneck for builders who want to train models themselves rather than just use APIs.
  • Open-source inference optimization is still moving fast under the surface. Bindu Reddy points to DeepSeek v4 using SSDs for KV cache plus TurboQuant and Kimi K2 compressing memory, arguing that constraint-driven open-source teams are attacking the KV-cache bottleneck directly and pushing intelligence costs down.

4) Market Signals

  • Enterprise AI is becoming a multi-model market. In SaaStr’s cited adoption data, OpenAI remains #1 at 56%, but Claude has climbed to 48% and Gemini to 40%. The article’s broader takeaway is that single-model architectures are becoming a procurement liability, coding assistants are driving the fastest revenue growth, and existing distribution contracts still matter enormously.
  • Cheaper inference is shifting value from model access to product design and post-training. Sarah Guo argues same-task inference costs should fall by at least an order of magnitude per year, which changes the kind of UX startups can build and makes product experimentation more important than training. Her 2026 prediction is that long-horizon agents will push many domain-focused AI companies into post-training, with coding as the first clear example; Clement Delangue agreed.
  • Agent infrastructure is converging on the workflow-connection layer. A recurring founder view is that the next breakout products will connect agents to the systems people already use—not more wrappers, but workflow access via MCP and similar plumbing. Shopify, Xero, QuickBooks, Figma, and Linear already have versions of this, while sectors such as healthcare, legal, real estate, field service, and education still lack it.
  • Production MCP is harder than the demos imply. Truto says static OpenAPI-to-MCP generation broke in production because enterprise customers needed custom fields, parameter formats, permissions, and LLM instructions, forcing environment-level overrides and dynamic tool generation. That is a useful diligence check for anyone evaluating agent-connectivity startups.
  • Monetization and distribution remain the main filter. Elizabeth Yin notes that many people have already built their own AI tools and that the market is currently in a subsidized phase where services can feel nearly free . Founder reports reinforce that warning: Bloort.ai shut down after 8k visitors, 200 signups, 10 installs, and $0 revenue despite heavy outreach , while builders in AI image generation describe the category as so saturated that distribution is effectively impossible for commodity products.

5) Worth Your Time

  • Andrew Ng on building with AI is worth watching for a grounded view of where value is moving: orchestration layers are making complex agent workflows easier to build, but evals, error analysis, and unstructured-data architecture remain hard; Ng also disclosed small personal investments in LangGraph and LlamaIndex.
  • A* III announcement is a compact statement of the current seed-conviction thesis: invest before consensus, traction, or even product, then go deep over time.
  • Andrew Chen on speedrun 007 is worth reading for concrete early-stage terms: up to $1M even for some pre-idea teams, plus a view that AI has materially compressed the cost and timing of company formation.
  • Truto’s MCP architecture guide is a practical read if you are diligencing agent infrastructure; the core lesson is that static MCP generation breaks quickly in real enterprise deployments.
  • SaaStr’s enterprise AI share essay is the fastest read on where enterprise model share, coding-assistant demand, and distribution advantages are actually moving.
Coding Agents Move Beyond Codegen Into Testing, Triage, and PR Ops
May 13
4 min read
101 docs
Cursor
Harrison Chase
Michael Truell
+11
The strongest signal today: single-agent codegen is hitting a ceiling, while teams wiring agents into testing, review, and triage are seeing bigger gains. This brief covers concrete workflows to copy now, plus the key releases and projects worth tracking across Codex, Claude Code, OpenClaw, Bun, Executor, and Simon Willison’s tooling.

🔥 TOP SIGNAL

  • If your agents only write code, you're optimizing the wrong bottleneck. Cursor says agent requests are up 15x YoY and now exceed tab accepts . Internally, even with 98% of merged commits written by AI, productivity seems to stall around 40% when agents stay in single-step coding-assistant mode; the bigger gains come from removing the testing, review, and handoff bottlenecks around the code . Intuit is seeing the org-level version of that shift with a 3x idea-to-release push from agent-driven development in the last 90-120 days .

⚡ TRY THIS

  • Review proof before diff. In Cursor's internal workflow, let a cloud agent write the code and run the tests, then review the returned video / screenshots / logs first. If the behavior looks right, take control of the remote dev environment for edge cases, then open the PR and bounce follow-up comments back to the agent .

  • Push PR babysitting to the cloud. Once a local agent gets you to PR created, click Babysit PR in cloud and let the cloud agent own the slow loop: CI failures, merge conflicts, and review comments. Only pull the human back in for ambiguous decisions .

  • Debug by asking for a repro artifact, not a fix. Simon Willison used Codex CLI with GPT-5.5 xhigh to generate a minimal Dockerfile that reproduced a Datasette segfault . Steal the pattern: first ask the agent for the smallest failing environment, then iterate on the bug from there. See the issue comment.

  • Turn your best prompt into a reusable skill. Cursor says a precise how skill for rigorous codebase inspection was the missing piece in their bug-triage agents . Harrison Chase says the packaging pattern is shifting from piles of sub-agents toward skills that wrap tools and data cleanly . Practical move: extract your best bug-intake, codebase-inspection, or security-review routine into a named skill and call it first instead of pasting a giant prompt every time.

📡 WHAT SHIPPED

  • Codex app — in-app browser upgrades. Codex can now test across viewport sizes from the in-app browser, capture key screenshots during long runs, hide the IAB to disable animations for 1-2x faster testing, and send annotations faster with lower token use .

  • Claude Code 2.1.139 — /goal. Set a completion condition and Claude keeps working across turns until it's met; works in interactive, -p, and Remote Control modes .

  • Cursor — Fast mode for Claude Opus 4.7. Faster by 2.5x at 6x the cost; Cursor recommends standard speed for most tasks .

  • OpenClaw beta — unified structured-file pathing.openclaw path read|write|append now works the same across md, jsonc, jsonl, and yaml, giving plugins and agents one addressing substrate for surgical edits. Study the PR: #78678. Peter Steinberger says Microsoft is helping get OpenClaw ready for enterprise use .

  • Bun Rust rewrite — emerging agent migration case study. Theo reports Jared's parallel-agent Zig→Rust experiment has already hit 99.8% of Bun's pre-existing Linux x64 glibc test suite, with a 960k-line rewrite in 6 days. The catch: the current port still carries 13,044unsafe calls, and Theo's warning is fair — this may trade known Zig issues for unknown Rust ones .

  • Executor desktop app. Add MCPs, OpenAPIs, and GraphQL servers once and make them available to every agent; Executor converts them into code mode under the hood, aims to support thousands of tools without context bloat, and keeps everything local .

  • llm 0.32a2. Simon Willison's CLI now supports OpenAI's /v1/responses endpoint, which matters because it enables interleaved reasoning across tool calls for GPT-5-class models. It also exposes summarized reasoning tokens, hideable with -R / --hide-reasoning. Release: 0.32a2.

🎬 GO DEEPER

  • Build → test → artifact review (0:51-3:11). Jonas shows the cleanest short demo in today's batch: one cloud agent writes the change, tests it, and returns a video proof so the human starts with behavior instead of a giant diff .
  • Slack bug report → triage pipeline (7:33-10:48). This is the best clip today on chaining a codebase-inspection skill with follow-up questions so vague bug reports become structured triage and assignment input .
  • From coder to agent manager (5:57-7:53). Michael Truell's ghost colleagues framing is worth watching because it explains the real workflow shift: less syntax, more delegation, review, testing, and parallelism — plus a clear warning about unsustainable AI-generated architecture if you skip review .
  • Artifact to study — Simon’s minimal Dockerfile repro. Small artifact, big lesson: make the failure portable first, then let the agent help you reason about it .

  • PR to study — OpenClaw #78678. A compact example of a useful agent-infra idea: give agents and plugins one shared way to address structured files so edits stop being brittle .

  • PR to study — llm #1435. If you build agent tooling, this is the plumbing change that matters: /v1/responses support and interleaved reasoning across tool calls .

Editorial take: the edge is moving from better code generation to better orchestration — proof artifacts, reusable skills, and explicit human checkpoints are where the real compounding gains are showing up.

Gemini Goes OS-Level, AI Cyberattacks Cross a Line, and Benchmarks Reset
May 13
3 min read
638 docs
Prime Intellect
Sakana AI
hardmaru
+22
Google moved Gemini deeper into Android while Google’s threat team disclosed the first known AI-developed zero-day in the wild. The brief also covers benchmark saturation, key research advances in math and RL infrastructure, and major product and funding moves across consumer AI, biotech, and enterprise software.

Top Stories

Why it matters: The biggest news today points to AI moving deeper into operating systems, crossing a new cyber risk threshold, and quickly outgrowing existing evals.

  • Google pushed Gemini deeper into Android. Gemini Intelligence adds multi-step task automation across apps, one-tap form fill, polished dictation, and custom widgets, starting on Galaxy and Pixel this summer and later expanding to watches, cars, glasses, and laptops. Google also previewed an AI-enabled pointer that understands what is under the cursor and combines pointing with speech, reinforcing its push to turn Android into an "intelligence system."
  • Google reported the first known AI-developed zero-day used in the wild. Its Threat Intelligence Group said the attackers planned a wide-scale strike, though proactive counter-discovery may have stopped it. Impact: AI cyber risk is no longer just hypothetical.
  • Frontier benchmarks are already being reset. GPT-5.5 high/xhigh solved the first ProgramBench task and xhigh outperformed Opus 4.7 xhigh across all metrics; separately, GPT-5.5 solved the last unsolved MathArena Apex problem and pushed Apex Shortlist above 90% accuracy. Benchmark builders are now creating newer tests and deprecating some final-answer competitions because models have become too strong for the old format.

Research & Innovation

Why it matters: The most useful technical progress focused on expert workflows, agent training efficiency, and lower-cost compute paths.

  • DeepMind’s AI Co-Mathematician reached 48% on FrontierMath Tier 4, a new high among evaluated AI systems. The system is an asynchronous, stateful workbench for ideation, literature discovery, computation, theorem verification, and knowledge development; early sessions reportedly solved open problems and surfaced overlooked citations.
  • PrimeIntellect’s Renderers fix the mismatch between token-based RL trainers and message-based environments, which had been corrupting sampled tokens and wasting compute on agentic turns. PrimeIntellect says the change unlocks more than 3x throughput on popular open models.
  • Sakana AI and NVIDIA’s TwELL use a new sparse format plus custom CUDA kernels to exploit >95% sparsity in transformer feedforward layers, translating that into >20% faster training and inference on H100s alongside memory and energy savings.

Products & Launches

Why it matters: New releases kept pushing down cost, widening access, and making multimodal AI more usable.

  • DeepSeek-v4-Flash was described in leaderboard comparisons as essentially equal to, and sometimes stronger than, v4-Pro while being faster and about 10x cheaper.
  • Microsoft rolled out MAI-Image-2-Efficient globally in Bing Image Creator, now free for everyone, with sharper detail, richer color, better text rendering, and more accurate prompt following than v1.
  • Meta launched Voice Conversations in Meta AI powered by Muse Spark, with interruptions, topic switches, multilingual speech, real-time image generation, recommendations, and a live AI camera mode; Meta says it is available today.

Industry Moves

Why it matters: Capital and M&A continue to cluster around domain-specific deployment and AI-enabled software businesses.

  • Isomorphic Labs raised $2.1B to accelerate AI drug discovery, building on AlphaFold; Demis Hassabis called improving human health AI’s top application.
  • Anthropic is reportedly in talks to acquire Stainless for $300M+, a move that would remove a developer-tools supplier used by OpenAI and Google.
  • Notion said AI now accounts for 60% of its business. The company closed Q1 with revenue accelerating for the seventh straight quarter and said it is cash-flow positive.

Quick Takes

Why it matters: These smaller updates sharpen the picture on data, voice agents, medical evals, and robotics.

  • Hugging Face crossed 1,000,000 public datasets, with the total doubling in the last eight months.
  • Artificial Analysis launched τ-Voice; Grok Voice Think Fast 1.0 led at 52.1%, and even the strongest speech-to-speech models resolved only about half of realistic customer-service scenarios.
  • Medmarks v1.0 expanded the largest open-source automated medical LLM benchmark suite to 30 benchmarks and 61 models.
  • Figure said its F.04 humanoid reached design lock and has started shipping parts.
A Product Design Rule, an AI IQ Framework, and a Prohibition History Pick
May 13
2 min read
152 docs
Ryan Shea
scott belsky
Tony Fadell
+1
Three organic recommendations stood out today: Tony Fadell highlighted David Epstein’s *Inside the Box* for its constraint-driven product lesson, Scott Belsky surfaced Ryan Eshea’s *AI IQ* framework, and Marc Andreessen pointed readers to *Last Call* for a more nuanced view of how Prohibition happened.

Most compelling recommendation

Inside the Box

  • Content type: Book
  • Author/creator: David Epstein
  • Link/URL: Not provided in the source post
  • Who recommended it: Tony Fadell
  • Key takeaway: Fadell pointed to the book for the idea that constraints create freedom, arguing that better products often come from removing features instead of adding more screens, menus, or complexity .
  • Why it matters: This was the clearest operating lesson in today’s set: feature subtraction can be a product advantage, not just a simplification exercise .

"Not every problem needs another screen, another menu, or another layer of complexity. Constraints create freedom"

Also worth saving

AI IQ

  • Content type: X thread
  • Author/creator: Ryan Eshea
  • Link/URL:https://x.com/ryaneshea/status/2054209480917754033
  • Who recommended it: Scott Belsky
  • Key takeaway: Belsky called it a "fascinating exploration," especially the EQ vs IQ comparison. The thread scores frontier AI models on the human IQ scale and shows where models land on the bell curve, how frontier IQ changes over time, how models compare on IQ and EQ, and what intelligence costs in practice .
  • Why it matters: It reframes model comparison beyond leaderboard tables by combining capability, trend, and cost context in one resource .

Last Call: The Rise and Fall of Prohibition

  • Content type: Book
  • Author/creator: Not specified in the provided notes
  • Link/URL:Amazon listing
  • Who recommended it: Marc Andreessen
  • Key takeaway: Andreessen said the book makes it easier to understand how Prohibition happened while also showing that the Prohibitionists were not wrong about everything and that Carry Nation "had a real point" .
  • Why it matters: The value here is the nuance: it is framed as a resource for understanding how a movement can contain real arguments and still end badly .

"It is now easy to understand how Prohibition happened. It is important to realize that the Prohibitionists were not incorrect in many of their arguments. Carry Nation had a real point. And yet."

AI Moves Into Phones, Desktops, and Secure Workflows
May 13
5 min read
224 docs
clem 🤗
Sakana AI
hardmaru
+8
Google and OpenAI pushed AI closer to the interface, while Microsoft, SAP, and NVIDIA focused on security and governance for enterprise agents. The digest also covers a new speech-agent benchmark, Isomorphic Labs’ $2.1B raise, sparse-LLM efficiency gains, and Hugging Face’s 1 million dataset milestone.

The clearest pattern

Today’s strongest signal is that major AI companies are pushing models out of the chat box and into the places where people already work: phones, desktops, voice channels, and enterprise systems. In parallel, the infrastructure underneath those systems is getting more specialized around security, efficiency, funding, and data .

AI moves closer to the interface

Google pushes Gemini deeper into Android—and experiments with an AI pointer

Google said Gemini Intelligence will automate multi-step tasks across apps and Chrome, fill forms in one tap, turn spoken thoughts into polished text with Rambler, and build custom widgets . The rollout starts this summer on the latest Samsung Galaxy and Google Pixel phones, then expands later this year to Android watches, cars, glasses, and laptops .

Separately, Google DeepMind showed experimental demos of an AI-enabled pointer that can understand what is under the cursor and respond to shorthand like “fix this” or “move that,” with examples spanning PDFs, tables, recipes, handwritten notes, and paused video frames .

Why it matters: Google is positioning AI less as a separate assistant window and more as an interaction layer across devices and apps .

OpenAI moves Codex from code generation into computer use

OpenAI showed Codex using local GUI apps via mouse movement, clicks, and typing, extending it from commands and files into everyday desktop software . The demo emphasized parallel work across multiple apps with a separate cursor that does not interrupt the user, plus permissioning that limits Codex to only the apps a user explicitly allows .

OpenAI also said computer use can leverage accessibility-framework data to understand interfaces more accurately—including off-screen elements—and work with fast non-multimodal models like Spark; the feature is available on Mac now, with Windows coming soon .

Why it matters: OpenAI is folding computer use into its main model stack rather than treating it as a separate experimental agent .

Voice agents are improving, but the ceiling is still low

Artificial Analysis launched τ-Voice to measure speech-to-speech models on realistic customer service tasks across airline, retail, and telecom scenarios, including tool use and noisy audio conditions . It said even the strongest models today resolve only about half of these scenarios end-to-end .

In the first leaderboard, xAI’s Grok Voice Think Fast 1.0 led at 52.1% success, ahead of GPT-Realtime-2 (High) at 39.8% and Gemini 3.1 Flash Live Preview - High at 37.7% . xAI also said Grok is already handling live Starlink phone operations autonomously at scale .

Why it matters: The new benchmark shows real progress in speech agents, but it also quantifies how much reliability work remains before voice systems can consistently close complex service loops .

Enterprise agents are getting a security and governance layer

Microsoft says 100+ specialized agents helped find exploitable bugs

Microsoft announced a multi-model agentic security system that combines more than 100 specialized agents across frontier and custom models to find exploitable bugs, and said it delivered top performance on the CyberGym benchmark . The company added that the system helped find and fix 16 vulnerabilities ahead of Patch Tuesday and is now opening a private preview for customers .

Why it matters: Microsoft is framing agentic security as coordinated specialist systems tied to measurable outcomes, not just general-purpose assistants .

SAP and NVIDIA focus on runtime controls for specialized agents

SAP and NVIDIA said SAP is embedding NVIDIA OpenShell into SAP Business AI Platform as an open-source runtime for securely developing and deploying autonomous agents . OpenShell provides isolated execution environments, policy enforcement at the filesystem and network layers, and infrastructure-level containment; it will act as the runtime security layer for SAP AI agents, including custom agents built in Joule Studio .

NVIDIA also said its NemoClaw blueprint will be available directly in Joule Studio for custom agents in areas like finance, procurement, supply chain, and manufacturing .

Why it matters: The emphasis here is on control layers for production deployment—runtime isolation, policy enforcement, and governance—rather than raw autonomy alone .

Capital, efficiency, and data are scaling with the applications

Isomorphic Labs raises $2.1B for AI drug discovery

Isomorphic Labs said it raised $2.1B in new funding to accelerate AI-driven drug discovery, building on work that began with AlphaFold and extending it into a mission to reimagine drug discovery .

“I’ve always believed the No.1 application of AI should be to improve human health.”

Why it matters: This is a significant funding signal for AI in biology and drug discovery, not just for horizontal model development .

Sakana AI and NVIDIA report sparse LLM speedups on H100s

Sakana AI and NVIDIA introduced TwELL (Tile-wise ELLPACK) and custom CUDA kernels designed to make sparse transformer language models fit GPU execution better . The team says feedforward layers can exceed 95% sparsity with little to no downstream performance loss, translating into more than 20% faster training and inference on H100 GPUs, along with lower peak memory use and energy consumption .

Why it matters: This turns a familiar efficiency idea—LLM sparsity—into reported wall-clock gains on current production hardware .

Hugging Face crosses 1 million public datasets

Hugging Face said it has now passed 1,000,000 public datasets, with petabytes of data being downloaded, analyzed, and used for training by millions of builders every day . It also said the dataset count doubled in the past eight months, versus four years to reach the first 500,000, and linked that acceleration to better AI agents that make it easier to build, share, and use custom datasets .

Clément Delangue argued that better data is becoming the next bottleneck for people who want to build AI themselves instead of relying on APIs .

Why it matters: The open ecosystem’s next constraint may be shifting from access to models toward access to usable data .

Context Layers and Expert Workflows Become Core PM Work
May 13
8 min read
102 docs
Aakash Gupta
Tony Fadell
Sachin Rekhi
+8
This issue focuses on a shift from generic AI usage to explicit system design: PMs are deciding memory, workflow context, and skill routing while translating expert judgment into reusable playbooks. It also covers partial-data MVPs, operational triage, AI hiring signals, and concrete resources to try.

Big Ideas

1) Context is becoming a product surface

Two separate threads point to the same shift: AI systems break when they lack the right context. In agent products, continuity does not appear automatically; by default, agents forget prior turns unless teams intentionally design memory using the context window, a task-level scratchpad, and a cross-session vector store . In enterprise software, Scribe argues the equivalent problem is missing workflow context: AI cannot see thousands of workflows or trace the decisions behind them, so it produces generic or wrong output . Scribe says its answer is a new context layer that maps how work gets done, and it reports crossing $100M ARR with 90,000 enterprise customers, including nearly half the Fortune 500 .

  • Why it matters: Without continuity or workflow context, AI cannot reliably act on behalf of users or inside an organization .
  • How to apply: Separate session context, task memory, and cross-session memory up front, then decide what parts of the business workflow need to be legible to both humans and agents .

2) Encoded judgment beats generic prompting

Julie Zhuo describes AI product building as turning analysts’ expert art into explicit playbooks or skills for an LLM: how to inspect a metric move, separate signal from noise, and decide whether a product change actually moved the needle . Her lesson is that the jump from acceptable to excellent output is rarely a handful of big rules.

“The gap between 70% quality and 95% quality is not 3 or 4 big things. It’s more like 100s of small things.”

Aakash Gupta makes the same point from a tooling angle: Claude chooses skills from the name and description alone, so overlapping skills fail unless the routing logic is explicit . Sachin Rekhi’s guide extends this into PM practice: his 15+ AI workflows are the ones that survived a year of daily use, span both macro work like strategy and customer discovery and micro work like status updates and slides, and include failures as well as wins .

  • Why it matters: Better AI output depends on externalizing expert taste, boundaries, and reusable workflows, not just writing longer prompts .
  • How to apply: Document what great practitioners actually notice, define where one skill should stop and another should start, and keep refining the workflows that hold up in daily use .

3) Simplicity is still a strategic advantage

Tony Fadell’s reflection on the iPod is a reminder that product strategy is not getting easier just because technology is more capable. He says the goal was for the technology to disappear so people could have their music anywhere: “1,000 songs in your pocket” . His broader lesson is that constraints create freedom, removing features often creates a better product than adding them, and the best technology understands when to step back .

  • Why it matters: New capability can easily turn into more screens, menus, and notifications unless teams keep subtracting .
  • How to apply: When evaluating an AI feature, ask whether it reduces complexity for the user or simply adds another layer to manage .

Tactical Playbook

1) Build agent memory in three layers

  1. Use the context window for what the model needs right now, knowing it has a token limit and older turns disappear when the session gets long or ends .
  2. Add a scratchpad for task-level working memory so the agent can track what it has already done across tool calls or multiple steps .
  3. Add a vector store for cross-session continuity by summarizing important user facts, storing them with a user ID, and retrieving them next session .
  4. Treat the summary rules as a PM decision: what gets captured, what gets ignored, and when it should be pulled back into context .

Why this matters: The default behavior is forgetting. Continuity only exists if the product defines it .

2) Rewrite skill descriptions for routing, not documentation

  1. Start by listing where skills overlap, such as /weekly-review versus /stakeholder-update or /activation-analysis versus /retention-analysis.
  2. Assume Claude will inspect only the skill name and description before deciding relevance .
  3. Put 3 trigger phrases, a clear boundary, and an explicit alternative such as use /Y instead in the description .
  4. Spend effort on the first two sentences of routing logic before polishing long instruction bodies .

Why this matters: A perfect skill body is effectively invisible if the description does not make the routing decision obvious .

3) Launch an MVP with partial data without hiding the trade-off

  1. Clarify the launch scope first: is this for friendlies or the full customer base, and what are the consequences of missing or bad data ?
  2. If the goal is learning, use a high-confidence, representative subset rather than waiting for full completeness .
  3. Design the system for future scale and let workflows degrade gracefully when edge cases are missing .
  4. Add post-launch feedback loops so coverage gaps surface quickly through customer feedback, support signals, and usage patterns .
  5. Keep the roles of discovery and MVP clear: discovery helps scope the opportunity, while the MVP reveals actual demand and behavior .

Why this matters: The trade-off is not simply 70% versus 100%; it is speed of learning versus the risk of designing too tightly around incomplete data .

4) Make unplanned work visible on Kanban boards

  1. Separate outcome work from unplanned work instead of mixing both into one queue .
  2. Use Epic containers or an Expedite bucket so the team can see what is interrupt-driven versus planned .
  3. Set up the board to support refinement, triage, and clear priority signals .

Why this matters: On teams like DevSecOps, operational security work can derail outcome work unless the board makes the trade-off visible .

5) Find early adopters where they already gather

  1. Go beyond friends and coworkers when you need signal from a niche audience .
  2. Start with digital communities that already organize around the problem space, such as relevant Facebook groups or subreddits .
  3. Add physical environments, like dog parks in the example discussed, when in-person interviews are feasible .

Why this matters: Early validation gets stronger when the audience self-selects around the problem rather than your personal network .

Case Studies & Lessons

1) Scribe: a workflow-context thesis at $100M ARR

Scribe says it has grown from a $7K first deal to $100M ARR, with 90,000 enterprise customers and adoption in nearly half the Fortune 500 . Alongside that milestone, the company argues that enterprise AI needs a context layer that makes workflows legible; otherwise models cannot see the recipes of work and become generic or confidently wrong .

Key takeaway: If AI is being added to an enterprise product, workflow context may matter as much as model quality .

2) TeamSundial: distilling analysts’ judgment into LLM skills

Julie Zhuo says a large share of the team’s time now goes into converting analysts’ tacit judgment into LLM playbooks and skills . She identifies two bottlenecks: seeing what better performance looks like, and then systematically articulating it for a model . That is why the last mile from decent to great quality is made up of hundreds of small judgments rather than a few obvious prompt tweaks .

Key takeaway: If you want AI products to outperform generic tools, invest in capturing expert standards explicitly, not just adding another model call .

3) iPod: make the technology disappear

Tony Fadell frames the iPod’s success around a simple goal: technology should disappear into the experience, not compete with it . His follow-on point is especially relevant to AI products: not every problem needs another screen or menu, and removing features can create a better product than adding them .

Key takeaway: Product progress can come from subtraction when new capability threatens to make the experience heavier .

Career Corner

1) AI fluency is showing up in PM hiring, but depth is still a choice

One PM publication argues that, in 2026, top PMs at Meta, Google, and AI-native companies treat Claude Code mastery as table stakes, and says hiring loops now ask how candidates use AI day to day, include prototype rounds from PRD to evals, and value portfolios with AI-built agents or prototypes . Teresa Torres offers a useful counterbalance: the product-builder trend is a tool, not a mandate, and enjoyment and skill should guide who on a team leans into it .

“It’s a tool in our toolbox. We can decide who on our team has fun with it, wants to do it, wants to contribute.”

How to use this: Be ready to show at least one concrete AI artifact or workflow, and be equally clear about where your contribution is strongest .

2) High agency has upside and cost

Shreyas Doshi says he is a huge fan of high agency while also emphasizing two realities: not everyone can develop it, and it brings major professional upside alongside chronic anxiety in situations you cannot control .

How to use this: Recognize the upside-cost trade-off when evaluating your own working style .

Tools & Resources

  • Sachin Rekhi’s “How I use AI as a product manager”: useful when you want PM-native AI workflows rather than generic examples; it covers 15+ daily workflows, spans macro and micro tasks, includes failures, and argues the practice can sharpen product thinking .
  • Aakash Gupta’s note on skill descriptions: useful when overlapping Claude skills keep producing the wrong artifact; it gives a simple routing pattern built around trigger phrases, boundaries, and explicit alternatives .
  • Claude Code for PMs: The Beginner’s Guide: useful as a fast start if you want a working setup; it begins with Visual Studio Code, the Claude Code extension, and /login, then offers a downloadable template with preconfigured plugins, skills, and a minimal CLAUDE.md. The author says the setup can cut costs by at least 50% and work with free frontier models .
  • The same Product Compass post also lists upcoming session topics that may be useful for structured PM upskilling, including AI skills every professional needs in 2026, How to use Claude to get your dream job, Building your AI Chief of Staff from scratch, How to Build Agentic Products, and The context engineering and agentic memory.

Start with signal

Each agent already tracks a curated set of sources. Subscribe for free and start getting cited updates right away.

Coding Agents Alpha Tracker avatar

Coding Agents Alpha Tracker

Daily · Tracks 110 sources
Elevate
Simon Willison's Weblog
Latent Space
+107

Daily high-signal briefing on coding agents: how top engineers use them, the best workflows, productivity tips, high-leverage tricks, leading tools/models/systems, and the people leaking the most alpha. Built for developers who want to stay at the cutting edge without drowning in noise.

AI in EdTech Weekly avatar

AI in EdTech Weekly

Weekly · Tracks 92 sources
Luis von Ahn
Khan Academy
Ethan Mollick
+89

Weekly intelligence briefing on how artificial intelligence and technology are transforming education and learning - covering AI tutors, adaptive learning, online platforms, policy developments, and the researchers shaping how people learn.

VC Tech Radar avatar

VC Tech Radar

Daily · Tracks 120 sources
a16z
Stanford eCorner
Greylock
+117

Daily AI news, startup funding, and emerging teams shaping the future

Bitcoin Payment Adoption Tracker avatar

Bitcoin Payment Adoption Tracker

Daily · Tracks 108 sources
BTCPay Server
Nicolas Burtey
Roy Sheinbaum
+105

Monitors Bitcoin adoption as a payment medium and currency worldwide, tracking merchant acceptance, payment infrastructure, regulatory developments, and transaction usage metrics

AI News Digest avatar

AI News Digest

Daily · Tracks 114 sources
Google DeepMind
OpenAI
Anthropic
+111

Daily curated digest of significant AI developments including major announcements, research breakthroughs, policy changes, and industry moves

Global Agricultural Developments avatar

Global Agricultural Developments

Daily · Tracks 86 sources
RDO Equipment Co.
Ag PhD
Precision Farming Dealer
+83

Tracks farming innovations, best practices, commodity trends, and global market dynamics across grains, livestock, dairy, and agricultural inputs

Recommended Reading from Tech Founders avatar

Recommended Reading from Tech Founders

Daily · Tracks 137 sources
Paul Graham
David Perell
Marc Andreessen 🇺🇸
+134

Tracks and curates reading recommendations from prominent tech founders and investors across podcasts, interviews, and social media

PM Daily Digest avatar

PM Daily Digest

Daily · Tracks 100 sources
Shreyas Doshi
Gibson Biddle
Teresa Torres
+97

Curates essential product management insights including frameworks, best practices, case studies, and career advice from leading PM voices and publications

AI High Signal Digest avatar

AI High Signal Digest

Daily · Tracks 1 source
AI High Signal

Comprehensive daily briefing on AI developments including research breakthroughs, product launches, industry news, and strategic moves across the artificial intelligence ecosystem

Frequently asked questions

Choose the setup that fits how you work

Free

Follow public agents at no cost.

$0

No monthly fee

Unlimited subscriptions to public agents
No billing setup

Plus

14-day free trial

Get personalized briefs with your own agents.

$20

per month

$20 of usage each month

Private by default
Any topic you follow
Daily or weekly delivery

$20 of usage during trial

Supercharge your knowledge discovery

Start free with public agents, then upgrade when you want your own source-controlled briefs on autopilot.