We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track
Your intelligence agent for what matters
Tell ZeroNoise what you want to stay on top of. It finds the right sources, follows them continuously, and sends you a cited daily or weekly brief.
Your time, back
An AI curator that monitors the web nonstop, lets you control every source and setting, and delivers verified daily or weekly briefs.
Save hours
AI monitors connected sources 24/7—YouTube, X, Substack, Reddit, RSS, people's appearances and more—condensing everything into one daily brief.
Full control over the agent
Add/remove sources. Set your agent's focus and style. Auto-embed clips from full episodes and videos. Control exactly how briefs are built.
Verify every claim
Citations link to the original source and the exact span.
Discover sources on autopilot
Your agent discovers relevant channels and profiles based on your goals. You get to decide what to keep.
Multi-media sources
Track YouTube channels, Podcasts, X accounts, Substack, Reddit, and Blogs. Plus, follow people across platforms to catch their appearances.
Private or Public
Create private agents for yourself, publish public ones, and subscribe to agents from others.
3 steps to your first brief
Describe your goal
Tell your AI agent what you want to track using natural language. Choose platforms for auto-discovery (YouTube, X, Substack, Reddit, RSS) or manually add sources later.
Review and launch
Your agent finds relevant channels and profiles based on your instructions. Review suggestions, keep what fits, remove what doesn't, add your own. Launch when ready—you can always adjust sources anytime.
Sam Altman
3Blue1Brown
Paul Graham
The Pragmatic Engineer
r/MachineLearning
Naval Ravikant
AI High Signal
Stratechery
Sam Altman
3Blue1Brown
Paul Graham
The Pragmatic Engineer
r/MachineLearning
Naval Ravikant
AI High Signal
Stratechery
Get your briefs
Get concise daily or weekly updates with precise citations directly in your inbox. You control the focus, style, and length.
Jerry Liu
sarah guo
Vinod Khosla
1) Funding & Deals
- Ethos — $22.75M Series A led by a16z. Ethos raised a $22.75M Series A with participation from General Catalyst, XTX Markets, Matt Evantic, and Common Magic . The product uses AI voice agents to capture expertise traditional profiles miss, then matches people into expert calls, research, AI training, fractional work, and full-time roles . The company says 35,000 people are joining weekly and users are making $10,000 per month on the platform . Ben Horowitz said he is “very excited to be working with James and the Ethos team” .
“AI shouldn’t replace you. It should make you irreplaceable.”
- VoriHQ — $22M Series B in grocery automation. Vori says it raised a $22M Series B to make every supermarket in America autonomous, targeting a $1.5T domestic grocery market it describes as still running on Reagan-era technology . Michael Seibel publicly backed the company, saying the team is “doing great work” .
2) Emerging Teams
Recursive Intelligence. Anna and Azalia launched Recursive Intelligence around a “recursive self improving loop” between AI and chip design . Their background includes AlphaChip, the deep-RL system used in multiple generations of Google TPUs, Axion CPUs, Pixel and AV chips, with external adoption by MediaTek . Phase one is faster physical design and verification, including a static timing analysis engine that correlates with commercial tools while running 1,000x faster; later phases move toward a workload-to-GDSII platform and, eventually, vertically integrated chips and models . The team combines ex-Claude, Gemini, and Groq LLM talent with chip-design specialists .
Unconventional AI. Naveen Rao—whose background includes a neuroscience PhD, an early AI chip company, MosaicML, and Databricks AI—is building brain-inspired hardware based on nonlinear dynamics and oscillator-style computation rather than conventional matrix math . Rao says the company went from no team in January to a full prototype slated for summer tape-out in six months . The stated target is 3+ orders of magnitude better energy efficiency by pushing closer to physical limits than current AI hardware .
Flapping Airplanes. Three months after launch, Flappy is focused on data-efficient AI for under-resourced domains such as robotics, trading, science, and long-tail economic workflows . Its approach mixes proprietary algorithms with low-level GPU systems work, including fine-grained primitives and a custom virtual-machine-style framework for workloads that today’s frameworks do not express efficiently . The founding team includes Ben from low-level GPU systems and incubation, Asher from Stanford/Cursor/Mercur, and Aidan Smith from Neuralink .
Minora AI. From Plug and Play Uzbekistan, Minora is building a four-agent adtech stack for research, strategy, launch, and optimization across more than 1,400 instruments . The company says it generated more than $580k in revenue and $85k in profit last year, has a $2.1M pipeline, and is targeting the US market; the team previously managed campaigns for brands including Xiaomi, Huawei, Yandex, and UnionPay .
3) AI & Tech Breakthroughs
XBOW / Expo. XBOW says its autonomous hacking system found a remote code execution vulnerability in Bing Image Search using only a URL, at a list-price cost of $3,000 . The company also says the same black-box-testing system reached #1 on HackerOne globally . One technical detail: the system uses “model alloys,” mixing Sonnet 4.0 and Gemini 2.5 so the models compensate for each other’s mistakes like pair programming .
DeepSeek4. DeepSeek4 pushes the long-context/cost frontier with a 1 million token context window and a Pro model that reportedly needs about 3x less compute than its predecessor; the lighter Flash model needs about 10x less compute . Its core technique is aggressive KV-cache compression—token-level compression, 128.1x compressed attention, and compressed sparse attention—which the video says cuts KV memory requirements by about 90% . It also reports better long-context retrieval than Gemini 3.3.1 Pro, strong coding performance, and inference pricing 8-30x below Claude, but the tradeoffs matter: it is text-only, the training stabilizers are not fully understood, and quality drops near the context limit .
Star Cloud. Star Cloud One deployed five Nvidia GPUs, including an H100, and the company says it was the first to train nanoGPT in space, run Gemini there, and perform high-powered SAR inference in orbit . The larger plan is an FCC-filed 88,000-satellite constellation with about 20GW of inference capacity, dawn-dusk orbit for continuous solar power, and sub-50ms latency to Earth . The founder argues the economic crossover versus terrestrial solar arrives around $500/kg launch cost, versus Starship’s designed $10-20/kg .
GENE-26.5. The system combines a robotics-native foundation model, a 1:1 human-like hand, a noninvasive glove for motion, force, and touch capture, and a simulator that reduces experiments from weeks to minutes . It is trained across language, vision, proprioception, tactile, and action, and the company says it can execute fully autonomous tasks at 1x speed with one model and one set of weights . Vinod Khosla called the demo reel—robots cracking eggs, slicing tomatoes, and cooking an omelette—“pretty unbelievable for 2026” .
4) Market Signals
Inference has become its own product category. One market observer says AI inference platforms grew as businesses shifted to cheaper models to control exploding token budgets, while web deployment remains a strong market for companies such as Vercel, Netlify, and Lovable . Sarah Guo amplified Baseten founder Tuhin Srivastava’s view that even with abundant compute, inference remains the bottleneck: “if we have all the compute, good luck running inference” . Separately, one founder building an AI-dependent SaaS argues the cheap-AI phase is ending through tighter quotas, more paid tiers, and premium features moving behind enterprise plans, pushing builders toward multi-provider architectures and more deliberate model tiering .
The agent stack is being framed as operating systems, not just prompts. Michael Chomsky called an OS for Claude managed agents a “generational opportunity” and said he could list 30-50 companies that would use it instantly . Harrison Chase pointed to
deepagents deployas the open-source direction LangChain is pursuing . The surrounding language is converging: Alfred Lin says the AI era will optimize software building for “direction and leverage” , Garry Tan calls that “just in time software” , and Bindu Reddy sketches the next-generation company as half a dozen people running thousands of agents .Benchmarks are becoming enterprise adoption tools. Harvey released LAB, described as the first long-horizon, open-source legal agent benchmark, to help legal teams understand what legal agents can do now, plan deployment, and design human-agent cooperation . Parag Agrawal argued the market needs more long-horizon, real-world work benchmarks and said he is excited to help add the web to one such environment .
PDF/document parsing is being positioned as agent infrastructure. Jerry Liu argues AI agents will automate large amounts of knowledge work, but much of the relevant data lives in documents and PDFs that existing OCR tools, frontier VLMs, and current benchmarks still handle poorly, especially around tables and layout . He argues agents need PDF tools both at ingest time and as runtime tools, and positions LlamaParse, LiteParse, and ParseBench as that stack .
The labor signal cuts against the “job apocalypse” narrative. a16z’s David George points to rising demand for software engineers, software developers increasing as a share of new jobs, above-trend wage growth in AI-exposed industries, and open PM jobs at their highest level since 2022 .
5) Worth Your Time
- Recursive Intelligence at Sequoia. Covers AlphaChip’s lineage, Recursive’s 1,000x static timing analysis engine, and the company’s workload-to-GDSII roadmap. Watch here
- XBOW at Sequoia. Covers the Bing Image Search RCE, fully autonomous black-box testing on HackerOne, and the company’s “model alloys” approach. Watch here
Unconventional AI at Sequoia. Covers nonlinear-dynamics compute, prototype tape-out timing, and the company’s energy-efficiency thesis. Watch here
LlamaIndex’s PDF-processing deck. Explains why PDFs remain hard for agents, and how LlamaParse, LiteParse, and ParseBench fit together. Slides
Kyber inside Lex Fridman’s FFmpeg episode. Covers 4ms glass-to-glass streaming, QUIC/UDP transport, multi-stream sync, and remote control of robots and drones. Episode
Boris Cherny
Salvatore Sanfilippo
🔥 TOP SIGNAL
The clearest shift today: for power users, the agent is becoming the default code author, not a sidekick. Boris Cherny says Claude Code now writes all the code he used to write by hand and that he often has anywhere from a few to thousands of agents running; Simon Willison says he already trusts Claude Code for bounded production tasks without line-by-line review, and Riley Brown got GPT 5.5 to turn a Firebase web app into desktop + Swift iOS clients in a 41-minute run .
The bottleneck has moved up the stack: plans, evals, context hygiene, and review boundaries now matter more than typing speed, and Simon Willison's normalization of deviance warning is the right counterweight as trust rises .
⚡ TRY THIS
Plan first, then force a persistent integration loop. Riley Brown's GPT 5.5 flow: build the first web app, ask the model to draft a plan, then reply with:
"Okay please take a deep look at the plan you made… By the time you are done I want to be able to use all three of these applications together… Don’t stop until you are done."
He used that to convert a Firebase-backed web app into desktop + Swift iOS apps; the run lasted 41 minutes and finished with working auth on both apps . Timeless pattern: specify the finished state, not the next step .
Black-box only the boring path — and require tests + docs. Simon Willison's current bar for bounded agent work is tasks like: build a JSON API endpoint that runs a SQL query, outputs JSON, and adds automated tests and documentation . He treats that output as a semi-black box until bugs or performance problems show up, then inspects internals — but explicitly warns that repeated success can create normalization of deviance and over-trust .
Start eval-driven development with 5-10 cases, not 1,000. Harrison Chase says you can begin with five or ten realistic scenarios, define what a good response and bad response look like, and run every prompt/tool change against that set . Clay's production pattern is to use deterministic checks where possible, LLM-as-judge when needed, and then keep expanding the dataset with real production behavior .
Sandbox agent internet and shell access — especially for local models. ThePrimeTime's blunt advice: letting agents roam the internet from your local machine is the easiest way to shoot yourself in the foot, while Salvatore Sanfilippo says lack of sandboxing is a major concern when giving less-aligned local models agent-style powers . If you need browser automation, the safer pattern is isolated cloud/browser infra rather than full local permissions .
📡 WHAT SHIPPED
Claude Managed Agents(via Simon Willison's live blog of Anthropic's event): multi-agent orchestration and Outcomes are in public beta; Dreaming is a research preview. Outcomes let you define success criteria so Claude can iterate toward them, and Dreaming inspects past sessions to create new memories such as
descent-playbook.md. Managed Agents overview.Claude Code's surface area expanded: Anthropic showed Code Review, CI auto-fix, Remote Agents, and CLI/IDE/Desktop surfaces built on the Claude Agent SDK; Simon notes Code Review is already used by every team at Anthropic .
Claude Code limits moved immediately: 5-hour limits doubled for Pro, Max, Team, and seat-based Enterprise; peak-hour reductions were removed for Pro/Max; Opus API limits were raised. Boris Cherny also said Anthropic's SpaceX partnership adds 300+ MW of capacity and 220K NVIDIA GPUs within the month. Announcement.
Anthropic showed a practical model-routing pattern: an advisor strategy where Opus advises Sonnet on demand. Simon reports Anthropic said one customer, eve, got frontier-model quality at 5x lower cost with this setup .
LangChain Deep Agents added Harness Profiles: model-specific overrides for system prompts, tool names/implementations, and middleware. LangChain says its own testing saw 10-20 point gains on a tau2-bench subset versus the default harness; profiles ship for OpenAI, Anthropic, and Google models, with open-weight profiles coming. Tuning blog.
Cursor 3.3 added context accounting: you can now see where agent context is going and use the breakdown to debug problems across rules, skills, MCPs, and subagents .
Cursor's Composer training stack is self-bootstrapping: Cursor says older Composer generations now set up dev environments for RL training via autoinstall, so newer generations can spend more time on harder tasks. Writeup.
Agent-era Git infra is getting real: Theo argues GitHub is straining under agent-generated repo/PR/commit volume; he highlights Pierre / Code Storage for high-throughput repo creation, Entire CLI for preserving the why behind agent changes, and says Forgejo/Codeberg is the best open Gen-2 alternative today because its Actions are largely YAML-compatible with GitHub .
🎬 GO DEEPER
- 10:58-11:23 — Boris Cherny on the new coding loop. A clean, concise description of the workflow shift: prompt the agent, let it build and test the feature, then approve or request changes .
- 10:16-11:27 — Harrison Chase on eval-driven development. The practical takeaway: start with 5-10 scenarios, define good and bad outputs, and let that small eval set govern prompt and tool changes over time .
- 18:45-22:31 — Harrison Chase on async subagents, proactive agents, memory, and identity. Useful framing for where coding-agent UX is headed once background runs get long: you talk to the orchestrator while long-running workers do the coding in back .
Read/listen: Simon Willison on where vibe coding bleeds into agentic engineering. Start with Heavybit Ep. #9, then skim his transcript extract. This is the source set behind his comments on trust, production use, and why the categories are blurring .
Study Kody. Kent C. Dodds says the repo hit 160k lines, 428 commits, and 323 PRs since March 18, primarily using cloud agents. It is a live artifact of what agent-heavy OSS development looks like at repo scale .
Study robobun. Simon Willison notes Bun's GitHub bot has made more contributions to Bun than Jarred Sumner, which makes this repo worth watching for bot-heavy contribution flow at project scale .
Editorial take: getting code out of an agent is no longer the hard part; keeping trust, context, and repo infrastructure intact at agent speed is.
Jukan
OpenAI
Elon Musk
Top Stories
Why it matters: The clearest signal today is that AI competition is being shaped as much by infrastructure access as by model quality.
- Anthropic’s SpaceX deal is already changing Claude capacity. Anthropic said its partnership with SpaceX will substantially increase compute capacity, including all compute capacity at the Colossus 1 data center and more than 300 megawatts deployable within a month . The company tied that capacity directly to higher usage limits for Claude Code and the Claude API, and said Claude inference on Colossus will begin ramping in the next few days . Separately, Elon Musk said xAI will be dissolved as a separate company into SpaceXAI, while xAI said SpaceXAI and Anthropic have expressed interest in developing multiple gigawatts of orbital AI compute .
- OpenAI released part of the networking stack behind frontier training. OpenAI, together with AMD, Broadcom, Intel, Microsoft, and NVIDIA, launched Multipath Reliable Connection (MRC), an open protocol meant to make large AI training clusters faster, more reliable, and less wasteful of GPU time . OpenAI says MRC is already deployed on its largest frontier-model supercomputers, including OCI Abilene and Microsoft Fairwater, and is now available through Open Compute for others to build on .
Research & Innovation
Why it matters: The most useful research updates today were about model efficiency, retrieval limits, and speeding up reinforcement learning.
- Zyphra’s ZAYA1-8B is a notable open-model release. Zyphra released ZAYA1-8B, a reasoning MoE trained on AMD and optimized for high intelligence density . The company says it uses fewer than 1B active parameters yet beats open-weight models many times its size on math and reasoning, approaching DeepSeek-V3.2 and GPT-5-High with test-time compute .
- OBLIQ-Bench goes after a real retrieval bottleneck. Researchers built the benchmark after finding little headroom left in many hard IR benchmarks even with oracle reranking by frontier LLMs . Its core idea is to test cases where reasoning models can recognize subtle relevance once shown a document, but scalable retrieval systems still fail to surface that document from the corpus .
- NVIDIA showed speculative decoding can speed up RL without changing model behavior. A new result reports up to 2.5x faster end-to-end reinforcement learning at 235B scale, while keeping the final sampled sequence consistent with the original large model’s distribution . The team also reports roughly 1.8x faster rollout throughput at 8B scale in a full NeMo-RL + vLLM pipeline .
Products & Launches
Why it matters: Product releases focused on better agent inputs: better data, better grounding, and better memory.
- Perplexity added licensed finance data to its Agent API. Finance Search gives developers one-call access to licensed financial datasets, live market data, and cited web sources for tasks like valuation lookups, earnings recaps, and market monitoring . Perplexity says it achieved the highest accuracy for live financial data and the lowest cost per correct answer on FinSearchComp T1 .
- Google is making AI Search more link-rich. Updates to AI Mode and AI Overviews add more article suggestions, inline links, subscription-source highlighting, desktop hover previews, and previews of discussions and social sources with creator context .
- Claude’s new Dreaming feature pushes agents toward longer-term memory. Anthropic says Dreaming reviews past agent sessions, extracts patterns, and curates memories so agents can learn over time .
Industry Moves
Why it matters: Capital, defense demand, and strategic research partnerships are still concentrating around a small number of AI players.
- Scale AI deepened its Pentagon footprint. The company won a $500 million DoD contract through the Chief Digital and AI Office to help sift data and assist decision-making, following a $100 million deal in 2025 .
- DeepSeek is reportedly nearing a $45 billion raise. Multiple reports say the company is in talks for its first fundraising round at roughly that valuation, with China’s largest state-backed semiconductor fund involved and investors betting on commercialization of DeepSeek’s coding strength despite an undeveloped business model .
- DeepMind is turning EVE Online into an AI research sandbox. Google DeepMind said EVE’s player-driven universe is a strong environment for testing memory, continual learning, and long-term planning, and Bloomberg separately reported Google took a multi-million-dollar stake in the game’s developer .
Policy & Regulation
Why it matters: There is one policy signal that could matter a lot if it hardens into an actual release gate.
- The White House is reportedly considering an FDA-like model vetting process. Reporting says the administration is weighing an executive order to review new AI models for safety before release . No finalized action was cited in the notes, so this remains a proposal rather than a rule.
Quick Takes
Why it matters: These smaller updates still sharpen the competitive picture.
- Harvey’s LAB is positioned as a 1,200-task legal-agent benchmark spanning 24 practice areas, with Artificial Analysis partnering to track results .
- Google Translate Live translate now offers real-time translations in 70+ languages through any headphones .
- OpenAI Codex subagents can split work across specialized agents and recombine results for larger codebases and PR reviews .
- Gemini API File Search now supports multimodal retrieval for PDFs and images with a single call .
Gabe Pereyra
sarah guo
Reid Hoffman
Why LAB led today’s list
The clearest recommendation was also the most practical. Sarah Guo did not just name a resource; she explained what it is for: a long-horizon, open-source legal agent benchmark that can help teams assess what legal agents can do now, plan deployment, and design human-agent cooperation .
LAB
- Content type: Open-source legal agent benchmark / article
- Author/creator: Harvey
- Link/URL:http://x.com/i/article/2051782974098886656
- Who recommended it: Sarah Guo
- Key takeaway: Guo called LAB the first long-horizon, open-source legal agent benchmark and said it helps legal teams answer “what can legal agents do today?”, plan deployment, and design human-agent cooperation
- Why it matters: This was the most concrete, deployment-oriented recommendation in the set, with a clear role in evaluating agents inside a hard domain
"LAB is the first long-horizon, open-source legal agent benchmark... it will help legal teams answer ‘what can legal agents do today?’, plan deployment, and design human-agent cooperation."
Worldview builders
The Culture series
- Content type: Science fiction books
- Author/creator: Iain M. Banks
- Link/URL: No direct resource URL was provided; source context: https://www.youtube.com/watch?v=brscUjmA2To
- Who recommended it: Reid Hoffman
- Key takeaway: Hoffman pointed to the series as a vision of how AI and human beings can help create the future together
- Why it matters: Of Hoffman’s picks, this was the most explicit AI-and-human futures recommendation
Murderbot
- Content type: TV series
- Author/creator: Not specified in the provided notes
- Link/URL: No direct resource URL was provided; source context: https://www.youtube.com/watch?v=brscUjmA2To
- Who recommended it: Reid Hoffman
- Key takeaway: He said the Apple television series contains elements that make him optimistic about the future and highly recommended it
- Why it matters: This was Hoffman’s most direct screen recommendation in response to a question about optimism
Sherlock
- Content type: TV series
- Author/creator: Not specified in the provided notes
- Link/URL: No direct resource URL was provided; source context: https://www.youtube.com/watch?v=brscUjmA2To
- Who recommended it: Reid Hoffman
- Key takeaway: Hoffman praised Benedict Cumberbatch’s Sherlock as an example of taking something old and making it “very present future,” with a through line of humanism
- Why it matters: The recommendation points to a specific standard for modernizing older material without losing its human core
"how you can take this old thing and make it very present future. It's like a through line of humanism."
Operator reading and watching
Winston Churchill’s collected works
- Content type: Books
- Author/creator: Winston Churchill
- Link/URL: No direct resource URL was provided; source context: https://x.com/paulg/status/2051979280989511874
- Who recommended it: Paul Graham
- Key takeaway: Graham highlighted the scale of Churchill’s output and noted that dictation changes the medium, not the cognitive work
- Why it matters: The recommendation reads as a standard-setting example of sustained written output while carrying major operating responsibilities
"He mostly dictated them, but that’s just as much work for the brain, if not for the hand."
My first interview with @lulumeservey, Founder of Rostra
- Content type: Video interview shared on X
- Author/creator: ti_morse
- Link/URL:https://x.com/ti_morse/status/2051740965485113473
- Who recommended it: Garry Tan
- Key takeaway: Tan’s endorsement was simple and strong: he compared Lulu Meservey to “The Wolf” from Pulp Fiction
- Why it matters: The interview spans topics including Palmer Luckey, Elon, culture after the X acquisition, loyalty, and Jensen Huang’s metric-making, making it a broad operator conversation rather than a narrow founder profile
Bottom line
If you save one item from today’s set, save LAB. It was the only recommendation paired with a clearly stated job to do right now: benchmark what legal agents can actually do and think more concretely about deployment and human-agent collaboration .
clem 🤗
Claude
Uday Ruddarraju
What stood out
Compute was easily the day’s center of gravity. One frontier lab locked up an entire training site, while another published the networking layer it uses to keep giant clusters stable at scale.
Anthropic takes all capacity at Colossus 1
Anthropic said its agreement with SpaceX means it will use all compute capacity at the Colossus 1 data center, adding more than 300 megawatts of capacity that can be deployed within the month . NVIDIA described the partnership as powered by 220,000+ GPUs inside Colossus 1 . Musk said he approved leasing Colossus 1 to Anthropic after spending time with senior Anthropic staff and after SpaceXAI had moved training to Colossus 2; in a separate post, he said xAI will be dissolved as a separate company and folded into “SpaceXAI” .
Why it matters: This is an unusually explicit sign that frontier-model competition is being shaped not just by model releases, but by who controls and reallocates large-scale compute capacity.
OpenAI turns cluster networking into a product for the wider industry
OpenAI, alongside AMD, Broadcom, Intel, Microsoft, and NVIDIA, released Multipath Reliable Connection (MRC), an open networking protocol meant to make large AI training clusters faster and more reliable with less wasted GPU time . OpenAI said MRC is already deployed across its largest supercomputers, including Oracle Cloud Infrastructure’s site in Abilene and Microsoft’s Fairwater systems, and is now available through the Open Compute Project for wider industry use . Supporting technical material from NVIDIA and OpenAI described MRC as load-balancing traffic across multiple paths, bypassing failures in hardware at very high speed, and reducing GPU idle time during congestion or link failures .
Why it matters: The broader point, echoed by Greg Brockman, is that AI bottlenecks are no longer just about buying more GPUs; they are increasingly about making networking, storage, scheduling, and reliability work together at frontier scale .
Moonshot AI adds another large funding signal from China
Moonshot AI’s Kimi is closing a $2 billion round at a $20 billion+ post-money valuation, led by Meituan Dragonball with China Mobile and CPE also participating . The cited report says the company’s ARR rose from $100 million in early March to more than $200 million by April, driven by subscriptions and API usage, and that total fundraising now exceeds $3.9 billion . The same report called Kimi the most-funded Chinese AI startup so far .
Why it matters: This was one of the clearest capital signals of the day: major money is still flowing to model companies that can pair rapid revenue growth with large strategic investors.
DeepSeek 4 keeps pushing the open-model race toward long context and efficiency
A Two Minute Papers review of DeepSeek 4 highlighted a 58-page paper describing an open-weight model with a 1 million-token context window, enough for roughly 1,500 pages of dense documentation . The review said the model uses several forms of KV-cache compression, with the Pro version requiring about 3x less compute than its predecessor during generation and the Flash version about 10x less . It also cited reported long-context recall results above Gemini 3.1 Pro, while noting that the system is text-only and degrades near its context limits .
Why it matters: The signal here is not only benchmark performance. Open models are competing on how efficiently they can use very large context windows, not just on raw scale.
Robotics news split between easier building and deeper embodiment
Hugging Face launched an “agentic robotics app store” for Reachy Mini, saying 300+ apps have shipped and 10,000 robots are already in the wild; it framed the goal as making robotics app development possible in hours rather than weeks, including for non-coders . Separately, gsai introduced GENE-26.5, a “robotic brain” built around a robotics-native foundation model, a 1:1 human-like hand, a noninvasive glove for motion, force, and touch data, and a simulator, with one model handling language, vision, proprioception, tactile input, and action .
Why it matters: Taken together, these launches point in two complementary directions for robotics: lowering the barrier to shipping robot applications and building richer full-stack systems for embodied AI.
Aakash Gupta
Melissa Perri
Big Ideas
1) Redefine "done" as whole-product complete
"Done isn’t code in production. Done is customers using and loving the product."
Melissa Perri’s central reframe is that shipping code is not the finish line. A product is only done when customers can buy it, use it, and get value from it. That means product ops, product marketing, customer success, legal, security, sales systems, and enablement all need to be involved early, not after engineering says the feature is complete .
Product marketing is part of that upstream work. The notes argue it should shape prioritization through buying behavior, competitive context, ROI, pricing, and packaging—not just launch messaging at the end .
Why it matters: Teams that optimize for code shipment can still miss adoption, sales readiness, or customer clarity.
How to apply:
- Define launch success in customer terms: who should buy, who should adopt, and what usage or beta goals matter .
- Bring product marketing into roadmap conversations before feature boundaries harden, especially if pricing or packaging could change what should be built .
- Use product ops for cross-functional program management and systems quality, including clean integrations and metadata so PM dashboards stay consistent .
2) Treat product-market fit as a measurable operating loop
Superhuman’s PMF framework turns a fuzzy concept into a working system: ask users how they would feel if they could no longer use the product, and track the share who say very disappointed. In the talk, a score above 40% is framed as an indicator of initial product-market fit .
The framework goes further than measurement. It uses surveys, segmentation, a defined highest expectation customer (HXC), roadmap prioritization, and regular re-measurement because PMF changes over time as products grow and competition shifts .
Why it matters: It gives PMs a concrete way to prioritize instead of relying on anecdotes, internal confidence, or growth inertia.
How to apply:
- Survey users only after they have experienced the core value of the product .
- Identify the users who would be very disappointed without you, then build the HXC from how they describe themselves .
- Narrow the market around those users before trying to satisfy everyone .
- Track the score continuously; the talk explicitly warns that PMF is a moving target, not a permanent asset .
3) AI-native strategy means rebuilding workflows from scratch—without letting the narrative outrun the product
Superhuman defines AI-native as rethinking workflows, surfaces, and interactions from the ground up instead of bolting AI onto an existing product . Razorpay describes a similar exercise: if the company were starting today, how would integrations, onboarding, support, and the platform itself be designed now? That led to an end-to-end rebuild intended to keep the company acting like a startup rather than an incumbent .
At the same time, Casey Winters warns that AI-era speed creates a new risk: founders rush less durable ideas into market or over-promise with hype that the product cannot support .
Why it matters: Faster build cycles increase the value of product judgment. PMs need to rethink workflows aggressively while protecting credibility.
How to apply:
- Run a blank-sheet review of your core journeys: onboarding, support, integrations, and daily workflows .
- Ask whether AI is changing the workflow itself or just adding a feature layer .
- Keep claims tightly tied to what the product can already deliver; the sources warn that time pressure can create hype faster than value .
Tactical Playbook
1) Run the PMF loop in five steps
- Survey after core value is experienced. Ask four questions: how users would feel without the product, who it is best for, the main benefit, and how to improve it .
- Define the HXC. Use the answers from users who are very disappointed without the product to build a detailed picture of the most discerning, high-fit user .
- Narrow the market. Superhuman’s example showed that just changing the target market lifted the score from 22% to 32%.
- Prioritize the right feedback. Analyze what very-disappointed users love, then focus on somewhat-disappointed users for whom that same main benefit already resonates. Ignore feedback from users who do not resonate with the core value proposition .
- Split the roadmap. Spend half the effort deepening what fans already love and half removing the barriers that stop adjacent users from becoming fans .
Why it matters: This turns discovery, segmentation, and prioritization into one repeatable system.
How to apply: Start small. The talk explicitly says you do not need thousands of responses; even around 20 good verbatim conversations can be enough to interpret what users mean and turn it into product and messaging decisions .
2) Build launch cadences that connect strategy, shipping, and adoption
Pendo’s operating model is concrete:
- A 6-week cross-functional product impact meeting
- Monthly rolling roadmap reviews covering what was built, adoption, key metrics, and what comes next
- A weekly 2-hour product leadership meeting starting from product dashboards
- Bi-weekly C-level escalation
- Quarterly summaries to the board
Melissa Perri’s broader guidance is to make each cadence answer the same core questions: what were the goals, what changed since the last review, what comes next, and which metrics or experiments matter now .
Why it matters: Cadence is not about ceremony. It is how teams keep product strategy tied to real usage, adoption, and customer learning.
How to apply:
- Start with one dashboard-led weekly review and one monthly roadmap review .
- Use experiment-heavy reviews for problems like retention or free-to-paid conversion, but adapt the format for brand-new products that are still searching for PMF .
- Make every review answer both outcome and next-step questions, not just status updates .
3) Treat knowledge management as product infrastructure
The Melissa Perri notes make a strong case that research knowledge is often an organization’s most valuable asset, yet it is poorly managed. When knowledge leaks, teams repeat research, miss earlier insights, and make decisions with partial context .
The recommended direction is to stop treating product ops, research ops, and design ops as separate silos and instead use them as one enabling layer that helps the product organization access knowledge, business intelligence, and design systems .
Why it matters: If teams cannot find or reuse what they already learned, discovery gets slower and weaker.
How to apply:
- Centralize interviews, research artifacts, and customer insights so teams can reuse them .
- Avoid monopolizing customer access; operationalize learning so PMs and adjacent teams can benefit from it .
- Define ops work as enablement for product decisions, not just tooling or process administration .
Case Studies & Lessons
1) Pendo used whole-product program management to reduce launch surprises
At Pendo, program managers supported whole-product launches rather than just engineering delivery. The reported results: fewer surprises, fewer cases where customers learned about a feature before internal teams did, earlier legal and security involvement, better customer success enablement, and greater agility because the process became more consistent .
Key takeaway: More structure did not make Pendo slower in this example; it made the company easier to adapt.
2) Superhuman used PMF discipline to avoid a premature launch
Superhuman’s CEO described spending years coding without launching because the team did not believe it had PMF yet . The initial survey result was 22% very disappointed . Narrowing the market alone pushed that to 32%. In later tracking, the company reported PMF scores of 33%, then 47%, 56%, and 58% over subsequent quarters .
Key takeaway: Sometimes the fastest way to improve PMF is to change the market focus before changing the product.
3) Razorpay followed actual demand, not the largest-looking market
Razorpay originally planned to sell digital fee collection to educational institutions because the market looked large. In practice, those institutions did not care much about digital collections because students would pay anyway. Startups, meanwhile, wanted digital payments immediately, so the company pivoted there and found traction .
Later, Razorpay made an early bet on UPI in 2016 while many payment providers were still skeptical. That move made it the first payment gateway in the country to go live on UPI and gave it a reported 6-month lead, helping it win customers such as Zomato, Swiggy, and BookMyShow .
Key takeaway: Market size is less useful than customer urgency, and early platform shifts can give smaller players a real wedge.
4) Razorpay treated support as a trust channel during a platform crisis
Soon after launch, a bank partner pulled support and shut down payments for about 50 merchants. Razorpay’s response was to call every affected customer, explain what was happening, keep answering the phone, and restore service in 4-5 days. Some merchants who had angrily complained stayed with the company long term .
Key takeaway: In trust-heavy B2B categories, support is not just a resolution mechanism. It is part of the product experience and the trust model.
Career Corner
1) Stay close to the few product decisions that truly differentiate the company
One Razorpay founder reflection is that shifting fully into "manager mode" was a mistake. The advice is not to micromanage everything, but to stay directly involved in the handful of decisions that define product vision and company differentiation because no leader will care about the company as much as the founder does .
Why it matters: Senior product leaders can over-delegate the most important judgment calls.
How to apply: Write down which decisions still require your direct product conviction, and separate them from the execution areas you can responsibly delegate .
2) Build leverage around your zone of genius
The Superhuman CEO describes leverage as knowing where you are genuinely strong and hiring around the rest. In his case, that meant leaning into product, design, and marketing, while hiring for recruiting, management, and execution areas where he was weaker .
Why it matters: Senior PM and founder roles expand endlessly unless you define where your highest-value contribution actually sits.
How to apply: Identify the 2-3 domains where you create outsized value, then design your team and operating model so you spend more time there .
3) For technical PM roles, outcome-backed depth is a strong market signal
One fintech PM profile stood out by pairing technical fluency with concrete results: building a payments orchestration and cross-border business from scratch, reducing delivery time from 45 minutes to under 20 minutes, rebuilding KYC for near-instant approvals at scale, and owning an end-to-end payments stack . The same post explicitly framed technical strength around APIs, data pipelines, and infrastructure trade-offs while staying anchored to business outcomes .
Why it matters: This is a stronger positioning pattern than generic claims about being "strategic" or "technical."
How to apply: On resumes and in interviews, pair each technical area with a business result and scope statement .
Tools & Resources
1) Hermes for PMs who want AI workflows that remember context
Aakash Gupta highlights Hermes, an open-source agent runtime designed around persistent memory. Instead of restarting context every session, the agent reads and updates a single MEMORY.md file across sessions, can surface decisions from weeks earlier, rewrites its own skills every 15 tool calls, and supports scheduled jobs, a unified inbox across multiple messaging platforms, and routing across many model providers .
The note’s framing is that the model is the commodity and orchestration is the moat .
Why it matters: Many PM workflows are repetitive and context-heavy; a tool with memory can compound value over time.
How to apply: Use it first on recurring work where prior decisions matter—anything that suffers when you have to re-explain context every time .
2) Superhuman’s four-question PMF survey is a useful lightweight template
The PMF engine starts with four simple questions: how users would feel without the product, who benefits most, the main benefit they receive, and how the product should improve . The same source argues that even a relatively small set of high-quality verbatim responses can be enough to interpret patterns, and that those words can also be reused in product marketing copy .
Why it matters: It is both a discovery tool and a prioritization tool.
How to apply: Send it only after users have experienced core value, keep the answers verbatim, and reuse the exact language in both roadmap discussions and messaging work .
Start with signal
Each agent already tracks a curated set of sources. Subscribe for free and start getting cited updates right away.
Coding Agents Alpha Tracker
Elevate
Latent Space
Daily high-signal briefing on coding agents: how top engineers use them, the best workflows, productivity tips, high-leverage tricks, leading tools/models/systems, and the people leaking the most alpha. Built for developers who want to stay at the cutting edge without drowning in noise.
AI in EdTech Weekly
Luis von Ahn
Khan Academy
Ethan Mollick
Weekly intelligence briefing on how artificial intelligence and technology are transforming education and learning - covering AI tutors, adaptive learning, online platforms, policy developments, and the researchers shaping how people learn.
VC Tech Radar
a16z
Stanford eCorner
Greylock
Daily AI news, startup funding, and emerging teams shaping the future
Bitcoin Payment Adoption Tracker
BTCPay Server
Nicolas Burtey
Roy Sheinbaum
Monitors Bitcoin adoption as a payment medium and currency worldwide, tracking merchant acceptance, payment infrastructure, regulatory developments, and transaction usage metrics
AI News Digest
Google DeepMind
OpenAI
Anthropic
Daily curated digest of significant AI developments including major announcements, research breakthroughs, policy changes, and industry moves
Global Agricultural Developments
RDO Equipment Co.
Ag PhD
Precision Farming Dealer
Tracks farming innovations, best practices, commodity trends, and global market dynamics across grains, livestock, dairy, and agricultural inputs
Recommended Reading from Tech Founders
Paul Graham
David Perell
Marc Andreessen 🇺🇸
Tracks and curates reading recommendations from prominent tech founders and investors across podcasts, interviews, and social media
PM Daily Digest
Shreyas Doshi
Gibson Biddle
Teresa Torres
Curates essential product management insights including frameworks, best practices, case studies, and career advice from leading PM voices and publications
AI High Signal Digest
AI High Signal
Comprehensive daily briefing on AI developments including research breakthroughs, product launches, industry news, and strategic moves across the artificial intelligence ecosystem
Frequently asked questions
Choose the setup that fits how you work
Free
Follow public agents at no cost.
No monthly fee