Hours of research in one daily brief–on your terms.

Tell us what you need to stay on top of. AI agents discover the best sources, monitor them 24/7, and deliver verified daily insights—so you never miss what's important.

Setup your daily brief agent
Discovering relevant sources...
Syncing sources 0/180...
Extracting information
Generating brief

Recent briefs

Cloudflare Pushes Code-Writing Agents as Claude Code and Cursor Tighten the Loop
Mar 25
6 min read
158 docs
Claude
Figma
Riley Brown
+12
Cloudflare's Dynamic Workers and Code Mode signal a more important shift than another model release: coding agents are moving from brittle tool menus toward fast, disposable code execution. Also inside: Claude Code auto mode, Cursor's Composer 2 report, Riley Brown's Codex/Xcode workflow, and production multi-agent patterns worth stealing.

🔥 TOP SIGNAL

The sharpest infra shift today is Cloudflare's push toward code-writing agents. Dynamic Workers are now in Open Beta with secure sandboxes that start ~100x faster than containers and use 1/10 the memory, and Kent C. Dodds says he's already using them to power his own AI assistant with CodeMode . The companion Code Mode thesis matters just as much: expose tools as an SDK and let the model write code to select and compose them, instead of stuffing more JSON-style tools into the context window .

"Agents should interact with the world by writing code, not tool calls."

🛠️ TOOLS & MODELS

  • Claude Code — auto mode. Claude can now decide file-write and bash permissions on the user's behalf, with safeguards checking each action before it runs. That replaces the old choice between constant approvals and --dangerously-skip-permissions.
  • Cursor Composer 2 — training report out. Cursor says Composer 2 was built with three main efforts: continued pretraining, reinforcement learning, and benchmark development to emulate the Cursor environment. The practical takeaways: continued pretraining kept improving downstream coding performance, the RL phase was critical enough that simple approaches worked best broadly, and CursorBench is their internal benchmark for more realistic software engineering problems. The report also covers distributed training / RL infrastructure and kernels they open-sourced. Read: Composer 2 report.
  • Cloudflare Dynamic Workers + Code Mode. Dynamic Workers entered Open Beta for paid Workers users, with disposable sandboxes that start ~100x faster than containers and use 1/10 the memory. The adjacent Code Mode posts surfaced three agent-facing building blocks: codemode for chewing through multiple tool calls in one/few shots, shell for a local filesystem backed by DO/D1/SQL/R2, and worker-bundler for packaging npm deps and serving full apps. Read: Dynamic Workers and Code Mode.
  • Cursor + Figma. Cursor can now create new components and frontends in Figma using your team's design system, including variables, tokens, and naming conventions. This landed alongside Figma's open beta for AI agents designing directly on canvas via the use_figma MCP tool .
  • Devin as reviewer, not just author. The clearest practitioner comparison today came from Gauri Gupta and swyx: both say Devin code review catches bugs Claude and Codex miss, and swyx says the review pass saves him 3-8x/day .

💡 WORKFLOWS & TRICKS

  • Replicable mobile-agent build loop from Riley Brown

    1. Create the app shell in Xcode, then open the top-level project folder in Codex and start a dedicated thread .
    2. Paste the full product spec and hit Shift+Tab for Plan mode .
    3. Run design in parallel in Paper.design via MCP while Codex builds .
    4. Steer inline while the thread is running, then paste API keys, Xcode logs, and screenshots back into the chat when builds fail .
    5. Test on a physical iPhone, and keep the generated app preview open so you can keep iterating without leaving the app .
    • Bonus: Brown adds Whisper-based voice input so he can keep refining the preview without opening the keyboard .
    • Timeless pattern: split planning, design, implementation, and debugging across parallel agent threads instead of forcing one giant chat .
  • Turn prompts into loops, not one-shots

    • Karpathy's Auto Research runs on a git feature branch: propose experiment, write code, run it, analyze results, commit improvements, repeat .
    • Matthew Berman says he applies the same shape to overnight fine-tuning of small open models against an Opus 4.6 baseline until a local model wins .
    • Minimax's M2.7 harness adds skills, memory, guardrails, and eval infra; it handled 30-50% of the workflow and reported a 30% performance gain on evals .
    • Anthropic says similar Claude loops now power almost all of its major agent loops, with humans reviewing solutions before final refinements .
  • Production pattern worth stealing from Moda

    • Use a cheap Haiku triage node to classify the job and preload task-specific Markdown skills .
    • Inject those skills as human messages, with caching breakpoints after the system prompt and after the skills block .
    • Keep only 12-15 core tools always in context; activate ~30 more on demand with RequestToolActivation.
    • For big projects, pass a high-level summary and let the agent pull details through tools; then use traces to tune prompt/tool changes, cost, cache hits, and failures .
  • Add a critic, not just another worker. swyx says he won't merge without a Devin review now, because it catches bugs Claude and Codex miss and acts like fresh eyes on a PR. Related design note: he points to Windsurf's "smart friend" pattern as a model for subagents that are more critical instead of weaker .

👤 PEOPLE TO WATCH

  • Kent C. Dodds — strongest practitioner confirmation on today's Cloudflare story: he's been using Dynamic Workers for a week to power his own AI assistant, and separately backs the "tools as SDK + agent-written code" pattern as the more effective direction .
  • Riley Brown — one of the most concrete end-to-end Codex workflows in the wild right now: Xcode + Codex + Claude Agent SDK + Vibe Code CLI + Paper.design, with screenshots and logs fed back into the loop .
  • swyx — short posts, strong signal: Devin review outperforming Claude/Codex in his day-to-day, plus the "smart friend" subagent pattern callout .
  • Ravi Parikh / Moda — rare production detail instead of benchmark chatter: triage, dynamic tool loading, prompt caching, context scaling, and tracing from a real multi-agent design product .
  • Anthropic Engineering — new post on a multi-agent harness for frontend design and long-running autonomous software engineering. Worth reading because it's from the team running the stack, not a third-party teardown: harness-design-long-running-apps.

🎬 WATCH & LISTEN

  • 1:56-4:28 — M2.7's harness loop. Best compact explanation today of the human/harness/agent split: human steers direction, the harness supplies skills/memory/guardrails/evals, and the agent writes, runs, analyzes, and improves experiments .
  • 3:36-5:43 — Riley Brown's Codex setup for "Jerry." Watch this if you want a replicable "agent builds the app that builds apps" flow: create the Xcode project, point Codex at the right folder, then fan out into parallel threads .
  • 25:16-26:17 — In-app preview + voice iteration. Shorter clip, but high leverage: Riley tweaks the generated app while the preview is open and adds Whisper-based voice input so the iteration loop never leaves the phone UI .

📊 PROJECTS & REPOS

  • karpathy/auto-research — GitHub repo for autonomous research loops on a git feature branch. Practical signal: Berman frames it as bringing frontier-style experiment loops to solo developers, and points to an overnight run that reached the fastest reported time to train one of these models .
  • langchain-ai/deepagents — better adoption signal than most new repos because Moda is already using Deep Agents in production for its Research Agent and Brand Kit Agent, with the Design Agent migration under evaluation. Repo: Deep Agents.

Editorial take: today's edge is loop design — faster execution sandboxes, smaller always-on tool sets, explicit review passes, and tighter human feedback beat one giant prompt.

1,000 True Fans Leads Today’s Operator, AI, and Engineering Picks
Mar 25
5 min read
191 docs
nick ercolano
martin_casado
Tim Ferriss
+5
Tim Ferriss provides the clearest high-conviction pick with Kevin Kelly’s 1,000 True Fans, while Martin Casado, Amjad Masad, Pratap Ranade, Vincent Weisser, and Ryan Hoover add useful resources on AI limits, creator leverage, verification, engineering history, and documentary storytelling.

Most compelling recommendation

Today’s clearest save is Kevin Kelly’s 1,000 True Fans. Tim Ferriss doesn’t just endorse it; he states the usable lesson: precise targeting makes a business smarter, faster, more agile, and less expensive .

  • Title:1,000 True Fans
    Content type: Essay / online article
    Author/creator: Kevin Kelly
    Link/URL: Not provided in the source material; Ferriss says it is free online
    Who recommended it: Tim Ferriss
    Key takeaway: Be very specific with your targeting
    Why it matters: Ferriss calls it "the most important marketing piece you may ever read," making it the highest-conviction recommendation in today’s set

"the most important marketing piece you may ever read is 1000 true fans"

Tim Ferriss’s operator shelf

  • Title:The 22 Immutable Laws of Marketing(original edition)
    Content type: Book
    Author/creator: Al Ries and Jack Trout
    Link/URL: Not provided in the source material
    Who recommended it: Tim Ferriss
    Key takeaway: Ferriss specifically recommends the older edition, not the internet version
    Why it matters: It is one of his core marketing recommendations, alongside 1,000 True Fans

  • Title:Secrets of Power Negotiating
    Content type: Book / audiobook
    Author/creator: Roger Dawson
    Link/URL: Not provided in the source material
    Who recommended it: Tim Ferriss
    Key takeaway: Ferriss says the audio version is especially good
    Why it matters: This is his clearest negotiation recommendation in the source material

  • Title:Built to Sell
    Content type: Book
    Author/creator: John Warrillow
    Link/URL: Not provided in the source material
    Who recommended it: Tim Ferriss
    Key takeaway: Think about selling your company from the very beginning and design it that way
    Why it matters: Ferriss presents it as a company-design lens, not just a late-stage exit book

  • Title:Topgrading(original edition)
    Content type: Book
    Author/creator: Bradford Smart
    Link/URL: Not provided in the source material
    Who recommended it: Tim Ferriss
    Key takeaway: Use the original edition; Ferriss says newer editions are mostly upsells
    Why it matters: He calls it the best book he has found on interviewing top talent

AI mental models and builder leverage

"Hence Shannon, not Kolmogorov. Lorentz, not Einstein."

  • Title:Video on AI and what’s happened over the last 6 months(title not specified in the source material)
    Content type: Video
    Author/creator: Nick Colano
    Link/URL:https://x.com/nickercolano/status/2036060436487926111
    Who recommended it: Amjad Masad
    Key takeaway: After hundreds of hours building with AI tools, Colano argues that the old barrier between creators and their ability to build products—price, speed, and vision—is basically gone
    Why it matters: Masad explicitly recommends it as a genuine, non-marketing account of how learning software changed a creator business

"the (former) barrier between creators and their ability to build products (price, speed, vision) is basically gone"

Weisser also named The Beginning of Infinity: Explanations That Transform the World and Superintelligence: Paths, Dangers, Strategies among the works shaping his thinking on intelligence and progress .

Engineering and verification deep cuts

  • Title:Cramming More Components Onto Integrated Circuits
    Content type: Research paper
    Author/creator: Gordon Moore
    Link/URL:https://hasler.ece.gatech.edu/Published_papers/Technology_overview/gordon_moore_1965_article.pdf
    Who recommended it: Pratap Ranade
    Key takeaway: Ranade points readers to the paper’s overlooked second half on analog silicon, which he says is even more fascinating today than the digital section
    Why it matters: It is a strong reminder to revisit canonical papers for the parts everyone skipped

  • Title:Asymmetry of verification and verifier’s rule
    Content type: Blog post
    Author/creator: Jason Wei
    Link/URL:https://www.jasonwei.net/blog/asymmetry-of-verification-and-verifiers-law
    Who recommended it: Pratap Ranade
    Key takeaway: Verifier’s law says that any task that is easy and fast to verify will be automated by AI
    Why it matters: Ranade uses it to explain why faster verification is the unlock for applying generative AI to EM design

  • Title:Skunk Works
    Content type: Book
    Author/creator: Ben Rich
    Link/URL:https://www.amazon.com/Skunk-Works-Personal-Memoir-Lockheed/dp/0316743003
    Who recommended it: Pratap Ranade
    Key takeaway: Ranade highlights its story of how the F-117 stealth bomber came to be through unconventional geometry
    Why it matters: It is a concrete engineering-history example of non-obvious design choices paying off

  • Title:Tuxedo Park
    Content type: Book
    Author/creator: Jennet Conant
    Link/URL:https://www.amazon.com/Tuxedo-Park-Street-Science-Changed/dp/0684872889
    Who recommended it: Pratap Ranade
    Key takeaway: Ranade recommends it for readers who want to learn more about WWII radar research at Tuxedo Park
    Why it matters: It adds historical context to the scientific lineage behind modern hardware ambition

One non-technical watch

  • Title:In Waves and War
    Content type: Documentary
    Author/creator: Not specified in the source material
    Link/URL: Not provided in the source material
    Who recommended it: Ryan Hoover
    Key takeaway: Hoover says the film is heartbreaking, cites roughly 20 veteran suicides a day, and says it showcases ibogaine’s positive benefits for three troubled Navy SEALs
    Why it matters: He says he teared up watching it, making this the strongest emotional recommendation in today’s set
LiteLLM Breach, OpenAI’s $1B Foundation, and an Opening at the AI Infrastructure Layer
Mar 25
5 min read
167 docs
Jim Fan
Daniel Hnyk
hardmaru
+14
A major supply-chain compromise put AI developer security at the center of the day. OpenAI launched a $1B foundation while shutting down the Sora app, and Modular, NVIDIA, and materials researchers pointed to deeper shifts in infrastructure and science.

Security wake-up call

LiteLLM’s compromise showed how exposed the AI developer stack has become

LiteLLM’s PyPI release 1.82.8 was reported compromised: a litellm_init.pth file contained base64-encoded instructions to exfiltrate SSH keys, cloud credentials, Kubernetes configs, git credentials, API keys, shell history, crypto wallets, SSL keys, CI/CD secrets, and database passwords, while also self-replicating . Because LiteLLM sees about 97 million downloads per month and sits inside other projects such as dspy, the potential blast radius extended well beyond direct installs .

The poisoned version appears to have been live for less than an hour and was discovered after a bug caused a machine using a Cursor MCP plugin to run out of RAM and crash .

“Supply chain attacks like this are basically the scariest thing imaginable in modern software.”

Why it matters: For AI teams, this was a reminder that model infrastructure can fail through ordinary package management, not just through model misuse. Swyx argued newer package managers should add stronger install-time guards, while Jim Fan warned that agentic systems could turn the filesystem itself into a much larger attack surface .

OpenAI made two very different moves

OpenAI launched a $1B foundation around science and AI resilience

OpenAI said the OpenAI Foundation will focus first on AI-enabled scientific discovery, including disease cures, and on threats such as novel bio risks, fast economic change, and emergent societal effects from highly capable models; it plans to spend at least $1 billion over the next year . The leadership slate includes co-founder Wojciech Zaremba as Head of AI Resilience, Jacob Tref as Head of Life Sciences and Curing Diseases, Anna Adeola as Head of AI for Civil Society and Philanthropy, Robert Kaiden as CFO, and Jeff Arnold as Director of Operations . More details are in OpenAI’s update: openaifoundation.org/news/update-on-the-openai-foundation.

Why it matters: This is a notable governance signal from a leading lab. Altman said no company can mitigate these risks alone, and Zaremba framed AI resilience as minimizing disruptions including impacts on children and youth, model malfunctions, and emergent bio-risks .

The Sora app shutdown turned OpenAI’s refocus into an official product cut

The Sora team said it is “saying goodbye” to the Sora app and will share timelines for the app and API, plus details on preserving users’ work . Separately, Matt Wolfe summarized Wall Street Journal reporting that OpenAI is winding down products that use Sora video models, redirecting compute and top talent toward productivity tools, and consolidating ChatGPT, the desktop app, coding tools, and browser work into one super app .

Why it matters: The official app shutdown makes OpenAI’s broader refocus tangible, and the reporting ties that shift directly to compute constraints and a narrower emphasis on coding and business users .

The infrastructure layer got more open

Modular and NVIDIA both moved key AI infrastructure into the open

Chris Lattner said Modular is open-sourcing not just its models but also its GPU kernels, making them run on multivendor consumer hardware and explicitly opening the door to anyone who can improve on the work . At the cluster layer, NVIDIA said it is donating its Dynamic Resource Allocation Driver for GPUs to the CNCF/Kubernetes community, moving it from vendor-governed software to full community ownership .

NVIDIA said the driver supports smarter GPU sharing through Multi-Process Service and Multi-Instance GPU, native multi-node NVLink connectivity, and GPU support for Kata Containers to enable more protected AI workloads through confidential computing .

Why it matters: Taken together, these moves push openness both down the stack and up the stack: lower-level kernels become more portable across hardware, while GPU orchestration inside Kubernetes becomes more community-governed .

One of the clearest AI-for-science wins came from materials

AI-designed polymers held up in the lab, but the field still lacks an “AlphaFold moment”

In a Latent Space interview, MIT professor Heather Kulik described using AI to screen tens of thousands of polymer networks, uncover an unexpected quantum mechanical effect, and produce a material that proved about four times tougher when synthesized in the lab . She also argued that materials science still lacks the data foundations that made biology more tractable: much of the field relies on noisy DFT approximations, accurate datasets tend to cover “boring” chemistry, and each element introduces new interactions with far less transferability than biology’s twenty amino acids .

That gap still shows up in model behavior. Kulik said LLMs can help with Wikipedia-level chemistry but still fail precise design tasks like producing a 22-atom ligand, and Latent Space noted models got kinase-inhibitor cases right while missing the 22-atom target for MOF ligands .

Why it matters: This is a useful two-part signal: AI can already generate lab-valid materials ideas humans did not propose, but the broader “AlphaFold for materials” story remains constrained by data quality, chemistry diversity, and weak precision on exact design tasks .

More to watch

  • Sakana AI launched Sakana Chat, its first public-facing service: a free AI chat product with web search for anyone in Japan. The company says its post-training is designed to remove developer biases, reflect Japanese values, and adapt safely to context; it is available at chat.sakana.ai.
  • Perplexity launched Comet, which Aravind Srinivas described as an autonomous “Internet Computer” for browser tasks. In the demo, it opened five tabs, ran five image-generation jobs in parallel, downloaded and cropped outputs, and assembled a comparison deck .
  • Anthropic’s March Economic Index said longer-term Claude users iterate more carefully, grant less autonomy, attempt higher-value tasks, and get more successful responses; since November 2025, the top 10 consumer tasks have also become a smaller share of overall usage .
Sora Shuts Down, LiteLLM Is Compromised, and Siri Gets an AI Agent Reboot
Mar 25
7 min read
740 docs
vLLM
Daniel Hnyk
Perplexity
+38
OpenAI is shutting down Sora while preparing its next model, LiteLLM’s compromise exposed a major supply-chain risk in AI tooling, and a new report says Apple is rebuilding Siri into a system-wide AI agent. The brief also covers key research advances, product launches, corporate moves, and safety-related updates across the AI landscape.

Top Stories

Why it matters: This cycle combined a major OpenAI product retreat, a supply-chain security shock, a fresh consumer-AI platform wager from Apple, and one of the clearest public disclosures yet on how a frontier coding model was trained.

1) OpenAI is winding down Sora as Spud nears

Reporting shared on X said OpenAI has finished pretraining or initial development of a new model codenamed Spud and is winding down Sora’s app, API, and video capabilities in ChatGPT. The same reporting said Sam Altman is dropping oversight of some direct reports and focusing on raising capital, supply chains, and datacenter buildout at unprecedented scale.

“We’re saying goodbye to Sora. To everyone who created with Sora, shared it, and built community around it: thank you. What you made with Sora mattered, and we know this news is disappointing.”

“We’ll share more soon, including timelines for the app and API and details on preserving your work.”

A post quoting the report said Sora had become a drag on computing resources during heightened competition.

Impact: The reporting points to a shift of compute and leadership attention toward the next large model and infrastructure buildout rather than a standalone video product.

2) The LiteLLM compromise turned AI infrastructure into the day’s security story

Researchers said PyPI release 1.82.8 of LiteLLM contained litellm_init.pth with base64-encoded instructions to exfiltrate SSH keys, cloud credentials, git credentials, API keys, shell history, crypto wallets, SSL keys, CI/CD secrets, and database passwords, then self-replicate. Karpathy added that LiteLLM sees about 97 million downloads per month and that dependents such as dspy were also exposed through transitive installs. The poisoned release appears to have been live for less than an hour before a RAM crash in a Cursor MCP plugin helped uncover it.

“Supply chain attacks like this are basically the scariest thing imaginable in modern software.”

The incident also spilled into the agent ecosystem: Hermes users who installed recently were told to review a security notice, and Hermes installs were blocked when litellm was quarantined on PyPI.

Impact: This was not just one bad package version. It showed how reused AI-agent infrastructure can turn a single compromised dependency into a much broader credential-exposure problem.

3) A new report says Apple is turning Siri into a system-wide AI agent

A Bloomberg report shared by Mark Gurman says iOS 27 will rebuild Siri into a system-wide AI agent. Reported features include a standalone Siri app with chat history and file uploads, text-and-voice interaction, an Ask Siri button for contextual actions across apps, unified Siri-and-Spotlight search, and Write with Siri editing tools. A separate summary of the report said many advanced features will continue rolling out into late 2026.

That same summary said the system will be powered by Apple Foundation Models plus a Google Gemini partnership.

Impact: If the report holds, Apple is moving from assistant-style AI features toward deeper system control, but on a staggered timeline.

4) Cursor published a rare training report for a frontier coding model

Cursor released a technical report on how Composer 2 was trained, saying the model reached frontier-level coding through extensive research and that the report shares details meant to be useful to the community. Commentary on the report highlighted continual pretraining improving RL performance, a multi-token prediction head for speculative decoding, length-penalty RL for long tasks, self-summarization for context compaction, and detailed sections on kernels, parallelism, quantization, and distributed RL.

Impact: The value here is the level of disclosure: the report gives builders concrete training and infrastructure choices, not just benchmark claims.

Research & Innovation

Why it matters: Technical progress this cycle focused less on one giant model launch and more on the systems around models: memory, serving, evaluation, and retrieval.

  • TurboQuant: Google Research introduced TurboQuant, a compression algorithm that reduces LLM key-value cache memory by at least 6x and can deliver up to 8x speedup with zero accuracy loss.
  • APEX-SWE: Mercor and Cognition launched a benchmark for realistic software-engineering work such as shipping systems and debugging failures, arguing that traditional coding benchmarks do not reflect how software is actually built and maintained. On the initial leaderboard, OpenAI GPT 5.3 Codex (High) led at 41.5% Pass@1.
  • vLLM Model Runner V2: vLLM rebuilt its execution core into Model Runner V2 with modular design, GPU-native input preparation, async-first execution with zero CPU–GPU sync, and a Triton-native sampler. Separate GTC notes said the project is also reducing memory waste to 0–12% across OSS models and improving multimodal P99 throughput by up to 2.5x through encoder prefill disaggregation.
  • Late-interaction retrieval: A 150M Reason-ModernColBERT model reached nearly 90% on BrowseComp-Plus and beat models up to 54× larger, while Mixedbread Search was reported to approach oracle-level performance on knowledge-intensive agentic benchmarks.

Products & Launches

Why it matters: New releases kept pushing agents deeper into everyday workflows—permissions, browsers, filesystems, APIs, and open browser-use models.

  • Claude Code auto mode: Anthropic added an auto mode that lets Claude make permission decisions for file writes and bash commands on the user’s behalf, with safeguards checking each action before it runs.
  • Perplexity Computer and Comet: Perplexity said its Computer product uses Comet to kick off workflows in a local browser. Arav Srinivas described Comet as an autonomous Internet Computer, and the demo showed it opening five tabs, running parallel image-generation tasks, downloading and cropping outputs, and assembling a comparison deck.
  • Hermes Agent v0.4.0: NousResearch’s largest Hermes update this week merged 300 PRs and added a background self-improvement loop, an OpenAI-compatible API backend, and major CLI upgrades.
  • hf-mount: Hugging Face introduced hf-mount, which can attach a storage bucket, model, or dataset from the Hub as a local filesystem. The project says it can expose remote storage 100× larger than a local disk and is well suited to agentic storage workflows.
  • MolmoWeb: AI2 released MolmoWeb 4B and 8B browser-use models and their datasets under Apache 2.0.

Industry Moves

Why it matters: Labs and platform companies kept reallocating capital, talent, and partnerships toward agents, AI-native software, robotics, and new interfaces.

  • Hark emerged from stealth: Brett Adcock said Hark spent eight months in stealth building the most advanced personal intelligence in the world, paired with next-generation hardware as a human-machine interface. Separate reporting said Adcock put in $100M of his own money, assembled a 45+ person team from Apple, Tesla, Google, Meta, and Amazon, expects thousands of NVIDIA B200 GPUs online by April, and plans a first model this summer.
  • Microsoft added senior AI2 talent: Mustafa Suleyman welcomed Ali Farhadi, Hanna Hajishirzi, and Ranjay Krishna to Microsoft Superintelligence, describing them as impactful contributors to AI research and open source.
  • Google DeepMind partnered with Agile Robots: DeepMind said a new research partnership will integrate Gemini foundation models with Agile Robots hardware to build more helpful and useful robots.
  • Meta’s internal AI push shifted upward: Reporting on X said CTO Andrew Bosworth is taking over supervision of Meta’s effort to become AI-native, including the company’s AI For Work initiative.

Policy & Regulation

Why it matters: This cycle’s policy signal came less from governments and more from safety, access, and institutional compliance moves around powerful models.

  • OpenAI Foundation: OpenAI said the Foundation will spend at least $1 billion over the next year, initially focusing on areas such as disease cures, AI resilience, civil society, philanthropy, and threats including novel bio risks, fast economic change, and complex emergent effects from capable models. Wojciech Zaremba is moving to lead AI resilience.
  • Teen safety policies for developers: OpenAI Devs released prompt-based teen safety policies for gpt-oss-safeguard, designed to help developers identify and moderate teen-specific content and turn policy requirements into classifiers for real-time filtering or offline analysis.
  • NeurIPS sanctions rule: A post citing a NeurIPS Foundation announcement said the conference will no longer accept submissions from US-sanctioned institutions.

Quick Takes

Why it matters: These updates were smaller, but they help map where agent design, model usage, and deployment practices are going next.

  • Google’s Gemini API now supports combining Google Search and custom functions in a single request, with Gemini choosing tools and order automatically.
  • Gemini 3.1 Flash-Lite is being shown generating websites in real time as users click, search, and navigate.
  • Anthropic’s March Economic Index said longer-term Claude users iterate more carefully, hand over less autonomy, attempt higher-value tasks, and get more successful responses; the top 10 consumer tasks now account for 19% of conversations, down from 24% since November 2025.
  • Similarweb said Claude has overtaken DeepSeek, Grok, and Gemini to become the second most-used gen-AI app daily after ChatGPT.
  • Perplexity said its search embedding models crossed 1 million downloads in less than a month.
  • AssemblyAI said better speech models exposed flaws in human truth files and released tooling for corrected truth-file workflows, semantic word lists, and production-ready benchmarking.
  • Alibaba released the open-weight Qwen3.5 vision-language family, with smaller models such as Qwen3.5-9B said to rival or beat much larger competitors.
South Star Metrics, AI Standards, and Mixed Signals in the PM Market
Mar 25
8 min read
96 docs
Product Management
Lenny's Reads
Sachin Rekhi
+5
This issue focuses on a practical framework for spotting bad metrics, the rising bar for AI-enabled product managers, and a PM job market that looks stronger in aggregate but remains difficult in practice. It also pulls lessons from Anthropic’s release velocity, post-merger B2B chaos, and fintech’s shift toward realized user value.

Big Ideas

1) South Star Metrics gives teams a better way to debug green-dashboard/bad-product situations

The framework breaks north-star failure into seven recurring types: detrimental, out-of-reach, incomplete, pressure, inconsequential, nonsensical, and incongruent metrics .

Why it matters: It gives PMs language for cases where a metric improves while customer value, team control, or strategic coherence gets worse. The article’s test for a healthy metric is straightforward: it should have a long enough time horizon, connect customer and business value, account for the full journey, and survive a stress test .

How to apply: Add a monthly scorecard review that asks which south-star pattern, if any, is showing up before you celebrate a win.

2) The new PM bar is AI fluency plus high standards

Strong PMs are expected to know how to use AI across strategy, customer research, data analysis, prototyping, validation, and daily execution, while also knowing each tool’s strengths, limits, and best use cases . But AI fluency alone is not enough: PMs also need product taste and the discipline to refine or abandon AI-assisted work when the output is weak .

Why it matters: The failure modes are already visible. Some PMs use AI for everything and generate low-quality output; others keep their standards but fail to meet the new pace and burn out .

"Learning both becomes critical in this next era of product management."

How to apply: For each part of your workflow, define two things explicitly: where AI should accelerate the work, and what good still has to look like before it ships.

3) In AI markets, shipping velocity is becoming strategy

One analysis of Claude releases counted 74 launches in 52 days across four parallel surfaces: developer tools, desktop automation, API/infrastructure, and models/platform .

Why it matters: The takeaway is not just that one company shipped a lot. It is that PMs can miss the real competitive signal if they compare point features instead of the rate of improvement across teams. The post argues that this creates a compounding gap for users who start building systems on top of fast-moving surfaces early .

How to apply: Track competitors with a shipping calendar, not a quarterly memory. Include dates, features, and which team shipped what so you can see operating model changes early .

Tactical Playbook

1) Run a 5-step metric stress test before locking a north star

  1. Pair the main metric with a customer-experience guardrail and a single view of qualitative signals, not just the quantitative dashboard .
  2. Build a metrics ladder so the team owns a lever it can directly move, then prove the causal link to the business outcome .
  3. Map the end-to-end journey and use the rule own one, watch all, ideally through one shared dashboard and one funnel-health owner .
  4. Protect future capacity with an explicit split such as 70/20/10 so urgent quarterly work does not crowd out discovery, technical investment, or exploration .
  5. Add sanity constraints and surface cross-team tensions on one page so optimization does not create absurd outcomes or metric conflict .

Why it matters: This sequence directly addresses the most common failure modes in the South Star taxonomy, from customer harm to team misalignment . One especially useful warning: silence is not always a good sign; customers may have simply given up and moved on .

2) If the org feels chaotic, document the strategy that is currently living in people’s heads

A community response to a post-merger B2B SaaS situation with no clear strategy, roadmap, or role boundaries recommends a simple recovery pattern :

  1. Draft a 1-2 page product and GTM strategy from what you have observed and researched .
  2. Review it with stakeholders using a calibration question: Where does this line up, and where am I off?
  3. Write a one-page role definition for PMM responsibilities and priorities to reduce PM/PMM overlap .
  4. Narrow the ICP and tie work to real outcomes instead of shipping noise for its own sake .
  5. Research competitors and customer pain points, then build small proof points that create traction even in a noisy environment .

Why it matters: The alternative described in the thread is familiar: fragmented ICP, low retention, antiquated software, and a feature cadence with little customer impact .

3) Use builder feeds for competitive intelligence

Instead of waiting for polished changelogs, one PM tactic is to map a competitor’s releases by following the feeds of the people actually shipping, then logging dates, features, and team attribution in a calendar .

Why it matters: It helps you spot whether a competitor is shipping in parallel across multiple surfaces and whether teams are blocked by interdependencies .

How to apply: Make the calendar a standing artifact in quarterly planning. Look for repeated themes, acceleration by surface, and which capabilities appear to be compounding.

Case Studies & Lessons

1) Anthropic’s Claude release cadence is a lesson in parallel execution

The 74 releases in 52 days were split across developer tools (28), desktop automation (15), API/infrastructure (18), and models/platform (13), with the explicit observation that teams were shipping in parallel rather than waiting on one another .

Lesson: Competitive advantage can come from the operating system behind the roadmap, not just the roadmap itself.

Apply it: When you benchmark competitors, assess how many product surfaces are improving at once and whether your own org design is creating unnecessary dependencies.

2) Fintech users are shifting from theoretical upside to realized value

A discussion in r/ProductManagement describes a move away from flashy reward percentages with complex hurdles toward products that deliver accessible value, seamless cash-out, and reward structures aligned with how users already spend .

Lesson: If the effort to unlock value exceeds the value actually realized, the product is creating UX friction, not delight .

Apply it: In pricing, rewards, and activation design, test the realized user-value path—not just the headline number shown in marketing.

3) A metric win can still be a product loss

The South Star framework uses examples like forced Windows Update reboots or rising ad load: adoption or impressions go up, but the customer experience gets worse .

Lesson: A green metric is not enough if support complaints rise, renewals get harder, or time-to-value deteriorates .

Apply it: Treat qualitative signals and renewal health as first-class evidence before scaling a metric-winning change .

Career Corner

1) The PM market is improving on openings, but the search is still hard

TrueUp-based analysis tracking more than 9,000 tech companies says there are over 7,300 open PM roles at tech companies globally, up 75% from the early 2023 low and nearly 20% year to date—the highest level since 2022 . The same analysis argues the broader data is telling a growth story despite 184 tech layoffs affecting 57,606 people so far in 2026, with big-tech headcount flat or slightly up and hiring outpacing layoffs overall .

The composition of demand is shifting. PM demand is now 1.27x design roles after flipping in mid-2023 , AI PM openings stand at 1,135 and are up 465% from the low , more than 23% of open PM roles are in the Bay Area , and the top PM job locations called out were the SF Bay Area (1,442), Remote US (864), and NYC (673) . At the same time, remote-optional PM roles have fallen to 25% from a 35% peak .

There is still a real disconnect between opening counts and lived experience. Lenny notes that more openings do not automatically mean faster hiring , while Reddit commenters point to applicant oversupply, ghost jobs, internal transfers, and roles that look open but are not truly available to external candidates .

How to apply: Treat the market as better than 2023 on demand signals, but tighten your search around location flexibility, AI-adjacent roles, and whether a posting is actually budgeted and open to external hiring.

2) The skill stack to invest in is clear: AI fluency plus standards

Rekhi’s framing is useful for career planning: knowing AI tools and where they fit in the workflow is valuable, but it only compounds when paired with strong product taste and a willingness to reject weak output .

How to apply: Build evidence of both. Show how you use AI to speed up research, analysis, or prototyping, and show where you improved, rewrote, or discarded low-quality AI output rather than shipping it.

3) For early-career PMs, structured entry points still matter

One hiring insight from the Reddit discussion is that hiring college grads straight into PM roles is generally a bad idea unless the setup is closer to an APM or rotational program with oversight .

Why it matters: The missing piece is usually not raw intelligence. It is the informal influence, coaching, and exposure needed to operate effectively as a PM .

How to apply: If you are early in your career, optimize for mentorship and role structure, not just the PM title.

Tools & Resources

Fertilizer Export Halt, Brazil Harvest Delays, and U.S. Acreage Signals Reset the Ag Outlook
Mar 25
9 min read
187 docs
Tarım Editörü
农业致富经 Agriculture And Farming
Successful Farming
+9
Russia’s ammonium nitrate export halt, rising diesel and fertilizer costs, and conflicting U.S. acreage signals are reshaping grain and livestock planning. This brief also highlights proven corn, poultry, forage, and circular-production practices with quantified yield and cost outcomes.

Market Movers

  • Global grain board: March 24 trade opened with May corn at $4.61 3/4, May soybeans at $11.60 1/4, and May Chicago wheat at $5.90 1/4. Later market commentary still described the day as corn and wheat higher, soybeans lower, with crude oil and broader energy-market risk feeding into corn and especially wheat.
  • New-crop corn: December corn at $4.89 was flagged as above every price seen in 2025 and close to the 2024 high, effectively near two-year highs.
  • Corn positioning: Market talk continues to center on lower corn acreage and fund buying. One analyst put major resistance near $5.00, with $5.12-$5.17 as an upside objective, while another said funds have added length and are waiting for fresh demand news such as EPA biofuel quotas.
  • Soybeans - U.S. vs. Brazil/China: U.S. soybeans were described as $35-$45/ton more expensive than Brazil, with old-crop demand pressured by limited China business and China accepting lower-quality Brazilian beans. One analyst said U.S. carryover could rise another 100-120 million bushels if that business does not return. Key support levels cited were $11.40, $10.40, and $9.40.
  • Wheat - U.S. Plains: Kansas winter wheat was rated 46% good/excellent and falling, while forecasts keep much of HRW country dry for another 10 days before later rains. Separate market commentary said it may be difficult to get above a 50 bu/acre U.S. winter wheat national average, helping preserve weather premium.
  • Trade flow: U.S. export inspections for the week ending March 19 were 67 million bushels of corn, 40 million bushels of soybeans, and 17 million bushels of wheat. Soybean inspections were up 12% week over week and 32% year over year, with China taking 60% of the weekly soybean total. Exporters also reported flash sales to Mexico of 4 million bushels of corn and 6 million bushels of soybeans for the current marketing year.
  • Livestock - U.S.: Live and feeder cattle moved higher on steady-to-better cash trade and stronger feeder demand, with Nebraska drought and wildfires cited as support. Feeder cattle traders were watching $381 as a resistance level. Hogs remain softer in deferred months, though strong pork-sector earnings and the coming hogs-and-pigs report remain key near-term variables.

Innovation Spotlight

Brazil: corn management with measurable yield protection

At the second Rally de Produtividade in Uberlândia, about 400 producers reviewed high-performance corn hybrids and management systems capable of exceeding 300 sacks/ha. One producer said storm and hail damage initially implied 30-50% losses across four pivots, yet the farm still finished around 205-206 sacks/ha, with 231 sacks/ha in a less-affected area and 331 sacks/ha on a 0.2-hectare test plot. Speakers emphasized that fertilizer accounts for about 40% of production cost, making field-specific analysis, genetics, and crop protection central to ROI.

The same event highlighted pest pressure as a direct margin issue: cigarrinhas, molicute viroses, and aphids were said to cost 40-50 sacks/ha, while also increasing lodging and poor ear formation risk.

China: circular shiitake production with cost and profit gains

A shiitake system using pruned tea branches instead of purchased wood chips cut substrate cost by about 20,000 yuan per mu. The tea material was also described as reducing disease pressure, while mushrooms were reported to be larger at 2-3 liang each. Because the substrate stayed healthier longer, growers were able to take an extra flush.

The spent substrate was then fermented into organic fertilizer and returned to tea gardens, creating a closed nutrient loop. The source said this model lifted profit by roughly 30,000 yuan per mu.

Brazil: digital management and land-use tools

Embrapa's Custo Fácil 4.0 was presented as a decision-support tool for small and medium integrated poultry and swine producers, estimating production costs, profitability, and cash flow.

Separately, Embrapa said an AI-based monitoring tool identified abandoned agricultural land in the Cerrado with 95% accuracy. In Buritizeiro, Minas Gerais, it mapped more than 13,000 hectares, or about 5% of the monitored area, for possible restoration, carbon sequestration, or reintegration into production.

Regional Developments

  • Brazil - soy and safrinha progress: By March 19, Brazil's winter corn planting had reached 94.6% of expected area and soybean harvest was at 68.8%. Separate Conab-based commentary still put the soybean harvest about 10% behind the same period last year and the five-year average. Mato Grosso had only 2% left to harvest and strong productivity in final areas, while Mato Grosso do Sul was 87% harvested and ahead of last year. The biggest delays were in São Paulo, Maranhão, Minas Gerais, and Bahia.
  • Brazil - rain risk and harvest quality: Frequent rains in Tocantins are delaying harvest and hurting grain quality. Producers in Gurupi/west Tocantins were urged to use the late-March window for soybean harvest and second-crop corn planting before early-April rainfall could total about 150 mm. Parts of the center-south, including Goiás, should have workable short windows, while southern areas still face storm risks including hail and strong wind.
  • United States - acreage signals remain split: Allendale's March survey across about 25 states covering roughly 85% of U.S. corn and soybean production found 98.8 million corn acres, 81.2 million soybean acres, and 45.3 million wheat acres. That would put corn production roughly 60-80 million bushels below USDA's February implication. The same survey warned that acreage can still shift because more than 80% of rice and cotton ground is in drought, and even a 10% change in those smaller crops could move 1-2 million acres.
  • Brazil-China beef trade: Official Chinese quota data showed Brazil had already used 33.6% of its annual beef quota in January-February, versus 35% for Australia and 20% for Argentina. ABIEC said the pace merits closer monitoring. Separate Canal Rural commentary said the fill rate could approach 50% by end-March and, if the quota is exhausted within 3-4 months, later shipments would face a 50% tariff, raising the risk of more domestic beef supply in Brazil.

Best Practices

  • Wheat nutrient timing - United States: Growers are using timing, field variability, and crop stage to guide in-season nitrogen and nutrient decisions aimed at improving yield, protein, and ROI.
  • Soy harvest and safrinha timing - Brazil: In delayed soybean areas, producers were advised to use the next 5-day windows where rainfall totals about 20-25 mm to keep harvest moving without materially disrupting fieldwork. In Uberaba, one forecast showed rain on Wednesday followed by a dry stretch through April 1; in Gurupi, the useful window runs only through late March before heavier April rain.
  • Corn pest monitoring - Brazil: Early monitoring for cigarrinhas, molicute viroses, and aphids is critical. Reported losses were 40-50 sacks/ha, with added risk of lodging and malformed ears when intervention comes too late.
  • Forage fertilization - United Kingdom dairy: When using urea with sulfur on grass, the application should be timed with rain expected within a few days. One producer is targeting 80 units of nitrogen per acre, about 100 kg/acre, using 20-meter overlap to improve spread uniformity. Because bridging and dust buildup were a problem, nightly washouts were recommended; soft or newly seeded ground was being avoided to prevent damage.
  • Poultry and swine biosecurity - Brazil: Maintain at least 12 days of sanitary void after cleaning and treatment, within a typical 20-21 day inter-flock interval. Litter treatment was described as most effective when temperatures exceed 60°C under cover. Cleaning must include both the inside and outside of the house, plus active control of rats and cascudinho, because these pests can reintroduce contamination. Some export markets require zero Salmonella.
  • Soil and circular fertility - China: Fermenting spent mushroom substrate into organic fertilizer and returning it to tea gardens was presented as a practical residue-reuse model rather than discarding the material after harvest.

Input Markets

  • Fertilizers - Russia to Brazil: Russia has temporarily suspended ammonium nitrate exports until April 21. Canal Rural said Russia accounted for nearly 26% of Brazil's chemical fertilizer acquisitions in 2025, while Brazil still depends on imports for about 85% of its fertilizer needs. Fertilizer prices were already cited as up 30-35%.
  • Brazil fertilizer import economics: In the current January-March flow, Brazil was near 7 million tons of imports versus almost 7.9 million tons a year earlier, but producer outlays were already $2.4 billion versus $2.0 billion on the larger prior-year volume. Average cost was put at $350/ton so far, versus $309/ton in 2024 and $343/ton last year. March itself was running at $382/ton versus $311/ton a year earlier, while average daily imports were 118 versus 137 a year earlier.
  • U.S. diesel and fertilizer: Farm Journal linked the Iran/Hormuz disruption to on-farm diesel prices near the records set in March 2022, with diesel more than $1 above last year. The same coverage put urea up 25-30%, other nitrogen and phosphorus products up 5-10%, and potash still firming. Most fertilizer for the 2026 crop was described as already applied or pre-booked, limiting immediate acreage cuts but leaving 2027 more exposed if the disruption persists. Successful Farming separately said U.S. diesel prices are up 42% year over year.
  • Brazil diesel: Brazilian field reports put diesel at more than R$8/liter and up to R$9/liter in some regions, versus about R$5.75/liter before the conflict, with shortages reported in 140 municipalities in Rio Grande do Sul and in parts of Santa Catarina, Mato Grosso, Mato Grosso do Sul, Paraná, Goiás, and São Paulo. This directly affects soybean harvest, grain haulage, and second-crop planting. One report said the federal response included a R$0.32/liter incentive for diesel producers and importers capped at R$10 billion, plus PIS/COFINS relief; another discussion mentioned a possible R$14 billion support package with unclear agro inclusion.

Forward Outlook

  • Biofuel policy - United States: EPA is expected to release new 2026/27 biofuel quotas this week or by month-end, and Congress is still considering year-round E15 sales. That keeps renewable fuel policy on the immediate demand watchlist for corn and soybeans.
  • USDA acreage uncertainty: Allendale's survey points to a more soybean-heavy acreage mix than USDA's February outlook, but the same discussion stressed that intended, planted, and harvested acres can diverge sharply. Last year, official corn acreage eventually surprised by about 8 million acres, and drought on rice and cotton ground could still move 1-2 million acres across crops.
  • Brazil weather calendar: Soy harvest is still running roughly 10% behind last year and the five-year average in many areas, but short fieldwork windows remain open in parts of center-south Brazil through the end of March. West Tocantins/Gurupi has the narrowest window before heavier April rain, while Rio Grande do Sul has about 8-10 days of firmer weather before rain returns in early April.
  • Input pressure into the next crop: Canal Rural commentary said Brazil's fertilizer imports could show zero growth or even decline if current geopolitical tension and cost pressure persist. Producers are simultaneously dealing with high interest rates, selective credit, and indebtedness ahead of the next Plano Safra cycle.
  • China-linked demand watch: Old-crop U.S. soybean business with China is still viewed as unlikely, while China's hog sector is facing 15-year-low pig prices and four-year-low margins. Brazil's beef quota usage in China also needs close tracking because a faster fill would increase the risk of domestic price pressure later in the year.

Your time, back.

An AI curator that monitors the web nonstop, lets you control every source and setting, and delivers one verified daily brief.

Save hours

AI monitors connected sources 24/7—YouTube, X, Substack, Reddit, RSS, people's appearances and more—condensing everything into one daily brief.

Full control over the agent

Add/remove sources. Set your agent's focus and style. Auto-embed clips from full episodes and videos. Control exactly how briefs are built.

Verify every claim

Citations link to the original source and the exact span.

Discover sources on autopilot

Your agent discovers relevant channels and profiles based on your goals. You get to decide what to keep.

Multi-media sources

Track YouTube channels, Podcasts, X accounts, Substack, Reddit, and Blogs. Plus, follow people across platforms to catch their appearances.

Private or Public

Create private agents for yourself, publish public ones, and subscribe to agents from others.

Get your briefs in 3 steps

1

Describe your goal

Tell your AI agent what you want to track using natural language. Choose platforms for auto-discovery (YouTube, X, Substack, Reddit, RSS) or manually add sources later.

Stay updated on space exploration and electric vehicle innovations
Daily newsletter on AI news and research
Track startup funding trends and venture capital insights
Latest research on longevity, health optimization, and wellness breakthroughs
Auto-discover sources

2

Confirm your sources and launch

Your agent finds relevant channels and profiles based on your instructions. Review suggestions, keep what fits, remove what doesn't, add your own. Launch when ready—you can always adjust sources anytime.

Discovering relevant sources...
Sam Altman Profile

Sam Altman

Profile
3Blue1Brown Avatar

3Blue1Brown

Channel
Paul Graham Avatar

Paul Graham

Account
Example Substack Avatar

The Pragmatic Engineer

Newsletter · Gergely Orosz
Reddit Machine Learning

r/MachineLearning

Community
Naval Ravikant Profile

Naval Ravikant

Profile ·
Example X List

AI High Signal

List
Example RSS Feed

Stratechery

RSS · Ben Thompson
Sam Altman Profile

Sam Altman

Profile
3Blue1Brown Avatar

3Blue1Brown

Channel
Paul Graham Avatar

Paul Graham

Account
Example Substack Avatar

The Pragmatic Engineer

Newsletter · Gergely Orosz
Reddit Machine Learning

r/MachineLearning

Community
Naval Ravikant Profile

Naval Ravikant

Profile ·
Example X List

AI High Signal

List
Example RSS Feed

Stratechery

RSS · Ben Thompson

3

Receive verified daily briefs

Get concise, daily updates with precise citations directly in your inbox. You control the focus, style, and length.

Cloudflare Pushes Code-Writing Agents as Claude Code and Cursor Tighten the Loop
Mar 25
6 min read
158 docs
Claude
Figma
Riley Brown
+12
Cloudflare's Dynamic Workers and Code Mode signal a more important shift than another model release: coding agents are moving from brittle tool menus toward fast, disposable code execution. Also inside: Claude Code auto mode, Cursor's Composer 2 report, Riley Brown's Codex/Xcode workflow, and production multi-agent patterns worth stealing.

🔥 TOP SIGNAL

The sharpest infra shift today is Cloudflare's push toward code-writing agents. Dynamic Workers are now in Open Beta with secure sandboxes that start ~100x faster than containers and use 1/10 the memory, and Kent C. Dodds says he's already using them to power his own AI assistant with CodeMode . The companion Code Mode thesis matters just as much: expose tools as an SDK and let the model write code to select and compose them, instead of stuffing more JSON-style tools into the context window .

"Agents should interact with the world by writing code, not tool calls."

🛠️ TOOLS & MODELS

  • Claude Code — auto mode. Claude can now decide file-write and bash permissions on the user's behalf, with safeguards checking each action before it runs. That replaces the old choice between constant approvals and --dangerously-skip-permissions.
  • Cursor Composer 2 — training report out. Cursor says Composer 2 was built with three main efforts: continued pretraining, reinforcement learning, and benchmark development to emulate the Cursor environment. The practical takeaways: continued pretraining kept improving downstream coding performance, the RL phase was critical enough that simple approaches worked best broadly, and CursorBench is their internal benchmark for more realistic software engineering problems. The report also covers distributed training / RL infrastructure and kernels they open-sourced. Read: Composer 2 report.
  • Cloudflare Dynamic Workers + Code Mode. Dynamic Workers entered Open Beta for paid Workers users, with disposable sandboxes that start ~100x faster than containers and use 1/10 the memory. The adjacent Code Mode posts surfaced three agent-facing building blocks: codemode for chewing through multiple tool calls in one/few shots, shell for a local filesystem backed by DO/D1/SQL/R2, and worker-bundler for packaging npm deps and serving full apps. Read: Dynamic Workers and Code Mode.
  • Cursor + Figma. Cursor can now create new components and frontends in Figma using your team's design system, including variables, tokens, and naming conventions. This landed alongside Figma's open beta for AI agents designing directly on canvas via the use_figma MCP tool .
  • Devin as reviewer, not just author. The clearest practitioner comparison today came from Gauri Gupta and swyx: both say Devin code review catches bugs Claude and Codex miss, and swyx says the review pass saves him 3-8x/day .

💡 WORKFLOWS & TRICKS

  • Replicable mobile-agent build loop from Riley Brown

    1. Create the app shell in Xcode, then open the top-level project folder in Codex and start a dedicated thread .
    2. Paste the full product spec and hit Shift+Tab for Plan mode .
    3. Run design in parallel in Paper.design via MCP while Codex builds .
    4. Steer inline while the thread is running, then paste API keys, Xcode logs, and screenshots back into the chat when builds fail .
    5. Test on a physical iPhone, and keep the generated app preview open so you can keep iterating without leaving the app .
    • Bonus: Brown adds Whisper-based voice input so he can keep refining the preview without opening the keyboard .
    • Timeless pattern: split planning, design, implementation, and debugging across parallel agent threads instead of forcing one giant chat .
  • Turn prompts into loops, not one-shots

    • Karpathy's Auto Research runs on a git feature branch: propose experiment, write code, run it, analyze results, commit improvements, repeat .
    • Matthew Berman says he applies the same shape to overnight fine-tuning of small open models against an Opus 4.6 baseline until a local model wins .
    • Minimax's M2.7 harness adds skills, memory, guardrails, and eval infra; it handled 30-50% of the workflow and reported a 30% performance gain on evals .
    • Anthropic says similar Claude loops now power almost all of its major agent loops, with humans reviewing solutions before final refinements .
  • Production pattern worth stealing from Moda

    • Use a cheap Haiku triage node to classify the job and preload task-specific Markdown skills .
    • Inject those skills as human messages, with caching breakpoints after the system prompt and after the skills block .
    • Keep only 12-15 core tools always in context; activate ~30 more on demand with RequestToolActivation.
    • For big projects, pass a high-level summary and let the agent pull details through tools; then use traces to tune prompt/tool changes, cost, cache hits, and failures .
  • Add a critic, not just another worker. swyx says he won't merge without a Devin review now, because it catches bugs Claude and Codex miss and acts like fresh eyes on a PR. Related design note: he points to Windsurf's "smart friend" pattern as a model for subagents that are more critical instead of weaker .

👤 PEOPLE TO WATCH

  • Kent C. Dodds — strongest practitioner confirmation on today's Cloudflare story: he's been using Dynamic Workers for a week to power his own AI assistant, and separately backs the "tools as SDK + agent-written code" pattern as the more effective direction .
  • Riley Brown — one of the most concrete end-to-end Codex workflows in the wild right now: Xcode + Codex + Claude Agent SDK + Vibe Code CLI + Paper.design, with screenshots and logs fed back into the loop .
  • swyx — short posts, strong signal: Devin review outperforming Claude/Codex in his day-to-day, plus the "smart friend" subagent pattern callout .
  • Ravi Parikh / Moda — rare production detail instead of benchmark chatter: triage, dynamic tool loading, prompt caching, context scaling, and tracing from a real multi-agent design product .
  • Anthropic Engineering — new post on a multi-agent harness for frontend design and long-running autonomous software engineering. Worth reading because it's from the team running the stack, not a third-party teardown: harness-design-long-running-apps.

🎬 WATCH & LISTEN

  • 1:56-4:28 — M2.7's harness loop. Best compact explanation today of the human/harness/agent split: human steers direction, the harness supplies skills/memory/guardrails/evals, and the agent writes, runs, analyzes, and improves experiments .
  • 3:36-5:43 — Riley Brown's Codex setup for "Jerry." Watch this if you want a replicable "agent builds the app that builds apps" flow: create the Xcode project, point Codex at the right folder, then fan out into parallel threads .
  • 25:16-26:17 — In-app preview + voice iteration. Shorter clip, but high leverage: Riley tweaks the generated app while the preview is open and adds Whisper-based voice input so the iteration loop never leaves the phone UI .

📊 PROJECTS & REPOS

  • karpathy/auto-research — GitHub repo for autonomous research loops on a git feature branch. Practical signal: Berman frames it as bringing frontier-style experiment loops to solo developers, and points to an overnight run that reached the fastest reported time to train one of these models .
  • langchain-ai/deepagents — better adoption signal than most new repos because Moda is already using Deep Agents in production for its Research Agent and Brand Kit Agent, with the Design Agent migration under evaluation. Repo: Deep Agents.

Editorial take: today's edge is loop design — faster execution sandboxes, smaller always-on tool sets, explicit review passes, and tighter human feedback beat one giant prompt.

1,000 True Fans Leads Today’s Operator, AI, and Engineering Picks
Mar 25
5 min read
191 docs
nick ercolano
martin_casado
Tim Ferriss
+5
Tim Ferriss provides the clearest high-conviction pick with Kevin Kelly’s 1,000 True Fans, while Martin Casado, Amjad Masad, Pratap Ranade, Vincent Weisser, and Ryan Hoover add useful resources on AI limits, creator leverage, verification, engineering history, and documentary storytelling.

Most compelling recommendation

Today’s clearest save is Kevin Kelly’s 1,000 True Fans. Tim Ferriss doesn’t just endorse it; he states the usable lesson: precise targeting makes a business smarter, faster, more agile, and less expensive .

  • Title:1,000 True Fans
    Content type: Essay / online article
    Author/creator: Kevin Kelly
    Link/URL: Not provided in the source material; Ferriss says it is free online
    Who recommended it: Tim Ferriss
    Key takeaway: Be very specific with your targeting
    Why it matters: Ferriss calls it "the most important marketing piece you may ever read," making it the highest-conviction recommendation in today’s set

"the most important marketing piece you may ever read is 1000 true fans"

Tim Ferriss’s operator shelf

  • Title:The 22 Immutable Laws of Marketing(original edition)
    Content type: Book
    Author/creator: Al Ries and Jack Trout
    Link/URL: Not provided in the source material
    Who recommended it: Tim Ferriss
    Key takeaway: Ferriss specifically recommends the older edition, not the internet version
    Why it matters: It is one of his core marketing recommendations, alongside 1,000 True Fans

  • Title:Secrets of Power Negotiating
    Content type: Book / audiobook
    Author/creator: Roger Dawson
    Link/URL: Not provided in the source material
    Who recommended it: Tim Ferriss
    Key takeaway: Ferriss says the audio version is especially good
    Why it matters: This is his clearest negotiation recommendation in the source material

  • Title:Built to Sell
    Content type: Book
    Author/creator: John Warrillow
    Link/URL: Not provided in the source material
    Who recommended it: Tim Ferriss
    Key takeaway: Think about selling your company from the very beginning and design it that way
    Why it matters: Ferriss presents it as a company-design lens, not just a late-stage exit book

  • Title:Topgrading(original edition)
    Content type: Book
    Author/creator: Bradford Smart
    Link/URL: Not provided in the source material
    Who recommended it: Tim Ferriss
    Key takeaway: Use the original edition; Ferriss says newer editions are mostly upsells
    Why it matters: He calls it the best book he has found on interviewing top talent

AI mental models and builder leverage

"Hence Shannon, not Kolmogorov. Lorentz, not Einstein."

  • Title:Video on AI and what’s happened over the last 6 months(title not specified in the source material)
    Content type: Video
    Author/creator: Nick Colano
    Link/URL:https://x.com/nickercolano/status/2036060436487926111
    Who recommended it: Amjad Masad
    Key takeaway: After hundreds of hours building with AI tools, Colano argues that the old barrier between creators and their ability to build products—price, speed, and vision—is basically gone
    Why it matters: Masad explicitly recommends it as a genuine, non-marketing account of how learning software changed a creator business

"the (former) barrier between creators and their ability to build products (price, speed, vision) is basically gone"

Weisser also named The Beginning of Infinity: Explanations That Transform the World and Superintelligence: Paths, Dangers, Strategies among the works shaping his thinking on intelligence and progress .

Engineering and verification deep cuts

  • Title:Cramming More Components Onto Integrated Circuits
    Content type: Research paper
    Author/creator: Gordon Moore
    Link/URL:https://hasler.ece.gatech.edu/Published_papers/Technology_overview/gordon_moore_1965_article.pdf
    Who recommended it: Pratap Ranade
    Key takeaway: Ranade points readers to the paper’s overlooked second half on analog silicon, which he says is even more fascinating today than the digital section
    Why it matters: It is a strong reminder to revisit canonical papers for the parts everyone skipped

  • Title:Asymmetry of verification and verifier’s rule
    Content type: Blog post
    Author/creator: Jason Wei
    Link/URL:https://www.jasonwei.net/blog/asymmetry-of-verification-and-verifiers-law
    Who recommended it: Pratap Ranade
    Key takeaway: Verifier’s law says that any task that is easy and fast to verify will be automated by AI
    Why it matters: Ranade uses it to explain why faster verification is the unlock for applying generative AI to EM design

  • Title:Skunk Works
    Content type: Book
    Author/creator: Ben Rich
    Link/URL:https://www.amazon.com/Skunk-Works-Personal-Memoir-Lockheed/dp/0316743003
    Who recommended it: Pratap Ranade
    Key takeaway: Ranade highlights its story of how the F-117 stealth bomber came to be through unconventional geometry
    Why it matters: It is a concrete engineering-history example of non-obvious design choices paying off

  • Title:Tuxedo Park
    Content type: Book
    Author/creator: Jennet Conant
    Link/URL:https://www.amazon.com/Tuxedo-Park-Street-Science-Changed/dp/0684872889
    Who recommended it: Pratap Ranade
    Key takeaway: Ranade recommends it for readers who want to learn more about WWII radar research at Tuxedo Park
    Why it matters: It adds historical context to the scientific lineage behind modern hardware ambition

One non-technical watch

  • Title:In Waves and War
    Content type: Documentary
    Author/creator: Not specified in the source material
    Link/URL: Not provided in the source material
    Who recommended it: Ryan Hoover
    Key takeaway: Hoover says the film is heartbreaking, cites roughly 20 veteran suicides a day, and says it showcases ibogaine’s positive benefits for three troubled Navy SEALs
    Why it matters: He says he teared up watching it, making this the strongest emotional recommendation in today’s set
LiteLLM Breach, OpenAI’s $1B Foundation, and an Opening at the AI Infrastructure Layer
Mar 25
5 min read
167 docs
Jim Fan
Daniel Hnyk
hardmaru
+14
A major supply-chain compromise put AI developer security at the center of the day. OpenAI launched a $1B foundation while shutting down the Sora app, and Modular, NVIDIA, and materials researchers pointed to deeper shifts in infrastructure and science.

Security wake-up call

LiteLLM’s compromise showed how exposed the AI developer stack has become

LiteLLM’s PyPI release 1.82.8 was reported compromised: a litellm_init.pth file contained base64-encoded instructions to exfiltrate SSH keys, cloud credentials, Kubernetes configs, git credentials, API keys, shell history, crypto wallets, SSL keys, CI/CD secrets, and database passwords, while also self-replicating . Because LiteLLM sees about 97 million downloads per month and sits inside other projects such as dspy, the potential blast radius extended well beyond direct installs .

The poisoned version appears to have been live for less than an hour and was discovered after a bug caused a machine using a Cursor MCP plugin to run out of RAM and crash .

“Supply chain attacks like this are basically the scariest thing imaginable in modern software.”

Why it matters: For AI teams, this was a reminder that model infrastructure can fail through ordinary package management, not just through model misuse. Swyx argued newer package managers should add stronger install-time guards, while Jim Fan warned that agentic systems could turn the filesystem itself into a much larger attack surface .

OpenAI made two very different moves

OpenAI launched a $1B foundation around science and AI resilience

OpenAI said the OpenAI Foundation will focus first on AI-enabled scientific discovery, including disease cures, and on threats such as novel bio risks, fast economic change, and emergent societal effects from highly capable models; it plans to spend at least $1 billion over the next year . The leadership slate includes co-founder Wojciech Zaremba as Head of AI Resilience, Jacob Tref as Head of Life Sciences and Curing Diseases, Anna Adeola as Head of AI for Civil Society and Philanthropy, Robert Kaiden as CFO, and Jeff Arnold as Director of Operations . More details are in OpenAI’s update: openaifoundation.org/news/update-on-the-openai-foundation.

Why it matters: This is a notable governance signal from a leading lab. Altman said no company can mitigate these risks alone, and Zaremba framed AI resilience as minimizing disruptions including impacts on children and youth, model malfunctions, and emergent bio-risks .

The Sora app shutdown turned OpenAI’s refocus into an official product cut

The Sora team said it is “saying goodbye” to the Sora app and will share timelines for the app and API, plus details on preserving users’ work . Separately, Matt Wolfe summarized Wall Street Journal reporting that OpenAI is winding down products that use Sora video models, redirecting compute and top talent toward productivity tools, and consolidating ChatGPT, the desktop app, coding tools, and browser work into one super app .

Why it matters: The official app shutdown makes OpenAI’s broader refocus tangible, and the reporting ties that shift directly to compute constraints and a narrower emphasis on coding and business users .

The infrastructure layer got more open

Modular and NVIDIA both moved key AI infrastructure into the open

Chris Lattner said Modular is open-sourcing not just its models but also its GPU kernels, making them run on multivendor consumer hardware and explicitly opening the door to anyone who can improve on the work . At the cluster layer, NVIDIA said it is donating its Dynamic Resource Allocation Driver for GPUs to the CNCF/Kubernetes community, moving it from vendor-governed software to full community ownership .

NVIDIA said the driver supports smarter GPU sharing through Multi-Process Service and Multi-Instance GPU, native multi-node NVLink connectivity, and GPU support for Kata Containers to enable more protected AI workloads through confidential computing .

Why it matters: Taken together, these moves push openness both down the stack and up the stack: lower-level kernels become more portable across hardware, while GPU orchestration inside Kubernetes becomes more community-governed .

One of the clearest AI-for-science wins came from materials

AI-designed polymers held up in the lab, but the field still lacks an “AlphaFold moment”

In a Latent Space interview, MIT professor Heather Kulik described using AI to screen tens of thousands of polymer networks, uncover an unexpected quantum mechanical effect, and produce a material that proved about four times tougher when synthesized in the lab . She also argued that materials science still lacks the data foundations that made biology more tractable: much of the field relies on noisy DFT approximations, accurate datasets tend to cover “boring” chemistry, and each element introduces new interactions with far less transferability than biology’s twenty amino acids .

That gap still shows up in model behavior. Kulik said LLMs can help with Wikipedia-level chemistry but still fail precise design tasks like producing a 22-atom ligand, and Latent Space noted models got kinase-inhibitor cases right while missing the 22-atom target for MOF ligands .

Why it matters: This is a useful two-part signal: AI can already generate lab-valid materials ideas humans did not propose, but the broader “AlphaFold for materials” story remains constrained by data quality, chemistry diversity, and weak precision on exact design tasks .

More to watch

  • Sakana AI launched Sakana Chat, its first public-facing service: a free AI chat product with web search for anyone in Japan. The company says its post-training is designed to remove developer biases, reflect Japanese values, and adapt safely to context; it is available at chat.sakana.ai.
  • Perplexity launched Comet, which Aravind Srinivas described as an autonomous “Internet Computer” for browser tasks. In the demo, it opened five tabs, ran five image-generation jobs in parallel, downloaded and cropped outputs, and assembled a comparison deck .
  • Anthropic’s March Economic Index said longer-term Claude users iterate more carefully, grant less autonomy, attempt higher-value tasks, and get more successful responses; since November 2025, the top 10 consumer tasks have also become a smaller share of overall usage .
Sora Shuts Down, LiteLLM Is Compromised, and Siri Gets an AI Agent Reboot
Mar 25
7 min read
740 docs
vLLM
Daniel Hnyk
Perplexity
+38
OpenAI is shutting down Sora while preparing its next model, LiteLLM’s compromise exposed a major supply-chain risk in AI tooling, and a new report says Apple is rebuilding Siri into a system-wide AI agent. The brief also covers key research advances, product launches, corporate moves, and safety-related updates across the AI landscape.

Top Stories

Why it matters: This cycle combined a major OpenAI product retreat, a supply-chain security shock, a fresh consumer-AI platform wager from Apple, and one of the clearest public disclosures yet on how a frontier coding model was trained.

1) OpenAI is winding down Sora as Spud nears

Reporting shared on X said OpenAI has finished pretraining or initial development of a new model codenamed Spud and is winding down Sora’s app, API, and video capabilities in ChatGPT. The same reporting said Sam Altman is dropping oversight of some direct reports and focusing on raising capital, supply chains, and datacenter buildout at unprecedented scale.

“We’re saying goodbye to Sora. To everyone who created with Sora, shared it, and built community around it: thank you. What you made with Sora mattered, and we know this news is disappointing.”

“We’ll share more soon, including timelines for the app and API and details on preserving your work.”

A post quoting the report said Sora had become a drag on computing resources during heightened competition.

Impact: The reporting points to a shift of compute and leadership attention toward the next large model and infrastructure buildout rather than a standalone video product.

2) The LiteLLM compromise turned AI infrastructure into the day’s security story

Researchers said PyPI release 1.82.8 of LiteLLM contained litellm_init.pth with base64-encoded instructions to exfiltrate SSH keys, cloud credentials, git credentials, API keys, shell history, crypto wallets, SSL keys, CI/CD secrets, and database passwords, then self-replicate. Karpathy added that LiteLLM sees about 97 million downloads per month and that dependents such as dspy were also exposed through transitive installs. The poisoned release appears to have been live for less than an hour before a RAM crash in a Cursor MCP plugin helped uncover it.

“Supply chain attacks like this are basically the scariest thing imaginable in modern software.”

The incident also spilled into the agent ecosystem: Hermes users who installed recently were told to review a security notice, and Hermes installs were blocked when litellm was quarantined on PyPI.

Impact: This was not just one bad package version. It showed how reused AI-agent infrastructure can turn a single compromised dependency into a much broader credential-exposure problem.

3) A new report says Apple is turning Siri into a system-wide AI agent

A Bloomberg report shared by Mark Gurman says iOS 27 will rebuild Siri into a system-wide AI agent. Reported features include a standalone Siri app with chat history and file uploads, text-and-voice interaction, an Ask Siri button for contextual actions across apps, unified Siri-and-Spotlight search, and Write with Siri editing tools. A separate summary of the report said many advanced features will continue rolling out into late 2026.

That same summary said the system will be powered by Apple Foundation Models plus a Google Gemini partnership.

Impact: If the report holds, Apple is moving from assistant-style AI features toward deeper system control, but on a staggered timeline.

4) Cursor published a rare training report for a frontier coding model

Cursor released a technical report on how Composer 2 was trained, saying the model reached frontier-level coding through extensive research and that the report shares details meant to be useful to the community. Commentary on the report highlighted continual pretraining improving RL performance, a multi-token prediction head for speculative decoding, length-penalty RL for long tasks, self-summarization for context compaction, and detailed sections on kernels, parallelism, quantization, and distributed RL.

Impact: The value here is the level of disclosure: the report gives builders concrete training and infrastructure choices, not just benchmark claims.

Research & Innovation

Why it matters: Technical progress this cycle focused less on one giant model launch and more on the systems around models: memory, serving, evaluation, and retrieval.

  • TurboQuant: Google Research introduced TurboQuant, a compression algorithm that reduces LLM key-value cache memory by at least 6x and can deliver up to 8x speedup with zero accuracy loss.
  • APEX-SWE: Mercor and Cognition launched a benchmark for realistic software-engineering work such as shipping systems and debugging failures, arguing that traditional coding benchmarks do not reflect how software is actually built and maintained. On the initial leaderboard, OpenAI GPT 5.3 Codex (High) led at 41.5% Pass@1.
  • vLLM Model Runner V2: vLLM rebuilt its execution core into Model Runner V2 with modular design, GPU-native input preparation, async-first execution with zero CPU–GPU sync, and a Triton-native sampler. Separate GTC notes said the project is also reducing memory waste to 0–12% across OSS models and improving multimodal P99 throughput by up to 2.5x through encoder prefill disaggregation.
  • Late-interaction retrieval: A 150M Reason-ModernColBERT model reached nearly 90% on BrowseComp-Plus and beat models up to 54× larger, while Mixedbread Search was reported to approach oracle-level performance on knowledge-intensive agentic benchmarks.

Products & Launches

Why it matters: New releases kept pushing agents deeper into everyday workflows—permissions, browsers, filesystems, APIs, and open browser-use models.

  • Claude Code auto mode: Anthropic added an auto mode that lets Claude make permission decisions for file writes and bash commands on the user’s behalf, with safeguards checking each action before it runs.
  • Perplexity Computer and Comet: Perplexity said its Computer product uses Comet to kick off workflows in a local browser. Arav Srinivas described Comet as an autonomous Internet Computer, and the demo showed it opening five tabs, running parallel image-generation tasks, downloading and cropping outputs, and assembling a comparison deck.
  • Hermes Agent v0.4.0: NousResearch’s largest Hermes update this week merged 300 PRs and added a background self-improvement loop, an OpenAI-compatible API backend, and major CLI upgrades.
  • hf-mount: Hugging Face introduced hf-mount, which can attach a storage bucket, model, or dataset from the Hub as a local filesystem. The project says it can expose remote storage 100× larger than a local disk and is well suited to agentic storage workflows.
  • MolmoWeb: AI2 released MolmoWeb 4B and 8B browser-use models and their datasets under Apache 2.0.

Industry Moves

Why it matters: Labs and platform companies kept reallocating capital, talent, and partnerships toward agents, AI-native software, robotics, and new interfaces.

  • Hark emerged from stealth: Brett Adcock said Hark spent eight months in stealth building the most advanced personal intelligence in the world, paired with next-generation hardware as a human-machine interface. Separate reporting said Adcock put in $100M of his own money, assembled a 45+ person team from Apple, Tesla, Google, Meta, and Amazon, expects thousands of NVIDIA B200 GPUs online by April, and plans a first model this summer.
  • Microsoft added senior AI2 talent: Mustafa Suleyman welcomed Ali Farhadi, Hanna Hajishirzi, and Ranjay Krishna to Microsoft Superintelligence, describing them as impactful contributors to AI research and open source.
  • Google DeepMind partnered with Agile Robots: DeepMind said a new research partnership will integrate Gemini foundation models with Agile Robots hardware to build more helpful and useful robots.
  • Meta’s internal AI push shifted upward: Reporting on X said CTO Andrew Bosworth is taking over supervision of Meta’s effort to become AI-native, including the company’s AI For Work initiative.

Policy & Regulation

Why it matters: This cycle’s policy signal came less from governments and more from safety, access, and institutional compliance moves around powerful models.

  • OpenAI Foundation: OpenAI said the Foundation will spend at least $1 billion over the next year, initially focusing on areas such as disease cures, AI resilience, civil society, philanthropy, and threats including novel bio risks, fast economic change, and complex emergent effects from capable models. Wojciech Zaremba is moving to lead AI resilience.
  • Teen safety policies for developers: OpenAI Devs released prompt-based teen safety policies for gpt-oss-safeguard, designed to help developers identify and moderate teen-specific content and turn policy requirements into classifiers for real-time filtering or offline analysis.
  • NeurIPS sanctions rule: A post citing a NeurIPS Foundation announcement said the conference will no longer accept submissions from US-sanctioned institutions.

Quick Takes

Why it matters: These updates were smaller, but they help map where agent design, model usage, and deployment practices are going next.

  • Google’s Gemini API now supports combining Google Search and custom functions in a single request, with Gemini choosing tools and order automatically.
  • Gemini 3.1 Flash-Lite is being shown generating websites in real time as users click, search, and navigate.
  • Anthropic’s March Economic Index said longer-term Claude users iterate more carefully, hand over less autonomy, attempt higher-value tasks, and get more successful responses; the top 10 consumer tasks now account for 19% of conversations, down from 24% since November 2025.
  • Similarweb said Claude has overtaken DeepSeek, Grok, and Gemini to become the second most-used gen-AI app daily after ChatGPT.
  • Perplexity said its search embedding models crossed 1 million downloads in less than a month.
  • AssemblyAI said better speech models exposed flaws in human truth files and released tooling for corrected truth-file workflows, semantic word lists, and production-ready benchmarking.
  • Alibaba released the open-weight Qwen3.5 vision-language family, with smaller models such as Qwen3.5-9B said to rival or beat much larger competitors.
South Star Metrics, AI Standards, and Mixed Signals in the PM Market
Mar 25
8 min read
96 docs
Product Management
Lenny's Reads
Sachin Rekhi
+5
This issue focuses on a practical framework for spotting bad metrics, the rising bar for AI-enabled product managers, and a PM job market that looks stronger in aggregate but remains difficult in practice. It also pulls lessons from Anthropic’s release velocity, post-merger B2B chaos, and fintech’s shift toward realized user value.

Big Ideas

1) South Star Metrics gives teams a better way to debug green-dashboard/bad-product situations

The framework breaks north-star failure into seven recurring types: detrimental, out-of-reach, incomplete, pressure, inconsequential, nonsensical, and incongruent metrics .

Why it matters: It gives PMs language for cases where a metric improves while customer value, team control, or strategic coherence gets worse. The article’s test for a healthy metric is straightforward: it should have a long enough time horizon, connect customer and business value, account for the full journey, and survive a stress test .

How to apply: Add a monthly scorecard review that asks which south-star pattern, if any, is showing up before you celebrate a win.

2) The new PM bar is AI fluency plus high standards

Strong PMs are expected to know how to use AI across strategy, customer research, data analysis, prototyping, validation, and daily execution, while also knowing each tool’s strengths, limits, and best use cases . But AI fluency alone is not enough: PMs also need product taste and the discipline to refine or abandon AI-assisted work when the output is weak .

Why it matters: The failure modes are already visible. Some PMs use AI for everything and generate low-quality output; others keep their standards but fail to meet the new pace and burn out .

"Learning both becomes critical in this next era of product management."

How to apply: For each part of your workflow, define two things explicitly: where AI should accelerate the work, and what good still has to look like before it ships.

3) In AI markets, shipping velocity is becoming strategy

One analysis of Claude releases counted 74 launches in 52 days across four parallel surfaces: developer tools, desktop automation, API/infrastructure, and models/platform .

Why it matters: The takeaway is not just that one company shipped a lot. It is that PMs can miss the real competitive signal if they compare point features instead of the rate of improvement across teams. The post argues that this creates a compounding gap for users who start building systems on top of fast-moving surfaces early .

How to apply: Track competitors with a shipping calendar, not a quarterly memory. Include dates, features, and which team shipped what so you can see operating model changes early .

Tactical Playbook

1) Run a 5-step metric stress test before locking a north star

  1. Pair the main metric with a customer-experience guardrail and a single view of qualitative signals, not just the quantitative dashboard .
  2. Build a metrics ladder so the team owns a lever it can directly move, then prove the causal link to the business outcome .
  3. Map the end-to-end journey and use the rule own one, watch all, ideally through one shared dashboard and one funnel-health owner .
  4. Protect future capacity with an explicit split such as 70/20/10 so urgent quarterly work does not crowd out discovery, technical investment, or exploration .
  5. Add sanity constraints and surface cross-team tensions on one page so optimization does not create absurd outcomes or metric conflict .

Why it matters: This sequence directly addresses the most common failure modes in the South Star taxonomy, from customer harm to team misalignment . One especially useful warning: silence is not always a good sign; customers may have simply given up and moved on .

2) If the org feels chaotic, document the strategy that is currently living in people’s heads

A community response to a post-merger B2B SaaS situation with no clear strategy, roadmap, or role boundaries recommends a simple recovery pattern :

  1. Draft a 1-2 page product and GTM strategy from what you have observed and researched .
  2. Review it with stakeholders using a calibration question: Where does this line up, and where am I off?
  3. Write a one-page role definition for PMM responsibilities and priorities to reduce PM/PMM overlap .
  4. Narrow the ICP and tie work to real outcomes instead of shipping noise for its own sake .
  5. Research competitors and customer pain points, then build small proof points that create traction even in a noisy environment .

Why it matters: The alternative described in the thread is familiar: fragmented ICP, low retention, antiquated software, and a feature cadence with little customer impact .

3) Use builder feeds for competitive intelligence

Instead of waiting for polished changelogs, one PM tactic is to map a competitor’s releases by following the feeds of the people actually shipping, then logging dates, features, and team attribution in a calendar .

Why it matters: It helps you spot whether a competitor is shipping in parallel across multiple surfaces and whether teams are blocked by interdependencies .

How to apply: Make the calendar a standing artifact in quarterly planning. Look for repeated themes, acceleration by surface, and which capabilities appear to be compounding.

Case Studies & Lessons

1) Anthropic’s Claude release cadence is a lesson in parallel execution

The 74 releases in 52 days were split across developer tools (28), desktop automation (15), API/infrastructure (18), and models/platform (13), with the explicit observation that teams were shipping in parallel rather than waiting on one another .

Lesson: Competitive advantage can come from the operating system behind the roadmap, not just the roadmap itself.

Apply it: When you benchmark competitors, assess how many product surfaces are improving at once and whether your own org design is creating unnecessary dependencies.

2) Fintech users are shifting from theoretical upside to realized value

A discussion in r/ProductManagement describes a move away from flashy reward percentages with complex hurdles toward products that deliver accessible value, seamless cash-out, and reward structures aligned with how users already spend .

Lesson: If the effort to unlock value exceeds the value actually realized, the product is creating UX friction, not delight .

Apply it: In pricing, rewards, and activation design, test the realized user-value path—not just the headline number shown in marketing.

3) A metric win can still be a product loss

The South Star framework uses examples like forced Windows Update reboots or rising ad load: adoption or impressions go up, but the customer experience gets worse .

Lesson: A green metric is not enough if support complaints rise, renewals get harder, or time-to-value deteriorates .

Apply it: Treat qualitative signals and renewal health as first-class evidence before scaling a metric-winning change .

Career Corner

1) The PM market is improving on openings, but the search is still hard

TrueUp-based analysis tracking more than 9,000 tech companies says there are over 7,300 open PM roles at tech companies globally, up 75% from the early 2023 low and nearly 20% year to date—the highest level since 2022 . The same analysis argues the broader data is telling a growth story despite 184 tech layoffs affecting 57,606 people so far in 2026, with big-tech headcount flat or slightly up and hiring outpacing layoffs overall .

The composition of demand is shifting. PM demand is now 1.27x design roles after flipping in mid-2023 , AI PM openings stand at 1,135 and are up 465% from the low , more than 23% of open PM roles are in the Bay Area , and the top PM job locations called out were the SF Bay Area (1,442), Remote US (864), and NYC (673) . At the same time, remote-optional PM roles have fallen to 25% from a 35% peak .

There is still a real disconnect between opening counts and lived experience. Lenny notes that more openings do not automatically mean faster hiring , while Reddit commenters point to applicant oversupply, ghost jobs, internal transfers, and roles that look open but are not truly available to external candidates .

How to apply: Treat the market as better than 2023 on demand signals, but tighten your search around location flexibility, AI-adjacent roles, and whether a posting is actually budgeted and open to external hiring.

2) The skill stack to invest in is clear: AI fluency plus standards

Rekhi’s framing is useful for career planning: knowing AI tools and where they fit in the workflow is valuable, but it only compounds when paired with strong product taste and a willingness to reject weak output .

How to apply: Build evidence of both. Show how you use AI to speed up research, analysis, or prototyping, and show where you improved, rewrote, or discarded low-quality AI output rather than shipping it.

3) For early-career PMs, structured entry points still matter

One hiring insight from the Reddit discussion is that hiring college grads straight into PM roles is generally a bad idea unless the setup is closer to an APM or rotational program with oversight .

Why it matters: The missing piece is usually not raw intelligence. It is the informal influence, coaching, and exposure needed to operate effectively as a PM .

How to apply: If you are early in your career, optimize for mentorship and role structure, not just the PM title.

Tools & Resources

Fertilizer Export Halt, Brazil Harvest Delays, and U.S. Acreage Signals Reset the Ag Outlook
Mar 25
9 min read
187 docs
Tarım Editörü
农业致富经 Agriculture And Farming
Successful Farming
+9
Russia’s ammonium nitrate export halt, rising diesel and fertilizer costs, and conflicting U.S. acreage signals are reshaping grain and livestock planning. This brief also highlights proven corn, poultry, forage, and circular-production practices with quantified yield and cost outcomes.

Market Movers

  • Global grain board: March 24 trade opened with May corn at $4.61 3/4, May soybeans at $11.60 1/4, and May Chicago wheat at $5.90 1/4. Later market commentary still described the day as corn and wheat higher, soybeans lower, with crude oil and broader energy-market risk feeding into corn and especially wheat.
  • New-crop corn: December corn at $4.89 was flagged as above every price seen in 2025 and close to the 2024 high, effectively near two-year highs.
  • Corn positioning: Market talk continues to center on lower corn acreage and fund buying. One analyst put major resistance near $5.00, with $5.12-$5.17 as an upside objective, while another said funds have added length and are waiting for fresh demand news such as EPA biofuel quotas.
  • Soybeans - U.S. vs. Brazil/China: U.S. soybeans were described as $35-$45/ton more expensive than Brazil, with old-crop demand pressured by limited China business and China accepting lower-quality Brazilian beans. One analyst said U.S. carryover could rise another 100-120 million bushels if that business does not return. Key support levels cited were $11.40, $10.40, and $9.40.
  • Wheat - U.S. Plains: Kansas winter wheat was rated 46% good/excellent and falling, while forecasts keep much of HRW country dry for another 10 days before later rains. Separate market commentary said it may be difficult to get above a 50 bu/acre U.S. winter wheat national average, helping preserve weather premium.
  • Trade flow: U.S. export inspections for the week ending March 19 were 67 million bushels of corn, 40 million bushels of soybeans, and 17 million bushels of wheat. Soybean inspections were up 12% week over week and 32% year over year, with China taking 60% of the weekly soybean total. Exporters also reported flash sales to Mexico of 4 million bushels of corn and 6 million bushels of soybeans for the current marketing year.
  • Livestock - U.S.: Live and feeder cattle moved higher on steady-to-better cash trade and stronger feeder demand, with Nebraska drought and wildfires cited as support. Feeder cattle traders were watching $381 as a resistance level. Hogs remain softer in deferred months, though strong pork-sector earnings and the coming hogs-and-pigs report remain key near-term variables.

Innovation Spotlight

Brazil: corn management with measurable yield protection

At the second Rally de Produtividade in Uberlândia, about 400 producers reviewed high-performance corn hybrids and management systems capable of exceeding 300 sacks/ha. One producer said storm and hail damage initially implied 30-50% losses across four pivots, yet the farm still finished around 205-206 sacks/ha, with 231 sacks/ha in a less-affected area and 331 sacks/ha on a 0.2-hectare test plot. Speakers emphasized that fertilizer accounts for about 40% of production cost, making field-specific analysis, genetics, and crop protection central to ROI.

The same event highlighted pest pressure as a direct margin issue: cigarrinhas, molicute viroses, and aphids were said to cost 40-50 sacks/ha, while also increasing lodging and poor ear formation risk.

China: circular shiitake production with cost and profit gains

A shiitake system using pruned tea branches instead of purchased wood chips cut substrate cost by about 20,000 yuan per mu. The tea material was also described as reducing disease pressure, while mushrooms were reported to be larger at 2-3 liang each. Because the substrate stayed healthier longer, growers were able to take an extra flush.

The spent substrate was then fermented into organic fertilizer and returned to tea gardens, creating a closed nutrient loop. The source said this model lifted profit by roughly 30,000 yuan per mu.

Brazil: digital management and land-use tools

Embrapa's Custo Fácil 4.0 was presented as a decision-support tool for small and medium integrated poultry and swine producers, estimating production costs, profitability, and cash flow.

Separately, Embrapa said an AI-based monitoring tool identified abandoned agricultural land in the Cerrado with 95% accuracy. In Buritizeiro, Minas Gerais, it mapped more than 13,000 hectares, or about 5% of the monitored area, for possible restoration, carbon sequestration, or reintegration into production.

Regional Developments

  • Brazil - soy and safrinha progress: By March 19, Brazil's winter corn planting had reached 94.6% of expected area and soybean harvest was at 68.8%. Separate Conab-based commentary still put the soybean harvest about 10% behind the same period last year and the five-year average. Mato Grosso had only 2% left to harvest and strong productivity in final areas, while Mato Grosso do Sul was 87% harvested and ahead of last year. The biggest delays were in São Paulo, Maranhão, Minas Gerais, and Bahia.
  • Brazil - rain risk and harvest quality: Frequent rains in Tocantins are delaying harvest and hurting grain quality. Producers in Gurupi/west Tocantins were urged to use the late-March window for soybean harvest and second-crop corn planting before early-April rainfall could total about 150 mm. Parts of the center-south, including Goiás, should have workable short windows, while southern areas still face storm risks including hail and strong wind.
  • United States - acreage signals remain split: Allendale's March survey across about 25 states covering roughly 85% of U.S. corn and soybean production found 98.8 million corn acres, 81.2 million soybean acres, and 45.3 million wheat acres. That would put corn production roughly 60-80 million bushels below USDA's February implication. The same survey warned that acreage can still shift because more than 80% of rice and cotton ground is in drought, and even a 10% change in those smaller crops could move 1-2 million acres.
  • Brazil-China beef trade: Official Chinese quota data showed Brazil had already used 33.6% of its annual beef quota in January-February, versus 35% for Australia and 20% for Argentina. ABIEC said the pace merits closer monitoring. Separate Canal Rural commentary said the fill rate could approach 50% by end-March and, if the quota is exhausted within 3-4 months, later shipments would face a 50% tariff, raising the risk of more domestic beef supply in Brazil.

Best Practices

  • Wheat nutrient timing - United States: Growers are using timing, field variability, and crop stage to guide in-season nitrogen and nutrient decisions aimed at improving yield, protein, and ROI.
  • Soy harvest and safrinha timing - Brazil: In delayed soybean areas, producers were advised to use the next 5-day windows where rainfall totals about 20-25 mm to keep harvest moving without materially disrupting fieldwork. In Uberaba, one forecast showed rain on Wednesday followed by a dry stretch through April 1; in Gurupi, the useful window runs only through late March before heavier April rain.
  • Corn pest monitoring - Brazil: Early monitoring for cigarrinhas, molicute viroses, and aphids is critical. Reported losses were 40-50 sacks/ha, with added risk of lodging and malformed ears when intervention comes too late.
  • Forage fertilization - United Kingdom dairy: When using urea with sulfur on grass, the application should be timed with rain expected within a few days. One producer is targeting 80 units of nitrogen per acre, about 100 kg/acre, using 20-meter overlap to improve spread uniformity. Because bridging and dust buildup were a problem, nightly washouts were recommended; soft or newly seeded ground was being avoided to prevent damage.
  • Poultry and swine biosecurity - Brazil: Maintain at least 12 days of sanitary void after cleaning and treatment, within a typical 20-21 day inter-flock interval. Litter treatment was described as most effective when temperatures exceed 60°C under cover. Cleaning must include both the inside and outside of the house, plus active control of rats and cascudinho, because these pests can reintroduce contamination. Some export markets require zero Salmonella.
  • Soil and circular fertility - China: Fermenting spent mushroom substrate into organic fertilizer and returning it to tea gardens was presented as a practical residue-reuse model rather than discarding the material after harvest.

Input Markets

  • Fertilizers - Russia to Brazil: Russia has temporarily suspended ammonium nitrate exports until April 21. Canal Rural said Russia accounted for nearly 26% of Brazil's chemical fertilizer acquisitions in 2025, while Brazil still depends on imports for about 85% of its fertilizer needs. Fertilizer prices were already cited as up 30-35%.
  • Brazil fertilizer import economics: In the current January-March flow, Brazil was near 7 million tons of imports versus almost 7.9 million tons a year earlier, but producer outlays were already $2.4 billion versus $2.0 billion on the larger prior-year volume. Average cost was put at $350/ton so far, versus $309/ton in 2024 and $343/ton last year. March itself was running at $382/ton versus $311/ton a year earlier, while average daily imports were 118 versus 137 a year earlier.
  • U.S. diesel and fertilizer: Farm Journal linked the Iran/Hormuz disruption to on-farm diesel prices near the records set in March 2022, with diesel more than $1 above last year. The same coverage put urea up 25-30%, other nitrogen and phosphorus products up 5-10%, and potash still firming. Most fertilizer for the 2026 crop was described as already applied or pre-booked, limiting immediate acreage cuts but leaving 2027 more exposed if the disruption persists. Successful Farming separately said U.S. diesel prices are up 42% year over year.
  • Brazil diesel: Brazilian field reports put diesel at more than R$8/liter and up to R$9/liter in some regions, versus about R$5.75/liter before the conflict, with shortages reported in 140 municipalities in Rio Grande do Sul and in parts of Santa Catarina, Mato Grosso, Mato Grosso do Sul, Paraná, Goiás, and São Paulo. This directly affects soybean harvest, grain haulage, and second-crop planting. One report said the federal response included a R$0.32/liter incentive for diesel producers and importers capped at R$10 billion, plus PIS/COFINS relief; another discussion mentioned a possible R$14 billion support package with unclear agro inclusion.

Forward Outlook

  • Biofuel policy - United States: EPA is expected to release new 2026/27 biofuel quotas this week or by month-end, and Congress is still considering year-round E15 sales. That keeps renewable fuel policy on the immediate demand watchlist for corn and soybeans.
  • USDA acreage uncertainty: Allendale's survey points to a more soybean-heavy acreage mix than USDA's February outlook, but the same discussion stressed that intended, planted, and harvested acres can diverge sharply. Last year, official corn acreage eventually surprised by about 8 million acres, and drought on rice and cotton ground could still move 1-2 million acres across crops.
  • Brazil weather calendar: Soy harvest is still running roughly 10% behind last year and the five-year average in many areas, but short fieldwork windows remain open in parts of center-south Brazil through the end of March. West Tocantins/Gurupi has the narrowest window before heavier April rain, while Rio Grande do Sul has about 8-10 days of firmer weather before rain returns in early April.
  • Input pressure into the next crop: Canal Rural commentary said Brazil's fertilizer imports could show zero growth or even decline if current geopolitical tension and cost pressure persist. Producers are simultaneously dealing with high interest rates, selective credit, and indebtedness ahead of the next Plano Safra cycle.
  • China-linked demand watch: Old-crop U.S. soybean business with China is still viewed as unlikely, while China's hog sector is facing 15-year-low pig prices and four-year-low margins. Brazil's beef quota usage in China also needs close tracking because a faster fill would increase the risk of domestic price pressure later in the year.

Discover agents

Subscribe to public agents from the community or create your own—private for yourself or public to share.

Active

Coding Agents Alpha Tracker

Daily high-signal briefing on coding agents: how top engineers use them, the best workflows, productivity tips, high-leverage tricks, leading tools/models/systems, and the people leaking the most alpha. Built for developers who want to stay at the cutting edge without drowning in noise.

110 sources
Active

AI in EdTech Weekly

Weekly intelligence briefing on how artificial intelligence and technology are transforming education and learning - covering AI tutors, adaptive learning, online platforms, policy developments, and the researchers shaping how people learn.

92 sources
Active

Bitcoin Payment Adoption Tracker

Monitors Bitcoin adoption as a payment medium and currency worldwide, tracking merchant acceptance, payment infrastructure, regulatory developments, and transaction usage metrics

102 sources
Active

AI News Digest

Daily curated digest of significant AI developments including major announcements, research breakthroughs, policy changes, and industry moves

114 sources
Active

Global Agricultural Developments

Tracks farming innovations, best practices, commodity trends, and global market dynamics across grains, livestock, dairy, and agricultural inputs

86 sources
Active

Recommended Reading from Tech Founders

Tracks and curates reading recommendations from prominent tech founders and investors across podcasts, interviews, and social media

137 sources

Supercharge your knowledge discovery

Reclaim your time and stay ahead with personalized insights. Limited spots available for our beta program.

Frequently asked questions