{
  "version": "https://jsonfeed.org/version/1.1",
  "title": "Sarmadi AI Digest",
  "home_page_url": "https://sarmadi.llc/ai-digest/",
  "feed_url": "https://sarmadi.llc/ai-digest/data/feeds/feed.json",
  "items": [
    {
      "id": "https://sarmadi.llc/ai-digest/2026-05-13.html",
      "url": "https://sarmadi.llc/ai-digest/2026-05-13.html",
      "title": "May 13, 2026 — Google rebrands Android around AI; OpenAI trial keeps unraveling",
      "date_published": "2026-05-13T11:50:00Z",
      "content_text": "Google used its pre-I/O Android showcase to rebuild the platform around AI — Googlebooks laptops, Gemini Intelligence across the phone, dictation in Gboard, and a 'Create My Widget' feature that turns vibe-coded UI into a system primitive. The OpenAI trial dominated the news cycle in parallel: Altman testified that Musk did 'huge damage' to the company and once floated handing it to his children, while Sutskever defended the 2023 ouster from the stand. Compute geography keeps escalating — Google and SpaceX are now reportedly in talks for orbital data centers, xAI is adding 19 gas turbines despite litigation, and one pitch is to host mini data centers inside private homes. Underneath, agent research is moving the safety frame from prompt to trajectory: papers on hidden multi-turn intent, on-policy self-evolution from failure trajectories, and privacy-aware device-cloud collaboration treat the agent's whole run as the alignment target. A 26M-parameter distillation of Gemini tool-calling, viral on Hacker News, is a useful reminder that the small-model frontier is moving faster than the headline-grabbing one."
    },
    {
      "id": "https://sarmadi.llc/ai-digest/2026-05-12.html",
      "url": "https://sarmadi.llc/ai-digest/2026-05-12.html",
      "title": "May 12, 2026 — Claude lands on AWS, attackers ship AI-built zero-days",
      "date_published": "2026-05-12T11:35:00Z",
      "content_text": "Anthropic took the Claude Platform live on AWS the same week OpenAI's DeployCo opened for business, turning enterprise distribution into a head-to-head channel fight rather than a model contest. Google's threat-intel team disclosed the first zero-day it has caught being developed with AI, and OpenAI answered with Daybreak — a defensive security program built around the same capability. The agent-research wave underneath these moves is no longer about raw ability but about boundaries: FORTIS measures whether agents stay inside their privilege; Shepherd makes their execution forkable and replayable; the AI Workflow Store argues that long-lived agents need software-engineering discipline, not on-the-fly synthesis. Quietly in the background, efficiency work — looped transformers with constant memory, queryable LoRA atoms, Muon fine-tuning of Adam-pretrained models — keeps narrowing the cost gap between frontier output and what a small team can run. The week's market story is older but louder: Sutskever defended his role in the Altman ouster on the witness stand, and Stratechery reframed the xAI–Anthropic deal as Musk choosing to serve other people's roadmaps."
    },
    {
      "id": "https://sarmadi.llc/ai-digest/2026-05-11.html",
      "url": "https://sarmadi.llc/ai-digest/2026-05-11.html",
      "title": "May 11, 2026 — Inference economics moves to center stage",
      "date_published": "2026-05-11T11:55:00Z",
      "content_text": "Monday's research and industry signals converge on one question: who pays for inference, and where does it run. Stratechery names the shift explicitly; CUDA's renewed framing as Nvidia's software moat and Nvidia's $40B in equity deals price it; a top-of-Hacker-News essay argues local AI should be the default. On the research side, three independent papers attack long-context inference cost from different layers — shallow-prefill KV visibility, block-iterative speculative decoding, and an analysis of state-tracking error control in linear models. Agent safety quietly graduated from manifesto to tooling: a red-team platform, a prefix-trace failure monitor, a skill compiler with built-in injection checks, and an SAE-based firewall for VLMs all arrived on the same day. A separate result reframes RL for reasoning as sparse policy selection rather than capability acquisition, recovering most of RL's gains at three orders of magnitude lower cost."
    },
    {
      "id": "https://sarmadi.llc/ai-digest/2026-05-10.html",
      "url": "https://sarmadi.llc/ai-digest/2026-05-10.html",
      "title": "May 10, 2026 — Long-context inference keeps getting cheaper; Anthropic blames the training corpus",
      "date_published": "2026-05-10T22:00:00Z",
      "content_text": "A Sunday with surprisingly concrete research throughput. Three separate papers attack the long-context inference bottleneck from three different angles — prefill sparsification, byte-level generation, and delta-rule linear attention — and the directions are not in conflict. Agents continue their march from demo to discipline: a survey crystallizes the memory subfield, a method auto-discovers test-time scaling strategies, and a new benchmark forces interleaved multimodal evidence. On the industry side, Anthropic argues — with a paper — that fictional portrayals of AI in the training data are a measurable cause of misbehavior, and an unusual xAI–Anthropic agreement leaves analysts puzzled. The grid story keeps mattering: Maryland is sending federal regulators a $2B bill."
    }
  ]
}