Daily digest
10 items · ~10 min · Week 2026-W18
Must-read (2)
DeepSeek V4: official open-source release with Day-0 adaptation for Huawei Ascend
DeepSeekDeepSeek officially released the V4 lineup in open-source under the MIT license on April 29. It includes DeepSeek-V4-Pro at 1.6T parameters (49B active) and DeepSeek-V4 at 284B (13B active) — both MoE models with native 1M token context. The release claims roughly a 9.5x reduction in memory requirements versus V3.2 and a near-closed gap with frontier closed models on reasoning benchmarks. A defining feature of the release is optimization for Chinese accelerators: Huawei Ascend, Cambricon, Hygon, and Moore Threads completed Day-0 adaptation on release day, with multi-deploy on Ascend 950 expected in the second half of the year.
GLM-5V-Turbo: a natively multimodal foundation model for agents
Z.aiZ.ai unveiled GLM-5V-Turbo, a multimodal foundation model in which visual perception is embedded as a first-class component of reasoning, planning, and tool use rather than bolted on after the fact. The model handles images, video, web pages, and documents; the authors report gains on multimodal coding, visual tool use, and agent tasks while preserving text-only quality. The role of end-to-end verification of agent trajectories during training is emphasized.
Worth knowing (5)
ElevenLabs launches ElevenMusic — a licensed platform for music generation, remix, and streaming
ElevenLabsElevenLabs unveiled an updated ElevenMusic — a product that combines music discovery, remixing existing tracks (genre swaps, tempo changes, reinterpretation), and creating original compositions from text, melody, or mood. The platform is built on a fully licensed music model; at launch it features more than 4,000 independent artists and a curated release, Eleven Album Vol. 2. It is positioned not as passive listening but as a fan-engagement layer with publishing and monetization options for creators.
Yandex Commerce Protocol: first retailers launch sales via Alice AI
YandexYandex disclosed the first partners of the Yandex Commerce Protocol (YCP) — a standard for integrating online stores with AI scenarios in Alice AI, Search, and Yandex Ritm. Going live with sales directly from chat with Alice AI are Stockmann, restore:, pharmacy chains Gorzdrav and 36.6, telecom operator Beeline, the brand The Act, and a number of other retailers; over 200 large online retailers and brands have begun YCP integration, and more than 1,600 additional stores have applied. The technology lets shoppers proceed to checkout directly from the assistant dialog without visiting the merchant's website — Alice AI acts as a transactional AI agent on top of partner catalogs.
Anthropic in talks for a round at over $900B valuation
AnthropicAnthropic has received preemptive offers to raise around $50B at a valuation in the $850–900B range, more than doubling its current capitalization and potentially putting the company ahead of OpenAI as the most valuable AI startup. Talks are at an early stage and no term sheet has been signed. In parallel, run-rate revenue is reported at >$30B versus ~$9B at the end of 2025.
Recursive Multi-Agent Systems: agent communication in latent space
Stanford UniversityRecursiveMAS replaces text exchange between agents with communication via latent representations connected by a lightweight RecursiveLink module, and trains the whole system jointly using a dedicated optimization algorithm. Across 9 benchmarks (math, science, medicine, search, code) the authors report +8.3% average accuracy, a 1.2–2.4x speedup in end-to-end inference, and a 34.6–75.6% reduction in token consumption versus text-based multi-agent baselines.
Mistral Workflows: public preview of a Temporal-based engine for enterprise AI orchestration
MistralMistral AI announced Workflows in public preview on April 29 — durable, observable AI orchestration in Studio and Le Chat. The architecture is built on Temporal with AI extensions: streaming, payload handling, and extended observability. The control plane runs on Mistral-managed infrastructure, while execution workers and data processing run inside the customer's environment. Workflows are written in Python, can be published to Le Chat to be triggered by non-technical users, and every step is traceable in Studio. According to VentureBeat, the engine is already handling millions of daily executions for early customers: ASML, ABANCA, CMA-CGM, France Travail, La Banque Postale.
For reference (3)
TIDE: cross-architecture distillation for diffusion LLMs
Peking UniversityTIDE is a distillation framework that transfers knowledge between different architectures for diffusion LLMs. It comprises three components: TIDAL (adaptive distillation strength by timestep), CompDemo (context via mask splitting), and Reverse CALM (cross-tokenizer objective). Teachers are a dense 8B and a 16B MoE; the student is a 0.6B diffusion model; the student's HumanEval score is 48.78 versus 32.3 for an AR baseline of the same size.
Programming with Data: test-driven data engineering for self-improving LLMs
OpenDataLabThe authors reframe data engineering for LLMs as software engineering: training data = source code of the model's behavioral spec, training = compilation, benchmarks = unit tests. If structured knowledge is extracted from the source corpus and used simultaneously for training and evaluation, model failures can be traced back to specific defects in the data and fixed surgically. The method is applied to 16 disciplines; a knowledge base, benchmarks, and training corpora are released.
OpenCode v1.14.30: Mistral Medium 3.5 with reasoning and Desktop session fixes
SSTSST released opencode v1.14.30 (April 29, 2026). Support for Mistral Medium 3.5 with reasoning mode was added, Azure response handling improved, and issues with sessions in the Desktop app and editor context across multiple directories were fixed. The April release cadence has been tight: v1.14.27 introduced a configurable default shell, v1.14.25 added Roslyn LSP for C#/Razor, and v1.14.21 brought improved compaction for long conversations.