AI News 2025-04-10

General

Research Insights

LLM

  • More progress in diffusion language models: Dream 7B: Introducing Dream 7B, the most powerful open diffusion large language model to date.
  • Meta releases Llama 4 series of MoE LLMs: Scout (109B, 17B active, 16 experts), Maverick (400B, 17B active, 128 experts), and Behemoth (2T, 288B active, 16 experts). These are MoE models with a 10M token context. The models appear to be competitive (nearing the state-of-the-art tradeoff curve for performance/price), and thus extremely impressive for open-source.
    • Independent evals (including follow-up) from Artificial Analysis show it performing well against non-reasoning models.
    • Evaluation of the 10M context on simple NIAH seem reasonable, but (reportedly) it does not fare as well on deeper understanding of long context.
  • Cloudflare launch an open beta for their AutoRAG solution.
  • Nvidia release Llama-3_1-Nemotron-Ultra-253B-v1, which seems to beat Llama 4 despite being based on Llama 3.1.
  • Amazon announces Nova Sonic speech-to-speech foundation models, for building conversational AI.
  • Agentica release open-source: DeepCoder-14B-Preview, a reasoning model optimized for coding (code, hf).
  • Anthropic announce a new “Max” plan for Claude ($100/month).
  • xAI release an API for Grok-3. Pricing appears relatively expensive (e.g. compared to Gemini models of better performance).
  • OpenAI adds an evals API, making it easier to programmatically define tests, evaluations, etc. This should make it faster/easier to test different prompts, LLMs, etc.
  • Bytedance release technical report for Seed-Thinking-v1.5, a 200B reasoning model.
  • OpenAI add a memory feature to ChatGPT, allowing it to reference all past chats in order to personalize responses.

AI Agents

Audio

Image Synthesis

Video

World Synthesis

Science

Brain

Hardware

Robots

This entry was posted in AI, News and tagged , , , , , , , , , , , . Bookmark the permalink.

Leave a Reply