AI News 2025-02-13

General

Research Insights

LLM

  • OpenAI announce that o1 and o3-mini now have file and image upload capabilities.
  • Distillation Scaling Laws. Is it better to directly train a small model, or to train a larger model and distill that into a smaller model? The answer is complicated. Roughly, if on a tight compute budget, then directly training a small model may be better. However, if the cost of the big model is “free” (you want to have the big model for other purposes, etc.) then distillation of course can be efficient.

Safety & Security

  • Auditing Prompt Caching in Language Model APIs. They use the response speed to detect whether a given input has been previously cached. This allows one to detect whether someone else has already input that prompt, which thereby leaks information between users. This has a similar flavor to other attacks based on timing or energy use; a system leaks information when it implements internal efficiencies. Leakage can be stopped, but only by giving up the efficiency/speed gains.

Voice

Video

Science

Hardware

  • Groq has secured $1.5B to expand AI inference infrastructure in Saudi Arabia.

Robots

This entry was posted in AI, News and tagged , , , , , , , . Bookmark the permalink.

Leave a Reply