Mistral Medium 3.5

· ai ai-agents systems · Source ↗

TLDR

  • Mistral releases Medium 3.5, a 128B dense open-weights model powering async cloud coding agents in Vibe and a new Work mode in Le Chat.

Key Takeaways

  • 128B dense model, 256k context window, self-hostable on 4 GPUs; scores 77.6% on SWE-Bench Verified, ahead of Devstral 2 and Qwen3.5 397B A17B.
  • Open weights on Hugging Face under a modified MIT license; API priced at $1.5/M input tokens and $7.5/M output tokens.
  • Vibe remote agents run async in the cloud with parallel sessions; ongoing local CLI sessions can be teleported to the cloud with full state preserved.
  • Work mode in Le Chat adds multi-step agentic execution with connectors on by default, integrating GitHub, Linear, Jira, Sentry, and Slack.
  • Reasoning effort is configurable per request, so the same weights handle quick chat replies and long-horizon agentic runs without model switching.

Hacker News Comment Review

  • Only one comment so far: a commenter is skeptical, citing GPT-5.5 as having raised the bar high enough that other frontier releases feel underwhelming by comparison.

Original | Discuss on HN