
"The launch comes as Mistral, which develops open-weight language models and a Europe-focused AI chatbot Le Chat, has appeared to be playing catch up with some of Silicon Valley's closed source frontier models. The two-year-old startup, founded by former DeepMind and Meta researchers, has raised roughly $2.7 billion to date at a $13.7 billion valuation - peanuts compared to the numbers competitors like OpenAI ($57 billion raised at a $500 billion valuation) and Anthropic ($45 billion raised at a $350 billion valuation) are pulling."
""Our customers are sometimes happy to start with a very large [closed] model that they don't have to fine-tune...but when they deploy it, they realize it's expensive, it's slow," Guillaume Lample, co-founder and chief scientist at Mistral, told TechCrunch. "Then they come to us to fine-tune small models to handle the use case [more efficiently]." "In practice, the huge majority of enterprise use cases are things that can be tackled by small models, especially if you fine tune them," Lample continued."
"Initial benchmark comparisons, which place Mistral's smaller models well behind its closed-source competitors, can be misleading, Lample said. Large closed-source models may perform better out-of-the-box, but the real gains happen when you customize. "In many cases, you can actually match or even out-perform closed source models," he said. Mistral's large frontier model, dubbed Mistral Large 3, catches up to some of the important capabilities that larger closed-source AI models like OpenAI's GPT-4o and Google's Gemini 2 boast, while also trading"
Mistral launched the Mistral 3 family, a 10-model open-weight lineup that includes a large multimodal, multilingual frontier model and nine smaller offline-capable, fully customizable models. The company has raised roughly $2.7 billion and holds a $13.7 billion valuation, substantially smaller than major competitors. The release emphasizes fine-tuning smaller models for enterprise deployments to reduce cost and latency. Many enterprise tasks can be handled by fine-tuned small models, which can sometimes match or outperform larger closed-source models after customization. Initial benchmarks may understate smaller models' potential because customization delivers the significant performance gains.
Read at TechCrunch
Unable to calculate read time
Collection
[
|
...
]