Guillaume Lample @ NeurIPS 2024
@GuillaumeLample
Cofounder & Chief Scientist http://Mistral.ai (@MistralAI). Working on LLMs. Ex @MetaAI | PhD @Sorbonne_Univ_ | MSc @CarnegieMellon | X11 @Polytechnique
Very excited to release our second model, Mixtral 8x7B, an open weight mixture of experts model. Mixtral matches or outperforms Llama 2 70B and GPT3.5 on most benchmarks, and has the inference speed of a 12B dense model. It supports a context length of 32k tokens. (1/n)
magnet:?xt=urn:btih:5546272da9065eddeb6fcd7ffddeef5b75be79a7&dn=mixtral-8x7b-32kseqlen&tr=udp%3A%2F%2Fopentracker.i2p.rocks%3A6969%2Fannounce&tr=http%3A%2F%https://t.co/g0m9cEUz0T%3A80%2Fannounce RELEASE a6bbd9affe0c2725c1b7410d66833e24
Happy to share that we’ve just signed a €100 million partnership with @cmacgm! @MistralAI is going to help CMA CGM adopt custom-designed AI solutions to support its shipping, logistics, and media activities. We will provide CMA CGM with our entire AI solutions portfolio and we…
Introducing powerful new features in Le Chat, making it more capable and more fun!
Introducing the world's best (and open) speech recognition models!
Introducing Devstral Small and Medium 2507! This latest update offers improved performance and cost efficiency, perfectly suited for coding agents and software engineering tasks.
New milestone for Project Numina and Kimi Moonshot! 🚀 We are open sourcing our KiminaProver-72B. This SotA theorem-proving model comes with Test-Time Reinforcement Learning Search and Error-Fixing Capability. We’re putting it to the test soon, with the IMO just around the corner…
Introducing Mistral Small 3.2, a small update to Mistral Small 3.1 to improve: - Instruction following: Small 3.2 is better at following precise instructions - Repetition errors: Small 3.2 produces less infinite generations or repetitive answers - Function calling: Small…
We're proud to announce Mistral Compute—an unprecedented AI infrastructure undertaking in Europe, and a strategic initiative that will ensure that all nation states, enterprises, and research labs globally remain at the forefront of AI innovation. Read more in the thread.
Today we are releasing ether0, our first scientific reasoning model. We trained Mistral 24B with RL on several molecular design tasks in chemistry. Remarkably, we found that LLMs can learn some scientific tasks more much data-efficiently than specialized models trained from…
Introducing Agents API: your go-to tool for building tailored agents to solve complex real-world problems! mistral.ai/news/agents-api
Meet Document AI, our end-to-end document processing solution powered by the world’s best OCR model! mistral.ai/solutions/docu…
Meet Devstral, our SOTA open model designed specifically for coding agents and developed with @allhands_ai mistral.ai/news/devstral
Introducing Mistral Medium 3: our new multimodal model offering SOTA performance at 8X lower cost. - A new class of models that balances performance, cost, and deployability. - High performance in coding and function-calling. - Full enterprise capabilities, including hybrid or…
Introducing Le Chat Enterprise, the most customizable and secure agent-powered AI assistant for businesses, making AI a real leverage for competitiveness. - Integration with your company knowledge (starting with Gmail, Google Drive, Sharepoint…) - Ability to add frequently used…