Mistral announces Large 2: flagship LLM with 123 billion parameters
Large 2 - the leading large language model (LLM) with significantly higher code generation, mathematical computation, and reasoning capabilities. Mistral has also added improved multi-language support and a range of advanced functions with the Large 2.
If you don't know, a large language model is a language model with general capabilities for language generation and other natural language processing tasks. LLM achieves this ability by learning statistical relationships from texts during highly computationally complex self-supervised and semi-supervised training.
The Mistral Large 2 model has 123 billion parameters, allowing it to run on a single H100 node with high throughput. This LLM has comprehensive support for French, German, Spanish, Italian, Portuguese, Arabic, Hindi, Russian, Chinese, Japanese and Korean. In terms of coding, Large 2 supports over 80 different programming languages, including Python, Java, C, C++, JavaScript and Bash…
Large 2 is currently available for open access, but is only made freely available by Mistral for research and non-commercial purposes. For commercial use, users need a specialized use license.
With 123 billion parameters (123B), Mistral Large 2's performance is comparable to GPT-4o, OpenAI's Claude Opus 3, and Meta's recently released Llama 3.1 405B in terms of encoding capabilities. On the Wild Bench, Arena Hard and MT Bench ratings, Large 2 outperformed Llama 3.1 405B and Claude 3 Opus. On the popular MMLU benchmark, this new model performs better than the Llama 3.1 70B and is comparable to the Llama 3.1 405B.
From a developers perspective, Mistral Large 2 now has improved function calling and retrieval skills. The model can now execute both parallel and sequential function calls, allowing developers to build complex business AI applications.
With the release of Large 2, Mistral's LLM ecosystem is now relatively diverse, including Mistral Nemo, Mistral Large, and two specialized models: Codestral and Embed. Mistral will discontinue the Apache models (Mistral 7B, Mistral 8x7B and 8x22B, Codestral Mamba, Mathstral) in the future.
Microsoft and Mistral have a partnership to integrate Mistral models on Azure. Today, Mistral is expanding its partnership with Google to bring its products to Google Cloud.
The consecutive releases of Mistral Large 2 and Llama 3.1 mark a major milestone for the open AI ecosystem, providing two powerful GPT-4 level models for research and development. This rapid progress drives growing momentum towards a more open and collaborative AI ecosystem.
You should read it
- How to send large video over the network?
- How to find large files on Windows 10
- How to send large files via Facebook quickly
- What is the Large Language Model (LLM)?
- How to Integrate Large Data Sets in Excel
- How to send large files, large videos via the Internet quickly and easily
- How to find files / folders taking up a large capacity on Windows
- Do you choose a large aperture or a large sensor when taking photos?
May be interested
- Apple may lose $ 9 billion because iPhone sales in some markets are not as expectedapple ceo tim cook said the decline in demand in large smartphone markets like china and other developing countries could seriously affect the company's profitability.
- Parameters in HTTPthis chapter lists some of the http protocol parameters and their syntax by which they are used in communication.
- Instructions for setting up room parameters in Among Usand to get the new game among us, you need to know how to adjust the parameters for the game. check out how to adjust parameters and some interesting settings below.
- The Earth has more than 8 billion tons of plastic, weighing about 1 billion elephantsfrom 1950 to 2015, in just over six decades, the world produced more than 8.3 billion tons of plastic, a total weight of 25,000 empire state buildings, 381m high in the us or 1 billion elephants.
- Get color parameters more easily with ColorSchemer Studioin the design world, color is a very important factor that directly affects the quality and aesthetics of the product being created. if you want to know these parameters to be able to use, color schemer studio will be a software worth your attention.
- IBM announces next generation Z processor: 7nm Telum chip, 22.5 billion transistors, 8 cores running on 5GHz clockibm has just released relatively detailed information about its next-generation telum chipset, which is part of the new z series of processors. the telum chip carries a completely new core architecture design, aimed at radically enhancing ai processing capabilities.
- Python function parameterin the previous article we learned about the built-in python function and the user-defined python function with customizable number of parameters. you will know how to define a function using the default parameters, keyword and custom parameters in this article.
- Review: Billion Road - Experience the giant Japanese streetcombining familiar billionaire chess play with unique and funny 'innovations', billion road promises to be a fun game for you.
- Which screen parameters are completely meaningless?let's take a closer look at these parameters to see what they mean.
- The number of Facebook and Instagram users is equal to nearly half of the world's populationaccording to meta's first quarter 2024 financial report dated april 24, the company's total daily active users (dap) currently reached 3.24 billion as of the end of march, an increase of 7% compared to 3.19 billion. billion by december 2023, accounting for 40% of the world's population (current world population is 8.1 billion).