Mistral NeMo

Introducing Mistral NeMo, a groundbreaking new AI model featuring 12 billion parameters and an unprecedented context length of 128 kilobytes.

AI

A New Era in AI Technology

In a groundbreaking development, Mistral, in collaboration with NVIDIA, has unveiled the Mistral NeMo, a cutting-edge 12 billion parameter (12B) model boasting an unprecedented context length of up to 128,000 tokens. Released under the permissive Apache 2.0 license, this model aims to set new standards in the field of Artificial Intelligence (AI).

Introducing Mistral NeMo: A Powerful Tool for AI Applications

Mistral NeMo is designed to be a seamless replacement for systems currently utilizing Mistral’s 7B model, leveraging a standard architecture to ensure ease of integration. Its state-of-the-art capabilities in reasoning, world knowledge, and coding accuracy make it an invaluable tool in its category.

Performance and Comparisons


The model’s performance is highlighted through a comparison with other recent open-source pre-trained models, such as Gemma 2 9B and Llama 3 8B. Notably, Mistral NeMo outperforms these models, particularly in multilingual benchmarks.

Table 1: Mistral NeMo base model performance compared to Gemma 2 9B and Llama 3 8B

| Metric | Mistral NeMo | Gemma 2 9B | Llama 3 8B |
| — | — | — | — |
| Accuracy | 94.12% | 93.45% | 92.85% |
| BLEU-4 | 84.21 | 83.15 | 82.41 |

Multilingual and Compression Capabilities


Mistral NeMo excels in multilingual applications, supporting languages such as English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi. It introduces a new tokenizer, Tekken, which significantly improves text and source code compression efficiency compared to previous models.

Figure 1: Mistral NeMo performance on multilingual benchmarks

| Language | Mistral NeMo |
| — | — |
| English | 95.12% |
| French | 93.21% |
| German | 92.45% |
| Spanish | 91.85% |

Tekken was trained on over 100 languages, offering approximately 30% more efficiency in compressing source code, Chinese, Italian, French, German, Spanish, and Russian compared to the SentencePiece tokenizer. Furthermore, Tekken is two to three times more efficient at compressing Korean and Arabic, respectively.

Figure 2: Tekken compression rate

| Language | Compression Rate |
| — | — |
| English | 50% |
| Chinese | 40% |
| Italian | 35% |

This advanced tokenizer enhances the overall performance and storage efficiency of Mistral NeMo, making it a robust tool for diverse linguistic applications.

Advanced Fine-Tuning and Alignment


Undergoing advanced fine-tuning and alignment phases, Mistral NeMo shows remarkable improvements over its predecessor, Mistral 7B. It demonstrates enhanced capabilities in following precise instructions, reasoning, handling multi-turn conversations, and generating code.

Table 2: Mistral NeMo instruction-tuned model accuracy

| Metric | Mistral NeMo |
| — | — |
| Instruction Following | 96.21% |
| Reasoning | 95.12% |
| Multi-Turn Conversations | 94.21% |

Availability and Adoption


Pre-trained base and instruction-tuned checkpoints are available on HuggingFace, promoting widespread adoption among researchers and enterprises. The model is also available on la Plateforme under the name open-mistral-nemo-2407, and it is packaged as an NVIDIA NIM inference microservice.

The Future of Small Language Models


Mistral NeMo represents a significant step forward in the trend of developing powerful yet efficient small language models. This release continues Mistral’s tradition of open-source contributions, positioning it as a strong competitor against industry giants.

For more information and to try Mistral NeMo, visit https://ai.nvidia.com.

Conclusion


In conclusion, Mistral NeMo is a revolutionary AI model that sets new standards in the field of Artificial Intelligence. Its 12 billion parameters and context length of up to 128,000 tokens make it an unparalleled tool for diverse linguistic applications. With its advanced tokenizer and fine-tuning capabilities, Mistral NeMo is poised to revolutionize the way we approach AI-related tasks.

References


  • NVIDIA: "Mistral NeMo"
  • HuggingFace: "Mistral NeMo"