{"id":59342,"date":"2024-10-17T08:55:43","date_gmt":"2024-10-17T07:55:43","guid":{"rendered":"https:\/\/dataconomy.ru\/?p=59342"},"modified":"2024-10-17T08:55:43","modified_gmt":"2024-10-17T07:55:43","slug":"les-ministraux-ministral-3b-and-8b-models-bring-genai-to-the-edge","status":"publish","type":"post","link":"https:\/\/dataconomy.ru\/2024\/10\/17\/les-ministraux-ministral-3b-and-8b-models-bring-genai-to-the-edge\/","title":{"rendered":"Les Ministraux: Ministral 3B and 8B models bring GenAI to the edge"},"content":{"rendered":"
In a world dominated by bloated AI models that live in the cloud, Mistral AI is flipping the script. The French startup just unleashed<\/a> two new models\u2014Ministral 3B and 8B\u2014that are designed to run on edge devices.<\/p>\n Mistral\u2019s new offerings, dubbed \u201cLes Ministraux,\u201d might sound like a French art-house film, but these models are poised to shake up the AI world. With just 3 billion and 8 billion parameters respectively, the Ministraux family is all about efficiency. Forget those resource-hogging AI models that require a data center to function.<\/p>\n \u201cOur most innovative customers and partners have increasingly been asking for local, privacy-first inference for critical applications,\u201d Mistral explained.<\/p>\n Here\u2019s where it gets really spicy: both the 3B and 8B models can handle a context window of 128,000 tokens. That\u2019s the equivalent of a 50-page book. For comparison, even OpenAI\u2019s GPT-4 Turbo<\/a> caps out around the same token count, and that\u2019s no small feat.<\/p>\n With this kind of capacity, the Ministraux models don\u2019t just outperform their predecessor, the Mistral 7B\u2014they\u2019re also eating Google\u2019s Gemma 2 2B<\/a> and Meta\u2019s Llama models<\/a> for breakfast.<\/p>\nLes Ministraux: Ministral 3B and 8B<\/h2>\n
Context length like never before<\/h2>\n