Skip to content

samouraiworld/awesome-mistral

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

6 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Awesome Mistral Awesome Last Updated

A curated list of awesome resources, tools, libraries, and projects for the Mistral AI ecosystem.

Mistral AI is a Paris-based AI company building open-weight, high-performance large language models. Founded in 2023, Mistral has quickly become a leading force in open-source AI, offering models that rival proprietary alternatives while remaining accessible to developers worldwide.

This repository maps and curates the entire Mistral.ai ecosystem for AI engineers, researchers, startup founders, and open-source contributors.

Legend:

  • 🧠 Official Mistral AI
  • 🌍 Community project
  • πŸ§ͺ Experimental

Contents


What's New (March 2026)

  • πŸš€ Mistral Small 4 – Hybrid MoE (119B/6.5B active) unifying reasoning, coding, and multimodal in one model.
  • πŸ§ͺ Leanstral – First open-source Lean 4 formal proof agent.
  • 🏒 Mistral Forge – Enterprise platform for training models on proprietary data.
  • πŸŽ™οΈ Voxtral Mini 4B Realtime – Real-time speech-to-text with sub-200ms latency.
  • πŸ›‘οΈ Mistral Moderation 2603 – Updated content moderation with jailbreaking detection.
  • πŸ–₯️ Mistral Compute – European-hosted GPU cloud.

Why Mistral?

Mistral AI offers a compelling alternative in the LLM landscape:

Aspect Mistral Advantage
Open Weights Models like Mistral Large 3, Small 4, and Ministral are fully open-weight (Apache 2.0), enabling local deployment and full control
Efficiency Mistral Small 4 (119B/6.5B active) and Large 3 (675B/41B active) use MoE parameter routing for high efficiency; Ministral 3B/8B/14B are optimized for edge
European Sovereignty Paris-based company offering GDPR-compliant, EU-hosted API options via Forge and Compute
Cost Efficiency Competitive API pricing; open models enable free self-hosting
Innovation Pioneered efficient MoE architectures, hybrid reasoning models, formal proof agents (Leanstral), and real-time speech AI
Full-Stack Platform Forge (enterprise model training) + Compute (European GPU cloud) + le Chat (AI assistant)

Official Mistral Resources

  • 🧠 Mistral AI – Official company website with product information and announcements.
  • 🧠 Mistral AI Documentation – Comprehensive API documentation, guides, and model specifications.
  • 🧠 AI Studio (la Plateforme) – Developer console for API keys, model access, and agent management.
  • 🧠 le Chat – AI assistant (web, iOS, Android) with free and Pro tiers.
  • 🧠 Mistral Forge – Enterprise platform for training frontier-grade models on proprietary data.
  • 🧠 Mistral Compute – European-hosted GPU cloud (NVIDIA Grace Blackwell).
  • 🧠 Mistral AI GitHub – Official GitHub organization with 24+ repositories.
  • 🧠 mistral-inference ⭐ 10k+ – Official inference library for running Mistral models.
  • 🧠 mistral-finetune ⭐ 3k+ – Official lightweight LoRA-based fine-tuning library.
  • 🧠 Mistral Cookbook ⭐ 2k+ – Official notebooks and examples for common use cases.
  • 🧠 mistral-common – Official tokenization and pre-processing library.
  • 🧠 Mistral Vibe – Native CLI coding assistant.
  • 🧠 Platform Docs Public – Open-source documentation repository.

Model Families

Flagship Models (API)

Model Context License Best For
Mistral Small 4 256k Apache 2.0 Hybrid reasoning + coding + multimodal (119B MoE / 6.5B active)
Mistral Large 3 256k Apache 2.0 Complex reasoning, multilingual, coding, vision (675B / 41B active)
Mistral Medium 3.1 128k Proprietary Prototype-to-production, balanced multimodal performance
Mistral Small 3.2 128k Apache 2.0 Low-latency, cost-sensitive applications (24B)
Mistral OCR 3 β€” Proprietary Document parsing, table reconstruction ($2/1k pages)

Open-Weight Models

General Purpose & Reasoning

  • 🧠 Mistral Small 4 – Hybrid MoE (119B / 6.5B active) unifying reasoning, coding, and multimodal. Configurable reasoning_effort.
  • 🧠 Mistral Large 3 – Flagship MoE (675B) with state-of-the-art reasoning and vision.
  • 🧠 Mistral Small 3.2 – High-performance dense 24B model (v3.2).
  • 🧠 Magistral Small 1.2 – Specialized 24B reasoning model with multimodality.
  • 🧠 Mixtral 8x22B – Legacy MoE workhorse (141B total / 39B active).

Edge & On-Device (Ministral)

  • 🧠 Ministral 14B – Dense edge model with vision (14B). Best-in-class at small scale.
  • 🧠 Ministral 8B – High-performance edge model (8B).
  • 🧠 Ministral 3B – Ultralight model for mobile/browser (3B).

Coding & Agentic (Devstral)

  • 🧠 Devstral 2 – 123B coding model (Modified MIT License). 72.2% SWE-bench Verified.
  • 🧠 Devstral Small 2 – 24B coding model (Apache 2.0) for local agents.
  • 🧠 Codestral 25.01 – Legacy code specialist.

Multimodal (Pixtral)

  • 🧠 Pixtral Large – 124B multimodal model building on Mistral Large 2.
  • 🧠 Pixtral 12B – Efficient vision-language model.

Specialized Models

  • 🧠 Mistral OCR 3 – Advanced document understanding and table reconstruction.
  • 🧠 Leanstral – First open-source Lean 4 formal proof agent (119B/6.5B active, Apache 2.0).
  • 🧠 Voxtral Mini 4B Realtime – Real-time speech-to-text, sub-200ms latency, 13 languages (Apache 2.0).
  • 🧠 Mistral Moderation 2603 – Content moderation with jailbreaking, dangerous, and criminal detection (3B, API only).

Community Fine-Tuned Models

High-quality community fine-tunes built on Mistral base models:

Instruction & Chat

Specialized

Quantized Model Collections

  • 🌍 TheBloke – Extensive GGUF/AWQ/GPTQ quantized model repository.
  • 🌍 bartowski – High-quality GGUF quantizations.

SDKs & APIs

Official SDKs

  • 🧠 client-python – Official Python client library.
  • 🧠 client-ts – Official TypeScript/JavaScript client library.
  • 🧠 @mistralai/mistralai – Official TypeScript/JavaScript SDK (npm).

Community SDKs

Official Libraries

  • 🧠 mistral-common – Official tokenization and pre-processing library.
  • 🧠 Mistral Vibe – Native CLI coding assistant powered by Devstral.

Inference & Deployment

High-Performance Inference

  • 🌍 vLLM ⭐ 35k+ – High-throughput with PagedAttention. Excellent Mistral support.
  • 🌍 Text Generation Inference – Hugging Face's production inference server.
  • 🌍 llama.cpp ⭐ 70k+ – CPU/GPU inference with GGUF quantization.
  • 🌍 ExLlamaV2 – Fast inference with EXL2 quantization.
  • 🌍 SGLang – Fast serving with RadixAttention.

Local Inference

  • 🌍 Ollama ⭐ 100k+ – Simple CLI for local Mistral models.
  • 🌍 LM Studio – Desktop GUI for local LLMs.
  • 🌍 Jan – Open-source ChatGPT alternative running locally.
  • 🌍 GPT4All – Local inference with Mistral support.
  • 🌍 Msty – Desktop app for running local LLMs.

Cloud & Container Deployment

  • 🌍 LocalAI ⭐ 25k+ – OpenAI-compatible local API server.
  • 🌍 SkyPilot – Run on any cloud with cost optimization.
  • 🌍 MLC LLM – Universal deployment (iOS/Android) perfect for Ministral 3B.
  • 🧠 TensorRT-LLM – Optimized inference for Mistral Large 3 on NVIDIA GPUs.

Fine-Tuning & Training

Fine-Tuning Frameworks

  • 🧠 mistral-finetune – Official LoRA fine-tuning library.
  • 🌍 Axolotl – Streamlined LoRA/QLoRA/full fine-tuning.
  • 🌍 Unsloth ⭐ 20k+ – 2-5x faster fine-tuning, 80% less memory.
  • 🌍 Hugging Face PEFT – Parameter-Efficient Fine-Tuning.
  • 🌍 Hugging Face TRL – RLHF and DPO training.
  • 🌍 LLaMA-Factory ⭐ 35k+ – Unified fine-tuning framework.
  • 🌍 torchtune – PyTorch-native fine-tuning.

Training Infrastructure


Model Merging & Quantization

Model Merging

  • 🌍 MergeKit ⭐ 5k+ – Toolkit for merging LLMs (SLERP, TIES, DARE).
  • 🌍 LazyMergeKit – Colab notebook for easy merging.

Quantization Tools

  • 🌍 llama.cpp – GGUF quantization (Q4, Q5, Q8).
  • 🌍 AutoGPTQ – GPTQ quantization.
  • 🌍 AutoAWQ – AWQ quantization.
  • 🌍 bitsandbytes – 4-bit and 8-bit quantization.
  • 🌍 GGUF – Quantization format specification.

Agents & Orchestration

Agent Frameworks

  • 🌍 LangChain ⭐ 95k+ – LLM app framework with native Mistral support.
  • 🌍 LlamaIndex ⭐ 37k+ – Data framework for RAG with Mistral.
  • 🌍 CrewAI ⭐ 20k+ – Multi-agent orchestration.
  • 🌍 AutoGen ⭐ 35k+ – Microsoft's multi-agent framework.
  • 🌍 Semantic Kernel – Microsoft's AI orchestration SDK.
  • 🌍 Haystack – End-to-end NLP framework.
  • 🌍 PydanticAI – Type-safe AI agent framework.

Function Calling & Structured Output

  • 🧠 Mistral Function Calling – Native function calling docs.
  • 🌍 Instructor ⭐ 8k+ – Structured outputs with Pydantic.
  • 🌍 Outlines ⭐ 10k+ – Guaranteed structured generation.
  • 🌍 Marvin – AI functions with type hints.

Tooling & Dev Experience

IDE Extensions & Code Assistants

  • 🧠 Zed Extensions – Official Mistral for Zed editor.
  • 🌍 Continue ⭐ 20k+ – Open-source AI code assistant (VSCode/JetBrains).
  • 🌍 Tabby ⭐ 22k+ – Self-hosted GitHub Copilot alternative.
  • 🌍 Aider ⭐ 20k+ – AI pair programming in terminal.
  • 🌍 Cody – AI coding assistant with codebase context.

Development Tools

  • 🌍 LiteLLM ⭐ 15k+ – Unified API for 100+ LLMs.
  • 🌍 Promptfoo ⭐ 5k+ – LLM evaluation and red-teaming.
  • 🌍 Langfuse ⭐ 7k+ – Open-source LLM observability.
  • 🌍 Phoenix – ML observability for LLM apps.
  • 🌍 Weights & Biases – Experiment tracking with LLM support.

Community Projects

Chat Interfaces

  • 🌍 Open WebUI ⭐ 50k+ – Self-hosted ChatGPT-like UI.
  • 🌍 LibreChat ⭐ 20k+ – Multi-model chat interface.
  • 🌍 Lobe Chat ⭐ 50k+ – Modern extensible chat framework.
  • 🌍 Chatbot UI – Open-source ChatGPT clone.
  • 🌍 BetterChatGPT – Enhanced chat interface.

RAG & Knowledge Management

  • 🌍 PrivateGPT ⭐ 55k+ – Private document Q&A.
  • 🌍 Danswer ⭐ 12k+ – Enterprise Q&A over internal docs.
  • 🌍 Quivr ⭐ 37k+ – Personal knowledge base.
  • 🌍 Khoj – AI second brain.
  • 🌍 LocalGPT – Chat with documents locally.

Specialized Applications

  • 🌍 Fabric ⭐ 25k+ – AI augmentation framework.
  • 🌍 GPT Researcher ⭐ 15k+ – Autonomous research agent.
  • 🌍 OpenDevin ⭐ 35k+ – AI software engineer.

Demos & Examples

Official Examples

Community Examples


Tutorials & Guides

Getting Started

Fine-Tuning Tutorials

RAG & Applications


Benchmarks & Evaluation

Leaderboards

Evaluation Frameworks

Code Benchmarks

  • 🌍 HumanEval – Code generation benchmark.
  • 🌍 BigCodeBench – Comprehensive code evaluation.
  • 🌍 EvalPlus – Rigorous code evaluation.

Research & Papers

Mistral Technical Reports

Related Research

  • 🌍 Sliding Window Attention – Longformer attention mechanism.
  • 🌍 LoRA – Low-Rank Adaptation paper.
  • 🌍 QLoRA – Quantized LoRA for efficient fine-tuning.
  • 🌍 DPO – Direct Preference Optimization.
  • 🌍 Mixture of Experts – MoE foundations.

Talks & Media

Official Channels

Conferences & Talks


Ecosystem & Community

Cloud Providers

Community Hubs

Partnerships


Contributing

Contributions are welcome! Please read the contribution guidelines before submitting a pull request.

Quick Guidelines

  1. Ensure all links point to real, existing resources
  2. Use consistent formatting: - 🧠/🌍/πŸ§ͺ [Name](url) – Brief description.
  3. Prefer high-signal, actively maintained projects
  4. Include star counts for major projects (⭐ 10k+)

License

CC0

This work is licensed under CC0 1.0 Universal.

About

A curated list of awesome resources, tools, libraries, and projects for the Mistral AI ecosystem.

Topics

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors