A curated list of awesome resources, tools, libraries, and projects for the Mistral AI ecosystem.
Mistral AI is a Paris-based AI company building open-weight, high-performance large language models. Founded in 2023, Mistral has quickly become a leading force in open-source AI, offering models that rival proprietary alternatives while remaining accessible to developers worldwide.
This repository maps and curates the entire Mistral.ai ecosystem for AI engineers, researchers, startup founders, and open-source contributors.
Legend:
- π§ Official Mistral AI
- π Community project
- π§ͺ Experimental
- What's New (March 2026)
- Why Mistral?
- Official Mistral Resources
- Models
- Community Fine-Tuned Models
- SDKs & APIs
- Inference & Deployment
- Fine-Tuning & Training
- Model Merging & Quantization
- Agents & Orchestration
- Tooling & Dev Experience
- Community Projects
- Demos & Examples
- Tutorials & Guides
- Benchmarks & Evaluation
- Research & Papers
- Talks & Media
- Ecosystem & Community
- Contributing
- License
- π Mistral Small 4 β Hybrid MoE (119B/6.5B active) unifying reasoning, coding, and multimodal in one model.
- π§ͺ Leanstral β First open-source Lean 4 formal proof agent.
- π’ Mistral Forge β Enterprise platform for training models on proprietary data.
- ποΈ Voxtral Mini 4B Realtime β Real-time speech-to-text with sub-200ms latency.
- π‘οΈ Mistral Moderation 2603 β Updated content moderation with jailbreaking detection.
- π₯οΈ Mistral Compute β European-hosted GPU cloud.
Mistral AI offers a compelling alternative in the LLM landscape:
| Aspect | Mistral Advantage |
|---|---|
| Open Weights | Models like Mistral Large 3, Small 4, and Ministral are fully open-weight (Apache 2.0), enabling local deployment and full control |
| Efficiency | Mistral Small 4 (119B/6.5B active) and Large 3 (675B/41B active) use MoE parameter routing for high efficiency; Ministral 3B/8B/14B are optimized for edge |
| European Sovereignty | Paris-based company offering GDPR-compliant, EU-hosted API options via Forge and Compute |
| Cost Efficiency | Competitive API pricing; open models enable free self-hosting |
| Innovation | Pioneered efficient MoE architectures, hybrid reasoning models, formal proof agents (Leanstral), and real-time speech AI |
| Full-Stack Platform | Forge (enterprise model training) + Compute (European GPU cloud) + le Chat (AI assistant) |
- π§ Mistral AI β Official company website with product information and announcements.
- π§ Mistral AI Documentation β Comprehensive API documentation, guides, and model specifications.
- π§ AI Studio (la Plateforme) β Developer console for API keys, model access, and agent management.
- π§ le Chat β AI assistant (web, iOS, Android) with free and Pro tiers.
- π§ Mistral Forge β Enterprise platform for training frontier-grade models on proprietary data.
- π§ Mistral Compute β European-hosted GPU cloud (NVIDIA Grace Blackwell).
- π§ Mistral AI GitHub β Official GitHub organization with 24+ repositories.
- π§ mistral-inference β 10k+ β Official inference library for running Mistral models.
- π§ mistral-finetune β 3k+ β Official lightweight LoRA-based fine-tuning library.
- π§ Mistral Cookbook β 2k+ β Official notebooks and examples for common use cases.
- π§ mistral-common β Official tokenization and pre-processing library.
- π§ Mistral Vibe β Native CLI coding assistant.
- π§ Platform Docs Public β Open-source documentation repository.
| Model | Context | License | Best For |
|---|---|---|---|
| Mistral Small 4 | 256k | Apache 2.0 | Hybrid reasoning + coding + multimodal (119B MoE / 6.5B active) |
| Mistral Large 3 | 256k | Apache 2.0 | Complex reasoning, multilingual, coding, vision (675B / 41B active) |
| Mistral Medium 3.1 | 128k | Proprietary | Prototype-to-production, balanced multimodal performance |
| Mistral Small 3.2 | 128k | Apache 2.0 | Low-latency, cost-sensitive applications (24B) |
| Mistral OCR 3 | β | Proprietary | Document parsing, table reconstruction ($2/1k pages) |
- π§ Mistral Small 4 β Hybrid MoE (119B / 6.5B active) unifying reasoning, coding, and multimodal. Configurable
reasoning_effort. - π§ Mistral Large 3 β Flagship MoE (675B) with state-of-the-art reasoning and vision.
- π§ Mistral Small 3.2 β High-performance dense 24B model (v3.2).
- π§ Magistral Small 1.2 β Specialized 24B reasoning model with multimodality.
- π§ Mixtral 8x22B β Legacy MoE workhorse (141B total / 39B active).
- π§ Ministral 14B β Dense edge model with vision (14B). Best-in-class at small scale.
- π§ Ministral 8B β High-performance edge model (8B).
- π§ Ministral 3B β Ultralight model for mobile/browser (3B).
- π§ Devstral 2 β 123B coding model (Modified MIT License). 72.2% SWE-bench Verified.
- π§ Devstral Small 2 β 24B coding model (Apache 2.0) for local agents.
- π§ Codestral 25.01 β Legacy code specialist.
- π§ Pixtral Large β 124B multimodal model building on Mistral Large 2.
- π§ Pixtral 12B β Efficient vision-language model.
- π§ Mistral OCR 3 β Advanced document understanding and table reconstruction.
- π§ Leanstral β First open-source Lean 4 formal proof agent (119B/6.5B active, Apache 2.0).
- π§ Voxtral Mini 4B Realtime β Real-time speech-to-text, sub-200ms latency, 13 languages (Apache 2.0).
- π§ Mistral Moderation 2603 β Content moderation with jailbreaking, dangerous, and criminal detection (3B, API only).
High-quality community fine-tunes built on Mistral base models:
- π OpenHermes-2.5-Mistral-7B β GPT-4 quality instruction-tuned by Teknium.
- π Zephyr-7B-beta β DPO-trained by HuggingFace H4, outperforms 70B on MT-Bench.
- π Nous-Hermes-2-Mistral-7B-DPO β DPO-enhanced with strong benchmark scores.
- π Hermes-2-Pro-Mistral-7B β Function calling and JSON mode specialist.
- π OpenChat-3.5-0106 β C-RLFT trained, ChatGPT-comparable performance.
- π Dolphin-2.8-Mistral-7B β Uncensored model by Eric Hartford.
- π MistralLite β AWS-optimized with 32k context window.
- π Mistral-7B-OpenOrca β Trained on OpenOrca dataset.
- π WizardMath-7B-V1.1 β Math-specialized Mistral fine-tune.
- π TheBloke β Extensive GGUF/AWQ/GPTQ quantized model repository.
- π bartowski β High-quality GGUF quantizations.
- π§ client-python β Official Python client library.
- π§ client-ts β Official TypeScript/JavaScript client library.
- π§ @mistralai/mistralai β Official TypeScript/JavaScript SDK (npm).
- π mistral.rs β Blazingly fast Rust inference with ISQ, LoRA, quantization.
- π mistral-go β Go client for Mistral AI API.
- π @ai-sdk/mistral β Vercel AI SDK provider.
- π @langchain/mistralai β LangChain.js integration.
- π§ mistral-common β Official tokenization and pre-processing library.
- π§ Mistral Vibe β Native CLI coding assistant powered by Devstral.
- π vLLM β 35k+ β High-throughput with PagedAttention. Excellent Mistral support.
- π Text Generation Inference β Hugging Face's production inference server.
- π llama.cpp β 70k+ β CPU/GPU inference with GGUF quantization.
- π ExLlamaV2 β Fast inference with EXL2 quantization.
- π SGLang β Fast serving with RadixAttention.
- π Ollama β 100k+ β Simple CLI for local Mistral models.
- π LM Studio β Desktop GUI for local LLMs.
- π Jan β Open-source ChatGPT alternative running locally.
- π GPT4All β Local inference with Mistral support.
- π Msty β Desktop app for running local LLMs.
- π LocalAI β 25k+ β OpenAI-compatible local API server.
- π SkyPilot β Run on any cloud with cost optimization.
- π MLC LLM β Universal deployment (iOS/Android) perfect for Ministral 3B.
- π§ TensorRT-LLM β Optimized inference for Mistral Large 3 on NVIDIA GPUs.
- π§ mistral-finetune β Official LoRA fine-tuning library.
- π Axolotl β Streamlined LoRA/QLoRA/full fine-tuning.
- π Unsloth β 20k+ β 2-5x faster fine-tuning, 80% less memory.
- π Hugging Face PEFT β Parameter-Efficient Fine-Tuning.
- π Hugging Face TRL β RLHF and DPO training.
- π LLaMA-Factory β 35k+ β Unified fine-tuning framework.
- π torchtune β PyTorch-native fine-tuning.
- π DeepSpeed β Distributed training optimization.
- π Hugging Face Accelerate β Simple distributed training.
- π MergeKit β 5k+ β Toolkit for merging LLMs (SLERP, TIES, DARE).
- π LazyMergeKit β Colab notebook for easy merging.
- π llama.cpp β GGUF quantization (Q4, Q5, Q8).
- π AutoGPTQ β GPTQ quantization.
- π AutoAWQ β AWQ quantization.
- π bitsandbytes β 4-bit and 8-bit quantization.
- π GGUF β Quantization format specification.
- π LangChain β 95k+ β LLM app framework with native Mistral support.
- π LlamaIndex β 37k+ β Data framework for RAG with Mistral.
- π CrewAI β 20k+ β Multi-agent orchestration.
- π AutoGen β 35k+ β Microsoft's multi-agent framework.
- π Semantic Kernel β Microsoft's AI orchestration SDK.
- π Haystack β End-to-end NLP framework.
- π PydanticAI β Type-safe AI agent framework.
- π§ Mistral Function Calling β Native function calling docs.
- π Instructor β 8k+ β Structured outputs with Pydantic.
- π Outlines β 10k+ β Guaranteed structured generation.
- π Marvin β AI functions with type hints.
- π§ Zed Extensions β Official Mistral for Zed editor.
- π Continue β 20k+ β Open-source AI code assistant (VSCode/JetBrains).
- π Tabby β 22k+ β Self-hosted GitHub Copilot alternative.
- π Aider β 20k+ β AI pair programming in terminal.
- π Cody β AI coding assistant with codebase context.
- π LiteLLM β 15k+ β Unified API for 100+ LLMs.
- π Promptfoo β 5k+ β LLM evaluation and red-teaming.
- π Langfuse β 7k+ β Open-source LLM observability.
- π Phoenix β ML observability for LLM apps.
- π Weights & Biases β Experiment tracking with LLM support.
- π Open WebUI β 50k+ β Self-hosted ChatGPT-like UI.
- π LibreChat β 20k+ β Multi-model chat interface.
- π Lobe Chat β 50k+ β Modern extensible chat framework.
- π Chatbot UI β Open-source ChatGPT clone.
- π BetterChatGPT β Enhanced chat interface.
- π PrivateGPT β 55k+ β Private document Q&A.
- π Danswer β 12k+ β Enterprise Q&A over internal docs.
- π Quivr β 37k+ β Personal knowledge base.
- π Khoj β AI second brain.
- π LocalGPT β Chat with documents locally.
- π Fabric β 25k+ β AI augmentation framework.
- π GPT Researcher β 15k+ β Autonomous research agent.
- π OpenDevin β 35k+ β AI software engineer.
- π§ Mistral Cookbook β RAG, function calling, embeddings, agents.
- π§ Fine-Tuning Guide β Official fine-tuning documentation.
- π§ API Examples β Complete API reference with examples.
- π Awesome-LLM β Curated LLM resources including Mistral.
- π LangChain Templates β Production-ready templates.
- π§ Mistral Quickstart β Official getting started guide.
- π§ Model Selection Guide β Choosing the right model.
- π Run Mistral Locally β Ollama setup guide.
- π§ Official Fine-Tuning β Mistral's fine-tuning guide.
- π Axolotl Mistral Examples β Config examples.
- π QLoRA Guide β 4-bit fine-tuning.
- π Unsloth Tutorial β Fast Mistral fine-tuning.
- π§ RAG with Mistral β Official RAG guide.
- π LlamaIndex + Mistral β RAG with LlamaIndex.
- π LangChain + Mistral β LangChain integration.
- π Open LLM Leaderboard β Hugging Face benchmarks.
- π Chatbot Arena β Human preference rankings.
- π Artificial Analysis β LLM quality and speed benchmarks.
- π lm-evaluation-harness β EleutherAI's eval framework.
- π HELM β Stanford's holistic evaluation.
- π OpenCompass β Comprehensive LLM evaluation.
- π HumanEval β Code generation benchmark.
- π BigCodeBench β Comprehensive code evaluation.
- π EvalPlus β Rigorous code evaluation.
- π§ Mistral 7B β Foundational 7B architecture paper.
- π§ Mixtral of Experts β Sparse MoE architecture.
- π§ Mistral Large 3 Blog β Technical announcement and benchmarks.
- π§ Mistral Small 4 Blog β Hybrid MoE architecture announcement.
- π§ Leanstral Blog β First open-source Lean 4 formal proof agent.
- π§ Forge Announcement β Enterprise model training platform.
- π§ Voxtral Blog β Real-time speech-to-text models.
- π§ Voxtral Mini Technical Report β Voxtral Mini 4B Realtime architecture paper.
- π Sliding Window Attention β Longformer attention mechanism.
- π LoRA β Low-Rank Adaptation paper.
- π QLoRA β Quantized LoRA for efficient fine-tuning.
- π DPO β Direct Preference Optimization.
- π Mixture of Experts β MoE foundations.
- π§ Mistral AI Blog β Official announcements.
- π§ le Chat β Official AI assistant.
- π§ Mistral AI Discord β Official community server.
- π§ Mistral AI Twitter/X β Official updates.
- π Hugging Face YouTube β Tutorials with Mistral.
- π AI Explained β Technical breakdowns.
- π Azure AI β Mistral on Azure AI Studio.
- π AWS Bedrock β Mistral via Amazon Bedrock.
- π Google Cloud Vertex AI β Mistral on GCP.
- π Groq β Ultra-fast Mistral inference.
- π Together AI β Mistral model hosting.
- π Replicate β Run Mistral via API.
- π Hugging Face Hub β Official model repository.
- π§ Mistral Discord β Official community.
- π r/LocalLLaMA β Local LLM community.
- π r/MistralAI β Mistral-focused subreddit.
- π§ Microsoft Azure Partnership β Strategic Azure partnership.
- π§ NVIDIA Nemotron Coalition β Founding member of NVIDIA's AI collaboration initiative.
- π§ AI Studio (la Plateforme) β Mistralβs developer and enterprise cloud platform.
- π§ Mistral Forge β Enterprise model training on proprietary data.
- π§ Mistral Compute β European-hosted GPU cloud infrastructure.
Contributions are welcome! Please read the contribution guidelines before submitting a pull request.
- Ensure all links point to real, existing resources
- Use consistent formatting:
- π§ /π/π§ͺ [Name](url) β Brief description. - Prefer high-signal, actively maintained projects
- Include star counts for major projects (β 10k+)
This work is licensed under CC0 1.0 Universal.