mistralai/ministral-8b-2512A balanced model in the Ministral 3 family, Ministral 3 8B is a powerful, efficient tiny language model with vision capabilities.
A balanced model in the Ministral 3 family, Ministral 3 8B is a powerful, efficient tiny language model with vision capabilities.
This model includes different quantization levels of the instruct post-trained version in GGUF, fine-tuned for instruction tasks, making it ideal for chat and instruction based use cases.
The Ministral 3 family is designed for edge deployment, capable of running on a wide range of hardware. Ministral 3 8B can even be deployed locally, capable of fitting in 12GB of VRAM in FP8, and less if further quantized.
Learn more in our blog post and paper.
We recommend deploying with the following best practices:
This model is licensed under the Apache 2.0 License.
You must not use this model in a manner that infringes, misappropriates, or otherwise violates any third party’s rights, including intellectual property rights.| Mode | chat |
| Context Window | 262K tokens |
| Max Output | 262K tokens |
| Function Calling | Supported |
| Vision | Supported |
| Reasoning | - |
| Web Search | - |
| Url Context | - |
| Base Model | mistralai/Ministral-3-8B-Instruct-2512 |
| Languages | en, fr, es, de, it, pt, nl, zh, ja, ko, ar |
| Library | vllm |
from openai import OpenAI
client = OpenAI(
base_url="https://api.haimaker.ai/v1",
api_key="YOUR_API_KEY",
)
response = client.chat.completions.create(
model="mistralai/ministral-8b-2512",
messages=[
{"role": "user", "content": "Hello, how are you?"}
],
)
print(response.choices[0].message.content)OpenAI-compatible endpoint. Start building in minutes.