Self-host Ministral 3 14B on dedicated GPU clusters

Run Mistral's compact text and vision model on bare-metal Kubernetes in EU data centers. Images, prompts, and generated responses stay inside your infrastructure boundary.

Top GPU offerings

No model/quant candidates pass the quality filter.

Modality
Text in Image Text out
Context
256k tokens
License
Recommended GPU
Fits in roughly 32 GB (single H200/B200 FP8 node)
Ministral 3 14B Inference
Powered by Asergo
ID: ministral-14b-2512

Send a prompt or image to start.

Use Cases

Screenshot and evidence indexing

Turn recurring screenshots, scans, and photo attachments into searchable records without sending images to an external vision API. Ministral tags what is visible, writes short operational summaries, and pushes the result into internal search and case systems so teams stop triaging evidence by filename and memory.

UI screenshots
Inspection photos
Scanned attachments
Email evidence

Run this as a steady background job on the same boundary as your file store. It is a compact, repeatable indexing workload, which is exactly where Ministral's footprint is an advantage.

Attachment API
Object Storage
Ministral 3 14B
Search Indexer
Search
Case Timeline
Ops Queue