Fine-tune & Run Mistral Small 3.1
Blog post from Unsloth
Mistral Small 3.1 is a new multimodal model capable of understanding both text and vision with a 128K context length, boasting 24 billion parameters that outperform GP4o on several benchmarks. Unsloth has optimized this model for fine-tuning, making it 1.8 times faster and reducing VRAM usage by 70%, while supporting longer contexts compared to environments using Flash Attention 2 on a 48GB GPU. The model and its various versions are available on Hugging Face, and fine-tuning can be conducted using Unsloth's dynamic 4-bit quants for better accuracy. Users can train their own reasoning models with Mistral Small 3 and access detailed instructions and resources through Unsloth's documentation. The announcement encourages joining their community on platforms like Reddit, Discord, and Twitter for support and updates.