Flash Sale 50% Off!

Don't miss out on our amazing 50% flash sale. Limited time only!

Sale ends in:

Get an additional 10% discount on any plan!

SPECIAL10
See Pricing
×

Daily Limit Reached

You have exhausted your limit of free daily generations. To get more free generations, consider upgrading to our unlimited plan for $4/month or come back tomorrow.

Get an additional 10% discount on any plan!

SPECIAL10
Upgrade Now
Save $385/Month - Unlock All AI Tools

Upgrade to Premium

Thank you for creating an account! To continue using AI4Chat's premium features, please upgrade to a paid plan.

Access to all premium features
Priority customer support
Regular updates and new features - See our changelog
View Pricing Plans
7-Day Money Back Guarantee
Not satisfied? Get a full refund, no questions asked.
×

Credits Exhausted

You have used up all your available credits. Upgrade to a paid plan to get more credits and continue generating content.

Upgrade Now

You do not have enough credits to generate this output.

High Reasoning

Mistral Small 3

Mistral Small 3 is a high-efficiency 24B AI model excelling in 80% of generative tasks with robust language understanding, superior 81% MMLU accuracy, and blazing 150 tokens/second latency for fast conversational assistance and local deployment. Perfect for low-latency function calling, fine-tuning into domain experts, and private inference on a single RTX 4090 or MacBook with 32GB RAM.

128k Context
High Intelligence
Oct '23 Knowledge

Available for Chat, Vision, and File Uploads.

Performance Benchmarks

MMLU
81.0%
HumanEval
88.4%
GPQA
46.0%

How do you want to interact?

Start a Conversation

Ask anything.
Have a natural conversation, brainstorm ideas, draft emails, or ask for advice.

Start Chatting

Use a Persona

Specialized Experts.
Instruct the AI to act as a Coding Tutor, Marketing Expert, or Travel Guide.

Pick a Persona

Why use Mistral Small 3?

Fast-response conversational assistance

Excels in quick, accurate responses for virtual assistants and real-time interactions

Low-latency function calling

Handles rapid function execution for automated and agentic workflows

Local inference

Runs efficiently on single GPU like RTX 4090 or MacBook with 32GB RAM for privacy

Capability Examples

Fast Conversational Assistance
Hey, quickly summarize the plot of Romeo and Juliet in 3 bullet points.
- Two star-crossed lovers from feuding families in Verona meet and fall in love. - They secretly marry but family rivalry leads to tragic misunderstandings and deaths. - Romeo and Juliet die, ending the feud between the Montagues and Capulets.
Low-Latency Function Calling
Call the weather API for current temperature in Paris and format as JSON.
{"function": "get_weather", "parameters": {"location": "Paris"}, "result": {"temperature": 22, "unit": "C", "condition": "sunny"}}

How to use

1
Go to Chat

Navigate to the "AI Chat" page.

2
Select Model

Ensure Mistral Small 3 is selected.

3
Type Prompt

Ask a question or paste code.

4
Interact

Refine the answer by replying to the AI.

Compare LLMs Side-by-Side

Is Mistral Small 3 better than Claude 3.5 or Gemini? Test same prompts simultaneously in the Chat Playground.

Open Chat Playground

Made with ❤ by AI4Chat