Mistral-Small is a balanced, efficient language model (llm) offering high performance across various tasks with lower latency and broad application potential
The maximum number of tokens to generate. Shorter token lengths will provide faster performance.
A decimal number that determines the degree of randomness in the response
An alternative to sampling with temperature, where the model considers the results of the tokens with top_p probability mass.
A system prompt sets the behavior and context for an AI assistant in a conversation, such as modifying its personality.
ResetModel loading...
Output
Notes
ID
Model Type ID
Text To Text
Input Type
text
Output Type
text
Description
Mistral-Small is a balanced, efficient language model (llm) offering high performance across various tasks with lower latency and broad application potential