SOLAR-10.7B-Instruct is a powerful 10.7 billion-parameter LLM with a unique depth up-scaling architecture, excelling in single-turn conversation tasks through advanced instruction fine-tuning methods.
The maximum number of tokens to generate. Shorter token lengths will provide faster performance.
A decimal number that determines the degree of randomness in the response
An alternative to sampling with temperature, where the model considers the results of the tokens with top_p probability mass.
The top_k parameter is used to limit the number of choices for the next predicted word or token.
ResetModel loading...
Output
Notes
ID
Model Type ID
Text To Text
Input Type
text
Output Type
text
Description
SOLAR-10.7B-Instruct is a powerful 10.7 billion-parameter LLM with a unique depth up-scaling architecture, excelling in single-turn conversation tasks through advanced instruction fine-tuning methods.