Llama-3-Instruct is an advanced, scalable llm designed for diverse applications, offering state-of-the-art performance in coding, reasoning, and multi-use conversational capabilities.
The maximum number of tokens to generate. Shorter token lengths will provide faster performance.
A decimal number that determines the degree of randomness in the response
An alternative to sampling with temperature, where the model considers the results of the tokens with top_p probability mass.
The top-k parameter limits the model's predictions to the top k most probable tokens at each step of generation.
ResetModel loading...
Output
Notes
ID
Model Type ID
Text To Text
Input Type
text
Output Type
text
Description
Llama-3-Instruct is an advanced, scalable llm designed for diverse applications, offering state-of-the-art performance in coding, reasoning, and multi-use conversational capabilities.