DeepSeek-Coder-33B-Instruct model is a SOTA 33 billion parameter code generation model, fine-tuned on 2 billion tokens of instruction data, offering superior performance in code completion and infilling tasks across more than 80 programming languages.
The maximum number of tokens to generate. Shorter token lengths will provide faster performance.
A decimal number that determines the degree of randomness in the response
An alternative to sampling with temperature, where the model considers the results of the tokens with top_p probability mass.
The top_k parameter is used to limit the number of choices for the next predicted word or token.
ResetModel loading...
Output
Notes
ID
Model Type ID
Text To Text
Input Type
text
Output Type
text
Description
DeepSeek-Coder-33B-Instruct model is a SOTA 33 billion parameter code generation model, fine-tuned on 2 billion tokens of instruction data, offering superior performance in code completion and infilling tasks across more than 80 programming languages.