[ollama conf] Adding num_predict
param for ollama models
#4271
Labels
area:configuration
Relates to configuration options
kind:enhancement
Indicates a new feature request, imrovement, or extension
"needs-triage"
Validations
Problem
I'm using ollama as the model provider, using the official Deepseek-R1 distilled 32B model on ollama.
The model config of continue is:
However, when I'm asking the model to edit a long piece of code, the response would cutoff suddenly.
I suspect that it is related to the ollama's limitation of the length of the model response, which is controlled by
num_predict
(see https://github.com/ollama/ollama/blob/main/docs/modelfile.md#parameter).although the ollama's default
num_predict
is infinite, I found that when I'm chatting with the model, the log shows the continue gives amaxTokens
of 4096:I think the
maxTokens
should also become a tunable parameter in the configuration of continue to allow for longer response. I really need to ask AI to help me to revise my LaTeX document!!Solution
The solution is just add
maxTokens
as a tunable parameter in the configuration of continue to allow for longer response :) Thank you for attention.The text was updated successfully, but these errors were encountered: