From bdb07753a30a280a152a2bdd29cfe35b87a071cb Mon Sep 17 00:00:00 2001 From: shaohuzhang1 Date: Thu, 24 Apr 2025 16:15:01 +0800 Subject: [PATCH] fix: Ollama maximum output token field --- .../impl/ollama_model_provider/credential/llm.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/apps/setting/models_provider/impl/ollama_model_provider/credential/llm.py b/apps/setting/models_provider/impl/ollama_model_provider/credential/llm.py index 0194d1f0d27..add06621937 100644 --- a/apps/setting/models_provider/impl/ollama_model_provider/credential/llm.py +++ b/apps/setting/models_provider/impl/ollama_model_provider/credential/llm.py @@ -25,7 +25,7 @@ class OllamaLLMModelParams(BaseForm): _step=0.01, precision=2) - max_tokens = forms.SliderField( + num_predict = forms.SliderField( TooltipLabel(_('Output the maximum Tokens'), _('Specify the maximum number of tokens that the model can generate')), required=True, default_value=1024,