From d6f1d25b59eef2b4657dfe98a70713b4692dfb7a Mon Sep 17 00:00:00 2001 From: shaohuzhang1 <80892890+shaohuzhang1@users.noreply.github.com> Date: Tue, 22 Apr 2025 18:10:49 +0800 Subject: [PATCH] fix: Change the max_token of the Qianfan large model to max_output_token (#2955) --- .../impl/wenxin_model_provider/credential/llm.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/apps/setting/models_provider/impl/wenxin_model_provider/credential/llm.py b/apps/setting/models_provider/impl/wenxin_model_provider/credential/llm.py index 06ec94aae..d4d379db3 100644 --- a/apps/setting/models_provider/impl/wenxin_model_provider/credential/llm.py +++ b/apps/setting/models_provider/impl/wenxin_model_provider/credential/llm.py @@ -27,7 +27,7 @@ class WenxinLLMModelParams(BaseForm): _step=0.01, precision=2) - max_tokens = forms.SliderField( + max_output_tokens = forms.SliderField( TooltipLabel(_('Output the maximum Tokens'), _('Specify the maximum number of tokens that the model can generate')), required=True, default_value=1024,