From 663a6cad43d818348613fda80af19bece5e44222 Mon Sep 17 00:00:00 2001 From: shaohuzhang1 <80892890+shaohuzhang1@users.noreply.github.com> Date: Thu, 23 Jan 2025 17:12:58 +0800 Subject: [PATCH] fix: Failed to save the Qianfan large model (#2092) --- .../impl/wenxin_model_provider/credential/llm.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/apps/setting/models_provider/impl/wenxin_model_provider/credential/llm.py b/apps/setting/models_provider/impl/wenxin_model_provider/credential/llm.py index 36df5e197..66f6c0e2f 100644 --- a/apps/setting/models_provider/impl/wenxin_model_provider/credential/llm.py +++ b/apps/setting/models_provider/impl/wenxin_model_provider/credential/llm.py @@ -45,7 +45,7 @@ class WenxinLLMModelCredential(BaseForm, BaseModelCredential): gettext('{model_type} Model type is not supported').format(model_type=model_type)) model = provider.get_model(model_type, model_name, model_credential, **model_params) model_info = [model.lower() for model in model.client.models()] - if not model_info.__containsgettext(model_name.lower()): + if not model_info.__contains__(model_name.lower()): raise AppApiException(ValidCode.valid_error.value, gettext('{model_name} The model does not support').format(model_name=model_name)) for key in ['api_key', 'secret_key']: