From a071d7c89b6f32b1b2456ec827760ac44962d4d0 Mon Sep 17 00:00:00 2001 From: wxg0103 <727495428@qq.com> Date: Mon, 17 Feb 2025 17:21:10 +0800 Subject: [PATCH] fix: Alibaba Cloud full modal docking error MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --bug=1052232 --user=王孝刚 【模型设置】-大语言模型/语音识别模型 对接不上阿里云的全模态模型 https://www.tapd.cn/57709429/s/1655034 --- .../impl/aliyun_bai_lian_model_provider/model/llm.py | 3 ++- .../impl/aliyun_bai_lian_model_provider/model/stt.py | 2 ++ 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/apps/setting/models_provider/impl/aliyun_bai_lian_model_provider/model/llm.py b/apps/setting/models_provider/impl/aliyun_bai_lian_model_provider/model/llm.py index 98f783ce5..6484e649e 100644 --- a/apps/setting/models_provider/impl/aliyun_bai_lian_model_provider/model/llm.py +++ b/apps/setting/models_provider/impl/aliyun_bai_lian_model_provider/model/llm.py @@ -5,7 +5,6 @@ from typing import Dict from setting.models_provider.base_model_provider import MaxKBBaseModel from setting.models_provider.impl.base_chat_open_ai import BaseChatOpenAI - class BaiLianChatModel(MaxKBBaseModel, BaseChatOpenAI): @staticmethod def is_cache_model(): @@ -14,6 +13,8 @@ class BaiLianChatModel(MaxKBBaseModel, BaseChatOpenAI): @staticmethod def new_instance(model_type, model_name, model_credential: Dict[str, object], **model_kwargs): optional_params = MaxKBBaseModel.filter_optional_params(model_kwargs) + if model_name == 'qwen-omni-turbo': + optional_params['streaming'] = True return BaiLianChatModel( model=model_name, openai_api_base=model_credential.get('api_base'), diff --git a/apps/setting/models_provider/impl/aliyun_bai_lian_model_provider/model/stt.py b/apps/setting/models_provider/impl/aliyun_bai_lian_model_provider/model/stt.py index e21f16376..ad17a32a9 100644 --- a/apps/setting/models_provider/impl/aliyun_bai_lian_model_provider/model/stt.py +++ b/apps/setting/models_provider/impl/aliyun_bai_lian_model_provider/model/stt.py @@ -26,6 +26,8 @@ class AliyunBaiLianSpeechToText(MaxKBBaseModel, BaseSpeechToText): optional_params['max_tokens'] = model_kwargs['max_tokens'] if 'temperature' in model_kwargs and model_kwargs['temperature'] is not None: optional_params['temperature'] = model_kwargs['temperature'] + if model_name == 'qwen-omni-turbo': + optional_params['streaming'] = True return AliyunBaiLianSpeechToText( model=model_name, api_key=model_credential.get('api_key'),