mirror of
https://github.com/1Panel-dev/MaxKB.git
synced 2025-12-26 01:33:05 +00:00
langchain-openai is not compatible with parameter Settings in ollama, such as num_ctx. Therefore, you need to create model instances using langchain-ollama
(cherry picked from commit 42ae7b443d)
This commit is contained in:
parent
a64adc2504
commit
ed8f8f8a3e
|
|
@ -10,7 +10,8 @@ from typing import List, Dict
|
||||||
from urllib.parse import urlparse, ParseResult
|
from urllib.parse import urlparse, ParseResult
|
||||||
|
|
||||||
from langchain_core.messages import BaseMessage, get_buffer_string
|
from langchain_core.messages import BaseMessage, get_buffer_string
|
||||||
from langchain_openai.chat_models import ChatOpenAI
|
from langchain_ollama.chat_models import ChatOllama
|
||||||
|
|
||||||
|
|
||||||
from common.config.tokenizer_manage_config import TokenizerManage
|
from common.config.tokenizer_manage_config import TokenizerManage
|
||||||
from setting.models_provider.base_model_provider import MaxKBBaseModel
|
from setting.models_provider.base_model_provider import MaxKBBaseModel
|
||||||
|
|
@ -24,7 +25,7 @@ def get_base_url(url: str):
|
||||||
return result_url[:-1] if result_url.endswith("/") else result_url
|
return result_url[:-1] if result_url.endswith("/") else result_url
|
||||||
|
|
||||||
|
|
||||||
class OllamaChatModel(MaxKBBaseModel, ChatOpenAI):
|
class OllamaChatModel(MaxKBBaseModel, ChatOllama):
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def is_cache_model():
|
def is_cache_model():
|
||||||
return False
|
return False
|
||||||
|
|
@ -33,12 +34,10 @@ class OllamaChatModel(MaxKBBaseModel, ChatOpenAI):
|
||||||
def new_instance(model_type, model_name, model_credential: Dict[str, object], **model_kwargs):
|
def new_instance(model_type, model_name, model_credential: Dict[str, object], **model_kwargs):
|
||||||
api_base = model_credential.get('api_base', '')
|
api_base = model_credential.get('api_base', '')
|
||||||
base_url = get_base_url(api_base)
|
base_url = get_base_url(api_base)
|
||||||
base_url = base_url if base_url.endswith('/v1') else (base_url + '/v1')
|
|
||||||
optional_params = MaxKBBaseModel.filter_optional_params(model_kwargs)
|
optional_params = MaxKBBaseModel.filter_optional_params(model_kwargs)
|
||||||
|
|
||||||
return OllamaChatModel(model=model_name, openai_api_base=base_url,
|
return OllamaChatModel(model=model_name, base_url=base_url,
|
||||||
openai_api_key=model_credential.get('api_key'),
|
stream=True, **optional_params)
|
||||||
stream_usage=True, **optional_params)
|
|
||||||
|
|
||||||
def get_num_tokens_from_messages(self, messages: List[BaseMessage]) -> int:
|
def get_num_tokens_from_messages(self, messages: List[BaseMessage]) -> int:
|
||||||
tokenizer = TokenizerManage.get_tokenizer()
|
tokenizer = TokenizerManage.get_tokenizer()
|
||||||
|
|
|
||||||
|
|
@ -28,6 +28,7 @@ pycryptodome = "^3.19.0"
|
||||||
beautifulsoup4 = "^4.12.2"
|
beautifulsoup4 = "^4.12.2"
|
||||||
html2text = "^2024.2.26"
|
html2text = "^2024.2.26"
|
||||||
langchain-openai = "^0.1.8"
|
langchain-openai = "^0.1.8"
|
||||||
|
langchain-ollama = "0.1.3"
|
||||||
django-ipware = "^6.0.4"
|
django-ipware = "^6.0.4"
|
||||||
django-apscheduler = "^0.6.2"
|
django-apscheduler = "^0.6.2"
|
||||||
pymupdf = "1.24.9"
|
pymupdf = "1.24.9"
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue