codefuse-chatbot/examples/utils.py

45 lines
1.6 KiB
Python
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

import os
from configs.model_config import ONLINE_LLM_MODEL
from configs.server_config import FSCHAT_MODEL_WORKERS
from configs.model_config import llm_model_dict, LLM_DEVICE
from loguru import logger
def get_model_worker_config(
model_name: str = None,
fastchat_mdoel_workers: dict = FSCHAT_MODEL_WORKERS,
online_llm_model: dict = ONLINE_LLM_MODEL,
llm_model_dict: dict = llm_model_dict,
llm_device: str = LLM_DEVICE
) -> dict:
'''
加载model worker的配置项。
优先级:FSCHAT_MODEL_WORKERS[model_name] > ONLINE_LLM_MODEL[model_name] > FSCHAT_MODEL_WORKERS["default"]
'''
from coagent.service import model_workers
config = fastchat_mdoel_workers.get("default", {}).copy()
config.update(online_llm_model.get(model_name, {}).copy())
config.update(fastchat_mdoel_workers.get(model_name, {}).copy())
if model_name in online_llm_model:
config["online_api"] = True
if provider := config.get("provider"):
try:
config["worker_class"] = getattr(model_workers, provider)
except Exception as e:
msg = f"在线模型 {model_name} 的provider没有正确配置"
logger.error(f'{e.__class__.__name__}: {msg}')
# 本地模型
if model_name in llm_model_dict:
path = llm_model_dict[model_name]["local_model_path"]
config["model_path"] = path
if path and os.path.isdir(path):
config["model_path_exists"] = True
config["device"] = llm_device
# logger.debug(f"config: {config}")
return config