45 lines
1.6 KiB
Python
45 lines
1.6 KiB
Python
import os
|
||
|
||
from configs.model_config import ONLINE_LLM_MODEL
|
||
from configs.server_config import FSCHAT_MODEL_WORKERS
|
||
from configs.model_config import llm_model_dict, LLM_DEVICE
|
||
|
||
from loguru import logger
|
||
|
||
|
||
|
||
def get_model_worker_config(
|
||
model_name: str = None,
|
||
fastchat_mdoel_workers: dict = FSCHAT_MODEL_WORKERS,
|
||
online_llm_model: dict = ONLINE_LLM_MODEL,
|
||
llm_model_dict: dict = llm_model_dict,
|
||
llm_device: str = LLM_DEVICE
|
||
) -> dict:
|
||
'''
|
||
加载model worker的配置项。
|
||
优先级:FSCHAT_MODEL_WORKERS[model_name] > ONLINE_LLM_MODEL[model_name] > FSCHAT_MODEL_WORKERS["default"]
|
||
'''
|
||
from coagent.service import model_workers
|
||
|
||
config = fastchat_mdoel_workers.get("default", {}).copy()
|
||
config.update(online_llm_model.get(model_name, {}).copy())
|
||
config.update(fastchat_mdoel_workers.get(model_name, {}).copy())
|
||
|
||
if model_name in online_llm_model:
|
||
config["online_api"] = True
|
||
if provider := config.get("provider"):
|
||
try:
|
||
config["worker_class"] = getattr(model_workers, provider)
|
||
except Exception as e:
|
||
msg = f"在线模型 ‘{model_name}’ 的provider没有正确配置"
|
||
logger.error(f'{e.__class__.__name__}: {msg}')
|
||
# 本地模型
|
||
if model_name in llm_model_dict:
|
||
path = llm_model_dict[model_name]["local_model_path"]
|
||
config["model_path"] = path
|
||
if path and os.path.isdir(path):
|
||
config["model_path_exists"] = True
|
||
config["device"] = llm_device
|
||
|
||
# logger.debug(f"config: {config}")
|
||
return config |