111 lines
3.8 KiB
Python
111 lines
3.8 KiB
Python
from fastchat.conversation import Conversation
|
||
from .base import *
|
||
from fastchat import conversation as conv
|
||
import sys
|
||
from typing import List, Dict, Iterator, Literal
|
||
from configs import logger, log_verbose
|
||
|
||
|
||
class ChatGLMWorker(ApiModelWorker):
|
||
DEFAULT_EMBED_MODEL = "text_embedding"
|
||
|
||
def __init__(
|
||
self,
|
||
*,
|
||
model_names: List[str] = ["zhipu-api"],
|
||
controller_addr: str = None,
|
||
worker_addr: str = None,
|
||
version: Literal["chatglm_turbo"] = "chatglm_turbo",
|
||
**kwargs,
|
||
):
|
||
kwargs.update(model_names=model_names, controller_addr=controller_addr, worker_addr=worker_addr)
|
||
kwargs.setdefault("context_len", 32768)
|
||
super().__init__(**kwargs)
|
||
self.version = version
|
||
|
||
def do_chat(self, params: ApiChatParams) -> Iterator[Dict]:
|
||
# TODO: 维护request_id
|
||
import zhipuai
|
||
|
||
params.load_config(self.model_names[0])
|
||
zhipuai.api_key = params.api_key
|
||
|
||
if log_verbose:
|
||
logger.info(f'{self.__class__.__name__}:params: {params}')
|
||
|
||
response = zhipuai.model_api.sse_invoke(
|
||
model=params.version,
|
||
prompt=params.messages,
|
||
temperature=params.temperature,
|
||
top_p=params.top_p,
|
||
incremental=False,
|
||
)
|
||
for e in response.events():
|
||
if e.event == "add":
|
||
yield {"error_code": 0, "text": e.data}
|
||
elif e.event in ["error", "interrupted"]:
|
||
data = {
|
||
"error_code": 500,
|
||
"text": str(e),
|
||
"error": {
|
||
"message": str(e),
|
||
"type": "invalid_request_error",
|
||
"param": None,
|
||
"code": None,
|
||
}
|
||
}
|
||
self.logger.error(f"请求智谱 API 时发生错误:{data}")
|
||
yield data
|
||
|
||
def do_embeddings(self, params: ApiEmbeddingsParams) -> Dict:
|
||
import zhipuai
|
||
|
||
params.load_config(self.model_names[0])
|
||
zhipuai.api_key = params.api_key
|
||
|
||
embeddings = []
|
||
try:
|
||
for t in params.texts:
|
||
response = zhipuai.model_api.invoke(model=params.embed_model or self.DEFAULT_EMBED_MODEL, prompt=t)
|
||
if response["code"] == 200:
|
||
embeddings.append(response["data"]["embedding"])
|
||
else:
|
||
self.logger.error(f"请求智谱 API 时发生错误:{response}")
|
||
return response # dict with code & msg
|
||
except Exception as e:
|
||
self.logger.error(f"请求智谱 API 时发生错误:{data}")
|
||
data = {"code": 500, "msg": f"对文本向量化时出错:{e}"}
|
||
return data
|
||
|
||
return {"code": 200, "data": embeddings}
|
||
|
||
def get_embeddings(self, params):
|
||
# TODO: 支持embeddings
|
||
print("embedding")
|
||
# print(params)
|
||
|
||
def make_conv_template(self, conv_template: str = None, model_path: str = None) -> Conversation:
|
||
# 这里的是chatglm api的模板,其它API的conv_template需要定制
|
||
return conv.Conversation(
|
||
name=self.model_names[0],
|
||
system_message="你是一个聪明的助手,请根据用户的提示来完成任务",
|
||
messages=[],
|
||
roles=["Human", "Assistant", "System"],
|
||
sep="\n###",
|
||
stop_str="###",
|
||
)
|
||
|
||
|
||
if __name__ == "__main__":
|
||
import uvicorn
|
||
from server.utils import MakeFastAPIOffline
|
||
from fastchat.serve.model_worker import app
|
||
|
||
worker = ChatGLMWorker(
|
||
controller_addr="http://127.0.0.1:20001",
|
||
worker_addr="http://127.0.0.1:21001",
|
||
)
|
||
sys.modules["fastchat.serve.model_worker"].worker = worker
|
||
MakeFastAPIOffline(app)
|
||
uvicorn.run(app, port=21001)
|