215 lines
9.5 KiB
Python
215 lines
9.5 KiB
Python
|
from typing import List, Union, Dict, Tuple
|
|||
|
import os
|
|||
|
import json
|
|||
|
import importlib
|
|||
|
import copy
|
|||
|
from loguru import logger
|
|||
|
|
|||
|
from dev_opsgpt.connector.agents import BaseAgent
|
|||
|
from dev_opsgpt.connector.chains import BaseChain
|
|||
|
from dev_opsgpt.tools.base_tool import BaseTools, Tool
|
|||
|
from dev_opsgpt.connector.shcema.memory import Memory
|
|||
|
from dev_opsgpt.connector.connector_schema import (
|
|||
|
Task, Env, Role, Message, Doc, Docs, AgentConfig, ChainConfig, PhaseConfig, CodeDoc,
|
|||
|
load_chain_configs, load_phase_configs, load_role_configs
|
|||
|
)
|
|||
|
from dev_opsgpt.connector.configs import AGETN_CONFIGS, CHAIN_CONFIGS, PHASE_CONFIGS
|
|||
|
from dev_opsgpt.tools import DDGSTool, DocRetrieval, CodeRetrieval
|
|||
|
|
|||
|
|
|||
|
role_configs = load_role_configs(AGETN_CONFIGS)
|
|||
|
chain_configs = load_chain_configs(CHAIN_CONFIGS)
|
|||
|
phase_configs = load_phase_configs(PHASE_CONFIGS)
|
|||
|
|
|||
|
|
|||
|
CUR_DIR = os.path.dirname(os.path.abspath(__file__))
|
|||
|
|
|||
|
|
|||
|
class BasePhase:
|
|||
|
|
|||
|
def __init__(
|
|||
|
self,
|
|||
|
phase_name: str,
|
|||
|
task: Task = None,
|
|||
|
do_summary: bool = False,
|
|||
|
do_search: bool = False,
|
|||
|
do_doc_retrieval: bool = False,
|
|||
|
do_code_retrieval: bool = False,
|
|||
|
do_tool_retrieval: bool = False,
|
|||
|
phase_config: Union[dict, str] = PHASE_CONFIGS,
|
|||
|
chain_config: Union[dict, str] = CHAIN_CONFIGS,
|
|||
|
role_config: Union[dict, str] = AGETN_CONFIGS,
|
|||
|
) -> None:
|
|||
|
self.conv_summary_agent = BaseAgent(role=role_configs["conv_summary"].role,
|
|||
|
task = None,
|
|||
|
memory = None,
|
|||
|
do_search = role_configs["conv_summary"].do_search,
|
|||
|
do_doc_retrieval = role_configs["conv_summary"].do_doc_retrieval,
|
|||
|
do_tool_retrieval = role_configs["conv_summary"].do_tool_retrieval,
|
|||
|
do_filter=False, do_use_self_memory=False)
|
|||
|
|
|||
|
self.chains: List[BaseChain] = self.init_chains(
|
|||
|
phase_name,
|
|||
|
task=task,
|
|||
|
memory=None,
|
|||
|
phase_config = phase_config,
|
|||
|
chain_config = chain_config,
|
|||
|
role_config = role_config,
|
|||
|
)
|
|||
|
self.phase_name = phase_name
|
|||
|
self.do_summary = do_summary
|
|||
|
self.do_search = do_search
|
|||
|
self.do_code_retrieval = do_code_retrieval
|
|||
|
self.do_doc_retrieval = do_doc_retrieval
|
|||
|
self.do_tool_retrieval = do_tool_retrieval
|
|||
|
|
|||
|
self.global_message = Memory([])
|
|||
|
# self.chain_message = Memory([])
|
|||
|
self.phase_memory: List[Memory] = []
|
|||
|
|
|||
|
def step(self, query: Message, history: Memory = None) -> Tuple[Message, Memory]:
|
|||
|
summary_message = None
|
|||
|
chain_message = Memory([])
|
|||
|
local_memory = Memory([])
|
|||
|
# do_search、do_doc_search、do_code_search
|
|||
|
query = self.get_extrainfo_step(query)
|
|||
|
input_message = copy.deepcopy(query)
|
|||
|
|
|||
|
self.global_message.append(input_message)
|
|||
|
for chain in self.chains:
|
|||
|
# chain can supply background and query to next chain
|
|||
|
output_message, chain_memory = chain.step(input_message, history, background=chain_message)
|
|||
|
output_message = self.inherit_extrainfo(input_message, output_message)
|
|||
|
input_message = output_message
|
|||
|
logger.info(f"{chain.chainConfig.chain_name} phase_step: {output_message.role_content}")
|
|||
|
|
|||
|
self.global_message.append(output_message)
|
|||
|
local_memory.extend(chain_memory)
|
|||
|
|
|||
|
# whether use summary_llm
|
|||
|
if self.do_summary:
|
|||
|
logger.info(f"{self.conv_summary_agent.role.role_name} input global memory: {self.global_message.to_str_messages(content_key='step_content')}")
|
|||
|
logger.info(f"{self.conv_summary_agent.role.role_name} input global memory: {self.global_message.to_str_messages(content_key='role_content')}")
|
|||
|
summary_message = self.conv_summary_agent.run(query, background=self.global_message)
|
|||
|
summary_message.role_name = chain.chainConfig.chain_name
|
|||
|
summary_message = self.conv_summary_agent.parser(summary_message)
|
|||
|
summary_message = self.conv_summary_agent.filter(summary_message)
|
|||
|
summary_message = self.inherit_extrainfo(output_message, summary_message)
|
|||
|
chain_message.append(summary_message)
|
|||
|
|
|||
|
# 由于不会存在多轮chain执行,所以直接保留memory即可
|
|||
|
for chain in self.chains:
|
|||
|
self.phase_memory.append(chain.global_memory)
|
|||
|
|
|||
|
message = summary_message or output_message
|
|||
|
message.role_name = self.phase_name
|
|||
|
# message.db_docs = query.db_docs
|
|||
|
# message.code_docs = query.code_docs
|
|||
|
# message.search_docs = query.search_docs
|
|||
|
return summary_message or output_message, local_memory
|
|||
|
|
|||
|
def init_chains(self, phase_name, phase_config, chain_config,
|
|||
|
role_config, task=None, memory=None) -> List[BaseChain]:
|
|||
|
# load config
|
|||
|
role_configs = load_role_configs(role_config)
|
|||
|
chain_configs = load_chain_configs(chain_config)
|
|||
|
phase_configs = load_phase_configs(phase_config)
|
|||
|
|
|||
|
chains = []
|
|||
|
self.chain_module = importlib.import_module("dev_opsgpt.connector.chains")
|
|||
|
self.agent_module = importlib.import_module("dev_opsgpt.connector.agents")
|
|||
|
phase = phase_configs.get(phase_name)
|
|||
|
for chain_name in phase.chains:
|
|||
|
logger.info(f"chain_name: {chain_name}")
|
|||
|
# chain_class = getattr(self.chain_module, chain_name)
|
|||
|
logger.debug(f"{chain_configs.keys()}")
|
|||
|
chain_config = chain_configs[chain_name]
|
|||
|
|
|||
|
agents = [
|
|||
|
getattr(self.agent_module, role_configs[agent_name].role.agent_type)(
|
|||
|
role_configs[agent_name].role,
|
|||
|
task = task,
|
|||
|
memory = memory,
|
|||
|
chat_turn=role_configs[agent_name].chat_turn,
|
|||
|
do_search = role_configs[agent_name].do_search,
|
|||
|
do_doc_retrieval = role_configs[agent_name].do_doc_retrieval,
|
|||
|
do_tool_retrieval = role_configs[agent_name].do_tool_retrieval,
|
|||
|
)
|
|||
|
for agent_name in chain_config.agents
|
|||
|
]
|
|||
|
chain_instance = BaseChain(
|
|||
|
chain_config, agents, chain_config.chat_turn,
|
|||
|
do_checker=chain_configs[chain_name].do_checker,
|
|||
|
do_code_exec=False,)
|
|||
|
chains.append(chain_instance)
|
|||
|
|
|||
|
return chains
|
|||
|
|
|||
|
def get_extrainfo_step(self, input_message):
|
|||
|
if self.do_doc_retrieval:
|
|||
|
input_message = self.get_doc_retrieval(input_message)
|
|||
|
|
|||
|
logger.debug(F"self.do_code_retrieval: {self.do_code_retrieval}")
|
|||
|
if self.do_code_retrieval:
|
|||
|
input_message = self.get_code_retrieval(input_message)
|
|||
|
|
|||
|
if self.do_search:
|
|||
|
input_message = self.get_search_retrieval(input_message)
|
|||
|
|
|||
|
return input_message
|
|||
|
|
|||
|
def inherit_extrainfo(self, input_message: Message, output_message: Message):
|
|||
|
output_message.db_docs = input_message.db_docs
|
|||
|
output_message.search_docs = input_message.search_docs
|
|||
|
output_message.code_docs = input_message.code_docs
|
|||
|
output_message.figures.update(input_message.figures)
|
|||
|
return output_message
|
|||
|
|
|||
|
def get_search_retrieval(self, message: Message,) -> Message:
|
|||
|
SEARCH_ENGINES = {"duckduckgo": DDGSTool}
|
|||
|
search_docs = []
|
|||
|
for idx, doc in enumerate(SEARCH_ENGINES["duckduckgo"].run(message.role_content, 3)):
|
|||
|
doc.update({"index": idx})
|
|||
|
search_docs.append(Doc(**doc))
|
|||
|
message.search_docs = search_docs
|
|||
|
return message
|
|||
|
|
|||
|
def get_doc_retrieval(self, message: Message) -> Message:
|
|||
|
query = message.role_content
|
|||
|
knowledge_basename = message.doc_engine_name
|
|||
|
top_k = message.top_k
|
|||
|
score_threshold = message.score_threshold
|
|||
|
if knowledge_basename:
|
|||
|
docs = DocRetrieval.run(query, knowledge_basename, top_k, score_threshold)
|
|||
|
message.db_docs = [Doc(**doc) for doc in docs]
|
|||
|
return message
|
|||
|
|
|||
|
def get_code_retrieval(self, message: Message) -> Message:
|
|||
|
# DocRetrieval.run("langchain是什么", "DSADSAD")
|
|||
|
query = message.input_query
|
|||
|
code_engine_name = message.code_engine_name
|
|||
|
history_node_list = message.history_node_list
|
|||
|
code_docs = CodeRetrieval.run(code_engine_name, query, code_limit=message.top_k, history_node_list=history_node_list)
|
|||
|
message.code_docs = [CodeDoc(**doc) for doc in code_docs]
|
|||
|
return message
|
|||
|
|
|||
|
def get_tool_retrieval(self, message: Message) -> Message:
|
|||
|
return message
|
|||
|
|
|||
|
def update(self) -> Memory:
|
|||
|
pass
|
|||
|
|
|||
|
def get_memory(self, ) -> Memory:
|
|||
|
return Memory.from_memory_list(
|
|||
|
[chain.get_memory() for chain in self.chains]
|
|||
|
)
|
|||
|
|
|||
|
def get_memory_str(self, do_all_memory=True, content_key="role_content") -> str:
|
|||
|
memory = self.global_message if do_all_memory else self.phase_memory
|
|||
|
return "\n".join([": ".join(i) for i in memory.to_tuple_messages(content_key=content_key)])
|
|||
|
|
|||
|
def get_chains_memory(self, content_key="role_content") -> List[Tuple]:
|
|||
|
return [memory.to_tuple_messages(content_key=content_key) for memory in self.phase_memory]
|
|||
|
|
|||
|
def get_chains_memory_str(self, content_key="role_content") -> str:
|
|||
|
return "************".join([f"{chain.chainConfig.chain_name}\n" + chain.get_memory_str(content_key=content_key) for chain in self.chains])
|