diff --git a/README.md b/README.md index 907cda4..4166303 100644 --- a/README.md +++ b/README.md @@ -139,6 +139,7 @@ git lfs clone https://huggingface.co/THUDM/chatglm2-6b # install Embedding-model git lfs clone https://huggingface.co/shibing624/text2vec-base-chinese +cp ~/shibing624/text2vec-base-chinese ~/codefuse-chatbot/embedding_models/ ``` @@ -155,18 +156,21 @@ os.environ["OPENAI_API_KEY"] = "sk-xxx" # 可自行替换自己需要的api_base_url os.environ["API_BASE_URL"] = "https://api.openai.com/v1" -# vi model_config#95 你需要选择的语言模型 +# vi model_config#105 你需要选择的语言模型 LLM_MODEL = "gpt-3.5-turbo" -# vi model_config#33 你需要选择的向量模型 +# vi model_config#43 你需要选择的向量模型 EMBEDDING_MODEL = "text2vec-base" -# vi model_config#19 修改成你的本地路径,如果能直接连接huggingface则无需修改 -"text2vec-base": "/home/user/xx/text2vec-base-chinese", +# vi model_config#25 修改成你的本地路径,如果能直接连接huggingface则无需修改 +"text2vec-base": "shibing624/text2vec-base-chinese", -# 是否启动本地的notebook用于代码解释,默认启动docker的notebook -# vi server_config#35,True启动docker的notebook,false启动local的notebook -"do_remote": False, / "do_remote": True, +# vi server_config#8~14, 推荐采用容器启动服务 +DOCKER_SERVICE = True +# 是否采用容器沙箱 +SANDBOX_DO_REMOTE = True +# 是否采用api服务来进行 +NO_REMOTE_API = True ``` 5、启动服务 @@ -182,9 +186,9 @@ python dev_opsgpt/service/llm_api.py ``` ```bash +# 配置好server_config.py后,可一键启动 cd examples -# python ../dev_opsgpt/service/llm_api.py 若需使用本地大语言模型,可执行该命令 -bash start_webui.sh +python start.py ``` ## 🤗 致谢 diff --git a/README_en.md b/README_en.md index 561f54a..5a60a0f 100644 --- a/README_en.md +++ b/README_en.md @@ -144,18 +144,21 @@ os.environ["OPENAI_API_KEY"] = "sk-xxx" # You can replace the api_base_url yourself os.environ["API_BASE_URL"] = "https://api.openai.com/v1" -# vi model_config#95 You need to choose the language model +# vi model_config#105 You need to choose the language model LLM_MODEL = "gpt-3.5-turbo" -# vi model_config#33 You need to choose the vector model +# vi model_config#43 You need to choose the vector model EMBEDDING_MODEL = "text2vec-base" -# vi model_config#19 Modify to your local path, if you can directly connect to huggingface, no modification is needed -"text2vec-base": "/home/user/xx/text2vec-base-chinese", +# vi model_config#25 Modify to your local path, if you can directly connect to huggingface, no modification is needed +"text2vec-base": "shibing624/text2vec-base-chinese", -# Whether to start the local notebook for code interpretation, start the docker notebook by default -# vi server_config#35, True to start the docker notebook, false to start the local notebook -"do_remote": False, / "do_remote": True, +# vi server_config#8~14, it is recommended to start the service using containers. +DOCKER_SERVICE = True +# Whether to use container sandboxing is up to your specific requirements and preferences +SANDBOX_DO_REMOTE = True +# Whether to use api-service to use chatbot +NO_REMOTE_API = True ``` 5. Start the Service @@ -171,8 +174,8 @@ python dev_opsgpt/service/llm_api.py ``` ```bash +# After configuring server_config.py, you can start with just one click. cd examples -# python ../dev_opsgpt/service/llm_api.py If you need to use the local large language model, you can execute this command bash start_webui.sh ```