From 25cf86dae6db43a76b42f34a82a7c4ab6ff0b14e Mon Sep 17 00:00:00 2001 From: binary-husky Date: Mon, 30 Oct 2023 10:59:08 +0800 Subject: [PATCH] =?UTF-8?q?=E4=BF=AE=E5=A4=8Dget=5Fconf=E6=8E=A5=E5=8F=A3?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- main.py | 2 +- request_llms/bridge_chatgpt.py | 1 + request_llms/bridge_qianfan.py | 2 +- request_llms/bridge_spark.py | 2 +- request_llms/bridge_stackclaude.py | 4 ++-- 5 files changed, 6 insertions(+), 5 deletions(-) diff --git a/main.py b/main.py index ee8f5cf7..bf843825 100644 --- a/main.py +++ b/main.py @@ -433,7 +433,7 @@ def main(): server_port=PORT, favicon_path=os.path.join(os.path.dirname(__file__), "docs/logo.png"), auth=AUTHENTICATION if len(AUTHENTICATION) != 0 else None, - blocked_paths=["config.py","config_private.py","docker-compose.yml","Dockerfile"]) + blocked_paths=["config.py","config_private.py","docker-compose.yml","Dockerfile","gpt_log/admin"]) # 如果需要在二级路径下运行 # CUSTOM_PATH = get_conf('CUSTOM_PATH') diff --git a/request_llms/bridge_chatgpt.py b/request_llms/bridge_chatgpt.py index 38933c60..c9572e19 100644 --- a/request_llms/bridge_chatgpt.py +++ b/request_llms/bridge_chatgpt.py @@ -120,6 +120,7 @@ def predict_no_ui_long_connection(inputs, llm_kwargs, history=[], sys_prompt="", raise ConnectionAbortedError("OpenAI拒绝了请求:" + error_msg) else: if "API key has been deactivated" in error_msg: report_invalid_key(api_key) + elif "exceeded your current quota" in error_msg: report_invalid_key(api_key) raise RuntimeError("OpenAI拒绝了请求:" + error_msg) if ('data: [DONE]' in chunk): break # api2d 正常完成 json_data = json.loads(chunk.lstrip('data:'))['choices'][0] diff --git a/request_llms/bridge_qianfan.py b/request_llms/bridge_qianfan.py index 99f0623f..81e7a9c1 100644 --- a/request_llms/bridge_qianfan.py +++ b/request_llms/bridge_qianfan.py @@ -75,7 +75,7 @@ def generate_message_payload(inputs, llm_kwargs, history, system_prompt): def generate_from_baidu_qianfan(inputs, llm_kwargs, history, system_prompt): - BAIDU_CLOUD_QIANFAN_MODEL, = get_conf('BAIDU_CLOUD_QIANFAN_MODEL') + BAIDU_CLOUD_QIANFAN_MODEL = get_conf('BAIDU_CLOUD_QIANFAN_MODEL') url_lib = { "ERNIE-Bot": "https://aip.baidubce.com/rpc/2.0/ai_custom/v1/wenxinworkshop/chat/completions" , diff --git a/request_llms/bridge_spark.py b/request_llms/bridge_spark.py index d6ff42fa..a3a3e091 100644 --- a/request_llms/bridge_spark.py +++ b/request_llms/bridge_spark.py @@ -8,7 +8,7 @@ from multiprocessing import Process, Pipe model_name = '星火认知大模型' def validate_key(): - XFYUN_APPID, = get_conf('XFYUN_APPID', ) + XFYUN_APPID = get_conf('XFYUN_APPID', ) if XFYUN_APPID == '00000000' or XFYUN_APPID == '': return False return True diff --git a/request_llms/bridge_stackclaude.py b/request_llms/bridge_stackclaude.py index a3939844..0b42a17c 100644 --- a/request_llms/bridge_stackclaude.py +++ b/request_llms/bridge_stackclaude.py @@ -36,7 +36,7 @@ try: CHANNEL_ID = None async def open_channel(self): - response = await self.conversations_open(users=get_conf('SLACK_CLAUDE_BOT_ID')[0]) + response = await self.conversations_open(users=get_conf('SLACK_CLAUDE_BOT_ID')) self.CHANNEL_ID = response["channel"]["id"] async def chat(self, text): @@ -51,7 +51,7 @@ try: # TODO:暂时不支持历史消息,因为在同一个频道里存在多人使用时历史消息渗透问题 resp = await self.conversations_history(channel=self.CHANNEL_ID, oldest=self.LAST_TS, limit=1) msg = [msg for msg in resp["messages"] - if msg.get("user") == get_conf('SLACK_CLAUDE_BOT_ID')[0]] + if msg.get("user") == get_conf('SLACK_CLAUDE_BOT_ID')] return msg except (SlackApiError, KeyError) as e: raise RuntimeError(f"获取Slack消息失败。")