镜像自地址
https://github.com/binary-husky/gpt_academic.git
已同步 2025-12-06 06:26:47 +00:00
修复Gemini对话错误问题(停用词数量为0的情况) (#2092)
这个提交包含在:
@@ -202,16 +202,29 @@ class GoogleChatInit:
|
|||||||
) # 处理 history
|
) # 处理 history
|
||||||
|
|
||||||
messages.append(self.__conversation_user(inputs, llm_kwargs, enable_multimodal_capacity)) # 处理用户对话
|
messages.append(self.__conversation_user(inputs, llm_kwargs, enable_multimodal_capacity)) # 处理用户对话
|
||||||
payload = {
|
stop_sequences = str(llm_kwargs.get("stop", "")).split(" ")
|
||||||
"contents": messages,
|
# 过滤空字符串并确保至少有一个停止序列
|
||||||
"generationConfig": {
|
stop_sequences = [s for s in stop_sequences if s]
|
||||||
# "maxOutputTokens": llm_kwargs.get("max_token", 1024),
|
if not stop_sequences:
|
||||||
"stopSequences": str(llm_kwargs.get("stop", "")).split(" "),
|
payload = {
|
||||||
"temperature": llm_kwargs.get("temperature", 1),
|
"contents": messages,
|
||||||
"topP": llm_kwargs.get("top_p", 0.8),
|
"generationConfig": {
|
||||||
"topK": 10,
|
"temperature": llm_kwargs.get("temperature", 1),
|
||||||
},
|
"topP": llm_kwargs.get("top_p", 0.8),
|
||||||
}
|
"topK": 10,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
else:
|
||||||
|
payload = {
|
||||||
|
"contents": messages,
|
||||||
|
"generationConfig": {
|
||||||
|
# "maxOutputTokens": llm_kwargs.get("max_token", 1024),
|
||||||
|
"stopSequences": stop_sequences,
|
||||||
|
"temperature": llm_kwargs.get("temperature", 1),
|
||||||
|
"topP": llm_kwargs.get("top_p", 0.8),
|
||||||
|
"topK": 10,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
return header, payload
|
return header, payload
|
||||||
|
|
||||||
|
|||||||
在新工单中引用
屏蔽一个用户