From 36c39974a73d6e2ae6cd284702ba1431028ba9c1 Mon Sep 17 00:00:00 2001 From: junleea <354425203@qq.com> Date: Tue, 13 May 2025 11:14:14 +0800 Subject: [PATCH] =?UTF-8?q?=E6=B7=BB=E5=8A=A0=E5=9B=BD=E5=A4=96=E6=9C=8D?= =?UTF-8?q?=E5=8A=A1=E5=99=A8=E4=B8=AD=E8=BD=ACai=E8=AF=B7=E6=B1=82?= =?UTF-8?q?=E6=95=B0=E6=8D=AE,=E6=B7=BB=E5=8A=A0gemini?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- proto/status.go | 1 + service/generalModel/openai.go | 3 +-- service/imService.go | 2 ++ 3 files changed, 4 insertions(+), 2 deletions(-) diff --git a/proto/status.go b/proto/status.go index 7b07867..1c8f29a 100644 --- a/proto/status.go +++ b/proto/status.go @@ -147,6 +147,7 @@ const ( ModelTypeQianfan = "qianfan" ModelTypeTongyi = "tongyi" ModelTypeHunyuan = "hunyuan" + ModelTypeGemini = "gemini" KnowledgeBaseServerResponseType = "kbase_query_resp" ) diff --git a/service/generalModel/openai.go b/service/generalModel/openai.go index 6445a5b..bf4c7ad 100644 --- a/service/generalModel/openai.go +++ b/service/generalModel/openai.go @@ -296,7 +296,6 @@ func OpenAIForeignAIResponseToGeneralMassageAndSendMsgQueue(data *proto.ReceiveM //PromptTokens = chunk.Usage.PromptTokens } //将消息存入数据库 - if chunk.Created < 0 { //这里由于没有返回结束标志,所以用该字段来判断 tokens := chunk.Usage.TotalTokens PromptTokens := chunk.Usage.PromptTokens @@ -304,6 +303,6 @@ func OpenAIForeignAIResponseToGeneralMassageAndSendMsgQueue(data *proto.ReceiveM DOSaveMessageAndToken(&data.IMParamContext, answer, "openai_"+data.PlatForm, tokens, PromptTokens) } else { //流式消息,暂存到redis - worker.SetRedis(data.IMParamContext.Channel+"_stream_msg", answer) + worker.SetRedisWithExpire(data.IMParamContext.Channel+"_stream_msg", answer, time.Second*30) } } diff --git a/service/imService.go b/service/imService.go index cb2295a..b894c0e 100644 --- a/service/imService.go +++ b/service/imService.go @@ -174,6 +174,8 @@ func WSReceiveMessageServiceV2(modelParam proto.ModelParam, ctx *proto.IMParamCo Tongyi(modelParam, ctx) case proto.ModelTypeHunyuan: generalModel.OpenAI(modelParam, ctx, ctx.ModelType) + case proto.ModelTypeGemini: + generalModel.OpenAIForeignAI(modelParam, ctx, ctx.ModelType) default: resErr, resID = errors.New("model type not exist:"+ctx.ModelType), 0 }