From 913f20aad2e12125aa4184d0bea4d5710d17c18e Mon Sep 17 00:00:00 2001 From: lvxiu_ext Date: Wed, 10 May 2023 12:09:18 +0800 Subject: [PATCH] =?UTF-8?q?http=20bug=E4=BF=AE=E5=A4=8D?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- main.go | 109 +++++++++++++++++++++++++++++--------------------------- 1 file changed, 56 insertions(+), 53 deletions(-) diff --git a/main.go b/main.go index ed518b6..e4a2096 100644 --- a/main.go +++ b/main.go @@ -329,7 +329,9 @@ func completions(c *gin.Context) { if err == io.EOF { break } - return err + log.Printf("httpError1 %v", err.Error()) + break + //return err } var chatResponse ChatResponse //去除回应中的data:前缀 @@ -345,6 +347,7 @@ func completions(c *gin.Context) { // 写回数据 _, err = c.Writer.Write([]byte(string(chunk) + "\n")) if err != nil { + log.Printf("httpError2 %v", err.Error()) return err } c.Writer.(http.Flusher).Flush() @@ -390,7 +393,7 @@ func completions(c *gin.Context) { func handleGetModels(c *gin.Context) { // BUGFIX: fix options request, see https://github.com/diemus/azure-openai-proxy/issues/3 //models := []string{"gpt-4", "gpt-4-0314", "gpt-4-32k", "gpt-4-32k-0314", "gpt-3.5-turbo", "gpt-3.5-turbo-0301", "text-davinci-003", "text-embedding-ada-002"} - models := []string{"gpt-3.5-turbo", "gpt-3.5-turbo-0301"} + models := []string{"gpt-3.5-turbo", "gpt-3.5-turbo-0301", "text-davinci-003", "text-embedding-ada-002"} result := ListModelResponse{ Object: "list", } @@ -501,57 +504,57 @@ func main() { Redis = InitRedis() //添加reids测试数据 - var serverInfo ServerInfo = ServerInfo{ - ServerAddress: "https://gptp.any-door.cn", - AvailableKey: "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,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,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", - } - var serverInfo2 ServerInfo = ServerInfo{ - ServerAddress: "https://azure.any-door.cn", - AvailableKey: "6c4d2c65970b40e482e7cd27adb0d119", - } - serverInfoStr, _ := json.Marshal(&serverInfo) - serverInfoStr2, _ := json.Marshal(&serverInfo2) - Redis.Set(context.Background(), "server:1", serverInfoStr, 0) - Redis.Set(context.Background(), "server:2", serverInfoStr2, 0) - - var modelInfo ModelInfo = ModelInfo{ - ModelName: "gpt-3.5-turbo", - ModelPrice: 0.0001, - ModelPrepayment: 4000, - } - modelInfoStr, _ := json.Marshal(&modelInfo) - var modelInfo2 ModelInfo = ModelInfo{ - ModelName: "text-davinci-003", - ModelPrice: 0.001, - ModelPrepayment: 4000, - } - modelInfoStr2, _ := json.Marshal(&modelInfo2) - var modelInfo3 ModelInfo = ModelInfo{ - ModelName: "text-davinci-003", - ModelPrice: 0.001, - ModelPrepayment: 4000, - } - modelInfoStr3, _ := json.Marshal(&modelInfo3) - Redis.Set(context.Background(), "model:gpt-3.5-turbo", modelInfoStr, 0) - Redis.Set(context.Background(), "model:gpt-3.5-turbo-0301", modelInfoStr, 0) - Redis.Set(context.Background(), "model:text-davinci-003", modelInfoStr2, 0) - Redis.Set(context.Background(), "model:text-embedding-ada-002", modelInfoStr3, 0) - - var userInfo UserInfo = UserInfo{ - UID: "1", - SID: "1", - } - var userInfo2 UserInfo = UserInfo{ - UID: "2", - SID: "2", - } - userInfoStr, _ := json.Marshal(&userInfo) - userInfoStr2, _ := json.Marshal(&userInfo2) - Redis.Set(context.Background(), "user:8aeb3747-715c-48e8-8b80-aec815949f22", userInfoStr, 0) - Redis.Set(context.Background(), "user:key2", userInfoStr2, 0) - - Redis.IncrByFloat(context.Background(), "user:1:balance", 1000).Result() - Redis.IncrByFloat(context.Background(), "user:2:balance", 1000).Result() + // var serverInfo ServerInfo = ServerInfo{ + // ServerAddress: "https://gptp.any-door.cn", + // AvailableKey: "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,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,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", + // } + // var serverInfo2 ServerInfo = ServerInfo{ + // ServerAddress: "https://azure.any-door.cn", + // AvailableKey: "6c4d2c65970b40e482e7cd27adb0d119", + // } + // serverInfoStr, _ := json.Marshal(&serverInfo) + // serverInfoStr2, _ := json.Marshal(&serverInfo2) + // Redis.Set(context.Background(), "server:1", serverInfoStr, 0) + // Redis.Set(context.Background(), "server:2", serverInfoStr2, 0) + + // var modelInfo ModelInfo = ModelInfo{ + // ModelName: "gpt-3.5-turbo", + // ModelPrice: 0.0001, + // ModelPrepayment: 4000, + // } + // modelInfoStr, _ := json.Marshal(&modelInfo) + // var modelInfo2 ModelInfo = ModelInfo{ + // ModelName: "text-davinci-003", + // ModelPrice: 0.001, + // ModelPrepayment: 4000, + // } + // modelInfoStr2, _ := json.Marshal(&modelInfo2) + // var modelInfo3 ModelInfo = ModelInfo{ + // ModelName: "text-davinci-003", + // ModelPrice: 0.001, + // ModelPrepayment: 4000, + // } + // modelInfoStr3, _ := json.Marshal(&modelInfo3) + // Redis.Set(context.Background(), "model:gpt-3.5-turbo", modelInfoStr, 0) + // Redis.Set(context.Background(), "model:gpt-3.5-turbo-0301", modelInfoStr, 0) + // Redis.Set(context.Background(), "model:text-davinci-003", modelInfoStr2, 0) + // Redis.Set(context.Background(), "model:text-embedding-ada-002", modelInfoStr3, 0) + + // var userInfo UserInfo = UserInfo{ + // UID: "1", + // SID: "1", + // } + // var userInfo2 UserInfo = UserInfo{ + // UID: "2", + // SID: "2", + // } + // userInfoStr, _ := json.Marshal(&userInfo) + // userInfoStr2, _ := json.Marshal(&userInfo2) + // Redis.Set(context.Background(), "user:8aeb3747-715c-48e8-8b80-aec815949f22", userInfoStr, 0) + // Redis.Set(context.Background(), "user:AK-7d8ab782-a152-4cc1-9972-568713465c96", userInfoStr2, 0) + + // Redis.IncrByFloat(context.Background(), "user:1:balance", 1000).Result() + // Redis.IncrByFloat(context.Background(), "user:2:balance", 1000).Result() //r.Run("127.0.0.1:8080") //docker下使用