🐛 fix: Clear empty messages.
This commit is contained in:
parent
8fc2f1a4a3
commit
df3b197b4b
@ -109,6 +109,7 @@ func (p *AliProvider) convertToChatOpenai(response *AliChatResponse, request *ty
|
||||
|
||||
// 阿里云聊天请求体
|
||||
func (p *AliProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *AliChatRequest {
|
||||
request.ClearEmptyMessages()
|
||||
messages := make([]AliMessage, 0, len(request.Messages))
|
||||
for i := 0; i < len(request.Messages); i++ {
|
||||
message := request.Messages[i]
|
||||
|
@ -62,6 +62,7 @@ func (p *BaichuanProvider) CreateChatCompletionStream(request *types.ChatComplet
|
||||
|
||||
// 获取聊天请求体
|
||||
func (p *BaichuanProvider) getChatRequestBody(request *types.ChatCompletionRequest) *BaichuanChatRequest {
|
||||
request.ClearEmptyMessages()
|
||||
messages := make([]BaichuanMessage, 0, len(request.Messages))
|
||||
for i := 0; i < len(request.Messages); i++ {
|
||||
message := request.Messages[i]
|
||||
|
@ -131,6 +131,7 @@ func (p *BaiduProvider) convertToChatOpenai(response *BaiduChatResponse, request
|
||||
}
|
||||
|
||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) *BaiduChatRequest {
|
||||
request.ClearEmptyMessages()
|
||||
baiduChatRequest := &BaiduChatRequest{
|
||||
Messages: make([]BaiduMessage, 0, len(request.Messages)),
|
||||
Temperature: request.Temperature,
|
||||
|
@ -94,6 +94,7 @@ func (p *ClaudeProvider) getChatRequest(request *types.ChatCompletionRequest) (*
|
||||
}
|
||||
|
||||
func ConvertFromChatOpenai(request *types.ChatCompletionRequest) (*ClaudeRequest, *types.OpenAIErrorWithStatusCode) {
|
||||
request.ClearEmptyMessages()
|
||||
claudeRequest := ClaudeRequest{
|
||||
Model: request.Model,
|
||||
Messages: []Message{},
|
||||
|
@ -109,6 +109,7 @@ func (p *CloudflareAIProvider) convertToChatOpenai(response *ChatRespone, reques
|
||||
}
|
||||
|
||||
func (p *CloudflareAIProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *ChatRequest {
|
||||
request.ClearEmptyMessages()
|
||||
chatRequest := &ChatRequest{
|
||||
Stream: request.Stream,
|
||||
MaxTokens: request.MaxTokens,
|
||||
|
@ -86,6 +86,7 @@ func (p *CohereProvider) getChatRequest(request *types.ChatCompletionRequest) (*
|
||||
}
|
||||
|
||||
func ConvertFromChatOpenai(request *types.ChatCompletionRequest) (*CohereRequest, *types.OpenAIErrorWithStatusCode) {
|
||||
request.ClearEmptyMessages()
|
||||
cohereRequest := CohereRequest{
|
||||
Model: request.Model,
|
||||
MaxTokens: request.MaxTokens,
|
||||
|
@ -86,6 +86,7 @@ func (p *GeminiProvider) getChatRequest(request *types.ChatCompletionRequest) (*
|
||||
}
|
||||
|
||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) (*GeminiChatRequest, *types.OpenAIErrorWithStatusCode) {
|
||||
request.ClearEmptyMessages()
|
||||
geminiRequest := GeminiChatRequest{
|
||||
Contents: make([]GeminiChatContent, 0, len(request.Messages)),
|
||||
SafetySettings: []GeminiChatSafetySettings{
|
||||
|
@ -132,6 +132,7 @@ func (p *MiniMaxProvider) convertToChatOpenai(response *MiniMaxChatResponse, req
|
||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) *MiniMaxChatRequest {
|
||||
var botSettings []MiniMaxBotSetting
|
||||
var messges []MiniMaxChatMessage
|
||||
request.ClearEmptyMessages()
|
||||
for _, message := range request.Messages {
|
||||
if message.Role == types.ChatMessageRoleSystem {
|
||||
botSettings = append(botSettings, MiniMaxBotSetting{
|
||||
|
@ -79,6 +79,7 @@ func (p *MistralProvider) getChatRequest(request *types.ChatCompletionRequest) (
|
||||
}
|
||||
|
||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) *MistralChatCompletionRequest {
|
||||
request.ClearEmptyMessages()
|
||||
mistralRequest := &MistralChatCompletionRequest{
|
||||
Model: request.Model,
|
||||
Messages: make([]types.ChatCompletionMessage, 0, len(request.Messages)),
|
||||
|
72
providers/moonshot/chat.go
Normal file
72
providers/moonshot/chat.go
Normal file
@ -0,0 +1,72 @@
|
||||
package moonshot
|
||||
|
||||
import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/requester"
|
||||
"one-api/providers/openai"
|
||||
"one-api/types"
|
||||
)
|
||||
|
||||
func (p *MoonshotProvider) CreateChatCompletion(request *types.ChatCompletionRequest) (openaiResponse *types.ChatCompletionResponse, errWithCode *types.OpenAIErrorWithStatusCode) {
|
||||
request.ClearEmptyMessages()
|
||||
req, errWithCode := p.GetRequestTextBody(common.RelayModeChatCompletions, request.Model, request)
|
||||
if errWithCode != nil {
|
||||
return nil, errWithCode
|
||||
}
|
||||
defer req.Body.Close()
|
||||
|
||||
response := &openai.OpenAIProviderChatResponse{}
|
||||
// 发送请求
|
||||
_, errWithCode = p.Requester.SendRequest(req, response, false)
|
||||
if errWithCode != nil {
|
||||
return nil, errWithCode
|
||||
}
|
||||
|
||||
// 检测是否错误
|
||||
openaiErr := openai.ErrorHandle(&response.OpenAIErrorResponse)
|
||||
if openaiErr != nil {
|
||||
errWithCode = &types.OpenAIErrorWithStatusCode{
|
||||
OpenAIError: *openaiErr,
|
||||
StatusCode: http.StatusBadRequest,
|
||||
}
|
||||
return nil, errWithCode
|
||||
}
|
||||
|
||||
if response.Usage == nil {
|
||||
response.Usage = &types.Usage{
|
||||
PromptTokens: p.Usage.PromptTokens,
|
||||
CompletionTokens: 0,
|
||||
TotalTokens: 0,
|
||||
}
|
||||
// 那么需要计算
|
||||
response.Usage.CompletionTokens = common.CountTokenText(response.GetContent(), request.Model)
|
||||
response.Usage.TotalTokens = response.Usage.PromptTokens + response.Usage.CompletionTokens
|
||||
}
|
||||
|
||||
*p.Usage = *response.Usage
|
||||
|
||||
return &response.ChatCompletionResponse, nil
|
||||
}
|
||||
|
||||
func (p *MoonshotProvider) CreateChatCompletionStream(request *types.ChatCompletionRequest) (requester.StreamReaderInterface[string], *types.OpenAIErrorWithStatusCode) {
|
||||
request.ClearEmptyMessages()
|
||||
req, errWithCode := p.GetRequestTextBody(common.RelayModeChatCompletions, request.Model, request)
|
||||
if errWithCode != nil {
|
||||
return nil, errWithCode
|
||||
}
|
||||
defer req.Body.Close()
|
||||
|
||||
// 发送请求
|
||||
resp, errWithCode := p.Requester.SendRequestRaw(req)
|
||||
if errWithCode != nil {
|
||||
return nil, errWithCode
|
||||
}
|
||||
|
||||
chatHandler := openai.OpenAIStreamHandler{
|
||||
Usage: p.Usage,
|
||||
ModelName: request.Model,
|
||||
}
|
||||
|
||||
return requester.RequestStream(p.Requester, resp, chatHandler.HandlerChatStream)
|
||||
}
|
@ -123,6 +123,7 @@ func (p *TencentProvider) convertToChatOpenai(response *TencentChatResponse, req
|
||||
}
|
||||
|
||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) *TencentChatRequest {
|
||||
request.ClearEmptyMessages()
|
||||
messages := make([]TencentMessage, 0, len(request.Messages))
|
||||
for i := 0; i < len(request.Messages); i++ {
|
||||
message := request.Messages[i]
|
||||
|
@ -74,6 +74,7 @@ func (p *XunfeiProvider) getChatRequest(request *types.ChatCompletionRequest) (*
|
||||
}
|
||||
|
||||
func (p *XunfeiProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *XunfeiChatRequest {
|
||||
request.ClearEmptyMessages()
|
||||
messages := make([]XunfeiMessage, 0, len(request.Messages))
|
||||
for _, message := range request.Messages {
|
||||
if message.FunctionCall != nil || message.ToolCalls != nil {
|
||||
|
@ -110,6 +110,7 @@ func (p *ZhipuProvider) convertToChatOpenai(response *ZhipuResponse, request *ty
|
||||
}
|
||||
|
||||
func (p *ZhipuProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *ZhipuRequest {
|
||||
request.ClearEmptyMessages()
|
||||
for i, _ := range request.Messages {
|
||||
request.Messages[i].Role = convertRole(request.Messages[i].Role)
|
||||
if request.Messages[i].FunctionCall != nil {
|
||||
|
@ -205,6 +205,16 @@ func (r *ChatCompletionRequest) GetFunctions() []*ChatCompletionFunction {
|
||||
return r.Functions
|
||||
}
|
||||
|
||||
func (r *ChatCompletionRequest) ClearEmptyMessages() {
|
||||
var messages []ChatCompletionMessage
|
||||
for _, message := range r.Messages {
|
||||
if message.StringContent() != "" || message.ToolCalls != nil || message.FunctionCall != nil {
|
||||
messages = append(messages, message)
|
||||
}
|
||||
}
|
||||
r.Messages = messages
|
||||
}
|
||||
|
||||
type ChatCompletionFunction struct {
|
||||
Name string `json:"name"`
|
||||
Description string `json:"description"`
|
||||
|
Loading…
Reference in New Issue
Block a user