🐛 fix: Clear empty messages.
This commit is contained in:
parent
8fc2f1a4a3
commit
df3b197b4b
@ -109,6 +109,7 @@ func (p *AliProvider) convertToChatOpenai(response *AliChatResponse, request *ty
|
|||||||
|
|
||||||
// 阿里云聊天请求体
|
// 阿里云聊天请求体
|
||||||
func (p *AliProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *AliChatRequest {
|
func (p *AliProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *AliChatRequest {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
messages := make([]AliMessage, 0, len(request.Messages))
|
messages := make([]AliMessage, 0, len(request.Messages))
|
||||||
for i := 0; i < len(request.Messages); i++ {
|
for i := 0; i < len(request.Messages); i++ {
|
||||||
message := request.Messages[i]
|
message := request.Messages[i]
|
||||||
|
@ -62,6 +62,7 @@ func (p *BaichuanProvider) CreateChatCompletionStream(request *types.ChatComplet
|
|||||||
|
|
||||||
// 获取聊天请求体
|
// 获取聊天请求体
|
||||||
func (p *BaichuanProvider) getChatRequestBody(request *types.ChatCompletionRequest) *BaichuanChatRequest {
|
func (p *BaichuanProvider) getChatRequestBody(request *types.ChatCompletionRequest) *BaichuanChatRequest {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
messages := make([]BaichuanMessage, 0, len(request.Messages))
|
messages := make([]BaichuanMessage, 0, len(request.Messages))
|
||||||
for i := 0; i < len(request.Messages); i++ {
|
for i := 0; i < len(request.Messages); i++ {
|
||||||
message := request.Messages[i]
|
message := request.Messages[i]
|
||||||
|
@ -131,6 +131,7 @@ func (p *BaiduProvider) convertToChatOpenai(response *BaiduChatResponse, request
|
|||||||
}
|
}
|
||||||
|
|
||||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) *BaiduChatRequest {
|
func convertFromChatOpenai(request *types.ChatCompletionRequest) *BaiduChatRequest {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
baiduChatRequest := &BaiduChatRequest{
|
baiduChatRequest := &BaiduChatRequest{
|
||||||
Messages: make([]BaiduMessage, 0, len(request.Messages)),
|
Messages: make([]BaiduMessage, 0, len(request.Messages)),
|
||||||
Temperature: request.Temperature,
|
Temperature: request.Temperature,
|
||||||
|
@ -94,6 +94,7 @@ func (p *ClaudeProvider) getChatRequest(request *types.ChatCompletionRequest) (*
|
|||||||
}
|
}
|
||||||
|
|
||||||
func ConvertFromChatOpenai(request *types.ChatCompletionRequest) (*ClaudeRequest, *types.OpenAIErrorWithStatusCode) {
|
func ConvertFromChatOpenai(request *types.ChatCompletionRequest) (*ClaudeRequest, *types.OpenAIErrorWithStatusCode) {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
claudeRequest := ClaudeRequest{
|
claudeRequest := ClaudeRequest{
|
||||||
Model: request.Model,
|
Model: request.Model,
|
||||||
Messages: []Message{},
|
Messages: []Message{},
|
||||||
|
@ -109,6 +109,7 @@ func (p *CloudflareAIProvider) convertToChatOpenai(response *ChatRespone, reques
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (p *CloudflareAIProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *ChatRequest {
|
func (p *CloudflareAIProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *ChatRequest {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
chatRequest := &ChatRequest{
|
chatRequest := &ChatRequest{
|
||||||
Stream: request.Stream,
|
Stream: request.Stream,
|
||||||
MaxTokens: request.MaxTokens,
|
MaxTokens: request.MaxTokens,
|
||||||
|
@ -86,6 +86,7 @@ func (p *CohereProvider) getChatRequest(request *types.ChatCompletionRequest) (*
|
|||||||
}
|
}
|
||||||
|
|
||||||
func ConvertFromChatOpenai(request *types.ChatCompletionRequest) (*CohereRequest, *types.OpenAIErrorWithStatusCode) {
|
func ConvertFromChatOpenai(request *types.ChatCompletionRequest) (*CohereRequest, *types.OpenAIErrorWithStatusCode) {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
cohereRequest := CohereRequest{
|
cohereRequest := CohereRequest{
|
||||||
Model: request.Model,
|
Model: request.Model,
|
||||||
MaxTokens: request.MaxTokens,
|
MaxTokens: request.MaxTokens,
|
||||||
|
@ -86,6 +86,7 @@ func (p *GeminiProvider) getChatRequest(request *types.ChatCompletionRequest) (*
|
|||||||
}
|
}
|
||||||
|
|
||||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) (*GeminiChatRequest, *types.OpenAIErrorWithStatusCode) {
|
func convertFromChatOpenai(request *types.ChatCompletionRequest) (*GeminiChatRequest, *types.OpenAIErrorWithStatusCode) {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
geminiRequest := GeminiChatRequest{
|
geminiRequest := GeminiChatRequest{
|
||||||
Contents: make([]GeminiChatContent, 0, len(request.Messages)),
|
Contents: make([]GeminiChatContent, 0, len(request.Messages)),
|
||||||
SafetySettings: []GeminiChatSafetySettings{
|
SafetySettings: []GeminiChatSafetySettings{
|
||||||
|
@ -132,6 +132,7 @@ func (p *MiniMaxProvider) convertToChatOpenai(response *MiniMaxChatResponse, req
|
|||||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) *MiniMaxChatRequest {
|
func convertFromChatOpenai(request *types.ChatCompletionRequest) *MiniMaxChatRequest {
|
||||||
var botSettings []MiniMaxBotSetting
|
var botSettings []MiniMaxBotSetting
|
||||||
var messges []MiniMaxChatMessage
|
var messges []MiniMaxChatMessage
|
||||||
|
request.ClearEmptyMessages()
|
||||||
for _, message := range request.Messages {
|
for _, message := range request.Messages {
|
||||||
if message.Role == types.ChatMessageRoleSystem {
|
if message.Role == types.ChatMessageRoleSystem {
|
||||||
botSettings = append(botSettings, MiniMaxBotSetting{
|
botSettings = append(botSettings, MiniMaxBotSetting{
|
||||||
|
@ -79,6 +79,7 @@ func (p *MistralProvider) getChatRequest(request *types.ChatCompletionRequest) (
|
|||||||
}
|
}
|
||||||
|
|
||||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) *MistralChatCompletionRequest {
|
func convertFromChatOpenai(request *types.ChatCompletionRequest) *MistralChatCompletionRequest {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
mistralRequest := &MistralChatCompletionRequest{
|
mistralRequest := &MistralChatCompletionRequest{
|
||||||
Model: request.Model,
|
Model: request.Model,
|
||||||
Messages: make([]types.ChatCompletionMessage, 0, len(request.Messages)),
|
Messages: make([]types.ChatCompletionMessage, 0, len(request.Messages)),
|
||||||
|
72
providers/moonshot/chat.go
Normal file
72
providers/moonshot/chat.go
Normal file
@ -0,0 +1,72 @@
|
|||||||
|
package moonshot
|
||||||
|
|
||||||
|
import (
|
||||||
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/requester"
|
||||||
|
"one-api/providers/openai"
|
||||||
|
"one-api/types"
|
||||||
|
)
|
||||||
|
|
||||||
|
func (p *MoonshotProvider) CreateChatCompletion(request *types.ChatCompletionRequest) (openaiResponse *types.ChatCompletionResponse, errWithCode *types.OpenAIErrorWithStatusCode) {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
|
req, errWithCode := p.GetRequestTextBody(common.RelayModeChatCompletions, request.Model, request)
|
||||||
|
if errWithCode != nil {
|
||||||
|
return nil, errWithCode
|
||||||
|
}
|
||||||
|
defer req.Body.Close()
|
||||||
|
|
||||||
|
response := &openai.OpenAIProviderChatResponse{}
|
||||||
|
// 发送请求
|
||||||
|
_, errWithCode = p.Requester.SendRequest(req, response, false)
|
||||||
|
if errWithCode != nil {
|
||||||
|
return nil, errWithCode
|
||||||
|
}
|
||||||
|
|
||||||
|
// 检测是否错误
|
||||||
|
openaiErr := openai.ErrorHandle(&response.OpenAIErrorResponse)
|
||||||
|
if openaiErr != nil {
|
||||||
|
errWithCode = &types.OpenAIErrorWithStatusCode{
|
||||||
|
OpenAIError: *openaiErr,
|
||||||
|
StatusCode: http.StatusBadRequest,
|
||||||
|
}
|
||||||
|
return nil, errWithCode
|
||||||
|
}
|
||||||
|
|
||||||
|
if response.Usage == nil {
|
||||||
|
response.Usage = &types.Usage{
|
||||||
|
PromptTokens: p.Usage.PromptTokens,
|
||||||
|
CompletionTokens: 0,
|
||||||
|
TotalTokens: 0,
|
||||||
|
}
|
||||||
|
// 那么需要计算
|
||||||
|
response.Usage.CompletionTokens = common.CountTokenText(response.GetContent(), request.Model)
|
||||||
|
response.Usage.TotalTokens = response.Usage.PromptTokens + response.Usage.CompletionTokens
|
||||||
|
}
|
||||||
|
|
||||||
|
*p.Usage = *response.Usage
|
||||||
|
|
||||||
|
return &response.ChatCompletionResponse, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *MoonshotProvider) CreateChatCompletionStream(request *types.ChatCompletionRequest) (requester.StreamReaderInterface[string], *types.OpenAIErrorWithStatusCode) {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
|
req, errWithCode := p.GetRequestTextBody(common.RelayModeChatCompletions, request.Model, request)
|
||||||
|
if errWithCode != nil {
|
||||||
|
return nil, errWithCode
|
||||||
|
}
|
||||||
|
defer req.Body.Close()
|
||||||
|
|
||||||
|
// 发送请求
|
||||||
|
resp, errWithCode := p.Requester.SendRequestRaw(req)
|
||||||
|
if errWithCode != nil {
|
||||||
|
return nil, errWithCode
|
||||||
|
}
|
||||||
|
|
||||||
|
chatHandler := openai.OpenAIStreamHandler{
|
||||||
|
Usage: p.Usage,
|
||||||
|
ModelName: request.Model,
|
||||||
|
}
|
||||||
|
|
||||||
|
return requester.RequestStream(p.Requester, resp, chatHandler.HandlerChatStream)
|
||||||
|
}
|
@ -123,6 +123,7 @@ func (p *TencentProvider) convertToChatOpenai(response *TencentChatResponse, req
|
|||||||
}
|
}
|
||||||
|
|
||||||
func convertFromChatOpenai(request *types.ChatCompletionRequest) *TencentChatRequest {
|
func convertFromChatOpenai(request *types.ChatCompletionRequest) *TencentChatRequest {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
messages := make([]TencentMessage, 0, len(request.Messages))
|
messages := make([]TencentMessage, 0, len(request.Messages))
|
||||||
for i := 0; i < len(request.Messages); i++ {
|
for i := 0; i < len(request.Messages); i++ {
|
||||||
message := request.Messages[i]
|
message := request.Messages[i]
|
||||||
|
@ -74,6 +74,7 @@ func (p *XunfeiProvider) getChatRequest(request *types.ChatCompletionRequest) (*
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (p *XunfeiProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *XunfeiChatRequest {
|
func (p *XunfeiProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *XunfeiChatRequest {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
messages := make([]XunfeiMessage, 0, len(request.Messages))
|
messages := make([]XunfeiMessage, 0, len(request.Messages))
|
||||||
for _, message := range request.Messages {
|
for _, message := range request.Messages {
|
||||||
if message.FunctionCall != nil || message.ToolCalls != nil {
|
if message.FunctionCall != nil || message.ToolCalls != nil {
|
||||||
|
@ -110,6 +110,7 @@ func (p *ZhipuProvider) convertToChatOpenai(response *ZhipuResponse, request *ty
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (p *ZhipuProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *ZhipuRequest {
|
func (p *ZhipuProvider) convertFromChatOpenai(request *types.ChatCompletionRequest) *ZhipuRequest {
|
||||||
|
request.ClearEmptyMessages()
|
||||||
for i, _ := range request.Messages {
|
for i, _ := range request.Messages {
|
||||||
request.Messages[i].Role = convertRole(request.Messages[i].Role)
|
request.Messages[i].Role = convertRole(request.Messages[i].Role)
|
||||||
if request.Messages[i].FunctionCall != nil {
|
if request.Messages[i].FunctionCall != nil {
|
||||||
|
@ -205,6 +205,16 @@ func (r *ChatCompletionRequest) GetFunctions() []*ChatCompletionFunction {
|
|||||||
return r.Functions
|
return r.Functions
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (r *ChatCompletionRequest) ClearEmptyMessages() {
|
||||||
|
var messages []ChatCompletionMessage
|
||||||
|
for _, message := range r.Messages {
|
||||||
|
if message.StringContent() != "" || message.ToolCalls != nil || message.FunctionCall != nil {
|
||||||
|
messages = append(messages, message)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
r.Messages = messages
|
||||||
|
}
|
||||||
|
|
||||||
type ChatCompletionFunction struct {
|
type ChatCompletionFunction struct {
|
||||||
Name string `json:"name"`
|
Name string `json:"name"`
|
||||||
Description string `json:"description"`
|
Description string `json:"description"`
|
||||||
|
Loading…
Reference in New Issue
Block a user