使用MaxTokens

This commit is contained in:
quzard 2023-06-17 17:36:23 +08:00
parent d624249f05
commit 625d0d4be5
3 changed files with 23 additions and 23 deletions

View File

@ -68,7 +68,7 @@ func buildTestRequest(c *gin.Context) *ChatRequest {
model_ := c.Query("model") model_ := c.Query("model")
testRequest := &ChatRequest{ testRequest := &ChatRequest{
Model: model_, Model: model_,
//MaxTokens: 1, MaxTokens: 1,
} }
testMessage := Message{ testMessage := Message{
Role: "user", Role: "user",

View File

@ -51,7 +51,7 @@ func relayPaLM(openAIRequest GeneralOpenAIRequest, c *gin.Context) *OpenAIErrorW
Temperature: openAIRequest.Temperature, Temperature: openAIRequest.Temperature,
CandidateCount: openAIRequest.N, CandidateCount: openAIRequest.N,
TopP: openAIRequest.TopP, TopP: openAIRequest.TopP,
//TopK: openAIRequest.MaxTokens, TopK: openAIRequest.MaxTokens,
} }
// TODO: forward request to PaLM & convert response // TODO: forward request to PaLM & convert response
fmt.Print(request) fmt.Print(request)

View File

@ -34,7 +34,7 @@ type GeneralOpenAIRequest struct {
Messages []Message `json:"messages"` Messages []Message `json:"messages"`
Prompt any `json:"prompt"` Prompt any `json:"prompt"`
Stream bool `json:"stream"` Stream bool `json:"stream"`
//MaxTokens int `json:"max_tokens"` MaxTokens int `json:"max_tokens"`
Temperature float64 `json:"temperature"` Temperature float64 `json:"temperature"`
TopP float64 `json:"top_p"` TopP float64 `json:"top_p"`
N int `json:"n"` N int `json:"n"`
@ -44,14 +44,14 @@ type GeneralOpenAIRequest struct {
type ChatRequest struct { type ChatRequest struct {
Model string `json:"model"` Model string `json:"model"`
Messages []Message `json:"messages"` Messages []Message `json:"messages"`
//MaxTokens int `json:"max_tokens"` MaxTokens int `json:"max_tokens"`
} }
type TextRequest struct { type TextRequest struct {
Model string `json:"model"` Model string `json:"model"`
Messages []Message `json:"messages"` Messages []Message `json:"messages"`
Prompt string `json:"prompt"` Prompt string `json:"prompt"`
//MaxTokens int `json:"max_tokens"` MaxTokens int `json:"max_tokens"`
//Stream bool `json:"stream"` //Stream bool `json:"stream"`
} }
@ -199,9 +199,9 @@ func relayHelper(c *gin.Context, relayMode int) *OpenAIErrorWithStatusCode {
promptTokens = countTokenInput(textRequest.Input, textRequest.Model) promptTokens = countTokenInput(textRequest.Input, textRequest.Model)
} }
preConsumedTokens := common.PreConsumedQuota preConsumedTokens := common.PreConsumedQuota
//if textRequest.MaxTokens != 0 { if textRequest.MaxTokens != 0 {
// preConsumedTokens = promptTokens + textRequest.MaxTokens preConsumedTokens = promptTokens + textRequest.MaxTokens
//} }
modelRatio := common.GetModelRatio(textRequest.Model) modelRatio := common.GetModelRatio(textRequest.Model)
groupRatio := common.GetGroupRatio(group) groupRatio := common.GetGroupRatio(group)
ratio := modelRatio * groupRatio ratio := modelRatio * groupRatio