package controller import ( "bufio" "bytes" "encoding/json" "errors" "fmt" "github.com/gin-gonic/gin" "github.com/pkoukk/tiktoken-go" "io" "net/http" "one-api/common" "one-api/model" "strings" ) type Message struct { Role string `json:"role"` Content string `json:"content"` } type ChatRequest struct { Model string `json:"model"` Messages []Message `json:"messages"` MaxTokens int `json:"max_tokens"` } type TextRequest struct { Model string `json:"model"` Messages []Message `json:"messages"` Prompt string `json:"prompt"` MaxTokens int `json:"max_tokens"` //Stream bool `json:"stream"` } type Usage struct { PromptTokens int `json:"prompt_tokens"` CompletionTokens int `json:"completion_tokens"` TotalTokens int `json:"total_tokens"` } type OpenAIError struct { Message string `json:"message"` Type string `json:"type"` Param string `json:"param"` Code string `json:"code"` } type TextResponse struct { Usage `json:"usage"` Error OpenAIError `json:"error"` } type StreamResponse struct { Choices []struct { Delta struct { Content string `json:"content"` } `json:"delta"` FinishReason string `json:"finish_reason"` } `json:"choices"` } var tokenEncoder, _ = tiktoken.GetEncoding("cl100k_base") func countToken(text string) int { token := tokenEncoder.Encode(text, nil, nil) return len(token) } func Relay(c *gin.Context) { err := relayHelper(c) if err != nil { c.JSON(http.StatusOK, gin.H{ "error": gin.H{ "message": err.Error(), "type": "one_api_error", }, }) if common.AutomaticDisableChannelEnabled { channelId := c.GetInt("channel_id") channelName := c.GetString("channel_name") disableChannel(channelId, channelName, err) } } } func relayHelper(c *gin.Context) error { channelType := c.GetInt("channel") tokenId := c.GetInt("token_id") consumeQuota := c.GetBool("consume_quota") var textRequest TextRequest if consumeQuota || channelType == common.ChannelTypeAzure { requestBody, err := io.ReadAll(c.Request.Body) if err != nil { return err } err = c.Request.Body.Close() if err != nil { return err } err = json.Unmarshal(requestBody, &textRequest) if err != nil { return err } // Reset request body c.Request.Body = io.NopCloser(bytes.NewBuffer(requestBody)) } baseURL := common.ChannelBaseURLs[channelType] requestURL := c.Request.URL.String() if channelType == common.ChannelTypeCustom { baseURL = c.GetString("base_url") } fullRequestURL := fmt.Sprintf("%s%s", baseURL, requestURL) if channelType == common.ChannelTypeAzure { // https://learn.microsoft.com/en-us/azure/cognitive-services/openai/chatgpt-quickstart?pivots=rest-api&tabs=command-line#rest-api query := c.Request.URL.Query() apiVersion := query.Get("api-version") if apiVersion == "" { apiVersion = c.GetString("api_version") } requestURL := strings.Split(requestURL, "?")[0] requestURL = fmt.Sprintf("%s?api-version=%s", requestURL, apiVersion) baseURL = c.GetString("base_url") task := strings.TrimPrefix(requestURL, "/v1/") model_ := textRequest.Model model_ = strings.Replace(model_, ".", "", -1) // https://github.com/songquanpeng/one-api/issues/67 model_ = strings.TrimSuffix(model_, "-0301") model_ = strings.TrimSuffix(model_, "-0314") fullRequestURL = fmt.Sprintf("%s/openai/deployments/%s/%s", baseURL, model_, task) } var promptText string for _, message := range textRequest.Messages { promptText += fmt.Sprintf("%s: %s\n", message.Role, message.Content) } promptTokens := countToken(promptText) + 3 preConsumedTokens := common.PreConsumedQuota if textRequest.MaxTokens != 0 { preConsumedTokens = promptTokens + textRequest.MaxTokens } ratio := common.GetModelRatio(textRequest.Model) preConsumedQuota := int(float64(preConsumedTokens) * ratio) if consumeQuota { err := model.PreConsumeTokenQuota(tokenId, preConsumedQuota) if err != nil { return err } } req, err := http.NewRequest(c.Request.Method, fullRequestURL, c.Request.Body) if err != nil { return err } if channelType == common.ChannelTypeAzure { key := c.Request.Header.Get("Authorization") key = strings.TrimPrefix(key, "Bearer ") req.Header.Set("api-key", key) } else { req.Header.Set("Authorization", c.Request.Header.Get("Authorization")) } req.Header.Set("Content-Type", c.Request.Header.Get("Content-Type")) req.Header.Set("Accept", c.Request.Header.Get("Accept")) req.Header.Set("Connection", c.Request.Header.Get("Connection")) client := &http.Client{} resp, err := client.Do(req) if err != nil { return err } err = req.Body.Close() if err != nil { return err } err = c.Request.Body.Close() if err != nil { return err } var textResponse TextResponse isStream := resp.Header.Get("Content-Type") == "text/event-stream" var streamResponseText string defer func() { if consumeQuota { quota := 0 usingGPT4 := strings.HasPrefix(textRequest.Model, "gpt-4") completionRatio := 1 if usingGPT4 { completionRatio = 2 } if isStream { completionText := fmt.Sprintf("%s: %s\n", "assistant", streamResponseText) quota = promptTokens + countToken(completionText)*completionRatio } else { quota = textResponse.Usage.PromptTokens + textResponse.Usage.CompletionTokens*completionRatio } quota = int(float64(quota) * ratio) quotaDelta := quota - preConsumedQuota err := model.PostConsumeTokenQuota(tokenId, quotaDelta) if err != nil { common.SysError("Error consuming token remain quota: " + err.Error()) } } }() if isStream { scanner := bufio.NewScanner(resp.Body) scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) { if atEOF && len(data) == 0 { return 0, nil, nil } if i := strings.Index(string(data), "\n\n"); i >= 0 { return i + 2, data[0:i], nil } if atEOF { return len(data), data, nil } return 0, nil, nil }) dataChan := make(chan string) stopChan := make(chan bool) go func() { for scanner.Scan() { data := scanner.Text() dataChan <- data data = data[6:] if !strings.HasPrefix(data, "[DONE]") { var streamResponse StreamResponse err = json.Unmarshal([]byte(data), &streamResponse) if err != nil { common.SysError("Error unmarshalling stream response: " + err.Error()) return } for _, choice := range streamResponse.Choices { streamResponseText += choice.Delta.Content } } } stopChan <- true }() c.Writer.Header().Set("Content-Type", "text/event-stream") c.Writer.Header().Set("Cache-Control", "no-cache") c.Writer.Header().Set("Connection", "keep-alive") c.Writer.Header().Set("Transfer-Encoding", "chunked") c.Stream(func(w io.Writer) bool { select { case data := <-dataChan: if strings.HasPrefix(data, "data: [DONE]") { data = data[:12] } c.Render(-1, common.CustomEvent{Data: data}) return true case <-stopChan: return false } }) err = resp.Body.Close() if err != nil { return err } return nil } else { for k, v := range resp.Header { c.Writer.Header().Set(k, v[0]) } if consumeQuota { responseBody, err := io.ReadAll(resp.Body) if err != nil { return err } err = resp.Body.Close() if err != nil { return err } err = json.Unmarshal(responseBody, &textResponse) if err != nil { return err } if textResponse.Error.Type != "" { return errors.New(fmt.Sprintf("type %s, code %s, message %s", textResponse.Error.Type, textResponse.Error.Code, textResponse.Error.Message)) } // Reset response body resp.Body = io.NopCloser(bytes.NewBuffer(responseBody)) } _, err = io.Copy(c.Writer, resp.Body) if err != nil { return err } err = resp.Body.Close() if err != nil { return err } return nil } } func RelayNotImplemented(c *gin.Context) { c.JSON(http.StatusOK, gin.H{ "error": gin.H{ "message": "Not Implemented", "type": "one_api_error", }, }) }