fix: fix usage & ratio
This commit is contained in:
parent
02539e8cd6
commit
17081393fc
@ -91,6 +91,7 @@ var ModelRatio = map[string]float64{
|
|||||||
"glm-4": 0.1 * RMB,
|
"glm-4": 0.1 * RMB,
|
||||||
"glm-4v": 0.1 * RMB,
|
"glm-4v": 0.1 * RMB,
|
||||||
"glm-3-turbo": 0.005 * RMB,
|
"glm-3-turbo": 0.005 * RMB,
|
||||||
|
"embedding-2": 0.0005 * RMB,
|
||||||
"chatglm_turbo": 0.3572, // ¥0.005 / 1k tokens
|
"chatglm_turbo": 0.3572, // ¥0.005 / 1k tokens
|
||||||
"chatglm_pro": 0.7143, // ¥0.01 / 1k tokens
|
"chatglm_pro": 0.7143, // ¥0.01 / 1k tokens
|
||||||
"chatglm_std": 0.3572, // ¥0.005 / 1k tokens
|
"chatglm_std": 0.3572, // ¥0.005 / 1k tokens
|
||||||
|
@ -285,7 +285,11 @@ func embeddingResponseZhipu2OpenAI(response *EmbeddingRespone) *openai.Embedding
|
|||||||
Object: "list",
|
Object: "list",
|
||||||
Data: make([]openai.EmbeddingResponseItem, 0, len(response.Embeddings)),
|
Data: make([]openai.EmbeddingResponseItem, 0, len(response.Embeddings)),
|
||||||
Model: response.Model,
|
Model: response.Model,
|
||||||
Usage: model.Usage{TotalTokens: response.Usage.TotalTokens},
|
Usage: model.Usage{
|
||||||
|
PromptTokens: response.PromptTokens,
|
||||||
|
CompletionTokens: response.CompletionTokens,
|
||||||
|
TotalTokens: response.Usage.TotalTokens,
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, item := range response.Embeddings {
|
for _, item := range response.Embeddings {
|
||||||
|
Loading…
Reference in New Issue
Block a user