mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-12 04:44:52 +00:00
Update the docs.
This commit is contained in:
parent
df287fe758
commit
d495a8ac3d
@ -23,9 +23,11 @@ docs/openapi.json:
|
||||
- '#/components/schemas/GenerateResponse/properties/details/nullable'
|
||||
- '#/components/schemas/StreamResponse/properties/details/nullable'
|
||||
- '#/components/schemas/ChatRequest/properties/response_format/nullable'
|
||||
- '#/components/schemas/ChatRequest/properties/stream_options/nullable'
|
||||
- '#/components/schemas/ChatRequest/properties/tool_choice/nullable'
|
||||
- '#/components/schemas/ToolChoice/nullable'
|
||||
- '#/components/schemas/ChatCompletionComplete/properties/logprobs/nullable'
|
||||
- '#/components/schemas/ChatCompletionChunk/properties/usage/nullable'
|
||||
- '#/components/schemas/ChatCompletionChoice/properties/logprobs/nullable'
|
||||
no-invalid-media-type-examples:
|
||||
- '#/paths/~1/post/responses/422/content/application~1json/example'
|
||||
|
@ -742,6 +742,14 @@
|
||||
},
|
||||
"system_fingerprint": {
|
||||
"type": "string"
|
||||
},
|
||||
"usage": {
|
||||
"allOf": [
|
||||
{
|
||||
"$ref": "#/components/schemas/Usage"
|
||||
}
|
||||
],
|
||||
"nullable": true
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -937,6 +945,14 @@
|
||||
"stream": {
|
||||
"type": "boolean"
|
||||
},
|
||||
"stream_options": {
|
||||
"allOf": [
|
||||
{
|
||||
"$ref": "#/components/schemas/StreamOptions"
|
||||
}
|
||||
],
|
||||
"nullable": true
|
||||
},
|
||||
"temperature": {
|
||||
"type": "number",
|
||||
"format": "float",
|
||||
@ -1912,6 +1928,19 @@
|
||||
}
|
||||
}
|
||||
},
|
||||
"StreamOptions": {
|
||||
"type": "object",
|
||||
"required": [
|
||||
"include_usage"
|
||||
],
|
||||
"properties": {
|
||||
"include_usage": {
|
||||
"type": "boolean",
|
||||
"description": "If set, an additional chunk will be streamed before the data: [DONE] message. The usage field on this chunk shows the token usage statistics for the entire request, and the choices field will always be an empty array. All other chunks will also include a usage field, but with a null value.",
|
||||
"example": "true"
|
||||
}
|
||||
}
|
||||
},
|
||||
"StreamResponse": {
|
||||
"type": "object",
|
||||
"required": [
|
||||
|
@ -886,7 +886,7 @@ pub(crate) struct ChatRequest {
|
||||
|
||||
/// Options for streaming response. Only set this when you set stream: true.
|
||||
#[serde(default)]
|
||||
#[schema(nullable = true, default = "null", example = "null")]
|
||||
#[schema(nullable = true, example = "null")]
|
||||
pub stream_options: Option<StreamOptions>,
|
||||
}
|
||||
|
||||
|
@ -13,8 +13,8 @@ use crate::{
|
||||
usage_stats, BestOfSequence, Details, ErrorResponse, FinishReason, FunctionName,
|
||||
GenerateParameters, GenerateRequest, GenerateResponse, GrammarType, HubModelInfo,
|
||||
HubProcessorConfig, HubTokenizerConfig, Info, Message, MessageChunk, MessageContent,
|
||||
OutputMessage, PrefillToken, SimpleToken, StreamDetails, StreamResponse, TextMessage, Token,
|
||||
TokenizeResponse, ToolCallDelta, ToolCallMessage, Url, Usage, Validation,
|
||||
OutputMessage, PrefillToken, SimpleToken, StreamDetails, StreamOptions, StreamResponse,
|
||||
TextMessage, Token, TokenizeResponse, ToolCallDelta, ToolCallMessage, Url, Usage, Validation,
|
||||
};
|
||||
use crate::{
|
||||
ChatCompletion, ChatCompletionChoice, ChatCompletionChunk, ChatCompletionComplete,
|
||||
@ -1688,6 +1688,7 @@ StreamDetails,
|
||||
ErrorResponse,
|
||||
GrammarType,
|
||||
Usage,
|
||||
StreamOptions,
|
||||
DeltaToolCall,
|
||||
ToolType,
|
||||
Tool,
|
||||
|
Loading…
Reference in New Issue
Block a user