mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-09 19:34:53 +00:00
formating
This commit is contained in:
parent
7d633582e4
commit
3fc811f596
@ -1,8 +1,8 @@
|
|||||||
/// HTTP Server logic
|
/// HTTP Server logic
|
||||||
use crate::infer::{InferError, InferStreamResponse};
|
use crate::infer::{InferError, InferStreamResponse};
|
||||||
use crate::{
|
use crate::{
|
||||||
Details, ErrorResponse, GenerateParameters, GenerateRequest, GenerateResponse, Infer, StreamResponse,
|
Details, ErrorResponse, GenerateParameters, GenerateRequest, GenerateResponse, Infer,
|
||||||
Validation,
|
StreamResponse, Validation,
|
||||||
};
|
};
|
||||||
use axum::extract::Extension;
|
use axum::extract::Extension;
|
||||||
use axum::http::{HeaderMap, StatusCode};
|
use axum::http::{HeaderMap, StatusCode};
|
||||||
|
@ -368,7 +368,9 @@ class CausalLM(Model):
|
|||||||
else:
|
else:
|
||||||
seed = None
|
seed = None
|
||||||
|
|
||||||
generated_text = GeneratedText(output_text, stopping_criteria.current_tokens, reason, seed)
|
generated_text = GeneratedText(
|
||||||
|
output_text, stopping_criteria.current_tokens, reason, seed
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
# Keep request in the batch
|
# Keep request in the batch
|
||||||
generated_text = None
|
generated_text = None
|
||||||
|
Loading…
Reference in New Issue
Block a user