mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 20:34:54 +00:00
Fix linting
This commit is contained in:
parent
4607b7e9c4
commit
c61013e6be
@ -975,4 +975,4 @@ class AsyncClient:
|
||||
except ValidationError:
|
||||
# If we failed to parse the payload, then it is an error payload
|
||||
raise parse_error(resp.status, json_payload)
|
||||
yield response
|
||||
yield response
|
||||
|
@ -453,4 +453,4 @@ class StreamResponse(BaseModel):
|
||||
# Inference API currently deployed model
|
||||
class DeployedModel(BaseModel):
|
||||
model_id: str
|
||||
sha: str
|
||||
sha: str
|
||||
|
@ -597,7 +597,7 @@ async fn completions(
|
||||
let span = tracing::Span::current();
|
||||
metrics::increment_counter!("tgi_request_count");
|
||||
|
||||
let CompletionRequest {
|
||||
let CompletionRequest {
|
||||
max_tokens,
|
||||
seed,
|
||||
stop,
|
||||
@ -642,13 +642,13 @@ async fn completions(
|
||||
}
|
||||
|
||||
let generate_requests: Vec<GenerateRequest> = req
|
||||
.prompt
|
||||
.prompt
|
||||
.iter()
|
||||
.map(|prompt| GenerateRequest {
|
||||
inputs: prompt.to_string(),
|
||||
parameters: GenerateParameters {
|
||||
best_of: None,
|
||||
temperature: temperature,
|
||||
temperature,
|
||||
repetition_penalty: req.repetition_penalty,
|
||||
frequency_penalty: req.frequency_penalty,
|
||||
top_k: None,
|
||||
|
Loading…
Reference in New Issue
Block a user