mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-23 16:02:10 +00:00
test(neuron): no error anymore when requesting too many tokens
This commit is contained in:
parent
d39e002fa5
commit
debf032ca3
@ -53,17 +53,6 @@ async def test_model_single_request(tgi_service):
|
|||||||
)
|
)
|
||||||
assert response.details.generated_tokens == 17
|
assert response.details.generated_tokens == 17
|
||||||
|
|
||||||
# check error
|
|
||||||
try:
|
|
||||||
await tgi_service.client.text_generation("What is Deep Learning?", max_new_tokens=170000)
|
|
||||||
except ValidationError:
|
|
||||||
pass
|
|
||||||
else:
|
|
||||||
raise AssertionError(
|
|
||||||
"The previous text generation request should have failed, "
|
|
||||||
"because too many tokens were requested, it succeeded"
|
|
||||||
)
|
|
||||||
|
|
||||||
# Sampling
|
# Sampling
|
||||||
await tgi_service.client.text_generation(
|
await tgi_service.client.text_generation(
|
||||||
"What is Deep Learning?",
|
"What is Deep Learning?",
|
||||||
|
Loading…
Reference in New Issue
Block a user