mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-24 16:32:12 +00:00
test(neuron): no error anymore when requesting too many tokens
This commit is contained in:
parent
4c0fa92cb4
commit
10b57727c2
@ -53,17 +53,6 @@ async def test_model_single_request(tgi_service):
|
||||
)
|
||||
assert response.details.generated_tokens == 17
|
||||
|
||||
# check error
|
||||
try:
|
||||
await tgi_service.client.text_generation("What is Deep Learning?", max_new_tokens=170000)
|
||||
except ValidationError:
|
||||
pass
|
||||
else:
|
||||
raise AssertionError(
|
||||
"The previous text generation request should have failed, "
|
||||
"because too many tokens were requested, it succeeded"
|
||||
)
|
||||
|
||||
# Sampling
|
||||
await tgi_service.client.text_generation(
|
||||
"What is Deep Learning?",
|
||||
|
Loading…
Reference in New Issue
Block a user