text-generation-inference/backends/v3/src
OlivierDehaene 8c3669b287
feat: auto max_new_tokens (#2803)
* feat: auto max_new_tokens

* update default

* Fixing the tests.

---------

Co-authored-by: Nicolas Patry <patry.nicolas@protonmail.com>
2024-12-06 05:50:35 +01:00
..
client Choosing input/total tokens automatically based on available VRAM? (#2673) 2024-10-28 04:59:49 +01:00
backend.rs fix: incomplete generations w/ single tokens generations and models that did not support chunking (#2770) 2024-11-21 16:37:55 +00:00
block_allocator.rs Lots of improvements (Still 2 allocators) (#2449) 2024-08-29 16:29:01 +02:00
lib.rs Choosing input/total tokens automatically based on available VRAM? (#2673) 2024-10-28 04:59:49 +01:00
main.rs feat: add payload limit (#2726) 2024-11-21 18:20:15 +00:00
queue.rs feat: auto max_new_tokens (#2803) 2024-12-06 05:50:35 +01:00
radix.rs Adding a test for FD. (#2516) 2024-09-16 17:00:54 +02:00