text-generation-inference/backends/v3/src
yuanwu 67ee45a270 Pass the max_batch_total_tokens to causal_lm
refine the warmup

Signed-off-by: yuanwu <yuan.wu@intel.com>
2024-10-23 08:28:26 +00:00
..
client Pass the max_batch_total_tokens to causal_lm 2024-10-23 08:28:26 +00:00
backend.rs Pass the max_batch_total_tokens to causal_lm 2024-10-23 08:28:26 +00:00
block_allocator.rs Lots of improvements (Still 2 allocators) (#2449) 2024-09-25 06:13:11 +00:00
lib.rs Pass the max_batch_total_tokens to causal_lm 2024-10-23 08:28:26 +00:00
main.rs Pr 2352 ci branch (#2382) 2024-09-25 06:01:59 +00:00
queue.rs Pass the max_batch_total_tokens to causal_lm 2024-10-23 08:28:26 +00:00
radix.rs Adding a test for FD. (#2516) 2024-09-25 06:17:09 +00:00