text-generation-inference/backends/llamacpp/src
2024-11-14 08:42:01 +01:00
..
backend.rs feat(backend): handle all the tokenization failure and send back to the client 2024-11-14 08:42:01 +01:00
lib.rs feat(backend): add early stopping criteria from TGI stream callback 2024-11-14 08:42:01 +01:00
main.rs feat(backend): expose tokenizer to the GenerationContext to decode token 2024-11-14 08:42:01 +01:00