text-generation-inference/server/text_generation
Nick Hill a172430d8b fix: Some small fixes
- Avoid theoretical hang in batcher loop
- Avoid a couple of clones in py server generate method
- Keep attention mask tensors as integers
2022-12-01 16:19:12 -08:00
..
models fix: Some small fixes 2022-12-01 16:19:12 -08:00
pb feat(server): Support all AutoModelForCausalLM on a best effort basis 2022-10-28 19:24:00 +02:00
__init__.py feat(server): Support all AutoModelForCausalLM on a best effort basis 2022-10-28 19:24:00 +02:00
cache.py feat(server): Support AutoModelForSeq2SeqLM 2022-11-04 18:03:04 +01:00
cli.py feat(server): Support all AutoModelForCausalLM on a best effort basis 2022-10-28 19:24:00 +02:00
server.py feat(server): Support AutoModelForSeq2SeqLM 2022-11-04 18:03:04 +01:00
utils.py feat(server): Support Galactica (#4) 2022-12-01 19:31:54 +01:00