text-generation-inference/router/src
Lucain 00f365353e
Update response type for /v1/chat/completions and /v1/completions (#1747)
`/v1/chat/completions` and `/v1/completions` have different output types
depending on the `stream` parameter. This PR aims at fixing the
inconsistency in the auto-generated
[openapi.json](https://huggingface.github.io/text-generation-inference/openapi.json)
specs.

cc @OlivierDehaene @drbh I reused what had been done for the `/`
endpoint but haven't tested anything myself. Could you confirm this is
the correct way of handling things?

Also, should I update the openapi.json file manually? If yes, how can I
do it?
2024-04-16 19:26:32 +02:00
..
config.rs Adding Llava-Next (Llava 1.6) with full support. (#1709) 2024-04-09 21:32:00 +02:00
health.rs Outlines guided generation (#1539) 2024-02-15 10:28:10 +01:00
infer.rs feat: improve tools to include name and add tests (#1693) 2024-04-16 09:02:46 -04:00
lib.rs feat: improve tools to include name and add tests (#1693) 2024-04-16 09:02:46 -04:00
main.rs Improve the defaults for the launcher (#1727) 2024-04-12 14:20:31 +02:00
queue.rs Improve the defaults for the launcher (#1727) 2024-04-12 14:20:31 +02:00
server.rs Update response type for /v1/chat/completions and /v1/completions (#1747) 2024-04-16 19:26:32 +02:00
validation.rs Improve the defaults for the launcher (#1727) 2024-04-12 14:20:31 +02:00