diff --git a/docs/source/basic_tutorials/consuming_tgi.md b/docs/source/basic_tutorials/consuming_tgi.md index fda01e2a..60b0d203 100644 --- a/docs/source/basic_tutorials/consuming_tgi.md +++ b/docs/source/basic_tutorials/consuming_tgi.md @@ -101,7 +101,7 @@ gr.ChatInterface( inference, chatbot=gr.Chatbot(height=300), textbox=gr.Textbox(placeholder="Chat with me!", container=False, scale=7), - description="This is the demo for Gradio UI consuming TGI endpoint with Falcon model.", + description="This is the demo for Gradio UI consuming TGI endpoint with LLaMA 2 7B-Chat model.", title="Gradio 🤝 TGI", examples=["Are tomatoes vegetables?"], retry_btn=None, @@ -119,10 +119,15 @@ The UI looks like this 👇 /> +You can try the demo directly here 👇 + + + + You can disable streaming mode using `return` instead of `yield` in your inference function. You can read more about how to customize a `ChatInterface` [here](https://www.gradio.app/guides/creating-a-chatbot-fast).