Introducing Gradio Clients
WatchIntroducing Gradio Clients
WatchImportant Note: if you are getting started, we recommend using the gr.ChatInterface
to create chatbots -- its a high-level abstraction that makes it possible to create beautiful chatbot applications fast, often with a single line of code. Read more about it here.
This tutorial will show how to make chatbot UIs from scratch with Gradio's low-level Blocks API. This will give you full control over your Chatbot UI. You'll start by first creating a a simple chatbot to display text, a second one to stream text responses, and finally a chatbot that can handle media files as well. The chatbot interface that we create will look something like this:
Prerequisite: We'll be using the gradio.Blocks
class to build our Chatbot demo.
You can read the Guide to Blocks first if you are not already familiar with it. Also please make sure you are using the latest version version of Gradio: pip install --upgrade gradio
.
Let's start with recreating the simple demo above. As you may have noticed, our bot simply randomly responds "How are you?", "Today is a great day", or "I'm very hungry" to any input. Here's the code to create this with Gradio:
import gradio as gr
import random
import time
with gr.Blocks() as demo:
chatbot = gr.Chatbot(type="messages")
msg = gr.Textbox()
clear = gr.ClearButton([msg, chatbot])
def respond(message, chat_history):
bot_message = random.choice(["How are you?", "Today is a great day", "I'm very hungry"])
chat_history.append({"role": "user", "content": message})
chat_history.append({"role": "assistant", "content": bot_message})
time.sleep(2)
return "", chat_history
msg.submit(respond, [msg, chatbot], [msg, chatbot])
demo.launch()
There are three Gradio components here:
Chatbot
, whose value stores the entire history of the conversation, as a list of response pairs between the user and bot.Textbox
where the user can type their message, and then hit enter/submit to trigger the chatbot responseClearButton
button to clear the Textbox and entire Chatbot historyWe have a single function, respond()
, which takes in the entire history of the chatbot, appends a random message, waits 1 second, and then returns the updated chat history. The respond()
function also clears the textbox when it returns.
Of course, in practice, you would replace respond()
with your own more complex function, which might call a pretrained model or an API, to generate a response.
Tip: For better type hinting and auto-completion in your IDE, you can use the `gr.ChatMessage` dataclass:
from gradio import ChatMessage
def chat_function(message, history):
history.append(ChatMessage(role="user", content=message))
history.append(ChatMessage(role="assistant", content="Hello, how can I help you?"))
return history
There are several ways we can improve the user experience of the chatbot above. First, we can stream responses so the user doesn't have to wait as long for a message to be generated. Second, we can have the user message appear immediately in the chat history, while the chatbot's response is being generated. Here's the code to achieve that:
import gradio as gr
import random
import time
with gr.Blocks() as demo:
chatbot = gr.Chatbot(type="messages")
msg = gr.Textbox()
clear = gr.Button("Clear")
def user(user_message, history: list):
return "", history + [{"role": "user", "content": user_message}]
def bot(history: list):
bot_message = random.choice(["How are you?", "I love you", "I'm very hungry"])
history.append({"role": "assistant", "content": ""})
for character in bot_message:
history[-1]['content'] += character
time.sleep(0.05)
yield history
msg.submit(user, [msg, chatbot], [msg, chatbot], queue=False).then(
bot, chatbot, chatbot
)
clear.click(lambda: None, None, chatbot, queue=False)
demo.launch()
You'll notice that when a user submits their message, we now chain two event events with .then()
:
The first method user()
updates the chatbot with the user message and clears the input field. Because we want this to happen instantly, we set queue=False
, which would skip any queue had it been enabled. The chatbot's history is appended with {"role": "user", "content": user_message}
.
The second method, bot()
updates the chatbot history with the bot's response. Finally, we construct the message character by character and yield
the intermediate outputs as they are being constructed. Gradio automatically turns any function with the yield
keyword into a streaming output interface.
Of course, in practice, you would replace bot()
with your own more complex function, which might call a pretrained model or an API, to generate a response.
The gr.Chatbot
component supports a subset of markdown including bold, italics, and code. For example, we could write a function that responds to a user's message, with a bold That's cool!, like this:
def bot(history):
response = {"role": "assistant", "content": "**That's cool!**"}
history.append(response)
return history
In addition, it can handle media files, such as images, audio, and video. You can use the MultimodalTextbox
component to easily upload all types of media files to your chatbot. To pass in a media file, we must pass in the file a dictionary with a path
key pointing to a local file and an alt_text
key. The alt_text
is optional, so you can also just pass in a tuple with a single element {"path": "filepath"}
, like this:
def add_message(history, message):
for x in message["files"]:
history.append({"role": "user", "content": {"path": x}})
if message["text"] is not None:
history.append({"role": "user", "content": message["text"]})
return history, gr.MultimodalTextbox(value=None, interactive=False, file_types=["image"])
Putting this together, we can create a multimodal chatbot with a multimodal textbox for a user to submit text and media files. The rest of the code looks pretty much the same as before:
import gradio as gr
import time
# Chatbot demo with multimodal input (text, markdown, LaTeX, code blocks, image, audio, & video). Plus shows support for streaming text.
def print_like_dislike(x: gr.LikeData):
print(x.index, x.value, x.liked)
def add_message(history, message):
for x in message["files"]:
history.append({"role": "user", "content": {"path": x}})
if message["text"] is not None:
history.append({"role": "user", "content": message["text"]})
return history, gr.MultimodalTextbox(value=None, interactive=False)
def bot(history: list):
response = "**That's cool!**"
history.append({"role": "assistant", "content": ""})
for character in response:
history[-1]["content"] += character
time.sleep(0.05)
yield history
with gr.Blocks() as demo:
chatbot = gr.Chatbot(elem_id="chatbot", bubble_full_width=False, type="messages")
chat_input = gr.MultimodalTextbox(
interactive=True,
file_count="multiple",
placeholder="Enter message or upload file...",
show_label=False,
)
chat_msg = chat_input.submit(
add_message, [chatbot, chat_input], [chatbot, chat_input]
)
bot_msg = chat_msg.then(bot, chatbot, chatbot, api_name="bot_response")
bot_msg.then(lambda: gr.MultimodalTextbox(interactive=True), None, [chat_input])
chatbot.like(print_like_dislike, None, None, like_user_message=True)
demo.launch()
And you're done! That's all the code you need to build an interface for your chatbot model. Finally, we'll end our Guide with some links to Chatbots that are running on Spaces so that you can get an idea of what else is possible: