One last tweak to our system should help users have just a bit more insight into what the AI system is doing and
Our system now streams in responses and allows configurable system prompts, but it’s missing an affordance to indicate that the AI is done. Fortunately, we can add that. We will add a boolean done
field to the message models. On AI responses, it will start false and will be updated. We can use that to provide visible feedback to the user:
rails g migration add_done_to_messages done:boolean!
<%= tag.div id: dom_id(message), class: [ "p-2 max-w-4xl", "border-2 border-gray-400": !message.done ] do %>
<div class="font-semibold"><%= message.role %>:</div>
<%= message.to_markdown %>
<%= tag.div "\"Thinking\" ..." unless message.done %>
<% end %>
This will give a border around the partial message and display “Thinking …” until complete. To wire it in, we’ll update the done
field while streaming in the messages:
# app/models/conversation.rb
class Conversation < ApplicationRecord
def send_thread
# collect the preceding messages
messages = self.messages.order(:created_at).map { {
role: _1.role,
content: _1.message
} }
# create the blank initial response
response = self.messages.create({
role: "assistant",
message: "",
done: false # 1
})
ollama_client.chat({
model: "llama3",
stream: true,
messages: messages
}) do |event, raw|
response.message += event["message"]["content"]
response.done = event["done"] # 2
response.save
end
end
end
We start with an incomplete message (#1), and updated it every time we get a response from the AI (#2). When the final response comes from the AI, the Turbo Stream broadcast will include the updated message, and the updated styling.
And with that, we’re able to clearly show the user when a message is pending or complete.