We tested our Ollama connection, now let’s actually wire it in. We will want create a Conversation
model that have many Message
s. Each message will be an individual piece of the conversation. Each message will have the message itself, and a role for who submitted it, the user, the AI, or the system. System Prompts are used to prime the AI.
rails g scaffold conversation title:string
rails g model message conversation:references role:string message:text
rails db:migrate
# app/models/conversation.rb
class Conversation < ApplicationRecord
has_many :messages
end
We will update the conversation view to include the messages and a form to post another message in the thread:
<!-- app/views/conversations/show.html.erb -->
<div class="mt-8">
<%= render "messages/form", message: @conversation.messages.build %>
</div>
<div class="divide-y-2 mt-8">
<%= render @conversation.messages.order(created_at: :desc) %>
</div>
<!-- app/views/messages/_message.html.erb -->
<%= tag.div id: dom_id(message), class: "p-2" do %>
<span class="font-semibold"><%= message.role %>:</span>
<%= simple_format(message.message) %>
<% end %>
<!-- app/views/messages/_form.html.erb -->
<%= form_with(model: [message.conversation, message]) do |form| %>
<div class="flex gap-x-2">
<%= form.text_field :message, class: "grow" %>
<%= form.submit nil, class: "bg-gray-300 hover:bg-gray-400 px-2 py-1 cursor-pointer" %>
</div>
<% end %>
This will require a new nested route for the messages and a controller:
# config/routes.rb
resources :conversations do
resources :messages, only: :create
end
# app/controller/messages_controller.rb
class MessagesController < ApplicationController
def create
@conversation = Conversation.find(params[:conversation_id])
@conversation.messages.create({
role: :user,
message: params[:message][:message]
})
redirect_to @conversation
end
end
This will store all the messages from the user, but never send them to the AI. Lets add a method to do that:
# app/models/conversation.rb
class Conversation < ApplicationRecord
def update_thread
messages = self.messages.order(:created_at).map { {
role: _1.role,
content: _1.message
} }
response = ollama_client.chat({
model: "llama3",
stream: false,
messages: messages
}).first["message"]
self.messages.create({
role: response["role"],
message: response["content"]
})
end
end
In this version, we are waiting for the entire response (stream: false
), then creating a new message with the role from the assistant. This doesn’t change our interface, but is needed for the AI to keep track of things.
Now we just need to use it in the message controller:
class MessagesController < ApplicationController
def create
# …
@conversation.update_thread
redirect_to @conversation
end
end
One thing to note is that this is slow. We can address this well by using Turbo to broadcast updates. But I think that is a topic for later.