mirror of
https://github.com/lingble/chatwoot.git
synced 2025-11-09 16:46:11 +00:00
Co-authored-by: Muhsin Keloth <muhsinkeramam@gmail.com> Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com> Co-authored-by: Sojan Jose <sojan@pepalo.com> Co-authored-by: Pranav <pranav@chatwoot.com> Co-authored-by: Sivin Varghese <64252451+iamsivin@users.noreply.github.com>
187 lines
5.0 KiB
Ruby
187 lines
5.0 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
require 'rails_helper'
|
|
|
|
RSpec.describe Concerns::Agentable do
|
|
let(:dummy_class) do
|
|
Class.new do
|
|
include Concerns::Agentable
|
|
|
|
attr_accessor :temperature
|
|
|
|
def initialize(name: 'Test Agent', temperature: 0.8)
|
|
@name = name
|
|
@temperature = temperature
|
|
end
|
|
|
|
def self.name
|
|
'DummyClass'
|
|
end
|
|
|
|
private
|
|
|
|
def agent_name
|
|
@name
|
|
end
|
|
|
|
def prompt_context
|
|
{ base_key: 'base_value' }
|
|
end
|
|
end
|
|
end
|
|
|
|
let(:dummy_instance) { dummy_class.new }
|
|
let(:mock_agents_agent) { instance_double(Agents::Agent) }
|
|
let(:mock_installation_config) { instance_double(InstallationConfig, value: 'gpt-4-turbo') }
|
|
|
|
before do
|
|
allow(Agents::Agent).to receive(:new).and_return(mock_agents_agent)
|
|
allow(InstallationConfig).to receive(:find_by).with(name: 'CAPTAIN_OPEN_AI_MODEL').and_return(mock_installation_config)
|
|
allow(Captain::PromptRenderer).to receive(:render).and_return('rendered_template')
|
|
end
|
|
|
|
describe '#agent' do
|
|
it 'creates an Agents::Agent with correct parameters' do
|
|
expect(Agents::Agent).to receive(:new).with(
|
|
name: 'Test Agent',
|
|
instructions: instance_of(Proc),
|
|
tools: [],
|
|
model: 'gpt-4-turbo',
|
|
temperature: 0.8,
|
|
response_schema: Captain::ResponseSchema
|
|
)
|
|
|
|
dummy_instance.agent
|
|
end
|
|
|
|
it 'converts nil temperature to 0.0' do
|
|
dummy_instance.temperature = nil
|
|
|
|
expect(Agents::Agent).to receive(:new).with(
|
|
hash_including(temperature: 0.0)
|
|
)
|
|
|
|
dummy_instance.agent
|
|
end
|
|
|
|
it 'converts temperature to float' do
|
|
dummy_instance.temperature = '0.5'
|
|
|
|
expect(Agents::Agent).to receive(:new).with(
|
|
hash_including(temperature: 0.5)
|
|
)
|
|
|
|
dummy_instance.agent
|
|
end
|
|
end
|
|
|
|
describe '#agent_instructions' do
|
|
it 'calls Captain::PromptRenderer with base context' do
|
|
expect(Captain::PromptRenderer).to receive(:render).with(
|
|
'dummy_class',
|
|
hash_including(base_key: 'base_value')
|
|
)
|
|
|
|
dummy_instance.agent_instructions
|
|
end
|
|
|
|
it 'merges context state when provided' do
|
|
context_double = instance_double(Agents::RunContext,
|
|
context: {
|
|
state: {
|
|
conversation: { id: 123 },
|
|
contact: { name: 'John' }
|
|
}
|
|
})
|
|
|
|
expected_context = {
|
|
base_key: 'base_value',
|
|
conversation: { id: 123 },
|
|
contact: { name: 'John' }
|
|
}
|
|
|
|
expect(Captain::PromptRenderer).to receive(:render).with(
|
|
'dummy_class',
|
|
hash_including(expected_context)
|
|
)
|
|
|
|
dummy_instance.agent_instructions(context_double)
|
|
end
|
|
|
|
it 'handles context without state' do
|
|
context_double = instance_double(Agents::RunContext, context: {})
|
|
|
|
expect(Captain::PromptRenderer).to receive(:render).with(
|
|
'dummy_class',
|
|
hash_including(
|
|
base_key: 'base_value',
|
|
conversation: {},
|
|
contact: {}
|
|
)
|
|
)
|
|
|
|
dummy_instance.agent_instructions(context_double)
|
|
end
|
|
end
|
|
|
|
describe '#template_name' do
|
|
it 'returns underscored class name' do
|
|
expect(dummy_instance.send(:template_name)).to eq('dummy_class')
|
|
end
|
|
end
|
|
|
|
describe '#agent_tools' do
|
|
it 'returns empty array by default' do
|
|
expect(dummy_instance.send(:agent_tools)).to eq([])
|
|
end
|
|
end
|
|
|
|
describe '#agent_model' do
|
|
it 'returns value from InstallationConfig when present' do
|
|
expect(dummy_instance.send(:agent_model)).to eq('gpt-4-turbo')
|
|
end
|
|
|
|
it 'returns default model when config not found' do
|
|
allow(InstallationConfig).to receive(:find_by).and_return(nil)
|
|
|
|
expect(dummy_instance.send(:agent_model)).to eq('gpt-4.1-mini')
|
|
end
|
|
|
|
it 'returns default model when config value is nil' do
|
|
allow(mock_installation_config).to receive(:value).and_return(nil)
|
|
|
|
expect(dummy_instance.send(:agent_model)).to eq('gpt-4.1-mini')
|
|
end
|
|
end
|
|
|
|
describe '#agent_response_schema' do
|
|
it 'returns Captain::ResponseSchema' do
|
|
expect(dummy_instance.send(:agent_response_schema)).to eq(Captain::ResponseSchema)
|
|
end
|
|
end
|
|
|
|
describe 'required methods' do
|
|
let(:incomplete_class) do
|
|
Class.new do
|
|
include Concerns::Agentable
|
|
end
|
|
end
|
|
|
|
let(:incomplete_instance) { incomplete_class.new }
|
|
|
|
describe '#agent_name' do
|
|
it 'raises NotImplementedError when not implemented' do
|
|
expect { incomplete_instance.send(:agent_name) }
|
|
.to raise_error(NotImplementedError, /must implement agent_name/)
|
|
end
|
|
end
|
|
|
|
describe '#prompt_context' do
|
|
it 'raises NotImplementedError when not implemented' do
|
|
expect { incomplete_instance.send(:prompt_context) }
|
|
.to raise_error(NotImplementedError, /must implement prompt_context/)
|
|
end
|
|
end
|
|
end
|
|
end
|