Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com> Co-authored-by: Shivam Mishra <scm.mymail@gmail.com>
40 lines
1.0 KiB
Ruby
40 lines
1.0 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
# Base service for LLM operations using RubyLLM.
|
|
# New features should inherit from this class.
|
|
class Llm::BaseAiService
|
|
DEFAULT_MODEL = Llm::Config::DEFAULT_MODEL
|
|
DEFAULT_TEMPERATURE = 1.0
|
|
|
|
attr_reader :model, :temperature
|
|
|
|
def initialize
|
|
Llm::Config.initialize!
|
|
setup_model
|
|
setup_temperature
|
|
end
|
|
|
|
def chat(model: @model, temperature: @temperature)
|
|
RubyLLM.chat(model: model).with_temperature(temperature)
|
|
end
|
|
|
|
private
|
|
|
|
# Strips markdown code fences (```json ... ``` or ``` ... ```) that some
|
|
# LLM providers/gateways wrap around JSON responses despite response_format hints.
|
|
def sanitize_json_response(response)
|
|
return response if response.nil?
|
|
|
|
response.strip.sub(/\A```(?:\w*)\s*\n?/, '').sub(/\n?\s*```\s*\z/, '').strip
|
|
end
|
|
|
|
def setup_model
|
|
config_value = InstallationConfig.find_by(name: 'CAPTAIN_OPEN_AI_MODEL')&.value
|
|
@model = (config_value.presence || DEFAULT_MODEL)
|
|
end
|
|
|
|
def setup_temperature
|
|
@temperature = DEFAULT_TEMPERATURE
|
|
end
|
|
end
|