PromptExecutorProxy

A wrapper around ai.koog.prompt.executor.model.PromptExecutor that allows for adding internal functionality to the executor to catch and log events related to LLM calls.

Constructors

Link copied to clipboard
constructor(executor: PromptExecutor, pipeline: AIAgentPipeline, runId: String)

Functions

Link copied to clipboard
open suspend override fun execute(prompt: Prompt, model: LLModel, tools: List<ToolDescriptor>): List<Message.Response>
Link copied to clipboard
open suspend override fun executeMultipleChoices(prompt: Prompt, model: LLModel, tools: List<ToolDescriptor>): List<LLMChoice>
Link copied to clipboard
open suspend override fun executeStreaming(prompt: Prompt, model: LLModel): Flow<String>
Link copied to clipboard
open suspend override fun moderate(prompt: Prompt, model: LLModel): ModerationResult