MultiLLMPromptExecutor
MultiLLMPromptExecutor is a class responsible for executing prompts across multiple Large Language Models (LLMs). This implementation supports direct execution with specific LLM clients or utilizes a fallback strategy if no primary LLM client is available for the requested provider.
Parameters
A map containing LLM providers associated with their respective LLMClients.
Optional settings to configure the fallback mechanism in case a specific provider is not directly available.
Constructors
Constructs an executor instance with a map of LLM providers associated with their respective clients.
Initializes a new instance of the MultiLLMPromptExecutor class with multiple LLM clients.
Secondary constructor for MultiLLMPromptExecutor that accepts a list of LLMClient instances. The provided clients are processed to create a mapping of LLMProvider to their respective LLMClient.
Secondary constructor for MultiLLMPromptExecutor that accepts a variable number of LLMClient instances. The provided clients are processed to create a mapping of LLMProvider to their respective LLMClient.
Types
Represents configuration for a fallback large language model (LLM) execution strategy.
Functions
Executes a given prompt using the specified tools and model, and returns a list of response messages.
Executes a given prompt using the specified tools and model and returns a list of model choices.
Executes the given prompt with the specified model and streams the response in chunks as a flow.
Executes a prompt with structured output, enhancing it with schema instructions or native structured output parameter, and parses the response into the defined structure.
Basic JSON schema generator required for the given model. Return BasicJsonSchemaGenerator by default.
Standard JSON schema generator required for the given model. Return StandardJsonSchemaGenerator by default.
Moderates the provided multi-modal content using the specified model.
Parses a structured response from the assistant message using the provided structured output configuration and language model. If a fixing parser is specified in the configuration, it will be used; otherwise, the structure will be parsed directly.