LiteLLM¶
LiteLLM implementation running the async API client.
Attributes¶
-
model: the model name to use for the LLM e.g. "gpt-3.5-turbo" or "mistral/mistral-large", etc.
-
verbose: whether to log the LiteLLM client's logs. Defaults to
False
. -
structured_output: a dictionary containing the structured output configuration configuration using
instructor
. You can take a look at the dictionary structure inInstructorStructuredOutputType
fromdistilabel.steps.tasks.structured_outputs.instructor
.
Runtime Parameters¶
- verbose: whether to log the LiteLLM client's logs. Defaults to
False
.