OllamaLLM¶
Ollama LLM implementation running the Async API client.
Attributes¶
- 
model: the model name to use for the LLM e.g. "notus". 
- 
host: the Ollama server host. 
- 
timeout: the timeout for the LLM. Defaults to 120.
- 
_aclient: the AsyncClientto use for the Ollama API. It is meant to be used internally. Set in theloadmethod.
Runtime Parameters¶
- 
host: the Ollama server host. 
- 
timeout: the client timeout for the Ollama API. Defaults to 120.