Ollama

Run Large Language Models locally with Ollama

Ollama is a fantastic tool for running models locally. Install ollama and run a model using

run modelserve

ollama run llama3.1

After you have the local model running, use the Ollama model to access them

Example

agent.py

from bitca.agent import Agent, RunResponse
from bitca.model.ollama import Ollama

agent = Agent(
    model=Ollama(id="llama3.1"),
    markdown=True
)

# Get the response in a variable
# run: RunResponse = agent.run("Share a 2 sentence horror story.")
# print(run.content)

# Print the response in the terminal
agent.print_response("Share a 2 sentence horror story.")

Params

Parameter
Type
Default
Description

id

str

"llama3.2"

The ID of the model to use.

name

str

"Ollama"

The name of the model.

provider

str

"Ollama llama3.2"

The provider of the model.

format

Optional[str]

None

The format of the response.

options

Optional[Any]

None

Additional options to pass to the model.

keep_alive

Optional[Union[float, str]]

None

The keep alive time for the model.

request_params

Optional[Dict[str, Any]]

None

Additional parameters to pass to the request.

host

Optional[str]

None

The host to connect to.

timeout

Optional[Any]

None

The timeout for the connection.

client_params

Optional[Dict[str, Any]]

None

Additional parameters to pass to the client.

client

Optional[OllamaClient]

None

A pre-configured instance of the Ollama client.

async_client

Optional[AsyncOllamaClient]

None

A pre-configured instance of the asynchronous Ollama client.

Last updated