OllamaLLM
You are currently on a page documenting the use of Ollama models as text completion models. Many popular Ollama models are chat completion models.
You may be looking for this page instead.
This page goes over how to use LangChain to interact with Ollama
models.
Installation
# install package
%pip install -U langchain-ollama
Setup
First, follow these instructions to set up and run a local Ollama instance:
- Download and install Ollama onto the available supported platforms (including Windows Subsystem for Linux)
- Fetch available LLM model via
ollama pull <name-of-model>
- View a list of available models via the model library
- e.g.,
ollama pull llama3
- This will download the default tagged version of the model. Typically, the default points to the latest, smallest sized-parameter model.
On Mac, the models will be download to
~/.ollama/models
On Linux (or WSL), the models will be stored at
/usr/share/ollama/.ollama/models
- Specify the exact version of the model of interest as such
ollama pull vicuna:13b-v1.5-16k-q4_0
(View the various tags for theVicuna
model in this instance) - To view all pulled models, use
ollama list
- To chat directly with a model from the command line, use
ollama run <name-of-model>
- View the Ollama documentation for more commands. Run
ollama help
in the terminal to see available commands too.
Usage
from langchain_core.prompts import ChatPromptTemplate
from langchain_ollama.llms import OllamaLLM
template = """Question: {question}
Answer: Let's think step by step."""
prompt = ChatPromptTemplate.from_template(template)
model = OllamaLLM(model="llama3.1")
chain = prompt | model
chain.invoke({"question": "What is LangChain?"})
ChatOllama
In some cases, you will need a model with a "Chat" interface. (for more info see - https://python.langchain.com/docs/concepts/chat_models/).
from langchain_core.messages import ChatMessage
from langchain_ollama.chat_models import ChatOllama
llm = ChatOllama(
base_url="http://localhost:11434",
model="granite3.2",
verbose=True,
format=None,
num_predict=20,
)
messages = [
ChatMessage( # We can use ChatMessage objects to pass custom roles to a model in Ollama.
role="control", content="thinking"
),
# ChatMessage(
# role="user",
# content="What is the meaning of life?"
# ),
{ # Dicts also work for messages with standard roles.
"role": "user",
"content": "What is the meaning of life?",
},
]
llm.invoke(messages)
Multi-modal
Ollama has support for multi-modal LLMs, such as bakllava and llava.
ollama pull bakllava
Be sure to update Ollama so that you have the most recent version to support multi-modal.
import base64
from io import BytesIO
from IPython.display import HTML, display
from PIL import Image
def convert_to_base64(pil_image):
"""
Convert PIL images to Base64 encoded strings
:param pil_image: PIL image
:return: Re-sized Base64 string
"""
buffered = BytesIO()
pil_image.save(buffered, format="JPEG") # You can change the format if needed
img_str = base64.b64encode(buffered.getvalue()).decode("utf-8")
return img_str
def plt_img_base64(img_base64):
"""
Display base64 encoded string as image
:param img_base64: Base64 string
"""
# Create an HTML img tag with the base64 string as the source
image_html = f'<img src="data:image/jpeg;base64,{img_base64}" />'
# Display the image by rendering the HTML
display(HTML(image_html))
file_path = "../../../static/img/ollama_example_img.jpg"
pil_image = Image.open(file_path)
image_b64 = convert_to_base64(pil_image)
plt_img_base64(image_b64)
from langchain_ollama import OllamaLLM
llm = OllamaLLM(model="bakllava")
llm_with_image_context = llm.bind(images=[image_b64])
llm_with_image_context.invoke("What is the dollar based gross retention rate:")
Related
- LLM conceptual guide
- LLM how-to guides