How to Run Llama-3.1🦙 locally in Python using Ollama, LangChain

emmakodes_

Emmanuel Onwuegbusi

Posted on August 2, 2024

How to Run Llama-3.1🦙 locally in Python using Ollama, LangChain

In this article, we will learn how to run Llama-3.1 model locally on our PC using Ollama and LangChain in Python

Outline

  • Install Ollama
  • Pull model
  • Serve model
  • Create a new folder, open it with a code editor
  • Create and activate Virtual environment
  • Install langchain-ollama
  • Run Ollama with model in Python
  • Conclusion

Install Ollama

Follow the instructions based on your OS type in its GitHub README to install Ollama:

https://github.com/ollama/ollama
Enter fullscreen mode Exit fullscreen mode

I am on a Linux-based PC, so I am going to run the following command in my terminal:

curl -fsSL https://ollama.com/install.sh | sh
Enter fullscreen mode Exit fullscreen mode

Pull model

Fetch the available LLM model via the following command:

ollama pull llama3.1
Enter fullscreen mode Exit fullscreen mode

This will download the default tagged version of the model. Typically, the default points to the latest, smallest sized-parameter model. In this case, it will be llama3.1:8b model.

To download another version of the model, you can go to: https://ollama.com/library/llama3.1 and select the version to install, and then run the ollama pull command with the model and its version number. Example: ollama pull llama3.1:70b

On Mac, the models will be downloaded to ~/.ollama/models

On Linux (or WSL), the models will be stored at /usr/share/ollama/.ollama/models

Serve model

Run the following command to start ollama without running the desktop application.

ollama serve
Enter fullscreen mode Exit fullscreen mode

All models are automatically served on localhost:11434

Create a new folder, open it with a code editor

Create a new folder on your computer and then open it with a code editor like VS Code.

Create and activate Virtual environment

Open the terminal. Use the following command to create a virtual environment .venv and activate it:

python3 -m venv .venv
Enter fullscreen mode Exit fullscreen mode
source .venv/bin/activate
Enter fullscreen mode Exit fullscreen mode

Install langchain-ollama

Run the following command to install langchain-ollama:

pip install -U langchain-ollama
Enter fullscreen mode Exit fullscreen mode

The above command will install or upgrade the LangChain Ollama package in Python. This package allows users to integrate and interact with Ollama models, which are open-source large language models, within the LangChain framework. The -U flag ensures that the package is upgraded to the latest version if it is already installed.

Run Ollama with model in Python

Create a Python file for example: main.py and add the following code:

from langchain_ollama import OllamaLLM

llm = OllamaLLM(model="llama3.1")

response = llm.invoke("The first man on the moon was ...")
print(response)
Enter fullscreen mode Exit fullscreen mode

The above code imports the OllamaLLM class from the LangChain library and initializes an instance of the language model "llama3.1". We pass a prompt about the first man on the moon, and store the generated response in the variable response. When we run the above code we get the following response from the model:

...Neil Armstrong!

On July 20, 1969, Neil Armstrong became the first person to set foot on the lunar surface during the Apollo 11 mission. As he stepped off the lunar module Eagle and onto the moon's surface, he famously declared: "That's one small step for man, one giant leap for mankind."
Enter fullscreen mode Exit fullscreen mode

Conclusion

Thanks for reading.
You can view the Ollama documentation for more commands.

💖 💪 🙅 🚩
emmakodes_
Emmanuel Onwuegbusi

Posted on August 2, 2024

Join Our Newsletter. No Spam, Only the good stuff.

Sign up to receive the latest update from our blog.

Related