Unlock the Power of Meta Llama LLM: Easy Guide to Hosting in Your Local Dev Environment
Bradston Henry
Posted on November 22, 2024
In my previous post I described how AI tools have revolutionized my Development workflow. Toward the end of the blog, I shared the step-by-step on how to host the Meta Llama LLM on your local machine.
Because it was alllll the way at the end of the blog, I figured I'd make a blog that showed those steps separately as many people have been finding value from it.
Check it out!
How do I Host my OWN local AI Chat Agent
So let me walk you through the few simple steps I used to start using Llama (and other available Open-Source AI Models):
Step 1: Download and Install Ollama on Your Local Machine
Navigate to the Official Ollama site and quickly download the Ollama for your Windows, Mac, or Linux Machine.
Ollama is light-weight tool that allows you to run Large Language Models on your local machine (e.g Llama 3.2, Mistral, and Gemma 2). Once installed, you are able to run and customize models on your local machine.
Step 2: Run the Ollama Setup Wizard
Once downloaded, open the Ollama Setup executable and navigate through the wizard to install the Ollama tools on your machine.
As you might expect, Ollama's Setup Wizard will quickly add all necessary file to run the Ollama tool
Step 3: Verify Ollama Installation
If Ollama was installed correctly, you should now be able to access it from your standard command prompt.
Sometimes, when installing Ollama it will automatically open a command prompt window, if not you will need to open it yourself.
On a Windows machine, search for "cmd" and you should be able to open Command Prompt.
Once open, verify the installation by typing the line below and pressing Enter/Return:
ollama
You should see an output like this:
Usage:
ollama [flags]
ollama [command]
Available Commands:
serve Start ollama
create Create a model from a Modelfile
show Show information for a model
run Run a model
stop Stop a running model
pull Pull a model from a registry
push Push a model to a registry
list List models
ps List running models
cp Copy a model
rm Remove a model
help Help about any command
Flags:
-h, --help help for ollama
-v, --version Show version information
Use "ollama [command] --help" for more information about a command.
Step 4: Install and Run a Large Language Model
With Ollama installed, we are only one step away from having our own local LLM on our machine.
To get started, let's install Meta's Llama3.2 model.
In our command prompt window, type:
ollama run llama3.2
This will do two things:
- Install Llama3.2 if it is not currently on our machine
- Run the Llama3.2 Model
Press return/enter and you should see something like this (if Llama3.2 is not installed):
pulling manifest
pulling dde5aa3fc5ff... 26% ▕████████████ ▏ 532 MB/2.0 GB 32 MB/s 48s
Ollama will now download and install the Llama3.2 model on your machine and once complete, run the llama3.2 model for you to interact with.
In your command line window, you should see something like this:
pulling manifest
pulling dde5aa3fc5ff... 100% ▕████████████████████████████████████████████████████████▏ 2.0 GB
pulling 966de95ca8a6... 100% ▕████████████████████████████████████████████████████████▏ 1.4 KB
pulling fcc5a6bec9da... 100% ▕████████████████████████████████████████████████████████▏ 7.7 KB
pulling a70ff7e570d9... 100% ▕████████████████████████████████████████████████████████▏ 6.0 KB
pulling 56bb8bd477a5... 100% ▕████████████████████████████████████████████████████████▏ 96 B
pulling 34bb5ab01051... 100% ▕████████████████████████████████████████████████████████▏ 561 B
verifying sha256 digest
writing manifest
success
>>> Send a message (/? for help)
And just like that, you now have a LLM on your local machine!!
Also, If you would like to install any other model available through Ollama, you would use the same command:
ollama run MODEL_NAME
Step 5: Test out your Large Language Model
Now that Llama3.2 is on our machine, let's test it out.
Let's ask it a simple question:
Hey Llama! Could you tell me a little bit about yourself?
You should then get a response similar to this:
Hey Llama! Could you tell me a little bit about yourself?
I'm happy to chat with you. I'm an artificial intelligence model known as Llama, which stands for "Large Language
Model Meta AI." My primary function is to process and generate human-like text based on the input I receive.
I was trained on a massive dataset of text from various sources, including books, articles, and online
conversations. This training allows me to understand and respond to a wide range of questions, topics, and styles.
Some key features of my abilities include:
.....
See! Wasn't that pretty easy!
Wrap Up
And just like that, Your Llama LLM should be up and running with ease on your local machine.
If you are interested in checking out Ollamas Docs you can find them here
If you are interested in checking out a GUI that helps you utilize Ollama even easier, check out Open Web UI. Hope to have a blog up in the near future with a step-by-step on how to set that up as well.
Until then, enjoy you own personal Llama personal assistant!
Photo Credits(Order of Appearance):
Cover Photo with Google Gemini
Prompt:
A llama with a confident smirk on its face, wearing a blue cape emblazoned with a infinity symbol, flies in against a sunset, ready to save the day in this cartoon style illustration.
Follow me on my Socials:
https://linktr.ee/bradstondev
Posted on November 22, 2024
Join Our Newsletter. No Spam, Only the good stuff.
Sign up to receive the latest update from our blog.