Pull to refresh

Run AI Locally: Llama 3.2 OpenWebUI Tutorial

Level of difficultyEasy
Reading time3 min
Views1.4K

Did you know you can run powerful AI models right on your computer? It's true! Today, I will show you how easy it is to get started with Llama 3.2 and OpenWebUI.

Watch on YouTube: Run AI On YOUR Computer

Running Llama 3.2 Locally: A Comprehensive Guide

Introduction to Llama 3.2

Llama 3.2 is the latest iteration of Meta's open-source language model, offering enhanced capabilities for text and image processing. It is designed to run efficiently on local devices, making it ideal for applications that require privacy and low latency. The model comes in various sizes, including 1B, 3B, and 11B parameters. In this tutorial, I'm going to use the 1B model, but you can download any you like.

Introduction to Llama 3.2
Introduction to Llama 3.2

Before setting up Llama 3.2 locally, ensure you have the following:

  • A computer with Windows, macOS, or Linux. I'm going to use Linux (Ubuntu).

  • Basic knowledge of using the terminal or command prompt.

Step-by-Step Installation Guide

Step 1: Install Docker

Docker is a tool that allows developers to package applications and their dependencies into a standardized unit called a container, which can run consistently across different computing environments. Unlike virtual machines, containers are lightweight and share the host system's operating system, making them more efficient and faster to start.

Docker
Docker

You can download the Docker Desktop application from the Docker website https://www.docker.com/ or via the terminal. Docker is available for Mac, Windows, and Linux.

How to Install Docker on Ubuntu

Download the package https://desktop.docker.com/linux/main/amd64/docker-desktop-amd64.deb

Then use these commands:

sudo apt-get update
sudo apt-get install [path to the docker package].deb

More details: https://docs.docker.com/desktop/install/linux/ubuntu/

Step 2: Install Ollama

Ollama is essential for running large language models like Llama 3.2 locally.

Install Ollama
Install Ollama

Follow these steps to install it:

  • Visit the official Ollama website and download the installer for your operating system.

  • Run the installer and follow the on-screen instructions to complete the installation.

  • Verify the installation by opening a terminal (or command prompt) and typing `ollama`. You should see a list of commands if installed correctly.

I'm going to use the terminal and this command:

curl -fsSL https://ollama.com/install.sh | sh

Step 3: Download a New Llama 3.2 model from Ollama website

Once Ollama is set up, you can download Llama 3.2 models. All the available models you can find here https://ollama.com/library

: Download a New Llama 3.2
: Download a New Llama 3.2

Since I want to use 1B parameters model I'll be using this command:

ollama run llama3.2:1b

Step 4: Install OpenWebUI with Default Configuration

OpenWebUI is a self-hosted, extensible web interface designed to interact entirely offline with large language models (LLMs). It offers a user-friendly experience similar to ChatGPT, supports integration with various LLMs, such as those compatible with OpenAI and Ollama, and provides features like markdown support, model management, and multi-user access.

Go to https://docs.openwebui.com/getting-started/ and find the section Quick Start with Docker. Copy the code and then run it in the terminal.

I'm going to run this code:

docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v ollama:/root/.ollama -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:ollama

We will run a Docker command. It is used to run a containerized version of the OpenWebUI application.

Once the installation is finished, open your browser and go to http://0.0.0.0:3000/. If everything is okay, you will see the website. Create an account, and then you will see the welcome screen, which is similar to ChatGPT.

Llama 3.2 in Browser
Llama 3.2 in Browser

Troubleshooting Common Issue

WebUI models not being pulled/applied

Some of you could face an issue when the models are not available in the drop-down. To fix it follow these instructions:

  1. Go to the settings at the bottom left corner.

    Settings
    Settings
  2. Click on "Admin Settings".

    Admin Settings
    Admin Settings
  3. Click on "Models."

    Models
    Models
  4. In the input field, write the name of the model you would like to pull and click on the "download" icon.

Pull the models
Pull the models

All the names you can find on https://ollama.com/library/llama3.2

This should fix the issue.

Conclusion

Running Llama 3.2 locally provides significant advantages regarding privacy and control over AI applications. But to have a smooth experience, you would need a powerful computer. :)

If you like this tutorial, please follow me on YouTube, join my Telegram, or support me on Patreon.

Thanks! :)

Tags:
Hubs:
-1
Comments0

Articles