Ad Code

How to: Install & Run A Standalone AI Model Locally with Llama 2

How to: Install & Running LlaMA 2 Locally

We will learn a simple way to install and use Llama 2 without setting up Python or any program. Just download the files and run a command in PowerShell.

 

New open source models like LLaMA 2 have become quite advanced and are free to use. You can use them commercially or fine-tune them on your own data to develop specialized versions. With their ease of use, you can now run them locally on your own device.

In this post, we will learn how to download the necessary files and the LLaMA 2 model to run the CLI program and interact with an AI assistant. The setup is simple enough that even non-technical users or students can get it running by following a few basic steps.

 

Downloading Llama.cpp for GPU machine

 

To install llama.cpp locally, the simplest method is to download the pre-built executable from the llama.cpp releases

To install it on Windows 11 with the NVIDIA GPU, we need to first download the llama-master-eb542d3-bin-win-cublas-[version]-x64.zip file. After downloading, extract it in the directory of your choice. It is recommended to create a new folder and extract all the files in it. 

Next, we will download the cuBLAS drivers cudart-llama-bin-win-[version]-x64.zip and extract them in the main directory. For using the GPU acceleration, you have two options: cuBLAS for NVIDIA GPUs and clBLAS for AMD GPUs.

 

Note: The [version] is the version of the CUDA installed on your local system. You can check it by running nvcc --version in the terminal.

 

A Simple Guide to Running LlaMA 2 Locally

 

Downloading the Model

 

To begin, create a folder named “Models” in the main directory. Within the Models folder, create a new folder named “llama2_7b”. Next, download the LLaMA 2 model file from the Hugging Face hub. You can choose any version you prefer, but for this guide, we will be downloading the llama-2-7b-chat.Q5_K_M.gguf file. Once the download is complete, move the file into the “llama2_7b” folder you just created.

 

A Simple Guide to Running LlaMA 2 Locally

 

Note: To avoid any errors, please make sure to download only the .gguf model files before running the mode.

 

Initiating the AI Assistant CLI Program

 

You can now open the terminal in the main directory. By right clicking and selecting “Open in Terminal” option. You can also open PowerShell and the us “cd” to change directory.

 

A Simple Guide to Running LlaMA 2 Locally

 

Copy and paste the command below and press “Enter”. We are executing the main.exe file with model directory location, gpu, color, and system prompt arguments. 

./main.exe -m .\Models\llama2_7b\llama-2-7b-chat.Q5_K_M.gguf -i --n-gpu-layers 32 -ins --color -p "<<SYS>> As an AI assistant, your core values include being supportive, considerate, and truthful, ensuring that every interaction is guided by clarity and helpfulness. <</SYS>>"

 

A Simple Guide to Running LlaMA 2 Locally

 

Our llama.ccp CLI program has been successfully initialized with the system prompt. It tells us it's a helpful AI assistant and shows various commands to use.

 

Using LLaMA 2 Locally in PowerShell

 

Let’s test out the LLaMA 2 in the PowerShell by providing the prompt. We have asked a simple question about the age of the earth.

The answer is accurate. Let’s ask a follow up question about earth. 

 

A Simple Guide to Running LlaMA 2 Locally

 

As you can see, the model has provided us with multiple interesting facts about our planet. 

 

A Simple Guide to Running LlaMA 2 Locally

 

You can ask the AI assistant to generate code and an explanation in the terminal, which you can easily copy and use in your IDE.

 

A Simple Guide to Running LlaMA 2 Locally

 

Perfect.

 

Conclusion

 

Running Llama 2 locally provides a powerful yet easy-to-use chatbot experience  that is customized to your needs. By following this simple guide, you can learn to build your own private chatbot set up in no time without needing to rely on paid services.

The main benefits of running LlaMA 2 locally are full control over your data and conversations as well as no usage limits. You can chat with your bot as much as you want and even tweak it to improve responses.

While less convenient than an instantly available cloud AI API, local setup brings peace of mind regarding data privacy

Post a Comment

0 Comments