Installing LLAMA2#
To install LLAMA 2, we will use ollama. If you use Windows, you can follow the instructions from ollama’s official Docker image. Note that when using Docker, the model will be running in a container. Without tuning, it is quite slow.
If you use Linux or Mac, download ollama from ollama’s download page and follow the installation instructions. You can use Homebrew on Mac:
Once installed, you should have access to the ollama
command line tool:
Ollama can run multiple models, as you can see on ollama’s library.
We will use the llama2
model, which is a GPT-2 model trained on a large corpus of text.
For that, you first need to start the ollama server:
Then, pull the model using:
Warning
The LLAMA2 model is quite large (> 3.8GB). Make sure you have enough disk space.
You can check the list of available models using:
Once pulled, we will be able to use it.