====== Ollama: Local Large Language Model Execution ====== Ollama is a tool that allows users to run large language models (LLMs) locally on their machines without relying on cloud services. This ensures greater privacy, data control, and offline usage capabilities. ===== Key Features ===== * **Local Model Execution:** Install and run AI models directly on your device, such as Llama 3.3, DeepSeek-R1, Phi-4, Mistral, and Gemma 2. * **Cross-Platform Compatibility:** Available for macOS, Linux, and Windows, making it accessible on multiple environments. * **Command Line Interface (CLI):** Operates through the terminal or command prompt, offering efficient interaction with installed models. * **Privacy and Data Control:** Since the tool runs locally, your data is not sent to external servers, ensuring enhanced security and privacy. ===== Installation and Basic Usage ===== 1. **Download and Install:** - Visit [Ollama Official Website](https://ollama.com/) and download the appropriate version for your operating system. - Follow the installer instructions to complete the setup. 2. **Using the Terminal:** - After installation, open your system's terminal or command prompt. - Run models using simple commands. For example, to run the Mistral model, use: ~$ ollama run mistral ===== Supported Models ===== Ollama supports several popular large language models, including but not limited to: * **Llama** (all versions) * **DeepSeek-R1** * **Phi-4** * **Mistral** * **Gemma 2** ===== Advantages of Ollama ===== - **Offline Functionality:** No internet connection is needed once models are installed. - **Data Security:** Data remains on the local device, eliminating the risk of data breaches from cloud services. - **High Performance:** Running models locally can offer faster responses depending on system specifications. ====== Model library ====== Ollama supports a list of models available on [ollama.com/library](https://ollama.com/library) Here are some example models that can be downloaded: ^ Model ^ Parameters ^ Size ^ Download Command ^ | DeepSeek-R1 | 7B | 4.7GB | `ollama run deepseek-r1` | | DeepSeek-R1 | 671B | 404GB | `ollama run deepseek-r1:671b` | | Llama 3.3 | 70B | 43GB | `ollama run llama3.3` | | Llama 3.2 | 3B | 2.0GB | `ollama run llama3.2` | | Llama 3.2 | 1B | 1.3GB | `ollama run llama3.2:1b` | | Llama 3.2 Vision | 11B | 7.9GB | `ollama run llama3.2-vision` | | Llama 3.2 Vision | 90B | 55GB | `ollama run llama3.2-vision:90b` | | Llama 3.1 | 8B | 4.7GB | `ollama run llama3.1` | | Llama 3.1 | 405B | 231GB | `ollama run llama3.1:405b` | | Phi 4 | 14B | 9.1GB | `ollama run phi4` | | Phi 3 Mini | 3.8B | 2.3GB | `ollama run phi3` | | Gemma 2 | 2B | 1.6GB | `ollama run gemma2:2b` | | Gemma 2 | 9B | 5.5GB | `ollama run gemma2` | | Gemma 2 | 27B | 16GB | `ollama run gemma2:27b` | | Mistral | 7B | 4.1GB | `ollama run mistral` | | Moondream 2 | 1.4B | 829MB | `ollama run moondream` | | Neural Chat | 7B | 4.1GB | `ollama run neural-chat` | | Starling | 7B | 4.1GB | `ollama run starling-lm` | | Code Llama | 7B | 3.8GB | `ollama run codellama` | | Llama 2 Uncensored | 7B | 3.8GB | `ollama run llama2-uncensored` | | LLaVA | 7B | 4.5GB | `ollama run llava` | | Solar | 10.7B | 6.1GB | `ollama run solar` | ==== Note ==== You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 13B models, and 32 GB to run the 33B models. ===== CLI Reference ===== Create a model ollama create is used to create a model from a Modelfile. ollama create mymodel -f ./Modelfile Pull a model ollama pull llama3.2 This command can also be used to update a local model. Only the diff will be pulled. Remove a model ollama rm llama3.2 Copy a model ollama cp llama3.2 my-model Multiline input For multiline input, you can wrap text with """: >>> """Hello, ... world! ... """ I'm a basic program that prints the famous "Hello, world!" message to the console. Multimodal models ollama run llava "What's in this image? /Users/jmorgan/Desktop/smile.png" Output: The image features a yellow smiley face, which is likely the central focus of the picture. Pass the prompt as an argument ollama run llama3.2 "Summarize this file: $(cat README.md)" Show model information ollama show llama3.2 List models on your computer ollama list List which models are currently loaded ollama ps Stop a model which is currently running ollama stop llama3.2 Start Ollama ollama serve is used when you want to start ollama without running the desktop application. ===== How to Run Ollama and Connect to the Service API Through Internal Network or Internet ===== Setting Environment Variables on Linux If Ollama is run as a systemd service, environment variables should be set using systemctl: Edit the Ollama Service File: Open the Ollama service configuration file with the following command: sudo systemctl edit ollama.service Add the Environment Variable: In the editor, add the following lines under the [Service] section: [Service] Environment="OLLAMA_HOST=0.0.0.0" Note #1: Sometimes, 0.0.0.0 does not work due to your environment setup. Instead, you can try setting it to your local ip address like 10.0.0.x or xxx.local, etc. Note #2: You should put this above this line ### Lines below this comment will be discarded. It should look something like this: ### Editing /etc/systemd/system/ollama.service.d/override.conf ### Anything between here and the comment below will become the new contents of the file [Service] Environment="OLLAMA_HOST=0.0.0.0" ### Lines below this comment will be discarded ### /etc/systemd/system/ollama.service # [Unit] # Description=Ollama Service # After=network-online.target # # [Service] # ExecStart=/usr/local/bin/ollama serve # User=ollama # Group=ollama # Restart=always # RestartSec=3 # Environment="PATH=/home/kimi/.nvm/versions/node/v20.5.0/bin:/home/kimi/.local/share/pnpm:/usr/local/sbin:/usr/local/bin:/usr/s> # # [Install] # WantedBy=default.target Restart the Service: After editing the file, reload the systemd daemon and restart the Ollama service: sudo systemctl daemon-reload sudo systemctl restart ollama ===== Learn More ===== For more detailed information and tutorials, visit [Ollama's official website](https://ollama.com/) or check out this [video overview](https://www.youtube.com/watch?v=wxyDEqR4KxM).