Ollama is an Open Source project by Meta that provides a platform for creating, running, and managing multiple different Large Language Models or LLM locally on a machine. Ollama acts as a bridge for LLM Model communication.
Ollama can be compared to Docker, Docker also allows to download of multiple images from the Docker Repository and runs in containers.
The latest versions of Ollama can be downloaded from the official website.
Jeffrey Morgan and Michael Chiang are the founders of Ollama
Ollama is available for Macintosh, Linux, and Windows operating systems. its command line Interface allows easy download, run, and manage Large Language Models locally.
Ollama meaning is to play with a rubber ball using the buttocks
Ollama can run on CPU
Ollama can run on Intel CPU also, but it should be avoided. Ollama is designed to give better performance but requires Nvidia or AMD GPUs, for development purposes, Ollama can run on CPU only, but it will not give better performance, even with 16 core CPU.
Enhanced Privacy
Ollama once downloaded the Lange Language Model from the internet, doesn't require a network connection. so all the communication and sensitive information never leaves your secure environment, on the other hand, Large Language Models provided by Other vendors collect data to train Models.
Zero Operational Cost
Large Language Models like Open AI's Chat GPT, or Google AI support millions of users simultaneously, which requires high processing power to run, and in return, they charge some amount based on consumption to support infrastructure, whereas Ollama runs on the local machine, handling a single user does not require that much high computation, hence the operation cost is none.
Ollama Model Library
Currently, Ollama supports the below Large Language Models.
Model | Parameters | Size | Download |
---|---|---|---|
Llama 3.2 | 3B | 2.0GB | ollama run llama3.2 |
Llama 3.2 | 1B | 1.3GB | ollama run llama3.2:1b |
Llama 3.1 | 8B | 4.7GB | ollama run llama3.1 |
Llama 3.1 | 70B | 40GB | ollama run llama3.1:70b |
Llama 3.1 | 405B | 231GB | ollama run llama3.1:405b |
Phi 3 Mini | 3.8B | 2.3GB | ollama run phi3 |
Phi 3 Medium | 14B | 7.9GB | ollama run phi3:medium |
Gemma 2 | 2B | 1.6GB | ollama run gemma2:2b |
Gemma 2 | 9B | 5.5GB | ollama run gemma2 |
Gemma 2 | 27B | 16GB | ollama run gemma2:27b |
Mistral | 7B | 4.1GB | ollama run mistral |
Moondream 2 | 1.4B | 829MB | ollama run moondream |
Neural Chat | 7B | 4.1GB | ollama run neural-chat |
Starling | 7B | 4.1GB | ollama run starling-lm |
Code Llama | 7B | 3.8GB | ollama run codellama |
Llama 2 Uncensored | 7B | 3.8GB | ollama run llama2-uncensored |
LLaVA | 7B | 4.5GB | ollama run llava |
Solar | 10.7B | 6.1GB | ollama run solar |