O

Ollama

4.5
Free

A tool for running large language models locally on your machine. Supports Llama, Mistral, Gemma, and dozens of other open-source models with a simple command-line interface and local API.

Key Features

  • Run LLMs locally
  • Simple CLI interface
  • Local API server
  • Custom model creation
  • GPU acceleration

Pros

  • + Completely free and local
  • + Dead-simple to use
  • + Great privacy — no data leaves your machine

Cons

  • Requires powerful hardware
  • Smaller models than cloud options
local open-source CLI privacy self-hosted