ollafit

command module
v0.0.0-...-249e08f Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Apr 1, 2026 License: MIT Imports: 1 Imported by: 0

README

ollafit

Go License Release

Check which Ollama models can run on your hardware. Detects your GPU, VRAM, RAM, and CPU, then shows a color-coded compatibility table for 370+ models.

Features

  • Interactive TUI with 4 tabs: Hardware, Compatibility, Models, Pull
  • Hardware detection for NVIDIA, AMD, Apple Silicon, and CPU-only setups
  • Model compatibility with color-coded verdicts (CAN RUN / DEGRADED / CAN'T RUN)
  • Model browser with search and local/remote toggle
  • Download models directly with progress bar
  • Cross-platform single binary (Linux, macOS, Windows) -- no CGO

Install

Download the latest binary from Releases and add it to your PATH.

From source
go install github.com/putzeys/ollafit@latest

Or clone and build:

git clone https://github.com/putzeys/ollafit.git
cd ollafit
go install .

Usage

Interactive TUI
ollafit

Opens a full-screen interface with 4 tabs:

Tab Description
1 Hardware Detected CPU, GPU, RAM
2 Compatibility Color-coded table for 370+ models
3 Models Browse remote or local models
4 Pull Check compatibility and download models
Keyboard shortcuts
Key Action
Tab / 1-4 Switch tab
j/k or Up/Down Navigate list
/ Search (Compatibility & Models)
q Cycle quantization: Q4_K_M, Q8_0, FP16 (Compatibility)
l Toggle local/remote (Models)
Enter Confirm / send to Pull
Esc Cancel
Ctrl+C Quit
CLI commands
# Detect hardware
ollafit scan

# Check model compatibility
ollafit check
ollafit check --search llama
ollafit check --quant FP16
ollafit check --json

# Browse models
ollafit models
ollafit models --search llama
ollafit models --local

# Download a model (requires Ollama running)
ollafit pull llama3.2:1b

Compatibility verdicts

Status Meaning
CAN RUN (green) Model fits in GPU VRAM
DEGRADED (yellow) Needs CPU/RAM offload (slower)
CAN'T RUN (red) Insufficient total memory

Configuration

Create ~/.config/ollafit/config.yaml:

ollama_host: "http://localhost:11434"
model_source: "ollamadb"
vram_overhead_percent: 20.0
gpu_memory_fraction: 0.75  # Apple Silicon: % of unified memory for GPU

Environment variables are also supported with the OLLAFIT_ prefix (e.g., OLLAFIT_OLLAMA_HOST).

Requirements

  • Ollama running locally for models --local and pull commands
  • GPU detection uses: nvidia-smi (NVIDIA), rocm-smi (AMD), or Apple Silicon system APIs
  • Go 1.24+ only if building from source

License

MIT

Documentation

The Go Gopher

There is no documentation for this package.

Directories

Path Synopsis
internal
ui

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL