Home/Docs/Ollama & LLaVA

Ollama & LLaVA Setup

Overview

Tag-AI's local image processing relies on two key components: Ollama, a lightweight local AI server, and LLaVA, a multimodal vision-language model.

Ollama and LLaVA Architecture

Together, these components enable completely local, private image analysis with no data sent to external services.

Automatic Setup

The Tag-AI setup wizard automatically handles Ollama and LLaVA installation:

Ollama Installation

During setup:

  1. Tag-AI detects if Ollama is already installed
  2. If not found, it downloads the appropriate installer for your OS
  3. The installer runs with appropriate permissions
  4. Tag-AI verifies Ollama is properly installed

LLaVA Model Download

After Ollama is installed:

  1. Tag-AI initiates the LLaVA model download
  2. Progress is displayed in the setup wizard
  3. The model is downloaded and prepared by Ollama (~4GB)
  4. Tag-AI verifies the model is properly installed

The automatic setup process requires internet connectivity for the initial download, but once completed, Tag-AI can process images entirely offline.

Manual Setup

If you need to manually set up Ollama and LLaVA:

Manual Ollama Installation

Windows

  1. Download the Ollama installer from ollama.ai
  2. Run the installer with administrator privileges
  3. Follow the installation prompts
  4. Verify Ollama is running in the system tray

macOS

  1. Download Ollama for Mac from ollama.ai
  2. Open the .dmg file
  3. Drag Ollama to the Applications folder
  4. Launch Ollama from Applications
  5. If prompted about an unidentified developer, open System Preferences → Security & Privacy and click "Open Anyway"

Linux

  1. Open a terminal
  2. Run: curl -fsSL https://ollama.ai/install.sh | sh
  3. Follow any additional prompts
  4. Verify installation with ollama --version

Manual LLaVA Model Installation

After installing Ollama:

  1. Open a terminal or command prompt
  2. Run: ollama pull llava
  3. Wait for the download to complete (~4GB)
  4. Verify with: ollama list (you should see llava in the list)

LLaVA Model

About LLaVA

LLaVA (Large Language and Vision Assistant) is a multimodal AI model that combines text and image understanding capabilities:

Model Location

The LLaVA model is stored at:

Model Versions

Tag-AI uses the standard LLaVA model by default, but Ollama offers several variants:

Alternative Models

Advanced users can use alternative vision-language models with Tag-AI:

Compatible Alternative Models

Installing Alternative Models

  1. Open a terminal or command prompt
  2. Run: ollama pull model_name (e.g., ollama pull bakllava)
  3. Wait for the download to complete

Configuring Tag-AI for Alternative Models

  1. Open the Configuration Editor (Actions → Edit Config)
  2. Locate the [tagger_ollama] section
  3. Change model_name to your desired model (e.g., bakllava)
  4. Save the configuration

Some alternative models may produce different tag styles or quantities. Test with a small batch of images before processing your entire library.

Advanced Configuration

Ollama API Endpoint

By default, Tag-AI connects to Ollama at http://localhost:11434/api/generate. If you've customized your Ollama setup:

  1. Open the Configuration Editor
  2. Locate the [tagger_ollama] section
  3. Modify the ollama_endpoint value
  4. Save the configuration

Custom Ollama Configuration

For advanced Ollama setup:

Model Parameters

Advanced users can modify the LLaVA prompting in the Tag-AI source code:

Troubleshooting

Ollama Not Running

If Tag-AI can't connect to Ollama:

  1. Check if Ollama is running in the system tray/notification area
  2. If not running, launch it manually:
    • Windows: Start Menu → Ollama
    • macOS: Applications → Ollama
    • Linux: Run ollama serve in a terminal
  3. Verify Ollama is listening on port 11434: curl http://localhost:11434/api/tags

LLaVA Model Issues

If the LLaVA model isn't working:

  1. Check if the model is installed: ollama list
  2. If missing, reinstall: ollama pull llava
  3. If the model seems corrupted, try: ollama rm llava followed by ollama pull llava

Memory Issues

If you encounter "out of memory" errors:

Slow Performance

If processing is unusually slow:

Logs and Diagnostics

For detailed troubleshooting: