Use Case: Local AI

Local AI Agent Without Docker.
Desktop App That Just Works.

Run Llama 3, Mistral, DeepSeek, or any Ollama model as a full desktop AI agent. No Docker. No terminal. No configuration. Double-click and you are running a local LLM in minutes.

The Docker problem

Every week on r/LocalLLaMA someone posts: “Tried to set up [local AI tool], gave up after 4 hours.” Docker Compose errors. Python version conflicts. WSL2 configuration. Port binding failures. CUDA driver mismatches. Most local AI projects are built by developers for developers.

The barrier is not the model - Ollama made downloading and running local LLMs trivially easy. The barrier is everything that comes after: building a useful interface, connecting it to your files and emails, making it remember context, and running it as a persistent agent rather than a chat session. That is what Skales solves.

What Skales gives you on top of Ollama

Ollama runs the model. Skales wraps it in a real agent with memory, tools, and a usable interface.

Double-click install

Download the installer. Double-click. That is it. No Homebrew, no pip, no npm, no Docker Compose, no WSL2, no Python environment, no port forwarding. Skales installs like any other desktop app.

Ollama integration built in

Install Ollama separately (also a double-click install), pull any model you want - Llama 3, Mistral, Phi-3, DeepSeek, Gemma, Qwen - and connect it to Skales in the settings. Takes under 5 minutes total.

Runs on your hardware

All inference happens on your CPU or GPU. No cloud API calls, no latency from a remote server, no API cost. Bring your own hardware - M-series Mac, modern Windows PC with a decent GPU, or even CPU-only.

Fully offline capable

Once models are downloaded, Skales with Ollama works with zero internet connection. No phone-home, no telemetry unless you opt in, no mandatory update checks. Works in airplane mode, on a local network, or air-gapped.

Your data never leaves your machine

No prompt is sent to OpenAI, Anthropic, or any cloud provider when using Ollama. Your conversations, documents, and context stay in ~/.skales-data on your own disk. Read the source code on GitHub to verify.

Cloud API as a fallback

Want the speed of GPT-4 or Claude for heavy tasks? Add an API key and Skales routes to it with a single setting. Use Ollama for everyday tasks, cloud for the occasional complex job. You choose per conversation.

From zero to local AI agent in 5 minutes

1

Install Ollama

Download from ollama.com. Double-click. Done. Ollama runs as a local server on port 11434. No configuration needed.

2

Pull a model

Open a terminal (just once): ollama pull llama3.2 or ollama pull mistral or whatever you want. Ollama handles the download.

3

Install Skales

Download the Skales installer from this page. Double-click. Follow the 30-second setup wizard.

4

Connect Ollama in settings

Open Skales Settings → Model → select Ollama → choose your model. Skales detects Ollama automatically if it is running.

5

Start using your local AI agent

Ask it anything. It has memory, tools, voice input, and persistent context. All running on your hardware. Zero cloud calls.

Works with any Ollama model

Llama 3.2
Llama 3.1
Mistral
Mistral Nemo
DeepSeek R1
Phi-4
Gemma 3
Qwen 2.5
Command R
Orca Mini
Vicuna
Any GGUF

If Ollama supports it, Skales supports it. New models appear automatically.

“I've tried 12 local AI setups. This is the first one I actually use every day.”

Free for personal use. Windows and macOS. No account required.