With OllaMan, Even Beginners Can Run LLMs
2026/01/15

With OllaMan, Even Beginners Can Run LLMs

A beginner-friendly guide to running AI models on your own computer. Get from zero to chatting with a local LLM in under 5 minutes using OllaMan's beautiful GUI.

You've probably heard of ChatGPT, Claude, and Gemini. They're powerful, but they all run in the cloud — meaning your conversations travel through someone else's servers.

What if you could run AI models entirely on your own machine? Local LLMs make this possible: complete privacy, no internet required, and zero API costs.

The catch? Setting up local models usually involves command lines, environment variables, and technical know-how that scares off most people.

That's where OllaMan comes in.

OllaMan Dashboard

What's Ollama? What's OllaMan?

Let's clear up these two terms first:

Ollama: The Engine

Ollama is the open-source project that actually runs AI models on your computer. It supports all the popular open-source models:

  • Llama 3 — Meta's flagship open model
  • Mistral — The lightweight European alternative
  • DeepSeek — Exceptional reasoning capabilities
  • Gemma — Google's efficient open model

Ollama is fantastic, but it only offers a command-line interface. Great for developers, intimidating for everyone else.

OllaMan: The Dashboard

Think of Ollama as the kitchen, and OllaMan as the restaurant's beautiful front-of-house.

OllaMan is a desktop app that wraps Ollama in a modern graphical interface. With it, you can:

  • 👀 Browse all your installed models at a glance
  • 🖱️ Download new models with a single click
  • 💬 Chat with models like you would with ChatGPT
  • 🎨 Enjoy polished dark and light themes

Ollama runs the models. OllaMan makes it delightful.

Getting Started: 5 Minutes to Your First Chat

Step 1: Install Ollama

First, get Ollama running on your machine:

  1. Visit ollama.ai
  2. Download the installer for your OS (macOS / Windows / Linux)
  3. Run the installer — it's a standard "Next, Next, Finish" setup

Once installed, Ollama runs silently in the background.

💡 Note: Don't expect a window to pop up — Ollama runs as a background service. That's normal.

Step 2: Install OllaMan

Next, grab OllaMan:

  1. Head to ollaman.com
  2. Download the app for your platform
  3. Install and launch OllaMan

OllaMan automatically detects your local Ollama service. If everything's working, you'll land on the dashboard.

Your First Conversation

Download a Model

A fresh Ollama installation has no models yet. Let's fix that:

  1. Click "Discover" in the left sidebar
  2. Browse the model library — you'll see dozens of options
  3. Pick something like Llama 3 or Mistral
  4. Click into the model details page
  5. Choose a size (we recommend 7B or 8B for beginners — lower hardware requirements)
  6. Hit the "Pull" button to start downloading

Model Library and Download

While downloading, you can:

  • Watch real-time progress on the Downloads page
  • See download speed and completion percentage
  • Queue up multiple models simultaneously

Download time: Depends on your internet speed and model size. A 4GB model takes roughly 5 minutes on a 100Mbps connection.

Start Chatting

Once downloaded, getting to your first conversation is straightforward:

  1. Click "Chat" in the sidebar
  2. Select your newly downloaded model from the top bar
  3. Type a message and hit Enter

That's it. You're now chatting with a local AI.

Chat Interface

Create an Agent for Repeated Tasks

After using OllaMan for a while, you might notice you're typing the same instructions repeatedly: "Act as a coding assistant" or "Always respond in a friendly tone."

Agents solve this. An Agent is a pre-configured AI persona with:

  • A system prompt (the AI's role)
  • A default model
  • Custom generation parameters

To create one:

  1. Go to the Chat page
  2. Click the current Agent card in the left sidebar
  3. Click the "+" button
  4. Set a name, icon, and system prompt
  5. Save

Creating an Agent

Here are some Agent ideas:

Agent NameUse CaseSystem Prompt Snippet
Code BuddyProgramming help"You're a patient coding mentor who explains concepts clearly..."
Writing CoachContent creation"You're a creative writing assistant who helps brainstorm and polish text..."
Study HelperLearning"You're a friendly tutor who breaks down complex topics into simple terms..."

Once created, switching Agents instantly changes your AI's personality and defaults.

Power User Tips

Tip 1: Attach Files Instead of Pasting

Need the AI to analyze code or a document? Skip the copy-paste.

Click the 📎 attachment button in the input area and select files directly. OllaMan supports:

  • Code files: .py, .js, .ts, .java, and more
  • Documents: .txt, .md, .json
  • Images (with vision models): .png, .jpg

Tip 2: Enable Thinking Mode

Some models (like DeepSeek R1 or QwQ) support "thinking mode" — they'll show their reasoning process before giving an answer.

If your model supports this, you'll see a "Think" toggle near the input. When enabled:

  • Responses split into "thinking" and "answer" sections
  • The thinking section is collapsible
  • Great for complex reasoning tasks

Tip 3: Tune Generation Parameters

The settings panel on the right side of each chat lets you adjust:

ParameterWhat It DoesRecommendations
TemperatureControls creativityCode/factual: 0.1-0.3
Creative writing: 0.8-1.2
Top PSampling rangeUsually keep at 0.9
Top KCandidate token countUsually keep at 40

Changes apply only to the current session — your Agent's defaults stay untouched.

Tip 4: Connect Multiple Servers

Got a beefy desktop at home and a thin laptop on the go? Great setup:

  1. Run Ollama on your powerful machine
  2. Connect to it remotely from OllaMan on any device

Just add the remote server address in Settings → Servers.

FAQ

Q: What specs do I need?

Quick reference:

Model SizeRecommended Setup
1B-3B8GB RAM — entry level
7B-8B16GB RAM — sweet spot
13B32GB RAM or 8GB VRAM
70B+Dedicated GPU required

💡 If unsure, start with a 7B model. It's the best balance of performance and quality.

Q: Where are models stored?

Ollama keeps models in:

  • macOS: ~/.ollama/models
  • Windows: C:\Users\<username>\.ollama\models
  • Linux: ~/.ollama/models

Q: Does it work offline?

Absolutely — that's the whole point!

  • Needs internet: Downloading models, browsing the model library
  • Works offline: Chatting with downloaded models

Once a model is on your machine, conversations happen entirely locally.

Wrapping Up

Running AI locally isn't just for power users anymore.

With Ollama + OllaMan:

  • No coding skills required
  • Your data never leaves your machine
  • Works without an internet connection

If you've been curious about local LLMs but intimidated by the terminal, now's the time.

5 minutes to install. An AI assistant that's truly yours.


📥 Download OllaMan: ollaman.com

📖 Documentation: ollaman.com/docs

Newsletter

Join the community

Subscribe to our newsletter for the latest news and updates