How to Use Local LLM with Ollama: A Complete Beginner's Guide (2025)
Learn how to run powerful AI models on your own computer with Ollama - no technical background required!

Running artificial intelligence models on your own computer might sound complicated, but with Ollama, it's surprisingly simple. This comprehensive guide will walk you through everything you need to know about setting up and using local Large Language Models (LLMs) on Windows, Mac, and Linux.
What is Ollama and Why Use Local LLMs?
Ollama is a free, open-source tool that makes it incredibly easy to run AI language models directly on your computer. Think of it as your personal AI assistant that works completely offline, without sending your data to external servers.
Benefits of using local LLMs:
- Complete privacy: Your conversations stay on your device
- No internet required: Work offline once models are downloaded
- No usage limits: Use AI as much as you want without restrictions
- Cost-effective: No monthly subscription fees
- Customizable: Choose from dozens of different AI models
System Requirements
Before we begin, ensure your computer meets these minimum requirements:
- RAM: At least 8GB (16GB recommended for better performance)
- Storage: 10-50GB free space (depending on model size)
- Operating System: Windows 10/11, macOS 10.15+, or Linux
- Internet: Required for initial download only
Installing Ollama on Different Operating Systems
Windows Installation
- Download Ollama for Windows
- Visit the official Ollama website at ollama.ai
- Click the "Download for Windows" button
- Save the installer file to your computer
- Run the Installation
- Double-click the downloaded installer file
- Follow the installation wizard prompts
- Choose your installation directory (default is recommended)
- Click "Install" and wait for completion
- Verify Installation
- Open Command Prompt (press Windows + R, type "cmd", press Enter)
- Type the following command and press Enter:
ollama --version
- You should see the Ollama version number
macOS Installation
- Download Ollama for Mac
- Go to ollama.ai in your web browser
- Click "Download for Mac"
- The .dmg file will download automatically
- Install the Application
- Open the downloaded .dmg file
- Drag the Ollama icon to your Applications folder
- Open Applications and double-click Ollama to launch it
- Verify Installation
- Open Terminal (press Cmd + Space, type "Terminal", press Enter)
- Type the following command and press Enter:
ollama --version
- The version number should appear
Linux Installation
- Install via Terminal
- Open your terminal application
- Run the following command:
curl -fsSL https://ollama.ai/install.sh | sh
- Alternative Manual Installation
- Download the Linux binary from ollama.ai
- Extract and move to your preferred location
- Add to your system PATH
- Verify Installation
- In terminal, type:
ollama --version
- Confirm the version displays correctly
- In terminal, type:
Getting Started with Qwen Model
Qwen is an excellent choice for beginners because it's relatively lightweight, highly capable, and supports multiple languages. Here's how to set it up:
Step 1: Download the Qwen Model
Open your terminal or command prompt and run:
ollama pull qwen:7b
This downloads the 7-billion parameter version of Qwen, which offers a great balance between performance and resource usage. The download may take 10-30 minutes depending on your internet speed.
Step 2: Start Your First Conversation
Once the download completes, start chatting with Qwen:
ollama run qwen:7b
You'll see a prompt where you can type your questions. Try asking something like:
- "Hello, can you introduce yourself?"
- "What can you help me with?"
- "Explain quantum physics in simple terms"
Step 3: Exit the Chat
To stop the conversation, simply type the following or press Ctrl+C (Cmd+C on Mac):
/bye
Essential Ollama Commands for Beginners
Here are the most important commands you'll use regularly:
List available models:
ollama list
Download a new model:
ollama pull [model-name]
Start a conversation:
ollama run [model-name]
Remove a model:
ollama rm [model-name]
Show model information:
ollama show [model-name]
Popular Models to Try After Qwen
Once you're comfortable with Qwen, explore these other excellent models:
- Llama 2 - Meta's popular open-source model
ollama pull llama2:7b
- Code Llama - Specialized for programming tasks
ollama pull codellama:7b
- Mistral - Fast and efficient European model
ollama pull mistral:7b
- Neural Chat - Great for conversational AI
ollama pull neural-chat:7b
Troubleshooting Common Issues
Model Won't Download
- Check your internet connection
- Ensure you have sufficient storage space
- Try downloading a smaller model first
Slow Performance
- Close other resource-intensive applications
- Consider upgrading your RAM
- Try a smaller model (3B or 1B parameters)
Command Not Found Error
- Restart your terminal/command prompt
- Check if Ollama is properly installed
- Verify your system PATH includes Ollama
Advanced Tips for Better Performance
Optimize Memory Usage
Most models work better with specific memory allocations. You can set this using:
OLLAMA_NUM_GPU=1 ollama run qwen:7b
Run Multiple Models
You can have multiple models installed and switch between them easily:
ollama run qwen:7b
/bye
ollama run llama2:7b
Create Custom Model Configurations
Advanced users can create custom model files with specific parameters and prompts tailored to their needs.
Use Cases for Local LLMs
Local LLMs are perfect for:
- Writing assistance: Draft emails, articles, or creative content
- Learning and research: Ask questions about any topic
- Code review: Get help with programming problems
- Language translation: Translate text between languages
- Data analysis: Analyze and interpret information
- Creative projects: Generate ideas, stories, or brainstorm solutions
Privacy and Security Benefits
Running AI models locally provides significant advantages:
- Your data never leaves your computer
- No account creation or login required
- Complete control over your AI interactions
- No usage tracking or data collection
- Works without internet connectivity
Prefer a Cloud-Based Solution?
If your computer doesn't meet the system requirements or you prefer the convenience of a managed cloud service, there are excellent free options available.
Explore our guide on How to Access Google Gemini for Free with Google AI Studio.