How to run OpenAI’s open weight models locally on your computer

Run open-weight reasoning models on your own computer with Ollama and LM Studio while balancing hardware limits, privacy, model size, and hosted testing options.

Chapters
10
Duration
1h 2m
Difficulty
Beginner
Updated
Aug 2025

What you'll learn

Pick the Right Model

Choose between GPT-OSS, hosted OpenAI models, and other open models based on privacy, offline access, cost, customization, and model quality.

Evaluate Hardware Fit

Assess whether GPT-OSS 20B or 120B fits your machine by comparing memory, parameter size, architecture, benchmarks, and practical performance needs.

Run Models with Ollama

Install Ollama and run GPT-OSS locally from the terminal, with smaller model alternatives ready when your computer cannot handle larger models.

Use LM Studio Locally

Use LM Studio to download local models, chat through a visual interface, and query local documents with retrieval-augmented generation.

Protect Private Data

Decide when fine-tuning, retrieval, or local execution is appropriate for sensitive legal, healthcare, finance, research, coding, or internal business workflows.

Experiment Safely

Test hosted demos safely for quick experimentation while keeping sensitive prompts and files reserved for local workflows.

Course curriculum

1 part · 10 chapters