The Day My AI Intern Burned a Hole in My Pocket
I once had a project that required summarizing about 10,000 customer reviews. A perfect job for an AI, right? I whipped up a quick script, pointed it at OpenAI’s latest and greatest model, and let it run overnight. I woke up feeling like a genius. Efficient. Automated. The future.
Then I saw the bill. A cool $450. For one job. My “genius” script had been about as efficient as hiring a team of calligraphers to write out the summaries on artisanal parchment.
That was the moment I got serious about bringing my AI workforce in-house. I needed the power, but without the meter running on every single thought. I needed an AI intern that lived in my computer, used my electricity, and kept my data to itself. Today, I’m going to show you how to hire your own, for free.
Why This Matters
In the world of AI, you can either rent or you can own. Using services like ChatGPT or the OpenAI API is renting. It’s easy, it’s powerful, but you pay for every use, and you’re sending your data—and potentially your customers’ data—to a third party. For many businesses, that’s a deal-breaker.
Running a Large Language Model (LLM) locally, on your own machine, is owning the factory. It means:
- Zero API Costs: After the initial hardware, running the model is free. You can ask it 10 questions or 10 million. The cost is the same.
- Total Data Privacy: Nothing ever leaves your computer. You can analyze sensitive financial reports, customer emails, or secret company roadmaps without a single byte leaking to the outside world.
- Infinite Customization: You can choose from dozens of powerful open-source models, each with different strengths. You control the entire environment.
This workflow replaces expensive API calls for repeatable tasks and eliminates the security risk of processing private data in the cloud. It’s the first step to building a truly sovereign automation system.
What This Tool / Workflow Actually Is
We’re going to set up a private, ChatGPT-style AI system using two key tools. Think of it like building a car. You need an engine and you need a dashboard.
The Engine: Ollama
Ollama is a beautifully simple tool that lets you download, manage, and run powerful open-source LLMs on your own computer. It handles all the complicated setup behind the scenes. You just type one command, like ollama run llama3, and it downloads the model and makes it ready for work. It’s the engine of our system, providing all the horsepower.
The Dashboard: Open WebUI
Open WebUI is a clean, fast, and feature-rich user interface that talks to Ollama. It looks and feels just like ChatGPT, but it runs entirely on your machine. It’s where you’ll chat with your local models, create prompt templates, and manage your conversations. It’s the steering wheel, pedals, and comfy leather seats for your AI engine.
Together, they give you a powerful, private, and free-to-run alternative to cloud-based AI services.
Prerequisites
Let’s be brutally honest. This isn’t magic. You need a reasonably decent computer. You’re building an AI factory, not a lemonade stand.
- A Modern Computer:
- Mac: An Apple Silicon Mac (M1, M2, M3) with at least 16GB of RAM is fantastic.
- PC (Windows/Linux): A modern CPU will work, but it’ll be slow. For real speed, you want a dedicated NVIDIA graphics card (a GPU) with at least 8GB of VRAM. The more VRAM, the bigger the models you can run.
- Docker Desktop: This sounds scary, but it’s not. Docker is a free application that lets you run software in isolated little boxes called containers. It makes installing complex things like Open WebUI a one-command affair. We won’t be coding with it, just using it to run one command. Go install it now.
- A Tiny Bit of Command Line Courage: You will need to open your computer’s terminal (Terminal on Mac/Linux, PowerShell or CMD on Windows) and copy-paste about three commands. That’s it. If you can type, you can do this.
Don’t panic. This is easier than assembling IKEA furniture.
Step-by-Step Tutorial
Follow these steps exactly. Do not skip ahead. The whole process should take less than 30 minutes, most of which is just waiting for things to download.
Step 1: Install Ollama (The Engine)
Ollama is the foundation. It manages and runs the AI models.
- On Mac or Linux: Open your terminal and paste this single command, then press Enter.
curl -fsSL https://ollama.com/install.sh | sh
- On Windows: Go to the Ollama website and download the Windows installer. Run it like any other application.
Ollama will now be running quietly in the background on your machine.
Step 2: Download Your First AI Model
Now we need to give our engine some fuel. We’ll download Llama 3, a fantastic model from Meta that’s small enough to run well on most modern machines. In your terminal, run this:
ollama run llama3:8b
Why this step exists: This command tells Ollama to find the Llama 3 model with 8 billion parameters (a good starting size), download it from the internet, and then immediately start a chat session with it in your terminal. You’ll see it download a file that’s a few gigabytes. Once it’s done, you can type messages right there. Type `/bye` to exit. You’ve just run your first local AI!
Step 3: Install Open WebUI (The Dashboard)
Make sure you have Docker Desktop installed and running. Check your system tray for the little whale icon. Now, open your terminal and paste this one, slightly longer, command.
docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main
Why this step exists: This command tells Docker to do a few things: download the Open WebUI application, run it in the background (`-d`), connect your computer’s port 3000 to the application’s port 8080 (`-p 3000:8080`), and give it a persistent storage volume (`-v open-webui…`) so your chats don’t disappear. It’s a fire-and-forget command.
Step 4: Access Your Private ChatGPT Clone
Open your web browser (Chrome, Firefox, Safari) and go to this address:
http://localhost:3000
You should see the Open WebUI sign-up screen. Click “Sign up” and create your first local user account. This doesn’t send anything to the internet; it’s just for your local database.
Congratulations. You’ve installed the dashboard. But it’s not connected to the engine yet. You’ll notice it can’t find any models.
Step 5: Connect WebUI to Ollama
Once you’re logged in, click the settings gear icon, go to the “Connections” tab, and set your Ollama API URL. By default, it looks for Ollama inside the Docker container, but our Ollama is running on our main computer. You need to tell it where to look.
The correct address to connect to Ollama from inside the Docker container is:
http://host.docker.internal:11434
Enter that into the API URL field and hit save. Now go back to the main chat screen. In the top left, click “Select a model”. You should see `llama3:8b` in the list! Select it and start chatting. It’s all running on your machine.
Complete Automation Example
Let’s use our new private AI to do a real business task: classifying customer feedback.
The Goal: Take unstructured customer feedback and categorize it into one of three buckets: ‘Bug Report’, ‘Feature Request’, or ‘General Praise’. This is a boring, manual job perfect for our new AI intern.
Step 1: Create a Reusable Prompt Template in Open WebUI
In Open WebUI, on the right side panel, you’ll see a section for Prompts. Click the ‘+’ to create a new one.
- Name it: `Feedback Classifier`
- Enter the prompt:
You are a helpful assistant for a software company. Your only job is to classify the following customer feedback into one of these three categories: Bug Report, Feature Request, or General Praise. Respond with ONLY the category name and nothing else.
Customer Feedback:
"""
{{feedback}}
"""
The `{{feedback}}` is a variable. Open WebUI will give us a box to paste text into for that variable. Save the prompt.
Step 2: Run the “Automation”
Now, on the main chat screen, you can apply your prompt. Type `#` in the chat box and select your `Feedback Classifier` prompt from the list. A box will appear for the `feedback` variable.
Let’s test it. Paste this into the box:
The export button in the dashboard is crashing the app every time I click it. It's really frustrating as I can't get my data out.
Hit enter. The AI should respond instantly with:
Bug Report
Now try another one. Paste this into the variable box:
I just wanted to say that your new UI update is fantastic. Everything is so much faster and easier to find. Keep up the great work!
The AI responds:
General Praise
You now have a repeatable, instant, and completely private system for triaging customer feedback. No data was sent to OpenAI. No API fees were incurred.
Real Business Use Cases
This exact setup can be used across hundreds of business functions.
- E-commerce Store: The owner can paste in a list of product reviews and use a prompt to have the local LLM extract common themes, pros, and cons to be used in marketing copy, all without sharing customer data.
- Marketing Agency: A copywriter can paste a client’s dry, technical brief into a “Rewrite for Humans” prompt template to instantly generate more engaging, benefit-focused marketing angles.
- Law Firm: A paralegal can paste paragraphs from a long contract into a prompt that extracts all dates, names, and monetary values. This is for initial review only, of course, but it saves hours of manual scanning and the sensitive document never leaves the firm’s computer.
- Startup Founder: A founder can paste a transcript from a user interview and use a prompt to summarize the user’s key pain points, goals, and direct quotes, helping them build their product roadmap faster.
- HR Department: An HR manager can take an internal job description and use a prompt to rewrite it for an external job board, making it more appealing to candidates while ensuring all key responsibilities are included.
Common Mistakes & Gotchas
- Using the Wrong Model for the Job: Don’t use a massive 70-billion-parameter model to classify emails. It’s like using a sledgehammer to crack a nut. Start with smaller models like Llama 3 8B or Phi-3 Mini. They are faster and more than capable for 90% of business tasks.
- Forgetting Docker is Running: Docker Desktop and the Open WebUI container use your computer’s resources. If you’re not using it, shut Docker Desktop down from your system tray to free up memory and CPU.
- The `localhost` vs `host.docker.internal` Trap: When one Docker container needs to talk to a service on your main computer (the “host”), it can’t just use `localhost`. It needs the special address `host.docker.internal`. This confuses almost everyone at first.
- Expecting GPT-4 Level Genius: Open-source models are incredible, but they may not match the absolute bleeding-edge reasoning of the most expensive proprietary models on every single task. For most business automation (summarizing, reformatting, classifying), they are more than good enough and the benefits of privacy and cost are immense.
How This Fits Into a Bigger Automation System
What we’ve built is a fantastic, interactive tool. But the real power is unlocked when you treat Ollama as a component in a larger machine.
Ollama exposes an API. That means other programs can send it requests automatically. The same request that Open WebUI sends when you chat can be sent by an automation platform like n8n, Make, or a custom Python script.
This is where the magic happens. You can build workflows that:
- Automatically get new emails from a Gmail account.
- Send the email content to your local Ollama model for classification.
- Based on the category (‘Bug Report’, ‘Sales Inquiry’), create a ticket in your CRM (like HubSpot or Salesforce) and assign it to the right team.
Your local LLM becomes the private, secure ‘brain’ in a fully automated business process. We’ve built the factory; next, we’ll build the assembly line.
What to Learn Next
You’ve done it. You have your own private AI powerhouse running on your machine. You’ve gone from a consumer of AI to an owner of an AI factory. Pat yourself on the back.
But right now, it’s a factory that requires a human operator to paste things into a web UI. That’s a great start, but it’s not true automation.
In the next lesson in this course, we’re going to connect this local AI brain to the outside world. I’ll show you, step-by-step, how to use a free tool called n8n to build a workflow that automatically reads your unread emails, sends them to your new local Llama 3 model for analysis, and then tags them in your inbox—no copying and pasting required. We’re going to build our first fully autonomous AI robot worker.
Stay tuned. This is where it gets really fun.
“,
“seo_tags”: “Ollama, Open WebUI, Local LLM, AI Automation, Self-hosting AI, Private AI, Llama 3, Free AI”,
“suggested_category”: “AI Automation Courses

