Zarrar Sheikh
February 9, 2026

Why I Run Local AI: Privacy in the Age of Cloud

Here is a scenario: A client hands me a CSV file containing sensitive customer financial data. They need a trend analysis. Do I upload that to ChatGPT? Absolutely not. That is the fastest way to get sued (and fired). This is where my background in hardware and "old school" tech comes in. I don't just […]

Here is a scenario: A client hands me a CSV file containing sensitive customer financial data. They need a trend analysis.

Do I upload that to ChatGPT? Absolutely not.
That is the fastest way to get sued (and fired).

This is where my background in hardware and "old school" tech comes in. I don't just rely on the cloud. I run Local LLMs (Large Language Models) right on my own machine.

The Stack: Llama 4 & Mistral

Using tools like *Ollama* or *LM Studio*, I run models like Meta's Llama 4 or Mistral's latest open-weight model locally.
* **No Internet Required:** The data never leaves my laptop.
* **Zero Leakage Risk:** OpenAI doesn't see it. Google doesn't see it.
* **Cost:** Free (after you buy the hardware).

Why This Matters for a BA

Trust is our currency. When I tell a stakeholder, "Your data is safe," I mean it literally. I can show them the air-gapped environment where the analysis is happening.

It’s also about speed. I can tweak the model's parameters (temperature, system prompts) to behave exactly how I need it to, without hitting API rate limits.

If your BA doesn't know what "running locally" means, be careful what data you give them.

Copyright © 2026 Zarrar Sheikh
linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram