AI Fine-Tuning for Non-Engineers

Compact AI models make fine-tuning possible for non-engineers. I tried it myself and learned how AI becomes portable, local and real.

AI Fine-Tuning for Non-Engineers

How I experimented with small models and learned a lot along the way.

I don’t have a background in engineering. My work is with content, language and communication. Still, I wanted to understand what “AI fine-tuning” means in practice. Not by reading theory, but by doing it myself.

It took some fiddling with details, but thanks to the new generation of smaller models, it was possible and very rewarding.

What is fine-tuning, really?

The term sounds technical, but the idea is familiar. Think of an old radio where you slowly turn the dial until the static clears and the station comes in. Or imagine a sound mixer with many knobs: you don’t rebuild the whole system, you adjust the balance so the music fits your taste.

Fine-tuning an AI model works in a similar way. You start with a base model that already “knows” a lot, and then you nudge it towards what matters for your context.

More emphasis here, less emphasis there, until it aligns better with your purpose.

Fine-tuning on a historic Motorola radio

Why smaller models matter

Large models are impressive, but they’re hard to run yourself. They need time, compute and technical expertise. Recently, though, compact models have arrived that you can work with on an ordinary computer.

I chose Gemma, a 270-million parameter model from Google. Small in AI terms, but powerful enough for meaningful experiments.

Introducing Gemma 3 270M: The compact model for hyper-efficient AI- Google Developers Blog
Explore Gemma 3 270M, a compact, energy-efficient AI model for task-specific fine-tuning, offering strong instruction-following and production-ready quantization.

My fine-tuning journey

With a method called LoRA (low-rank adaptation), I fine-tuned the model on my own dataset. Think of it as adding a lightweight custom layer without retraining the whole model from scratch.

At the end, you get a portable file that you can run locally, on your laptop or even on a mobile phone. That was already exciting: I wasn’t just reading about AI, I had my own adapted model in hand.

Making it smaller: quantisation

The next step was quantisation. Like compressing an image, you reduce precision to make the model lighter. Instead of storing numbers in 32 or 16 bits, you shrink them down to 4.

The result is a model that runs on much smaller devices. You lose a bit of accuracy, but you gain portability. For me, that was the real eye-opener: suddenly, AI wasn’t abstract anymore. It was something I could carry.

Running a Local LLM on Your iPhone
I explored how far mobile AI has come by running LLMs directly on my iPhone. No cloud, no upload. Here’s what I learned from testing Haplo AI.

What I learned

The biggest lesson wasn’t any single technique, but the way the steps connect. Starting with the base model, applying LoRA, quantising, and then running it locally: each piece only really makes sense once you try it.

Beyond the process, there’s a broader insight: AI is becoming more personal and portable. Models that once required massive infrastructure can now run privately, locally, and on the edge. That changes how we can all experiment with them, even if we’re not engineers.

Try it yourself

If you want to see the process in action, here is the original thread with the step-by-step instructions. It walks you through the setup on Google Colab so you can try fine-tuning a model yourself.

Check the guide by Paul Couvert here


Llama 3.2 on a Mac
I tested Meta’s Lama 3.2 LLM on my Mac Mini, setting it up via Docker. It’s fast, private, and generates code, but lacks memory and multimodal features like ChatGPT.
Chrome, Gemini Nano, and the Browser as AI Platform
The AI race is moving into the browser. Experiments with local models helped me recognise Chrome’s Gemini Nano for what it is.