Aryan Pathak
← Back to writing

Small Language Models: Smaller Than You Think, Smarter Than You Expect

Why small, well-trained language models deserve more credit — and when they outperform their larger cousins.

Honestly, I went into this week a little skeptical. I had been defaulting to larger models for most of my tasks and had not seriously revisited the smaller end of the model spectrum in a while. But after reading about some of the newer compact models and seeing benchmark claims I found hard to believe, I decided to run my own comparisons.

I set up tests across tasks I actually care about — structured data extraction, document classification, and short-form summarization — and evaluated a 3B and a 7B parameter model against a much larger baseline after targeted fine-tuning. The results genuinely surprised me. On most tasks the fine-tuned small models were competitive, and on a couple they actually edged ahead. The latency difference was substantial, and the memory footprint was small enough to run comfortably on a single consumer GPU.

My takeaway is that the assumption that bigger is always better is getting harder to defend for the majority of real-world use cases. There is an obvious place for frontier 100B+ models when you genuinely need that capability ceiling, but for the bulk of production tasks, a well-fine-tuned small model is often faster, cheaper, and surprisingly competitive. Model size selection should be a deliberate, task-specific decision — not a reflexive default to whatever is largest.

Small Language Models: Smaller Than You Think, Smarter Than You Expect illustration 1Small Language Models: Smaller Than You Think, Smarter Than You Expect illustration 2Small Language Models: Smaller Than You Think, Smarter Than You Expect illustration 3Small Language Models: Smaller Than You Think, Smarter Than You Expect illustration 4Small Language Models: Smaller Than You Think, Smarter Than You Expect illustration 5