Introduction: Defining Small in a World That Celebrates Big
In the AI industry, small is a relative term. Understanding what actually makes a language model small, and what that means for its capabilities and appropriate use cases, is the foundation for everything else in this series.

The Parameter Question
SLMs typically range from a few hundred million to around seven billion parameters. Large Language Models, by contrast, operate from tens of billions to hundreds of billions of parameters.
But parameter count alone is a blunt instrument. What matters more is the architecture, the training data quality, the fine-tuning approach, and the specific tasks the model is designed to handle. A well-designed three-billion parameter model fine-tuned on domain-specific data can outperform a generic hundred-billion parameter model on the tasks it was built for.
Capabilities: What SLMs Do Well
SLMs handle a narrower range of tasks than their larger counterparts, but they handle their specialised tasks with impressive reliability. The capabilities where SLMs consistently perform well include document summarisation, text classification, code completion, sentiment analysis, and on-device conversational assistance.
Key Differences from LLMs
Efficiency is the most significant difference: SLMs run with lower latency and can operate on edge devices, embedded systems, and mobile hardware. Cost is substantially lower for both training and inference. The energy footprint is dramatically smaller. Privacy is perhaps the most strategically important difference, as SLMs can run locally without requiring cloud connectivity.
The Trade School Analogy
The analogy that captures the SLM-LLM distinction most intuitively is the difference between a general-purpose university and a specialised trade school. Both have genuine value. Many businesses do not need the university. They need the trade school.
Conclusion
Choosing an SLM over an LLM for the right use case is not settling for less. It is making a deliberate strategic choice for speed, cost, privacy, and fit-for-purpose performance.