A neural network to understand and generate language using far fewer parameters and computing resources than large language models.
A Small Language Model (SLM) is a language model with fewer parameters than Large Language Models (LLMs, typically 10B+) but more parameters than Tiny Language Models (TLMs, often under 1B). SLMs generally range from hundreds of millions to a few billion parameters (e.g., 1B–4B), enabling strong performance on diverse language tasks with far less compute than massive LLMs.
Rather than covering every possible topic, SLMs are often optimized for specific tasks such as classification, summarization, translation, or domain-specific assistance. This specialization allows them to deliver strong performance with lower latency and reduced computing costs.
Phi-3 Mini (3.8B) — High-quality synthetic data-trained model from Microsoft, optimized for logical reasoning and coding in constrained environments.
Gemma 2B — Lightweight Google release focused on safety alignments and fast inference for developer tools and mobile deployment.
Mistral 7B (7B) — French AI lab’s breakthrough in sliding window attention, powering multilingual translation and creative writing tasks.
For instance, the MedMobile clinical assistant uses the Phi-3 Mini model on smartphones to help doctors review medical cases and retrieve guidance without relying on cloud infrastructure.
In localization, SLMs can help identify text that needs translation, classify strings such as UI labels or error messages, generate draft translations for review, summarize source content for translators, and flag possible terminology inconsistencies.
Compared with Tiny Language Models (TLMs), SLMs are typically used for heavier language tasks such as translation, summarization, or code assistance that run on servers or developer machines. In contrast, TLMs are usually deployed directly inside applications or devices where memory and computing power are very limited.
Learn more about how AI models can be used efficiently in translation and localization in our blog.