Small
Language Models (SLM) are the mini version of Large language Models. SLMs have
less number of parameters compare to LLMs. These models are designed to perform
tasks such as sentiment analysis, text generation, etc. similar to LLM but with
less number of parameters. The less parameters helps to improve to
computational efficiency, accessibility and adaptability.
Some
examples of SLM are
- Llama 2 7B - Released by
Meta in July'23, it has 7 billion parameter
- Phi2 - 2.7 billion
parameter model, developed by Microsoft
- Stable Beluga 7B -
Developed by Stability AI , it is auto regressive language model fine
tuned on llam2
- Xgen - Developed by
Salesforce. It is a smaller scale model that is customized for particular
domains.
- Alibaba's Qwen - Developed
by Alibaba cloud
- Alpaca 7B - Fine tuned on
Meta Llama 7B model on 52 K instruction-following demonstration.
- Falcon 7B – This model is 7B parameters causal
decoder-only model developed by TII and trained on 1,500B token.
- MPT - Developed by MosaicML
Foundation
- Zephyr - Zephyr 7B is a
model created by the HuggingFace H4 (Helpful, Honest, Harmless, Huggy)
team with an objective to create a smaller language model that is aligned
with user intent and outperforms even bigger models.
- MobileBERT
- GPT-Neo and GPT-J
- T5-Small - Text-To-Text
Transfer Transformer (T5) is a pre-trained encoder-decoder model handling
all NLP tasks as a unified text-to-text-format where the input and output
are always text strings. T5-Small is the checkpoint with 60 million
parameters.