Jump to content

Small language model

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Cloud29371 (talk | contribs) at 10:45, 13 April 2025 (Removed redirect to Large language model). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Small language model are artificial intelligence language model that designed for human natural language processing including language and text generation. Unlike large language model, small language model are much more smaller in scale and scope.

Normally, a large language model training parameter has more than hundreds of billions, some are even more than trillion parameter. The scale large language model is so large make the model has huge amount of information and able to generate better content, however, this requires huge amount of computation power, make individual impossible to train a large language model using a single or few computer and GPU.

Small language model, however, just use much fewer parameters, normally from a few million to a few billion, make it more feasible to train and host in a resource constrained environments such as single computer or even a mobile device. [1][2][3][4]

References

  1. ^ Rina Diane Caballar. "What are small language models?". IBM.
  2. ^ John JOhnson. "Small Language Models (SLM): A Comprehensive Overview". Huggingface.
  3. ^ Kate Whiting. "What is a small language model and how can businesses leverage this AI tool?". The World Economic Forum.
  4. ^ "SLM (Small Language Model) with your Data". Microsoft.

See also