Small language model: Difference between revisions
Redirected page to Large language model via Articles for Creation (you can help!) (AFCRHS.js) Tag: New redirect |
Cloud29371 (talk | contribs) ←Removed redirect to Large language model Tag: Removed redirect |
||
Line 1: | Line 1: | ||
'''Small language model''' are [[artificial intelligence]] [[language model]] that designed for human [[natural language processing]] including [[Generative artificial intelligence|language and text generation]]. Unlike [[large language model]], small language model are much more smaller in scale and scope. |
|||
#REDIRECT [[Large language model]] |
|||
{{R from antonym}} |
|||
Normally, a large language model training parameter has more than hundreds of billions, some are even more than trillion parameter. The scale large language model is so large make the model has huge amount of information and able to generate better content, however, this requires huge amount of computation power, make individual impossible to train a large language model using a single or few computer and [[GPU]]. |
|||
Small language model, however, just use much fewer parameters, normally from a few million to a few billion, make it more feasible to train and host in a resource constrained environments such as single computer or even a mobile device. <ref>{{cite web|url=https://www.ibm.com/think/topics/small-language-models|title= What are small language models? |author=Rina Diane Caballar|publisher=IBM}}</ref><ref>{{cite web|url=https://huggingface.co/blog/jjokah/small-language-model|title=Small Language Models (SLM): A Comprehensive Overview|author=John JOhnson|publisher=Huggingface}}</ref><ref>{{cite web|url=https://www.weforum.org/stories/2025/01/ai-small-language-models/|title=What is a small language model and how can businesses leverage this AI tool?|author=Kate Whiting|publisher=The World Economic Forum}}</ref><ref>{{cite web|url=https://learn.microsoft.com/en-us/shows/data-exposed/slm-small-language-model-with-your-data-data-exposed|title=SLM (Small Language Model) with your Data|publisher=Microsoft}}</ref> |
|||
== References == |
|||
<references /> |
|||
== See also== |
|||
* [[Large Language Model]] |
|||
* [[Language Model]] |
|||
* [[Artificial intelligence]] |
|||
[[Category:Language modeling]] |
|||
[[Category:Statistical natural language processing]] |
Revision as of 10:45, 13 April 2025
Small language model are artificial intelligence language model that designed for human natural language processing including language and text generation. Unlike large language model, small language model are much more smaller in scale and scope.
Normally, a large language model training parameter has more than hundreds of billions, some are even more than trillion parameter. The scale large language model is so large make the model has huge amount of information and able to generate better content, however, this requires huge amount of computation power, make individual impossible to train a large language model using a single or few computer and GPU.
Small language model, however, just use much fewer parameters, normally from a few million to a few billion, make it more feasible to train and host in a resource constrained environments such as single computer or even a mobile device. [1][2][3][4]
References
- ^ Rina Diane Caballar. "What are small language models?". IBM.
- ^ John JOhnson. "Small Language Models (SLM): A Comprehensive Overview". Huggingface.
- ^ Kate Whiting. "What is a small language model and how can businesses leverage this AI tool?". The World Economic Forum.
- ^ "SLM (Small Language Model) with your Data". Microsoft.