large language model family
C16400
concept
A large language model family is a group of related neural network models that share a common architecture and training paradigm but vary in size, capabilities, and specialization to handle diverse natural language understanding and generation tasks.
Observed surface forms (9)
- language model ×8
- pretrained language model ×3
- Generative Pre-trained Transformer ×1
- XLM-RoBERTa architecture ×1
- autoregressive language model family ×1
- long-sequence transformer ×1
- neural language model ×1
- open-source language model ×1
- transformer-based language model architecture ×1
Instances (20)
- GPT series
- Elmo via concept surface "language model"
- GPT-Neo via concept surface "autoregressive language model family"
- RoBERTa via concept surface "language model"
- DistilBERT via concept surface "pretrained language model"
- XLNet via concept surface "language model"
- T5 via concept surface "neural language model"
- ALBERT via concept surface "language model"
- DeBERTa via concept surface "pretrained language model"
- LLaMA
- OPT
- Falcon
- XLM-R via concept surface "XLM-RoBERTa architecture"
- BigBird via concept surface "transformer-based language model architecture"
- LayoutLM via concept surface "pretrained language model"
- BERT via concept surface "language model"
- GPT
- GPT-1 via concept surface "Generative Pre-trained Transformer"
- Llama
- Gee language model series