Home/Glossary/Large Language Model
Models

Large Language Model

A neural network with billions of parameters trained on text to understand and generate language.

Full Definition

A large language model (LLM) is a transformer-based neural network trained on internet-scale text corpora to predict the next token in a sequence. 'Large' refers to both parameter count (billions to trillions) and training data scale (trillions of tokens). LLMs develop emergent capabilities — reasoning, translation, code generation — that are not explicitly programmed and that only appear beyond certain scale thresholds. They are the foundation for virtually every modern AI language application. The term encompasses both base models and instruction-tuned chat models. Key LLM families include GPT, Claude, Gemini, Llama, and Mistral.

Examples

1

GPT-4, with an estimated 1.8 trillion parameters, generating coherent long-form essays on complex topics.

2

Llama 3 70B running on a single high-end GPU server and answering coding questions in real time.

Apply this in your prompts

PromptITIN automatically uses techniques like Large Language Model to build better prompts for you.

✦ Try it free

Related Terms

Transformer

The neural network architecture that underpins all modern large language models,

View →

Foundation Model

A large model trained on broad data that can be adapted to many downstream tasks

View →

Neural Network

A computational model loosely inspired by the brain, made of interconnected laye

View →
← Browse all 100 terms