AI Glossary
From tokens to transformers, chain-of-thought to RAG — every term you need to know to work effectively with AI.
What is the Glossary? A dictionary of AI and prompt engineering terms — from tokens and temperature to RAG and chain-of-thought. Each entry has a plain-English definition, real examples, and links to related terms. Use it when you hit an unfamiliar word.
Adversarial Prompting
SafetyCrafting inputs specifically designed to cause a model to behave incorrectly or unsafely.
Agent
ConceptsAn AI system that autonomously plans and executes multi-step tasks using tools and environment feedback.
Agentic Workflow
ConceptsA multi-step pipeline where an AI agent plans, acts, observes, and iterates to complete a complex task.
AI Alignment
SafetyThe research field focused on ensuring AI systems pursue goals and values intended by their designers.
AI Safety
SafetyThe interdisciplinary field studying how to develop AI systems that are safe, reliable, and beneficial.
API
TechnicalAn Application Programming Interface that lets developers call AI model capabilities programmatically.
Attention Mechanism
TechnicalThe core transformer operation that weighs the relevance of each token to every other token in a sequence.
Autonomous AI
ConceptsAn AI system capable of pursuing goals and executing long-horizon tasks with minimal human oversight.
Base Model
ModelsA model trained only on next-token prediction over a large corpus, before any instruction tuning.
Benchmark
ConceptsA standardised test suite used to measure and compare AI model capabilities across tasks.
Bias in AI
SafetySystematic errors in model outputs that unfairly favour or disadvantage certain groups or perspectives.
Chain-of-Thought
PromptingA prompting technique that asks the model to reason step-by-step before giving a final answer.
Chat Model
ModelsA model fine-tuned to handle multi-turn conversational exchanges between a user and an assistant.
Claude
ModelsAnthropic's family of large language models, known for safety, long context, and instruction-following.
Closed-Source Model
ModelsAn AI model whose weights and training details are proprietary and not publicly released.
Code Model
ModelsA language model specifically trained or fine-tuned to understand and generate programming code.
Completion Model
ModelsA model that continues text from a given prefix, used before chat-style interfaces became standard.
Constitutional AI
SafetyAnthropic's technique for training helpful, harmless AI using a set of written principles as the training signal.
Content Moderation
SafetyAutomated or human review of AI inputs and outputs to prevent harmful, illegal, or policy-violating content.
Context Length
TechnicalThe maximum number of tokens a model can handle in a single forward pass.
Context Window
PromptingThe maximum number of tokens a model can process in a single input-output interaction.
Dataset
TrainingA structured collection of data examples used to train, validate, or evaluate a machine learning model.
Directional Stimulus
PromptingA hint or nudge embedded in the prompt that steers the model toward a desired answer style.
DPO (Direct Preference Optimization)
TrainingA training method that aligns models to human preferences without requiring a separate reward model.
Embedding
TechnicalA dense numerical vector that represents a token, sentence, or document in a continuous semantic space.
Emergent Behaviour
ConceptsCapabilities that appear suddenly in large models without being explicitly trained for.
Evaluation
ConceptsThe systematic process of measuring AI model quality, safety, and alignment against defined criteria.
Fairness
SafetyThe property of an AI system treating individuals and groups equitably and without unjust discrimination.
Few-Shot Learning
PromptingProviding a small number of input-output examples in the prompt to teach the model a task pattern.
Fine-Tuned Model
ModelsA pretrained model whose weights have been updated on a specific dataset for a targeted task.
Fine-Tuning
TrainingContinuing training of a pretrained model on a smaller, task-specific dataset to specialise its behaviour.
Foundation Model
ModelsA large model trained on broad data that can be adapted to many downstream tasks.
Function Calling
TechnicalA model feature that allows it to request the execution of developer-defined functions with structured arguments.
Gemini
ModelsGoogle DeepMind's family of multimodal large language models, available in Ultra, Pro, and Nano tiers.
Generated Knowledge
PromptingA technique where the model first generates relevant facts, then uses them to answer the question.
GPT-4
ModelsOpenAI's large multimodal model that set the benchmark standard for general-purpose AI capability.
Grounding
TechnicalConnecting model outputs to verifiable external sources to reduce hallucination and improve accuracy.
Guardrails
SafetyProgrammatic constraints that prevent an AI application from producing or acting on harmful outputs.
Inference
TechnicalThe process of running a trained model to generate predictions or responses from new inputs.
Instruction Prompting
PromptingDirectly telling the model what to do using clear imperative commands.
Instruction Tuning
TrainingSupervised fine-tuning on diverse instruction-response pairs to improve a model's ability to follow commands.
Instruction-Tuned Model
ModelsA model fine-tuned on instruction-response pairs to follow natural-language directives reliably.
Iterative Prompting
PromptingRefining a prompt through multiple rounds of feedback until the output meets requirements.
Large Language Model
ModelsA neural network with billions of parameters trained on text to understand and generate language.
Latency
TechnicalThe time delay between sending a request to a model and receiving its first token in response.
Llama
ModelsMeta's open-weight large language model family, the dominant open-source alternative to GPT and Claude.
Logit
TechnicalThe raw, unnormalised score a model assigns to each vocabulary token before converting to probabilities.
LoRA (Low-Rank Adaptation)
TrainingA parameter-efficient fine-tuning method that updates only small low-rank matrices injected into model layers.
Memory in AI
ConceptsMechanisms that allow AI agents to store, retrieve, and act on information across interactions.
Meta-Prompt
PromptingA prompt that instructs the model to generate or improve other prompts.
Mistral
ModelsA French AI company producing efficient open-weight models that punch above their size class.
Multimodal Model
ModelsA model that can process and generate across multiple data types such as text, images, and audio.
Parameter-Efficient Fine-Tuning
TrainingA family of methods that fine-tune large models by updating only a small fraction of their parameters.
Planning in AI
ConceptsThe process by which an AI agent determines a sequence of actions to achieve a goal.
Positional Encoding
TechnicalA mechanism that injects token position information into transformer inputs, since attention is order-agnostic.
Pretraining
TrainingThe initial phase of training a model on massive text data to learn general language representations.
Prompt
PromptingThe input text sent to an AI model to elicit a specific response or behaviour.
Prompt Chaining
PromptingPassing the output of one model call as the input to a subsequent call to complete complex tasks.
Prompt Compression
PromptingReducing prompt length while preserving the information needed for accurate responses.
Prompt Injection
PromptingAn attack where malicious text in external data hijacks the model's instruction-following behaviour.
Prompt Template
PromptingA reusable prompt structure with placeholders that are filled in at runtime.
RAG (Retrieval-Augmented Generation)
TechnicalAugmenting model responses by retrieving relevant documents from an external knowledge base at inference time.
ReAct Prompting
PromptingInterleaving reasoning traces and actions so a model can use tools and reflect on their results.
Reasoning
ConceptsThe capacity of an AI model to derive conclusions from premises through logical or analogical inference.
Reasoning Model
ModelsA model trained to perform extended internal reasoning before producing a response.
Red Teaming
SafetySystematically testing an AI system by attempting to elicit harmful or unintended behaviours before deployment.
Reinforcement Learning from Human Feedback
TrainingA training technique that uses human preference ratings to align model outputs with human values.
Responsible AI
SafetyThe practice of developing and deploying AI systems ethically, transparently, and with accountability.
RLHF
TrainingShorthand for Reinforcement Learning from Human Feedback — the alignment training paradigm.
Role Prompting
PromptingAssigning the model a persona or expert identity to shape the style and depth of its responses.
Scaling Laws
ConceptsEmpirical relationships describing how model performance improves predictably with more compute, data, or parameters.
Self-Attention
TechnicalAn attention operation where a sequence attends to itself, allowing each token to gather context from all others.
Self-Consistency
PromptingSampling multiple reasoning paths and selecting the most common answer to improve reliability.
Softmax
TechnicalA function that converts a vector of real numbers into a probability distribution summing to 1.
Streaming
TechnicalSending model output tokens to the client incrementally as they are generated rather than all at once.
Structured Output
PromptingConstraining model generation to produce valid, machine-parseable formats like JSON or XML.
Supervised Learning
TrainingTraining a model on input-output pairs where the correct output is provided as a label.
Synthetic Data
TrainingTraining data generated by AI models rather than collected from human-created sources.
System Prompt
PromptingA privileged instruction block sent before the conversation that sets the model's persona and rules.
Temperature
PromptingA sampling parameter that controls the randomness and creativity of model outputs.
Throughput
TechnicalThe number of tokens or requests an inference system can process per unit of time.
Token
TechnicalThe basic unit of text a language model processes, roughly corresponding to a word or word fragment.
Tokenisation
TechnicalThe process of splitting text into tokens using a vocabulary and encoding algorithm like BPE.
Tool Use
TechnicalThe ability of a model to invoke external tools — APIs, code executors, search — to augment its capabilities.
Top-P (Nucleus Sampling)
PromptingA sampling strategy that limits token selection to the smallest set covering a cumulative probability.
Training Data
TrainingThe corpus of examples a model learns from during its training process.
Transfer Learning
TrainingReusing a model trained on one task as the starting point for a related task.
Transformer
ModelsThe neural network architecture that underpins all modern large language models, based on self-attention.
Tree of Thoughts
PromptingA framework that explores multiple reasoning branches in parallel and selects the most promising path.
Use PromptITIN to build perfectly structured prompts using these techniques automatically.
Start free — no credit card