AI Glossary
100 essential AI & machine learning terms explained clearly โ from neural networks to RAG, LLMs, and beyond.
Artificial Intelligence
Core ConceptsThe simulation of human intelligence by machines โ enabling computers to learn, reason, and make decisions.
Learn more โAI Agent
ApplicationsAn AI system that can perceive its environment, make decisions, use tools, and take autonomous actions to achieve a goal.
Learn more โAttention Mechanism
Models & ArchitectureA neural network technique that lets a model focus on the most relevant parts of an input when producing an output.
Learn more โActivation Function
Models & ArchitectureA mathematical function applied inside neural networks that introduces nonlinearity and lets models learn complex patterns.
Learn more โAI Alignment
Safety & AlignmentThe effort to make AI systems behave in ways that are helpful, safe, and consistent with human goals and values.
Learn more โAutoencoder
Models & ArchitectureA neural network that learns to compress data into a low-dimensional representation and reconstruct it back to the original.
Learn more โAdversarial Attack
Safety & AlignmentAn input crafted to fool a machine learning model into making a wrong prediction, often imperceptible to humans.
Learn more โBackpropagation
TrainingThe algorithm used to compute how each model parameter contributed to error so the network can update itself during training.
Learn more โBatch Size
Training & LearningThe number of training examples processed together in one forward/backward pass of the model.
Learn more โBERT
Models & ArchitectureBidirectional Encoder Representations from Transformers โ a Google-built language model that reads text in both directions.
Learn more โBatch Normalization
Training & LearningA layer that normalizes activations across a batch to stabilize and speed up training of deep neural networks.
Learn more โContext Window
Inference & GenerationThe maximum amount of text (measured in tokens) that an AI model can process in a single interaction.
Learn more โCLIP
Models & ArchitectureA vision-language model that learns shared representations of images and text so they can be compared in the same embedding space.
Learn more โComputer Vision
Core ConceptsThe field of AI focused on enabling machines to interpret and understand images and video.
Learn more โConvolutional Neural Network
Models & ArchitectureA neural network architecture specialized for processing grid-like data such as images using convolutional filters.
Learn more โChain of Thought
Techniques & MethodsA prompting technique that asks an LLM to reason step by step before giving a final answer, improving complex reasoning.
Learn more โChunking
RAG & RetrievalThe process of splitting long documents into smaller pieces that fit into a language model's context window.
Learn more โCosine Similarity
RAG & RetrievalA metric that measures similarity between two vectors based on the cosine of the angle between them, commonly used for embeddings.
Learn more โCross-Attention
Models & ArchitectureAn attention mechanism where queries from one sequence attend to keys and values from a different sequence.
Learn more โCross-Validation
Evaluation & MetricsA technique for evaluating model performance by splitting data into multiple folds and testing on each fold in turn.
Learn more โConstitutional AI
Safety & AlignmentAn alignment technique developed by Anthropic where an AI model critiques and revises its own outputs using a set of principles.
Learn more โDeep Learning
Core ConceptsA subset of machine learning using multi-layered neural networks to learn complex patterns from large datasets.
Learn more โDiffusion Model
Models & ArchitectureAn AI model that generates images or other media by learning to reverse a gradual noise-adding process.
Learn more โDistillation
Techniques & MethodsA technique for transferring knowledge from a large "teacher" model to a smaller "student" model that can run faster and cheaper.
Learn more โDropout
Training & LearningA regularization technique that randomly deactivates a fraction of neurons during training to prevent co-adaptation and overfitting.
Learn more โData Augmentation
Training & LearningTechniques that expand training data by creating modified versions of existing examples โ like rotating images or paraphrasing text.
Learn more โEmbedding
Models & ArchitectureA numerical vector representation of text, images, or other data that captures semantic meaning in a high-dimensional space.
Learn more โEpoch
Training & LearningOne complete pass through the entire training dataset during model training.
Learn more โExplainability
Safety & AlignmentThe degree to which humans can understand why a model made a particular prediction or decision.
Learn more โFine-tuning
TrainingTraining a pre-trained model further on a smaller, task-specific dataset to adapt it for a particular use case.
Learn more โFunction Calling
ApplicationsA capability that lets language models request structured tool or API calls instead of only generating plain text.
Learn more โFew-Shot Learning
ApplicationsA prompting or training approach where a model is shown a small number of examples before handling a new task.
Learn more โF1 Score
Evaluation & MetricsA classification metric that combines precision and recall into a single score, balancing false positives and false negatives.
Learn more โFlashAttention
Inference & OptimizationA memory-efficient attention algorithm that speeds up transformer training and inference by avoiding materialization of the full attention matrix.
Learn more โFoundation Model
Core ConceptsA large-scale model trained on broad data that can be adapted to many downstream tasks โ like GPT-4, Claude, or Gemini.
Learn more โGenerative AI
Core ConceptsAI that can create new content โ text, images, audio, video, and code โ rather than just classifying or predicting.
Learn more โGradient Descent
TrainingAn optimization method that updates model parameters in the direction that most reduces prediction error.
Learn more โGenerative Adversarial Network
Models & ArchitectureA generative model architecture where two neural networks compete: a generator creates data and a discriminator tries to detect fakes.
Learn more โInference
Inference & GenerationThe process of using a trained AI model to generate predictions, classifications, or responses on new input data.
Learn more โIn-Context Learning
ApplicationsThe ability of a model to learn patterns from instructions and examples provided inside the current prompt without updating its weights.
Learn more โInstruction Tuning
TrainingA fine-tuning approach where a model is trained on many instruction-and-response examples to improve its ability to follow user requests.
Learn more โKnowledge Graph
DataA structured representation of entities and the relationships between them, used to organize and reason over information.
Learn more โKV Cache
Inference & OptimizationA memory structure that stores previously computed attention keys and values, allowing LLMs to generate tokens without recomputing from scratch.
Learn more โLarge Language Model
Models & ArchitectureA massive AI model trained on text data that can generate, summarize, translate, and reason about language.
Learn more โLoss Function
TrainingA mathematical measure of how wrong a modelโs predictions are during training.
Learn more โLoRA
TrainingLow-Rank Adaptation, a parameter-efficient fine-tuning method that updates a small set of low-rank matrices instead of the full model.
Learn more โLearning Rate
Training & LearningA hyperparameter that controls how much the model weights change with each update during training.
Learn more โLSTM
Models & ArchitectureLong Short-Term Memory โ a type of recurrent neural network designed to learn long-range dependencies in sequential data.
Learn more โLayer Normalization
Training & LearningA normalization technique that stabilizes training by normalizing activations across features within each sample.
Learn more โMachine Learning
Core ConceptsA subset of AI where systems learn from data to improve performance without being explicitly programmed.
Learn more โMultimodal AI
Core ConceptsAI systems that can process and generate multiple types of data โ such as text, images, audio, and video โ in a unified model.
Learn more โMulti-Head Attention
Models & ArchitectureA transformer technique that runs multiple attention operations in parallel so the model can capture different kinds of relationships at once.
Learn more โModel Weights
Models & ArchitectureThe learned parameter values in a neural network that determine how input signals are transformed into outputs.
Learn more โModel Context Protocol
AI AgentsAn open protocol for connecting AI assistants to external tools, data sources, and systems in a standardized way.
Learn more โMixture of Experts
Models & ArchitectureAn architecture where a gating network routes each input to a small subset of specialized sub-models (experts), enabling massive parameter counts efficiently.
Learn more โModel Card
Safety & AlignmentA standardized document that describes a model's purpose, capabilities, limitations, training data, and intended use.
Learn more โPrompt Engineering
ApplicationsThe practice of crafting inputs to AI models to elicit better, more accurate, or more useful outputs.
Learn more โParameters
Models & ArchitectureThe learned numerical values inside a neural network that store what the model has learned from training data.
Learn more โPrompt Chaining
ApplicationsA workflow pattern where multiple prompts are linked together so the output of one step becomes the input to the next.
Learn more โPre-trained Model
TrainingA model that has already been trained on broad data and can then be adapted or used for downstream tasks.
Learn more โPerplexity
Evaluation & MetricsA metric that measures how well a language model predicts text โ lower perplexity means better predictions.
Learn more โPositional Encoding
Models & ArchitectureA technique for injecting information about token positions into transformer models, which otherwise have no notion of order.
Learn more โPretraining
Training & LearningThe initial training phase where a model learns general patterns from large amounts of raw data before being fine-tuned for specific tasks.
Learn more โPrompt Injection
Safety & AlignmentAn attack where malicious instructions in user input override an AI system's original instructions.
Learn more โQLoRA
TrainingA LoRA-based fine-tuning method that combines low-rank adapters with quantized base models to reduce memory requirements even further.
Learn more โQuantization
Inference & GenerationA technique that reduces model size and inference cost by storing weights and activations with lower numerical precision.
Learn more โRetrieval-Augmented Generation
ApplicationsA technique that enhances LLM outputs by first retrieving relevant documents from an external knowledge base before generating a response.
Learn more โRLHF
TrainingReinforcement Learning from Human Feedback โ a training technique that aligns AI models with human preferences using human ratings.
Learn more โReinforcement Learning
Training & LearningA machine learning paradigm where an agent learns to take actions in an environment to maximize cumulative reward.
Learn more โRecurrent Neural Network
Models & ArchitectureA neural network architecture with connections that loop back, allowing it to process sequences and maintain memory of past inputs.
Learn more โRAG Pipeline
RAG & RetrievalThe end-to-end architecture for retrieval-augmented generation, from query through retrieval to final LLM response.
Learn more โRegularization
Training & LearningTechniques that prevent models from overfitting training data by penalizing complexity or introducing noise.
Learn more โResidual Connection
Models & ArchitectureA shortcut that adds a layer's input to its output, enabling much deeper networks by preserving gradient flow.
Learn more โReranking
RAG & RetrievalA second-stage retrieval step that reorders initial search results using a more accurate but slower model to improve relevance.
Learn more โSelf-Attention
Models & ArchitectureA form of attention where each token in a sequence looks at every other token in the same sequence to build context-aware representations.
Learn more โSystem Prompt
ApplicationsA high-priority instruction that sets the role, behavior, constraints, and goals for an AI model within an application.
Learn more โSemantic Search
DataA search approach that finds results based on meaning and intent rather than exact keyword matches.
Learn more โSupervised Learning
Training & LearningA machine learning approach where models learn from labeled input-output pairs to predict outcomes on new data.
Learn more โSelf-Supervised Learning
Training & LearningA form of learning where the model creates its own labels from raw data, enabling training on massive unlabeled datasets.
Learn more โSpeculative Decoding
Inference & OptimizationAn inference acceleration technique where a small draft model predicts multiple tokens that a larger model then verifies in parallel.
Learn more โSynthetic Data
Training & LearningArtificially generated data used to train or evaluate AI models, often created by other models or simulations.
Learn more โTransformer
Models & ArchitectureThe neural network architecture behind most modern AI โ uses attention mechanisms to process sequences in parallel.
Learn more โTokenization
Language & TextThe process of splitting text into smaller units (tokens) that a language model can process.
Learn more โTemperature
Inference & GenerationA parameter that controls the randomness of an AI model's outputs โ lower values are more deterministic, higher values are more creative.
Learn more โToken
Language & TextThe basic unit of text processed by a language model, often representing a word, subword, punctuation mark, or symbol.
Learn more โText-to-Image
ApplicationsAI generation that creates images from natural language prompts.
Learn more โText-to-Video
ApplicationsAI generation that creates video clips from natural language prompts.
Learn more โTransfer Learning
Training & LearningThe practice of reusing knowledge from a model trained on one task to accelerate learning on a different but related task.
Learn more โTraining Data
Training & LearningThe dataset used to teach a machine learning model the patterns it needs to make predictions or generate outputs.
Learn more โTree of Thoughts
Techniques & MethodsAn advanced reasoning technique where the model explores multiple reasoning paths in a tree structure before choosing the best.
Learn more โTop-K Sampling
Inference & OptimizationA text generation strategy that restricts sampling to the K most likely next tokens at each step.
Learn more โTop-P Sampling
Inference & OptimizationA text generation strategy that samples from the smallest set of tokens whose cumulative probability exceeds P.
Learn more โ