Skip to main content

AI glossary

TermDefinition
AI AlignmentEnsuring AI behavior aligns with human goals and values.
API LatencyTime delay between request and response in a deployed AI service.
AUC (Area Under Curve)Measure of model’s ability to distinguish between classes.
AccuracyProportion of correct predictions made by a model.
Activation FunctionFunction that introduces non-linearity into a neural network (e.g., ReLU, Sigmoid).
Artificial Intelligence (AI)Simulation of human intelligence in machines programmed to think and learn.
Attention MechanismMechanism that enables models to focus on relevant parts of the input.
BackpropagationAlgorithm for updating neural network weights.
Batch NormalizationNormalizes activations to improve training speed and stability.
BiasSystematic error introduced by assumptions in data or model.
Chain-of-Thought (CoT)Technique where models generate intermediate reasoning steps.
Computer VisionAI field that enables machines to interpret and make decisions based on visual data.
Confusion MatrixTable showing true vs predicted classifications.
Data PrivacyProtection of sensitive user data during model training and usage.
Deep LearningSubset of ML using neural networks with many layers to model complex patterns.
DropoutRegularization technique that randomly drops units in a neural network during training.
EmbeddingNumerical representation of data, often used for similarity or search.
EpochOne complete pass through the training dataset.
Explainable AI (XAI)Techniques to interpret and understand model predictions.
F1 ScoreHarmonic mean of precision and recall.
Few-shot LearningModel learns from a few labeled examples.
Fine-TuningTraining a pre-trained model on a specific task or dataset.
Generative AIAI models that can generate new content like text, images, audio, or code.
Gradient DescentOptimization algorithm for minimizing loss function.
HallucinationGenerated output that is fluent but factually incorrect.
HuggingFaceEcosystem for pretrained NLP models and transformers.
InferenceRunning a trained model to make predictions.
LangChainFramework for building LLM-powered applications using composable chains.
Large Language Model (LLM)A transformer-based model trained on large corpora of text data.
LlamaIndexTool for indexing and querying external data with LLMs.
LoRA (Low-Rank Adaptation)Efficient method for fine-tuning large models with fewer parameters.
Loss FunctionFunction that measures error between predicted and actual values.
Machine Learning (ML)Subset of AI that allows systems to learn from data without explicit programming.
Mixture of ExpertsModel architecture that routes input through subsets of expert networks.
Model CheckpointingSaving model states during training to resume or analyze progress.
Model FairnessEnsuring model performance does not discriminate against subgroups.
Model ServingHosting and providing access to ML models for inference.
Multimodal AIModels that process and generate across multiple data types (e.g., text, image, audio).
Natural Language Processing (NLP)AI branch that deals with understanding and generation of human language.
Neural NetworkComputational model inspired by the human brain, used in deep learning.
ONNXOpen Neural Network Exchange; format for model interoperability.
Open Weight ModelAI model with publicly available weights for reuse and fine-tuning.
OverfittingModel learns training data too well and fails to generalize.
Positional EncodingAdds information about the position of tokens in sequences to transformer models.
Pre-trainingInitial training of a model on a large generic dataset.
PrecisionProportion of true positive predictions among all positive predictions.
Prompt EngineeringCrafting effective prompts to guide LLM responses.
Proprietary ModelModel with restricted access, typically hosted and maintained by a company.
PyTorchPopular deep learning library developed by Facebook.
ROC CurveGraphical plot showing performance of classification model.
RecallProportion of true positives among all actual positives.
Reinforcement LearningLearning method where agents learn by taking actions and receiving rewards.
Residual ConnectionSkip connections in neural networks that help prevent vanishing gradients.
Responsible AIEthical and accountable development and deployment of AI.
Retrieval-Augmented Generation (RAG)LLM approach that augments prompts with relevant context from a document store.
Scikit-learnPython library for traditional machine learning.
Self-AttentionMechanism allowing models to weigh importance of different parts of input.
Semi-Supervised LearningUses a small amount of labeled data with a large amount of unlabeled data.
Supervised LearningTraining a model on labeled data.
TensorFlowGoogle’s open-source deep learning framework.
TensorRTNVIDIA platform for optimizing and deploying deep learning models.
TokenizationBreaking text into smaller units (tokens) for processing.
TransformerDeep learning model architecture that uses self-attention for sequence tasks.
UnderfittingModel is too simple to capture underlying patterns in data.
Unsupervised LearningTraining a model on data without labels to find patterns.
VarianceModel's sensitivity to small fluctuations in training data.
Vector DatabaseDatabase optimized for storing and querying high-dimensional embeddings.
Zero-shot LearningModel predicts on tasks without having seen labeled examples.