Results for "training loss"

188 results

Train/Validation/Test Split Intermediate

Separating data into training (fit), validation (tune), and test (final estimate) to avoid leakage and optimism bias.

Evaluation & Benchmarking
ReLU Intermediate

Activation max(0, x); improves gradient flow and training speed in deep nets.

Foundations & Theory
Exploding Gradient Intermediate

Gradients grow too large, causing divergence; mitigated by clipping, normalization, careful init.

Foundations & Theory
Weight Initialization Intermediate

Methods to set starting weights to preserve signal/gradient scales across layers.

Foundations & Theory
Dropout Intermediate

Randomly zeroing activations during training to reduce co-adaptation and overfitting.

Foundations & Theory
Data Augmentation Intermediate

Expanding training data via transformations (flips, noise, paraphrases) to improve robustness.

Foundations & Theory
Compute Intermediate

Hardware resources used for training/inference; constrained by memory bandwidth, FLOPs, and parallelism.

Foundations & Theory
Data Poisoning Intermediate

Maliciously inserting or altering training data to implant backdoors or degrade performance.

Foundations & Theory
Privacy Attack Intermediate

Attacks that infer whether specific records were in training data, or reconstruct sensitive training examples.

Foundations & Theory
Variance Term Intermediate

Error due to sensitivity to fluctuations in the training dataset.

AI Economics & Strategy
Learning Rate Schedule Intermediate

Adjusting learning rate over training to improve convergence.

AI Economics & Strategy
Generative Model Advanced

Models that learn to generate samples resembling training data.

Diffusion & Generative Models
Data Scaling Intermediate

Increasing performance via more data.

AI Economics & Strategy
Chinchilla Scaling Intermediate

Scaling law optimizing compute vs data.

AI Economics & Strategy
Adaptive Optimization Intermediate

Methods like Adam adjusting learning rates dynamically.

Foundations & Theory
Distribution Shift Intermediate

Train/test environment mismatch.

Model Failure Modes
Domain Randomization Advanced

Randomizing simulation parameters to improve real-world transfer.

Simulation & Sim-to-Real
Supervised Learning Intermediate

Learning a function from input-output pairs (labeled data), optimizing performance on predicting outputs for unseen inputs.

Machine Learning
Online Learning Intermediate

Learning where data arrives sequentially and the model updates continuously, often under changing distributions.

Machine Learning
Transfer Learning Intermediate

Reusing knowledge from a source task/domain to improve learning on a target task/domain, typically via pretrained models.

Machine Learning
Representation Learning Intermediate

Automatically learning useful internal features (latent variables) that capture salient structure for downstream tasks.

Machine Learning
Calibration Intermediate

The degree to which predicted probabilities match true frequencies (e.g., 0.8 means ~80% correct).

Foundations & Theory
Neural Network Intermediate

A parameterized function composed of interconnected units organized in layers with nonlinear activations.

Neural Networks
Context Window Intermediate

Maximum number of tokens the model can attend to in one forward pass; constrains long-document reasoning.

Transformers & LLMs
Fine-Tuning Intermediate

Updating a pretrained model’s weights on task-specific data to improve performance or adapt style/behavior.

Large Language Models
LIME Intermediate

Local surrogate explanation method approximating model behavior near a specific input.

Foundations & Theory
Differential Privacy Intermediate

A formal privacy framework ensuring outputs do not reveal much about any single individual’s data contribution.

Security & Privacy
Convex Optimization Intermediate

Optimization problems where any local minimum is global.

AI Economics & Strategy
Pruning Intermediate

Removing weights or neurons to shrink models and improve efficiency; can be structured or unstructured.

Foundations & Theory
Non-Convex Optimization Intermediate

Optimization with multiple local minima/saddle points; typical in neural networks.

AI Economics & Strategy

Welcome to AI Glossary

The free, curated AI dictionary built from real, established terms and designed for a clean reading experience.

Search

Type any question or keyword into the search bar at the top.

Browse

Tap a letter in the A–Z bar to browse terms alphabetically, or filter by domain, industry, or difficulty level.

3D WordGraph

Fly around the interactive 3D graph to explore how AI concepts connect. Click any word to read its full definition.