Results for "optimization"
Convex Optimization
IntermediateOptimization problems where any local minimum is global.
Convex Optimization is like finding the lowest point in a bowl-shaped curve. When you have a problem where you want to minimize something, like costs or errors, and the shape of the problem is nice and smooth (convex), it means that if you find a low point, it’s the lowest point possible. This ma...
Lowest possible loss.
A subfield of AI where models learn patterns from data to make predictions or decisions, improving with experience rather than explicit rule-coding.
The field of building systems that perform tasks associated with human intelligence—perception, reasoning, language, planning, and decision-making—via algori...
Training with a small labeled dataset plus a larger unlabeled dataset, leveraging assumptions like smoothness/cluster structure.
A branch of ML using multi-layer neural networks to learn hierarchical representations, often excelling in vision, speech, and language.
Methods that learn training procedures or initializations so models can adapt quickly to new tasks with little data.
Learning where data arrives sequentially and the model updates continuously, often under changing distributions.
The learned numeric values of a model adjusted during training to minimize a loss function.
Automatically learning useful internal features (latent variables) that capture salient structure for downstream tasks.
A scalar measure optimized during training, typically expected loss over data, sometimes with regularization terms.
Average of squared residuals; common regression objective.
Iterative method that updates parameters in the direction of negative gradient to minimize loss.
A gradient method using random minibatches for efficient training on large datasets.
Popular optimizer combining momentum and per-parameter adaptive step sizes via first/second moment estimates.
One complete traversal of the training dataset during training.
A parameterized function composed of interconnected units organized in layers with nonlinear activations.
Nonlinear functions enabling networks to approximate complex mappings; ReLU variants dominate modern DL.
Gradients grow too large, causing divergence; mitigated by clipping, normalization, careful init.
An RNN variant using gates to mitigate vanishing gradients and capture longer context.
Constraining outputs to retrieved or provided sources, often with citation, to improve factual reliability.
Updating a pretrained model’s weights on task-specific data to improve performance or adapt style/behavior.
A preference-based training method optimizing policies directly from pairwise comparisons without explicit RL loops.
Model trained to predict human preferences (or utility) for candidate outputs; used in RLHF-style pipelines.
Ensuring model behavior matches human goals, norms, and constraints, including reducing harmful or deceptive outputs.
Ordering training samples from easier to harder to improve convergence or generalization.
Training across many devices/silos without centralizing raw data; aggregates updates, not data.
Practices for operationalizing ML: versioning, CI/CD, monitoring, retraining, and reliable production management.
Converts logits to probabilities by exponentiation and normalization; common in classification and LMs.
Removing weights or neurons to shrink models and improve efficiency; can be structured or unstructured.
Inputs crafted to cause model errors or unsafe behavior, often imperceptible in vision or subtle in text.