Results for "dynamics learning"
Ensuring learned behavior matches intended objective.
Using limited human feedback to guide large models.
Model behaves well during training but not deployment.
Asking model to review and improve output.
Applying learned patterns incorrectly.
Train/test environment mismatch.
Model relies on irrelevant signals.
Startup latency for services.
Running models locally.
AI systems that perceive and act in the physical world through sensors and actuators.
Algorithm computing control actions.
Directly optimizing control policies.
Reward only given upon task completion.
Control shared between human and agent.
Inferring human goals from behavior.
Automated assistance identifying disease indicators.
AI supporting legal research, drafting, and analysis.
AI-assisted review of legal documents.
AI selecting next experiments.
AI tacitly coordinating prices.
Research ensuring AI remains safe.
A continuous vector encoding of an item (word, image, user) such that semantic similarity corresponds to geometric closeness.
The learned numeric values of a model adjusted during training to minimize a loss function.
Minimizing average loss on training data; can overfit when data is limited or biased.
When a model fits noise/idiosyncrasies of training data and performs poorly on unseen data.
When a model cannot capture underlying structure, performing poorly on both training and test data.
How well a model performs on new data drawn from the same (or similar) distribution as training.
Separating data into training (fit), validation (tune), and test (final estimate) to avoid leakage and optimism bias.
A robust evaluation technique that trains/evaluates across multiple splits to estimate performance variability.
A table summarizing classification outcomes, foundational for metrics like precision, recall, specificity.