Results for "model-based"
Model-Based RL
AdvancedRL using learned or known environment models.
Model-based reinforcement learning is like having a map while exploring a new city. Instead of wandering around aimlessly, you can look at the map to plan your route and make better decisions about where to go next. In this type of learning, an AI agent first learns how the environment works—like...
Imagined future trajectories.
Grouping patients by predicted outcomes.
Predicting borrower default risk.
Ensuring models comply with lending fairness laws.
Returns above benchmark.
Effect of trades on prices.
Early signals disproportionately influence outcomes.
When a model fits noise/idiosyncrasies of training data and performs poorly on unseen data.
Separating data into training (fit), validation (tune), and test (final estimate) to avoid leakage and optimism bias.
A robust evaluation technique that trains/evaluates across multiple splits to estimate performance variability.
A high-priority instruction layer setting overarching behavior constraints for a chat model.
Studying internal mechanisms or input influence on outputs (e.g., saliency maps, SHAP, attention analysis).
Training across many devices/silos without centralizing raw data; aggregates updates, not data.
System for running consistent evaluations across tasks, versions, prompts, and model settings.
Hidden behavior activated by specific triggers, causing targeted mispredictions or undesired outputs.
Error due to sensitivity to fluctuations in the training dataset.
The range of functions a model can represent.
Embedding signals to prove model ownership.
Probabilistic model for sequential data with latent states.
Maps audio signals to linguistic units.
Generative model that learns to reverse a gradual noise process.
Shift in model outputs.
Cost of model training.
Models whose weights are publicly available.
Models accessible only via service APIs.
Task instruction without examples.
One example included to guide output.
Requirement to reveal AI usage in legal decisions.
The field of building systems that perform tasks associated with human intelligence—perception, reasoning, language, planning, and decision-making—via algori...
A subfield of AI where models learn patterns from data to make predictions or decisions, improving with experience rather than explicit rule-coding.