Results for "aligned agents"
Agent calls external tools dynamically.
Interleaving reasoning and tool use.
Agent reasoning about future outcomes.
AI systems that perceive and act in the physical world through sensors and actuators.
Agents optimize collective outcomes.
Agents have opposing objectives.
Market reacting strategically to AI.
Collective behavior without central control.
Agents copy others’ actions.
The text (and possibly other modalities) given to an LLM to condition its output behavior.
A high-priority instruction layer setting overarching behavior constraints for a chat model.
Continuous cycle of observation, reasoning, action, and feedback.
Expected cumulative reward from a state or state-action pair.
Separates planning from execution in agent architectures.
Learning from data generated by a different policy.
Models evaluating and improving their own outputs.
Using production outcomes to improve models.
Internal sensing of joint positions, velocities, and forces.
External sensing of surroundings (vision, audio, lidar).
Randomizing simulation parameters to improve real-world transfer.
Predicts next state given current state and action.
Directly optimizing control policies.
Modifying reward to accelerate learning.
Reward only given upon task completion.
Modeling environment evolution in latent space.
Imagined future trajectories.
Acting to minimize surprise or free energy.
Perceived actions an environment allows.
Human-like understanding of physical behavior.
Humans assist or override autonomous behavior.