Results for "agents"
Expected cumulative reward from a state or state-action pair.
Continuous cycle of observation, reasoning, action, and feedback.
Learning from data generated by a different policy.
Separates planning from execution in agent architectures.
Models evaluating and improving their own outputs.
Internal sensing of joint positions, velocities, and forces.
Using production outcomes to improve models.
External sensing of surroundings (vision, audio, lidar).
Randomizing simulation parameters to improve real-world transfer.
Predicts next state given current state and action.
Directly optimizing control policies.
Modifying reward to accelerate learning.
Reward only given upon task completion.
Modeling environment evolution in latent space.
Acting to minimize surprise or free energy.
Imagined future trajectories.
Perceived actions an environment allows.
Human-like understanding of physical behavior.
Humans assist or override autonomous behavior.
Intelligence emerges from interaction with the physical world.
Robots learning via exploration and growth.
Learning without catastrophic forgetting.
Combination of cooperation and competition.
No agent benefits from unilateral deviation.
Rules governing auctions.
Designing efficient marketplaces.
System-level behavior arising from interactions.
Early signals disproportionately influence outcomes.
Existential risk from AI systems.
Internal representation of the agent itself.