Retour aux articles
IAOpenAI News

Benchmarking safe exploration in deep reinforcement learning

OpenAI November 21, 2019 Publication Benchmarking safe exploration in deep reinforcement learning Read paper (opens in a new window) Loading… Share Abstract Reinforcement learning (RL) agents need to explo...

Le flux RSS ne fournissait qu'un extrait. FlowMarket a récupéré le contenu public disponible depuis la page originale, sans contourner les contenus réservés.

November 21, 2019

Benchmarking safe exploration in deep reinforcement learning

Benchmarking Safe Exploration In Deep Reinforcement Learning

Abstract

Reinforcement learning (RL) agents need to explore their environments in order to learn optimal policies by trial and error. In many environments, safety is a critical concern and certain errors are unacceptable: for example, robotics systems that interact with humans should never cause injury to the humans while exploring. While it is currently typical to train RL agents mostly or entirely in simulation, where safety concerns are minimal, we anticipate that challenges in simulating the complexities of the real world (such as human-AI interactions) will cause a shift towards training RL agents directly in the real world, where safety concerns are paramount. Consequently we take the position that safe exploration should be viewed as a critical focus area for RL research, and in this work we make three contributions to advance the study of safe exploration. First, building on a wide range of prior work on safe reinforcement learning, we propose to standardize constrained RL as the main formalism for safe exploration. Second, we present the Safety Gym benchmark suite, a new slate of high-dimensional continuous control environments for measuring research progress on constrained RL. Finally, we benchmark several constrained deep RL algorithms on Safety Gym environments to establish baselines that future work can build on.

  • Ethics & Safety

Authors

Related articles

Disrupting malicious > media

Security Feb 14, 2024

Image de l'article

Publication Jan 31, 2024

Democratic Inputs To AI Grant Program Update

Safety Jan 16, 2024

Besoin d'un workflow n8n ou d'aide pour l'installer ?

Après la veille, passez à l'action : trouvez un template n8n ou un créateur capable de l'adapter à vos outils.

Source

OpenAI News - openai.com

Voir la publication originale