Back to articles
AIOpenAI News

Benchmarking safe exploration in deep reinforcement learning

OpenAI November 21, 2019 Publication Benchmarking safe exploration in deep reinforcement learning Read paper (opens in a new window) Loading… Share Abstract Reinforcement learning (RL) agents need to explo...

The RSS feed only provided an excerpt. FlowMarket recovered the public content available from the original page without bypassing restricted content.

November 21, 2019

Benchmarking safe exploration in deep reinforcement learning

Benchmarking Safe Exploration In Deep Reinforcement Learning

Abstract

Reinforcement learning (RL) agents need to explore their environments in order to learn optimal policies by trial and error. In many environments, safety is a critical concern and certain errors are unacceptable: for example, robotics systems that interact with humans should never cause injury to the humans while exploring. While it is currently typical to train RL agents mostly or entirely in simulation, where safety concerns are minimal, we anticipate that challenges in simulating the complexities of the real world (such as human-AI interactions) will cause a shift towards training RL agents directly in the real world, where safety concerns are paramount. Consequently we take the position that safe exploration should be viewed as a critical focus area for RL research, and in this work we make three contributions to advance the study of safe exploration. First, building on a wide range of prior work on safe reinforcement learning, we propose to standardize constrained RL as the main formalism for safe exploration. Second, we present the Safety Gym benchmark suite, a new slate of high-dimensional continuous control environments for measuring research progress on constrained RL. Finally, we benchmark several constrained deep RL algorithms on Safety Gym environments to establish baselines that future work can build on.

  • Ethics & Safety

Authors

Related articles

Disrupting malicious > media

Security Feb 14, 2024

Image de l'article

Publication Jan 31, 2024

Democratic Inputs To AI Grant Program Update

Safety Jan 16, 2024

Need an n8n workflow or help installing it?

After the briefing, move to execution: find an n8n template or a creator who can adapt it to your tools.

Source

OpenAI News - openai.com

View original publication