Retour aux articles
IAOpenAI News

Requests for Research 2.0

We’re releasing a new batch of seven unsolved problems which have come up in the course of our research at OpenAI.

Le flux RSS ne fournissait qu'un extrait. FlowMarket a récupéré le contenu public disponible depuis la page originale, sans contourner les contenus réservés.

January 31, 2018

Requests for Research 2.0

We’re releasing a new batch of seven unsolved problems which have come up in the course of our research at OpenAI.

Requests For Research 20

Like our original  Requests for Research ⁠ (opens in a new window) (which  resulted ⁠ (opens in a new window)   in ⁠ (opens in a new window)   several ⁠ (opens in a new window)   papers ⁠ (opens in a new window) ), we expect these problems to be a fun and meaningful way for new people to enter the field, as well as for practitioners to hone their skills (it’s also a great way to get a  job ⁠ (opens in a new window)  at OpenAI). Many will require inventing new ideas. Please  email ⁠  us with questions or solutions you’d like us to publicize!

(Also, if you don’t have deep learning background but want to learn to solve problems like these, please apply for our  Fellowship ⁠ (opens in a new window)  program!)

Warmups

If you’re not sure where to begin, here are some solved starter problems.

⭐ Train an LSTM to solve the  XOR  problem: that is, given a sequence of bits, determine its parity. The  LSTM ⁠ (opens in a new window)  should consume the sequence, one bit at a time, and then output the correct answer at the sequence’s end. Test the two approaches below:

  • Generate a dataset of random 100,000 binary strings of length 50. Train the LSTM; what performance do you get?
  • Generate a dataset of random 100,000 binary strings, where the length of each string is independently and randomly chosen between 1 and 50. Train the LSTM. Does it succeed? What explains the difference?

⭐ Implement a clone of the classic  Snake ⁠ (opens in a new window)  game as a  Gym ⁠ (opens in a new window)  environment, and solve it with a  reinforcement learning ⁠ (opens in a new window)  algorithm of your choice.  Tweet ⁠ (opens in a new window)  us videos of the agent playing. Were you able to train a policy that wins the game?

Requests for Research

⭐⭐  Slitherin’.  Implement and solve a multiplayer clone of the classic  Snake ⁠ (opens in a new window)  game (see  slither.io ⁠ (opens in a new window)  for inspiration) as a  Gym ⁠ (opens in a new window)  environment.

  • Environment: have a reasonably large field with multiple snakes; snakes grow when eating randomly-appearing fruit; a snake dies when colliding with another snake, itself, or the wall; and the game ends when all snakes die. Start with two snakes, and scale from there.
  • Agent: solve the environment using self-play with an RL algorithm of  your ⁠   choice ⁠ (opens in a new window) . You’ll need to experiment with various approaches to overcome self-play instability (which resembles the instability people see with GANs). For example, try training your current policy against a distribution of past policies. Which approach works best?
  • Inspect the learned behavior: does the agent learn to competently pursue food and avoid other snakes? Does the agent learn to attack, trap, or gang up against the competing snakes? Tweet us videos of the learned policies!

⭐⭐⭐  Parameter Averaging in Distributed RL.  Explore the effect of parameter averaging schemes on  sample complexity ⁠ (opens in a new window)  and amount of communication in RL algorithms. While the simplest solution is to average the gradients from every worker on every update, you can  save ⁠ (opens in a new window)  on communication bandwidth by independently updating workers and then infrequently averaging parameters. In RL, this may have another benefit: at any given time we’ll have agents with different parameters, which could lead to better exploration behavior. Another possibility is use algorithms like  EASGD ⁠ (opens in a new window)  that bring parameters partly together each update.

⭐⭐⭐  Transfer Learning Between Different Games via Generative Models.  Proceed as follows:

  • Train 11 good policies for 11  Atari ⁠ (opens in a new window)  games. Generate 10,000 trajectories of 1,000 steps each from the policy for each game.
  • Fit a generative model (such as the  Transformer ⁠ (opens in a new window) ) to the trajectories produced by 10 of the games.
  • Then fine-tune that model on the 11th game.
  • Your goal is to quantify the benefit from pre-training on the 10 games. How large does the model need to be for the pre-training to be useful? How does the size of the effect change when the amount of data from the 11th game is reduced by 10x? By 100x?

⭐⭐⭐  Transformers with Linear Attention.  The  Transformer ⁠ (opens in a new window)  model uses soft attention with softmax. If we could instead use linear attention (which can be converted into an RNN that uses  fast weights ⁠ (opens in a new window) ), we could use the resulting model for RL. Specifically, an RL rollout with a transformer over a huge context would be impractical, but running an RNN with fast weights would be very feasible. Your goal: take any language modeling task; train a transformer; then find a way to get the same bits per character/word using a linear-attention transformer with different hyperparameters, without increasing the total number of parameters by much. Only one caveat: this may turn out to be impossible. But one potentially helpful hint: it is likely that transformers with linear attention require much higher dimensional key/value vectors compared to attention that uses the softmax, which can be done without significantly increasing the number of parameters.

⭐⭐⭐  Learned Data Augmentation.  You could use a learned  VAE ⁠ (opens in a new window)  of data, to perform “learned data augmentation”. One would first train a VAE on input data, then each training point would be transformed by encoding to a latent space, then applying a simple (e.g. Gaussian) perturbation in latent space, then decoding back to observed space. Could we use such an approach to obtain improved generalization? A potential benefit of such data augmentation is that it could include many nonlinear transformations like viewpoint changes and changes in scene lightning. Can we approximate the set of transformations to which the label is invariant? Check out the  existing ⁠ (opens in a new window)   work ⁠ (opens in a new window)   on ⁠ (opens in a new window)   this ⁠ (opens in a new window)   topic ⁠ (opens in a new window)   if ⁠ (opens in a new window)   you ⁠ (opens in a new window)   want ⁠ (opens in a new window)  a place to get started.

⭐⭐⭐⭐  Regularization in Reinforcement Learning.  Experimentally investigate (and qualitatively explain) the effect of different regularization methods on an RL algorithm of choice. In supervised deep learning, regularization is extremely important for  improving optimization ⁠ (opens in a new window)  and for preventing overfitting, with very successful methods like  dropout ⁠ (opens in a new window) ,  batch normalization ⁠ (opens in a new window) , and  L2 regularization ⁠ (opens in a new window) . However, people haven’t benefited from regularization with reinforcement learning algorithms such as  policy gradients ⁠ (opens in a new window)  and  Q-learning ⁠ (opens in a new window) . Incidentally, people generally use much smaller models in RL than in supervised learning, as large models perform worse — perhaps because they overfit to recent experience. To get started,  here ⁠ (opens in a new window)  is a relevant but older theoretical study.

⭐⭐⭐⭐⭐  Automated Solutions of Olympiad Inequality Problems.  Olympiad inequality problems are simple to express, but  solving ⁠ (opens in a new window)  them often requires clever manipulations. Build a dataset of olympiad inequality problems and write a program that can solve a large fraction of them. It’s not clear whether machine learning will be useful here, but you could potentially use a learned policy to reduce the branching factor.

Want to work on problems like these professionally?  Apply ⁠  to OpenAI!

  • Language
  • Learning Paradigms

Authors

Related articles

Frontier Risk And Preparedness

Safety Oct 26, 2023

Red Teaming Network

Safety Sep 19, 2023

Confidence-Building Measures for Artificial Intelligence Workshop proceedings

Conclusion Aug 1, 2023

Besoin d'un workflow n8n ou d'aide pour l'installer ?

Après la veille, passez à l'action : trouvez un template n8n ou un créateur capable de l'adapter à vos outils.

Source

OpenAI News - openai.com

Voir la publication originale