Abstract:
In this paper, we study transformers for text-based games. As a promising replacement of recurrent modules in Natural Language Processing (NLP) tasks, the transformer arc...Show MoreMetadata
Abstract:
In this paper, we study transformers for text-based games. As a promising replacement of recurrent modules in Natural Language Processing (NLP) tasks, the transformer architecture could be treated as a powerful state representation generator for reinforcement learning. However, the vanilla transformer is neither effective nor efficient to learn with a huge amount of weight parameters. Unlike existing research that encodes states using LSTMs or GRUs, we develop a novel lightweight transformer-based representation generator featured with reordered layer normalization, weight sharing and block-wise aggregation. The experimental results show that our proposed model not only solves single games with much fewer interactions, but also achieves better generalization on a set of unseen games. Furthermore, our model outperforms state-of-the-art agents in a variety of man-made games.
Published in: 2020 IEEE Conference on Games (CoG)
Date of Conference: 24-27 August 2020
Date Added to IEEE Xplore: 20 October 2020
ISBN Information:
ISSN Information:
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Deep Reinforcement Learning ,
- Natural Language ,
- Normalization Layer ,
- State Representation ,
- Variety Of Agents ,
- Fewer Interactions ,
- Natural Language Processing Tasks ,
- Transformer Architecture ,
- Game Setting ,
- Single Game ,
- Convolutional Neural Network ,
- Adam Optimizer ,
- Attention Mechanism ,
- Multilayer Perceptron ,
- Feed-forward Network ,
- Language Model ,
- Word Embedding ,
- Residual Connection ,
- Identity Mapping ,
- Output Of Block ,
- Transformer Block ,
- Reinforcement Learning Task ,
- Pre-trained Language Models ,
- Gate Layer ,
- Multiple Games ,
- Partial Observation ,
- Number Of Heads ,
- Input Stream ,
- Sum Of Rewards
- Author Keywords
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Deep Reinforcement Learning ,
- Natural Language ,
- Normalization Layer ,
- State Representation ,
- Variety Of Agents ,
- Fewer Interactions ,
- Natural Language Processing Tasks ,
- Transformer Architecture ,
- Game Setting ,
- Single Game ,
- Convolutional Neural Network ,
- Adam Optimizer ,
- Attention Mechanism ,
- Multilayer Perceptron ,
- Feed-forward Network ,
- Language Model ,
- Word Embedding ,
- Residual Connection ,
- Identity Mapping ,
- Output Of Block ,
- Transformer Block ,
- Reinforcement Learning Task ,
- Pre-trained Language Models ,
- Gate Layer ,
- Multiple Games ,
- Partial Observation ,
- Number Of Heads ,
- Input Stream ,
- Sum Of Rewards
- Author Keywords