-
An End-to-End Deep RL Framework for Task Arrangement in Crowdsourcing Platforms
A Deep Reinforcement Learning framework for task arrangement in crowdsourcing platforms. -
BabyAI-PickUpDist-v0
The dataset used in the paper is the BabyAI environment 'BabyAI-PickUpDist-v0' with a one-pickup-per-episode wrapper. -
SHARING LIFELONG REINFORCEMENT LEARNING KNOWLEDGE VIA MODULATING MASKS
The CT-graph and Minigrid environments are used to evaluate lifelong reinforcement learning approaches. -
Pybullet Multigoal
A dataset for robotic manipulation tasks, including ChestPush, ChestPickAndPlace, and BlockStack. -
Atari 2600 domain
The Atari 2600 domain dataset, used for training and testing reinforcement learning algorithms. -
Real-World RL Challenge
The dataset used in the paper is the Real-World RL Challenge dataset. It contains a set of continuous control tasks. -
DeepMind Control Suite and Real-World RL Experiments
The dataset used in the paper is the DeepMind Control Suite and Real-World RL Experiments. It contains a set of continuous control tasks based on MuJoCo. -
CartPole and Blackjack environments
The dataset used in this paper is the CartPole and Blackjack environments from OpenAI Gym. -
PyBulletGym tasks
The dataset used in the paper is a collection of experiences sampled from a replay buffer, used to train and evaluate the proposed Multi-step DDPG (MDDPG) and Mixed Multi-step... -
Visual CartPole
A visual version of the CartPole environment from OpenAI Gym. -
Contra State Dataset
The dataset used in the paper is a collection of instruction sets and states for the Contra game, used to train a language model and a reinforcement learning policy. -
Contra Instruction Dataset
The dataset used in the paper is a collection of instruction sets and states for the Contra game, used to train a language model and a reinforcement learning policy. -
Contra Dataset
The dataset used in the paper is a collection of instruction sets and states for the Contra game, used to train a language model and a reinforcement learning policy.