HyperAIHyperAI

Atari Games On Atari 2600 Pitfall

Metrics

Score

Results

Performance results of various models on this benchmark

Model Name
Score
Paper TitleRepository
NoisyNet-Dueling0Noisy Networks for Exploration-
Go-Explore6954First return, then explore-
POP3D0Policy Optimization With Penalized Point Probability Distance: An Alternative To Proximal Policy Optimization-
QR-DQN-10Distributional Reinforcement Learning with Quantile Regression-
IQN0Implicit Quantile Networks for Distributional Reinforcement Learning-
DNA0DNA: Proximal Policy Optimization with a Dual Network Architecture-
Advantage Learning0Increasing the Action Gap: New Operators for Reinforcement Learning-
MuZero (Res2 Adam)0Online and Offline Reinforcement Learning by Planning with a Learned Model-
SND-V0Self-supervised network distillation: an effective approach to exploration in sparse reward environments-
MuZero0.00Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model-
SND-VIC0Self-supervised network distillation: an effective approach to exploration in sparse reward environments-
DreamerV20Mastering Atari with Discrete World Models-
CGP0Evolving simple programs for playing Atari games-
GDI-H3-4.345Generalized Data Distribution Iteration-
Ape-X-0.6Distributed Prioritized Experience Replay-
Go-Explore102571Go-Explore: a New Approach for Hard-Exploration Problems-
ASL DDQN0Train a Real-world Local Path Planner in One Hour via Partially Decoupled Reinforcement Learning and Vectorized Diversity-
IMPALA (deep)-1.66IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures-
R2D20.0Recurrent Experience Replay in Distributed Reinforcement Learning-
RND-3Exploration by Random Network Distillation-
0 of 23 row(s) selected.
Atari Games On Atari 2600 Pitfall | SOTA | HyperAI