HyperAI超神経

Atari Games On Atari 2600 River Raid

評価指標

Score

評価結果

このベンチマークにおける各モデルのパフォーマンス結果

比較表
モデル名Score
asynchronous-methods-for-deep-reinforcement6591.9
generalized-data-distribution-iteration28075
dueling-network-architectures-for-deep21162.6
massively-parallel-methods-for-deep5310.3
human-level-control-through-deep8316.0
mastering-atari-go-chess-and-shogi-by323417.18
a-distributional-perspective-on-reinforcement17322.0
learning-values-across-many-orders-of12530.8
deep-reinforcement-learning-with-double-q4748.5
the-arcade-learning-environment-an-evaluation4449
dueling-network-architectures-for-deep14884.5
dueling-network-architectures-for-deep20607.6
dna-proximal-policy-optimization-with-a-dual16789
asynchronous-methods-for-deep-reinforcement12201.8
recurrent-experience-replay-in-distributed45632.1
asynchronous-methods-for-deep-reinforcement10001.2
deep-reinforcement-learning-with-double-q10838.4
fully-parameterized-quantile-function-for23560.7
generalized-data-distribution-iteration28349
evolution-strategies-as-a-scalable5009.0
agent57-outperforming-the-atari-human63318.67
モデル 222650.0
implicit-quantile-networks-for-distributional17765
policy-optimization-with-penalized-point8052.23
deep-exploration-via-bootstrapped-dqn12845
mastering-atari-with-discrete-world-models-116351
online-and-offline-reinforcement-learning-by171673.78
model-free-episodic-control-with-state3868
train-a-real-world-local-path-planner-in-one24445
deep-reinforcement-learning-with-double-q7377.6
self-imitation-learning14306.1
prioritized-experience-replay11807.2
impala-scalable-distributed-deep-rl-with29608.05
increasing-the-action-gap-new-operators-for10585.12
distributed-prioritized-experience-replay63864.4
deep-reinforcement-learning-with-double-q16496.8
dueling-network-architectures-for-deep16569.4
prioritized-experience-replay14522.3
evolving-simple-programs-for-playing-atari2914
the-arcade-learning-environment-an-evaluation1904.3
distributional-reinforcement-learning-with-117571