HyperAI超神経

Atari Games On Atari 2600 Battle Zone

評価指標

Score

評価結果

このベンチマークにおける各モデルのパフォーマンス結果

比較表
モデル名Score
soft-actor-critic-for-discrete-action4386.7
policy-optimization-with-penalized-point15466.67
deep-exploration-via-bootstrapped-dqn38666.7
mastering-atari-go-chess-and-shogi-by848623.00
dueling-network-architectures-for-deep31320.0
impala-scalable-distributed-deep-rl-with20885.00
a-distributional-perspective-on-reinforcement28742.0
asynchronous-methods-for-deep-reinforcement12950.0
prioritized-experience-replay25520.0
recurrent-experience-replay-in-distributed751880.0
dueling-network-architectures-for-deep37150.0
learning-values-across-many-orders-of8220.0
distributed-prioritized-experience-replay98895
noisy-networks-for-exploration52262
agent57-outperforming-the-atari-human934134.88
dna-proximal-policy-optimization-with-a-dual71003
mastering-atari-with-discrete-world-models-140325
evolving-simple-programs-for-playing-atari34200
fully-parameterized-quantile-function-for87928.6
implicit-quantile-networks-for-distributional42244
train-a-real-world-local-path-planner-in-one38986
prioritized-experience-replay31530.0
evolution-strategies-as-a-scalable16600.0
deep-reinforcement-learning-with-double-q30650.0
asynchronous-methods-for-deep-reinforcement11340.0
human-level-control-through-deep26300.0
the-reactor-a-fast-and-sample-efficient-actor64070.0
massively-parallel-methods-for-deep19938.0
generalized-data-distribution-iteration478830
increasing-the-action-gap-new-operators-for28789.29
deep-reinforcement-learning-with-double-q29900.0
the-arcade-learning-environment-an-evaluation70333.3
distributional-reinforcement-learning-with-139268
deep-reinforcement-learning-with-double-q23750.0
dueling-network-architectures-for-deep35520.0
generalized-data-distribution-iteration824360
recurrent-rational-networks23403
recurrent-rational-networks25749
self-imitation-learning25075
curl-contrastive-unsupervised-representations11208
dueling-network-architectures-for-deep31700.0
online-and-offline-reinforcement-learning-by178716.9
asynchronous-methods-for-deep-reinforcement20760.0
deep-reinforcement-learning-with-double-q24740.0
モデル 4516.2
increasing-the-action-gap-new-operators-for34583.07
the-arcade-learning-environment-an-evaluation15819.7