HyperAI초신경
홈
뉴스
최신 연구 논문
튜토리얼
데이터셋
백과사전
SOTA
LLM 모델
GPU 랭킹
컨퍼런스
전체 검색
소개
한국어
HyperAI초신경
Toggle sidebar
전체 사이트 검색...
⌘
K
홈
SOTA
Atari Games
Atari Games On Atari 2600 Solaris
Atari Games On Atari 2600 Solaris
평가 지표
Score
평가 결과
이 벤치마크에서 각 모델의 성능 결과
Columns
모델 이름
Score
Paper Title
Repository
CGP
8324
Evolving simple programs for playing Atari games
R2D2
3787.2
Recurrent Experience Replay in Distributed Reinforcement Learning
-
MuZero (Res2 Adam)
5132.95
Online and Offline Reinforcement Learning by Planning with a Learned Model
DNA
2225
DNA: Proximal Policy Optimization with a Dual Network Architecture
NoisyNet-Dueling
6522
Noisy Networks for Exploration
SND-VIC
11865
Self-supervised network distillation: an effective approach to exploration in sparse reward environments
GDI-H3
9105
Generalized Data Distribution Iteration
-
MuZero
56.62
Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model
IMPALA (deep)
2365.00
IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures
GDI-I3
11074
Generalized Data Distribution Iteration
-
GDI-I3
11074
GDI: Rethinking What Makes Reinforcement Learning Different From Supervised Learning
-
RND
3282
Exploration by Random Network Distillation
Advantage Learning
4785.16
Increasing the Action Gap: New Operators for Reinforcement Learning
SND-STD
12460
Self-supervised network distillation: an effective approach to exploration in sparse reward environments
Ape-X
2892.9
Distributed Prioritized Experience Replay
IQN
8007
Implicit Quantile Networks for Distributional Reinforcement Learning
ASL DDQN
3506.8
Train a Real-world Local Path Planner in One Hour via Partially Decoupled Reinforcement Learning and Vectorized Diversity
-
DreamerV2
922
Mastering Atari with Discrete World Models
SND-V
11582
Self-supervised network distillation: an effective approach to exploration in sparse reward environments
Go-Explore
19671
First return, then explore
0 of 23 row(s) selected.
Previous
Next