HyperAI초신경
홈
뉴스
최신 연구 논문
튜토리얼
데이터셋
백과사전
SOTA
LLM 모델
GPU 랭킹
컨퍼런스
전체 검색
소개
한국어
HyperAI초신경
Toggle sidebar
전체 사이트 검색...
⌘
K
홈
SOTA
Smac 1
Smac On Smac Def Armored Sequential
Smac On Smac Def Armored Sequential
평가 지표
Median Win Rate
평가 결과
이 벤치마크에서 각 모델의 성능 결과
Columns
모델 이름
Median Win Rate
Paper Title
Repository
QTRAN
93.8
QTRAN: Learning to Factorize with Transformation for Cooperative Multi-Agent Reinforcement Learning
VDN
96.9
Value-Decomposition Networks For Cooperative Multi-Agent Learning
QMIX
0.0
QMIX: Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning
MASAC
0.0
Decomposed Soft Actor-Critic Method for Cooperative Multi-Agent Reinforcement Learning
DRIMA
100
Disentangling Sources of Risk for Distributional Multi-Agent Reinforcement Learning
-
IQL
9.4
The StarCraft Multi-Agent Challenges+ : Learning of Multi-Stage Tasks and Environmental Factors without Precise Reward Functions
DDN
71.9
DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning
COMA
0.0
Counterfactual Multi-Agent Policy Gradients
MADDPG
90.6
Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments
DIQL
53.1
DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning
DMIX
81.3
DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning
0 of 11 row(s) selected.
Previous
Next