HyperAIHyperAI초신경
홈뉴스연구 논문튜토리얼데이터셋백과사전SOTALLM 모델GPU 랭킹컨퍼런스
전체 검색
소개
한국어
HyperAIHyperAI초신경
  1. 홈
  2. SOTA
  3. 질문 응답
  4. Question Answering On Strategyqa

Question Answering On Strategyqa

평가 지표

Accuracy

평가 결과

이 벤치마크에서 각 모델의 성능 결과

모델 이름
Accuracy
Paper TitleRepository
Rethinking with retrieval (GPT-3)77.73Rethinking with Retrieval: Faithful Large Language Model Inference
SearchChain-Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models
PaLM 540B76.4Transcending Scaling Laws with 0.1% Extra Compute-
Least-to-Most-Least-to-Most Prompting Enables Complex Reasoning in Large Language Models
Self-Evaluation Guided Decoding (Codex, CoT, single reasoning chain, 6-shot gen, 4-shot eval)77.2--
SearchChain-Search-in-the-Chain: Interactively Enhancing Large Language Models with Search for Knowledge-intensive Tasks
CoA w/o actions-Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models
CoA-Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models
Least-to-Most-Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models
U-PaLM 540B76.6Transcending Scaling Laws with 0.1% Extra Compute-
Minerva 540B61.9Transcending Scaling Laws with 0.1% Extra Compute-
PaLM 2 (few-shot, CoT, SC)90.4PaLM 2 Technical Report
0 of 12 row(s) selected.
HyperAI

학습, 이해, 실천, 커뮤니티와 함께 인공지능의 미래를 구축하다

한국어

소개

회사 소개데이터셋 도움말

제품

뉴스튜토리얼데이터셋백과사전

링크

TVM 한국어Apache TVMOpenBayes

© HyperAI초신경

TwitterBilibili