Code Generation On Turbulence
평가 지표
CorrSc
평가 결과
이 벤치마크에서 각 모델의 성능 결과
모델 이름 | CorrSc | Paper Title | Repository |
---|---|---|---|
GPT-3.5-Turbo | 0.617 | Turbulence: Systematically and Automatically Testing Instruction-Tuned Large Language Models for Code | - |
CodeLlama:13B-4bit-quantised | 0.327 | Turbulence: Systematically and Automatically Testing Instruction-Tuned Large Language Models for Code | - |
GPT-4 | 0.848 | Turbulence: Systematically and Automatically Testing Instruction-Tuned Large Language Models for Code | - |
Command | 0.063 | Turbulence: Systematically and Automatically Testing Instruction-Tuned Large Language Models for Code | - |
CodeLlama:7B-4bit-quantised | 0.289 | Turbulence: Systematically and Automatically Testing Instruction-Tuned Large Language Models for Code | - |
0 of 5 row(s) selected.