LLM ReferenceLLM Reference
HumanEval+activeCoding

HumanEval+

Metric: Pass@1 (higher is better)Introduced: 2023

About

HumanEval+ extends HumanEval with an average of 764 test cases per problem (vs 9.6 in original), greatly increasing evaluation rigor for edge case coverage.