We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 89b4c1d commit 5d9a4c7Copy full SHA for 5d9a4c7
bigcode_eval/tasks/mbppplus.py
@@ -10,6 +10,8 @@
10
Homepage: https://github.com/evalplus/evalplus
11
"""
12
13
+import os
14
+
15
from bigcode_eval.tasks.mbpp import MBPP
16
from bigcode_eval.tasks.custom_metrics.code_eval import compute_code_eval
17
@@ -46,6 +48,9 @@ def get_prompt(self, doc):
46
48
# is different from HumanEval(+) which further requires a `check` func
47
49
def get_reference(self, doc):
50
"""Builds the reference solution for the doc (sample from the test dataset)."""
51
+ use_mbpp_tests = os.getenv("MBBPPLUS_USE_MBPP_TESTS", "0")
52
+ if use_mbpp_tests == "1":
53
+ return "\n".join(doc["test_list"])
54
return "\n" + doc["test"]
55
56
def get_dataset(self):
0 commit comments