You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for developing this fantastic dataset for code generation!
I had a quick question about the evaluation script (compute_metric.py). In my experience, it takes about 17 hours complete, even with the provided example generation. When I tried switching to parallel evaluation, the process finished much faster (in seconds), but all the results returned as zero. During evaluation, I used the original code an generated solutions provided in this repo.
Is this expected behavior, or am I missing something? I would appreciate any guidance you could provide.
The text was updated successfully, but these errors were encountered:
Hi,
Thank you for developing this fantastic dataset for code generation!
I had a quick question about the evaluation script (
compute_metric.py
). In my experience, it takes about 17 hours complete, even with the provided example generation. When I tried switching to parallel evaluation, the process finished much faster (in seconds), but all the results returned as zero. During evaluation, I used the original code an generated solutions provided in this repo.Is this expected behavior, or am I missing something? I would appreciate any guidance you could provide.
The text was updated successfully, but these errors were encountered: