Add description to card metadata

#1
by julien-c HF staff - opened

This metric implements the evaluation harness for the HumanEval problem solving dataset
described in the paper "Evaluating Large Language Models Trained on Code"
(https://arxiv.org/abs/2107.03374).

julien-c changed pull request status to merged

Sign up or log in to comment