Spaces:
Runtime error
Runtime error
Refactor app templates
Browse files
app.py
CHANGED
@@ -49,8 +49,8 @@ TASK_TO_DEFAULT_METRICS = {
|
|
49 |
],
|
50 |
"entity_extraction": ["precision", "recall", "f1", "accuracy"],
|
51 |
"extractive_question_answering": [],
|
52 |
-
"translation": ["sacrebleu"
|
53 |
-
"summarization": ["rouge1", "rouge2", "rougeL", "rougeLsum"
|
54 |
}
|
55 |
|
56 |
SUPPORTED_TASKS = list(TASK_TO_ID.keys())
|
@@ -96,10 +96,11 @@ st.title("Evaluation as a Service")
|
|
96 |
st.markdown(
|
97 |
"""
|
98 |
Welcome to Hugging Face's Evaluation as a Service! This application allows
|
99 |
-
you to evaluate
|
100 |
-
|
101 |
-
|
102 |
-
[
|
|
|
103 |
"""
|
104 |
)
|
105 |
|
@@ -302,6 +303,7 @@ with st.expander("Advanced configuration"):
|
|
302 |
col_mapping[answers_text_col] = "answers.text"
|
303 |
col_mapping[answers_start_col] = "answers.answer_start"
|
304 |
|
|
|
305 |
st.markdown("**Select metrics**")
|
306 |
st.markdown("The following metrics will be computed")
|
307 |
html_string = " ".join(
|
@@ -327,7 +329,6 @@ with st.expander("Advanced configuration"):
|
|
327 |
with st.form(key="form"):
|
328 |
|
329 |
compatible_models = get_compatible_models(selected_task, selected_dataset)
|
330 |
-
|
331 |
selected_models = st.multiselect("Select the models you wish to evaluate", compatible_models)
|
332 |
print("Selected models:", selected_models)
|
333 |
|
@@ -345,7 +346,7 @@ with st.form(key="form"):
|
|
345 |
|
346 |
if submit_button:
|
347 |
if len(selected_models) > 0:
|
348 |
-
project_id = str(uuid.uuid4())
|
349 |
payload = {
|
350 |
"username": AUTOTRAIN_USERNAME,
|
351 |
"proj_name": f"eval-project-{project_id}",
|
|
|
49 |
],
|
50 |
"entity_extraction": ["precision", "recall", "f1", "accuracy"],
|
51 |
"extractive_question_answering": [],
|
52 |
+
"translation": ["sacrebleu"],
|
53 |
+
"summarization": ["rouge1", "rouge2", "rougeL", "rougeLsum"],
|
54 |
}
|
55 |
|
56 |
SUPPORTED_TASKS = list(TASK_TO_ID.keys())
|
|
|
96 |
st.markdown(
|
97 |
"""
|
98 |
Welcome to Hugging Face's Evaluation as a Service! This application allows
|
99 |
+
you to evaluate [Transformers
|
100 |
+
models](https://huggingface.co/models?library=transformers&sort=downloads)
|
101 |
+
with a dataset on the Hub. Please select the dataset and configuration
|
102 |
+
below. The results of your evaluation will be displayed on the [public
|
103 |
+
leaderboard](https://huggingface.co/spaces/autoevaluate/leaderboards).
|
104 |
"""
|
105 |
)
|
106 |
|
|
|
303 |
col_mapping[answers_text_col] = "answers.text"
|
304 |
col_mapping[answers_start_col] = "answers.answer_start"
|
305 |
|
306 |
+
# Select metrics
|
307 |
st.markdown("**Select metrics**")
|
308 |
st.markdown("The following metrics will be computed")
|
309 |
html_string = " ".join(
|
|
|
329 |
with st.form(key="form"):
|
330 |
|
331 |
compatible_models = get_compatible_models(selected_task, selected_dataset)
|
|
|
332 |
selected_models = st.multiselect("Select the models you wish to evaluate", compatible_models)
|
333 |
print("Selected models:", selected_models)
|
334 |
|
|
|
346 |
|
347 |
if submit_button:
|
348 |
if len(selected_models) > 0:
|
349 |
+
project_id = str(uuid.uuid4())[:8]
|
350 |
payload = {
|
351 |
"username": AUTOTRAIN_USERNAME,
|
352 |
"proj_name": f"eval-project-{project_id}",
|