Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
refactor: rename the variables
Browse files- app.py +16 -16
- src/loaders.py +8 -8
- src/models.py +2 -2
app.py
CHANGED
@@ -114,7 +114,7 @@ def update_datastore(version):
|
|
114 |
selected_domains = get_domain_dropdown(QABenchmarks[datastore.slug])
|
115 |
selected_langs = get_language_dropdown(QABenchmarks[datastore.slug])
|
116 |
selected_rerankings = get_reranking_dropdown(datastore.reranking_models)
|
117 |
-
leaderboard_table = get_leaderboard_table(datastore.
|
118 |
hidden_leaderboard_table = \
|
119 |
get_leaderboard_table(datastore.raw_df_qa, datastore.types_qa, visible=False)
|
120 |
return selected_domains, selected_langs, selected_rerankings, leaderboard_table, hidden_leaderboard_table
|
@@ -127,7 +127,7 @@ def update_datastore_long_doc(version):
|
|
127 |
selected_domains = get_domain_dropdown(LongDocBenchmarks[datastore.slug])
|
128 |
selected_langs = get_language_dropdown(LongDocBenchmarks[datastore.slug])
|
129 |
selected_rerankings = get_reranking_dropdown(datastore.reranking_models)
|
130 |
-
leaderboard_table = get_leaderboard_table(datastore.
|
131 |
hidden_leaderboard_table = \
|
132 |
get_leaderboard_table(datastore.raw_df_long_doc, datastore.types_long_doc, visible=False)
|
133 |
return selected_domains, selected_langs, selected_rerankings, leaderboard_table, hidden_leaderboard_table
|
@@ -170,7 +170,7 @@ with demo:
|
|
170 |
with gr.Column():
|
171 |
selected_rerankings = get_reranking_dropdown(datastore.reranking_models)
|
172 |
# shown_table
|
173 |
-
lb_table = get_leaderboard_table(datastore.
|
174 |
# Dummy leaderboard for handling the case when the user uses backspace key
|
175 |
hidden_lb_table = get_leaderboard_table(datastore.raw_df_qa, datastore.types_qa, visible=False)
|
176 |
|
@@ -218,9 +218,9 @@ with demo:
|
|
218 |
with gr.Column(scale=1):
|
219 |
selected_noreranker = get_noreranking_dropdown()
|
220 |
|
221 |
-
lb_df_retriever = datastore.
|
222 |
-
datastore.
|
223 |
-
|
224 |
lb_df_retriever = reset_rank(lb_df_retriever)
|
225 |
lb_table_retriever = get_leaderboard_table(lb_df_retriever, datastore.types_qa)
|
226 |
|
@@ -276,9 +276,9 @@ with demo:
|
|
276 |
queue=True,
|
277 |
)
|
278 |
with gr.TabItem("Reranking Only", id=12):
|
279 |
-
lb_df_reranker = datastore.
|
280 |
-
datastore.
|
281 |
-
|
282 |
lb_df_reranker = reset_rank(lb_df_reranker)
|
283 |
reranking_models_reranker = (
|
284 |
lb_df_reranker[COL_NAME_RERANKING_MODEL].apply(remove_html).unique().tolist()
|
@@ -365,7 +365,7 @@ with demo:
|
|
365 |
selected_rerankings = get_reranking_dropdown(datastore.reranking_models)
|
366 |
|
367 |
lb_table_long_doc = get_leaderboard_table(
|
368 |
-
datastore.
|
369 |
)
|
370 |
|
371 |
# Dummy leaderboard for handling the case when the user uses backspace key
|
@@ -421,9 +421,9 @@ with demo:
|
|
421 |
search_bar_retriever = get_search_bar()
|
422 |
with gr.Column(scale=1):
|
423 |
selected_noreranker = get_noreranking_dropdown()
|
424 |
-
lb_df_retriever_long_doc = datastore.
|
425 |
-
datastore.
|
426 |
-
|
427 |
lb_df_retriever_long_doc = reset_rank(lb_df_retriever_long_doc)
|
428 |
lb_table_retriever_long_doc = get_leaderboard_table(
|
429 |
lb_df_retriever_long_doc, datastore.types_long_doc
|
@@ -479,9 +479,9 @@ with demo:
|
|
479 |
queue=True,
|
480 |
)
|
481 |
with gr.TabItem("Reranking Only", id=22):
|
482 |
-
lb_df_reranker_ldoc = datastore.
|
483 |
-
datastore.
|
484 |
-
|
485 |
lb_df_reranker_ldoc = reset_rank(lb_df_reranker_ldoc)
|
486 |
reranking_models_reranker_ldoc = (
|
487 |
lb_df_reranker_ldoc[COL_NAME_RERANKING_MODEL].apply(remove_html).unique().tolist()
|
|
|
114 |
selected_domains = get_domain_dropdown(QABenchmarks[datastore.slug])
|
115 |
selected_langs = get_language_dropdown(QABenchmarks[datastore.slug])
|
116 |
selected_rerankings = get_reranking_dropdown(datastore.reranking_models)
|
117 |
+
leaderboard_table = get_leaderboard_table(datastore.formatted_df_qa, datastore.types_qa)
|
118 |
hidden_leaderboard_table = \
|
119 |
get_leaderboard_table(datastore.raw_df_qa, datastore.types_qa, visible=False)
|
120 |
return selected_domains, selected_langs, selected_rerankings, leaderboard_table, hidden_leaderboard_table
|
|
|
127 |
selected_domains = get_domain_dropdown(LongDocBenchmarks[datastore.slug])
|
128 |
selected_langs = get_language_dropdown(LongDocBenchmarks[datastore.slug])
|
129 |
selected_rerankings = get_reranking_dropdown(datastore.reranking_models)
|
130 |
+
leaderboard_table = get_leaderboard_table(datastore.formatted_df_ldoc, datastore.types_long_doc)
|
131 |
hidden_leaderboard_table = \
|
132 |
get_leaderboard_table(datastore.raw_df_long_doc, datastore.types_long_doc, visible=False)
|
133 |
return selected_domains, selected_langs, selected_rerankings, leaderboard_table, hidden_leaderboard_table
|
|
|
170 |
with gr.Column():
|
171 |
selected_rerankings = get_reranking_dropdown(datastore.reranking_models)
|
172 |
# shown_table
|
173 |
+
lb_table = get_leaderboard_table(datastore.formatted_df_qa, datastore.types_qa)
|
174 |
# Dummy leaderboard for handling the case when the user uses backspace key
|
175 |
hidden_lb_table = get_leaderboard_table(datastore.raw_df_qa, datastore.types_qa, visible=False)
|
176 |
|
|
|
218 |
with gr.Column(scale=1):
|
219 |
selected_noreranker = get_noreranking_dropdown()
|
220 |
|
221 |
+
lb_df_retriever = datastore.formatted_df_qa[
|
222 |
+
datastore.formatted_df_qa[COL_NAME_RERANKING_MODEL] == "NoReranker"
|
223 |
+
]
|
224 |
lb_df_retriever = reset_rank(lb_df_retriever)
|
225 |
lb_table_retriever = get_leaderboard_table(lb_df_retriever, datastore.types_qa)
|
226 |
|
|
|
276 |
queue=True,
|
277 |
)
|
278 |
with gr.TabItem("Reranking Only", id=12):
|
279 |
+
lb_df_reranker = datastore.formatted_df_qa[
|
280 |
+
datastore.formatted_df_qa[COL_NAME_RETRIEVAL_MODEL] == BM25_LINK
|
281 |
+
]
|
282 |
lb_df_reranker = reset_rank(lb_df_reranker)
|
283 |
reranking_models_reranker = (
|
284 |
lb_df_reranker[COL_NAME_RERANKING_MODEL].apply(remove_html).unique().tolist()
|
|
|
365 |
selected_rerankings = get_reranking_dropdown(datastore.reranking_models)
|
366 |
|
367 |
lb_table_long_doc = get_leaderboard_table(
|
368 |
+
datastore.formatted_df_ldoc, datastore.types_long_doc
|
369 |
)
|
370 |
|
371 |
# Dummy leaderboard for handling the case when the user uses backspace key
|
|
|
421 |
search_bar_retriever = get_search_bar()
|
422 |
with gr.Column(scale=1):
|
423 |
selected_noreranker = get_noreranking_dropdown()
|
424 |
+
lb_df_retriever_long_doc = datastore.formatted_df_ldoc[
|
425 |
+
datastore.formatted_df_ldoc[COL_NAME_RERANKING_MODEL] == "NoReranker"
|
426 |
+
]
|
427 |
lb_df_retriever_long_doc = reset_rank(lb_df_retriever_long_doc)
|
428 |
lb_table_retriever_long_doc = get_leaderboard_table(
|
429 |
lb_df_retriever_long_doc, datastore.types_long_doc
|
|
|
479 |
queue=True,
|
480 |
)
|
481 |
with gr.TabItem("Reranking Only", id=22):
|
482 |
+
lb_df_reranker_ldoc = datastore.formatted_df_ldoc[
|
483 |
+
datastore.formatted_df_ldoc[COL_NAME_RETRIEVAL_MODEL] == BM25_LINK
|
484 |
+
]
|
485 |
lb_df_reranker_ldoc = reset_rank(lb_df_reranker_ldoc)
|
486 |
reranking_models_reranker_ldoc = (
|
487 |
lb_df_reranker_ldoc[COL_NAME_RERANKING_MODEL].apply(remove_html).unique().tolist()
|
src/loaders.py
CHANGED
@@ -70,23 +70,23 @@ def load_leaderboard_datastore(file_path, version) -> LeaderboardDataStore:
|
|
70 |
|
71 |
lb_data_store.raw_df_qa = get_leaderboard_df(lb_data_store, task="qa", metric=DEFAULT_METRIC_QA)
|
72 |
print(f"QA data loaded: {lb_data_store.raw_df_qa.shape}")
|
73 |
-
lb_data_store.
|
74 |
shown_columns_qa, types_qa = get_default_cols("qa", lb_data_store.slug, add_fix_cols=True)
|
75 |
lb_data_store.types_qa = types_qa
|
76 |
-
lb_data_store.
|
77 |
-
~lb_data_store.
|
78 |
][shown_columns_qa]
|
79 |
-
lb_data_store.
|
80 |
|
81 |
lb_data_store.raw_df_long_doc = get_leaderboard_df(lb_data_store, task="long-doc", metric=DEFAULT_METRIC_LONG_DOC)
|
82 |
print(f"Long-Doc data loaded: {len(lb_data_store.raw_df_long_doc)}")
|
83 |
-
lb_data_store.
|
84 |
shown_columns_long_doc, types_long_doc = get_default_cols("long-doc", lb_data_store.slug, add_fix_cols=True)
|
85 |
lb_data_store.types_long_doc = types_long_doc
|
86 |
-
lb_data_store.
|
87 |
-
~lb_data_store.
|
88 |
][shown_columns_long_doc]
|
89 |
-
lb_data_store.
|
90 |
|
91 |
lb_data_store.reranking_models = sorted(
|
92 |
list(frozenset([eval_result.reranking_model for eval_result in lb_data_store.raw_data]))
|
|
|
70 |
|
71 |
lb_data_store.raw_df_qa = get_leaderboard_df(lb_data_store, task="qa", metric=DEFAULT_METRIC_QA)
|
72 |
print(f"QA data loaded: {lb_data_store.raw_df_qa.shape}")
|
73 |
+
lb_data_store.formatted_df_qa = lb_data_store.raw_df_qa.copy()
|
74 |
shown_columns_qa, types_qa = get_default_cols("qa", lb_data_store.slug, add_fix_cols=True)
|
75 |
lb_data_store.types_qa = types_qa
|
76 |
+
lb_data_store.formatted_df_qa = lb_data_store.formatted_df_qa[
|
77 |
+
~lb_data_store.formatted_df_qa[COL_NAME_IS_ANONYMOUS]
|
78 |
][shown_columns_qa]
|
79 |
+
lb_data_store.formatted_df_qa.drop([COL_NAME_REVISION, COL_NAME_TIMESTAMP], axis=1, inplace=True)
|
80 |
|
81 |
lb_data_store.raw_df_long_doc = get_leaderboard_df(lb_data_store, task="long-doc", metric=DEFAULT_METRIC_LONG_DOC)
|
82 |
print(f"Long-Doc data loaded: {len(lb_data_store.raw_df_long_doc)}")
|
83 |
+
lb_data_store.formatted_df_ldoc = lb_data_store.raw_df_long_doc.copy()
|
84 |
shown_columns_long_doc, types_long_doc = get_default_cols("long-doc", lb_data_store.slug, add_fix_cols=True)
|
85 |
lb_data_store.types_long_doc = types_long_doc
|
86 |
+
lb_data_store.formatted_df_ldoc = lb_data_store.formatted_df_ldoc[
|
87 |
+
~lb_data_store.formatted_df_ldoc[COL_NAME_IS_ANONYMOUS]
|
88 |
][shown_columns_long_doc]
|
89 |
+
lb_data_store.formatted_df_ldoc.drop([COL_NAME_REVISION, COL_NAME_TIMESTAMP], axis=1, inplace=True)
|
90 |
|
91 |
lb_data_store.reranking_models = sorted(
|
92 |
list(frozenset([eval_result.reranking_model for eval_result in lb_data_store.raw_data]))
|
src/models.py
CHANGED
@@ -143,8 +143,8 @@ class LeaderboardDataStore:
|
|
143 |
raw_data: Optional[list]
|
144 |
raw_df_qa: Optional[pd.DataFrame]
|
145 |
raw_df_long_doc: Optional[pd.DataFrame]
|
146 |
-
|
147 |
-
|
148 |
reranking_models: Optional[list]
|
149 |
types_qa: Optional[list]
|
150 |
types_long_doc: Optional[list]
|
|
|
143 |
raw_data: Optional[list]
|
144 |
raw_df_qa: Optional[pd.DataFrame]
|
145 |
raw_df_long_doc: Optional[pd.DataFrame]
|
146 |
+
formatted_df_qa: Optional[pd.DataFrame]
|
147 |
+
formatted_df_ldoc: Optional[pd.DataFrame]
|
148 |
reranking_models: Optional[list]
|
149 |
types_qa: Optional[list]
|
150 |
types_long_doc: Optional[list]
|