clefourrier HF staff multimodalart HF staff commited on
Commit
d2179b0
1 Parent(s): ed1fdef

Add a model size filter ✨ (#218)

Browse files

- Add a model size filter ✨ (64f1a6e6406aec4f07846a78fe8528b3c1d71c4c)
- New style 😎 (31944c88cfa84efcfedcd257f460932059d08ca6)


Co-authored-by: Apolinário from multimodal AI art <[email protected]>

Files changed (2) hide show
  1. app.py +72 -16
  2. src/assets/css_html_js.py +33 -1
app.py CHANGED
@@ -294,7 +294,30 @@ def filter_items(df, leaderboard_table, query):
294
  if AutoEvalColumn.model_type_symbol.name in leaderboard_table.columns:
295
  filtered_df = df[(df[AutoEvalColumn.model_type_symbol.name] == query)]
296
  else:
297
- return leaderboard_table.columns
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
298
  return filtered_df[leaderboard_table.columns]
299
 
300
  def change_tab(query_param):
@@ -310,6 +333,10 @@ def change_tab(query_param):
310
  else:
311
  return gr.Tabs.update(selected=0)
312
 
 
 
 
 
313
 
314
  demo = gr.Blocks(css=custom_css)
315
  with demo:
@@ -332,18 +359,44 @@ with demo:
332
  show_label=False,
333
  elem_id="search-bar",
334
  )
335
- filter_columns = gr.Radio(
336
- label="⏚ Filter model types",
337
- choices = [
338
- "all",
339
- ModelType.PT.to_str(),
340
- ModelType.FT.to_str(),
341
- ModelType.IFT.to_str(),
342
- ModelType.RL.to_str(),
343
- ],
344
- value="all",
345
- elem_id="filter-columns"
346
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
347
  leaderboard_table = gr.components.Dataframe(
348
  value=leaderboard_df[[AutoEvalColumn.model_type_symbol.name, AutoEvalColumn.model.name] + shown_columns.value + [AutoEvalColumn.dummy.name]],
349
  headers=[AutoEvalColumn.model_type_symbol.name, AutoEvalColumn.model.name] + shown_columns.value + [AutoEvalColumn.dummy.name],
@@ -367,8 +420,11 @@ with demo:
367
  [hidden_leaderboard_table_for_search, leaderboard_table, search_bar],
368
  leaderboard_table,
369
  )
370
- shown_columns.change(select_columns, [hidden_leaderboard_table_for_search, shown_columns], leaderboard_table)
371
- filter_columns.change(filter_items, [hidden_leaderboard_table_for_search, leaderboard_table, filter_columns], leaderboard_table)
 
 
 
372
  with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
373
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
374
 
@@ -495,4 +551,4 @@ with demo:
495
  scheduler = BackgroundScheduler()
496
  scheduler.add_job(restart_space, "interval", seconds=3600)
497
  scheduler.start()
498
- demo.queue(concurrency_count=40).launch()
 
294
  if AutoEvalColumn.model_type_symbol.name in leaderboard_table.columns:
295
  filtered_df = df[(df[AutoEvalColumn.model_type_symbol.name] == query)]
296
  else:
297
+ return filtered_df[leaderboard_table.columns]
298
+ return filtered_df[leaderboard_table.columns]
299
+
300
+ def filter_items_size(df, leaderboard_table, query):
301
+ numeric_intervals = {
302
+ "all": None,
303
+ "< 1B": (0, 1),
304
+ "~3B": (1, 5),
305
+ "~7B": (6, 11),
306
+ "~13B": (12, 15),
307
+ "~35B": (16, 55),
308
+ "60B+": (55, 1000)
309
+ }
310
+
311
+ if query == "all":
312
+ return df[leaderboard_table.columns]
313
+
314
+ numeric_interval = numeric_intervals[query]
315
+
316
+ if AutoEvalColumn.params.name in leaderboard_table.columns:
317
+ params_column = pd.to_numeric(df[AutoEvalColumn.params.name], errors='coerce')
318
+ filtered_df = df[params_column.between(*numeric_interval)]
319
+ else:
320
+ return filtered_df[leaderboard_table.columns]
321
  return filtered_df[leaderboard_table.columns]
322
 
323
  def change_tab(query_param):
 
333
  else:
334
  return gr.Tabs.update(selected=0)
335
 
336
+ def update_filter_type(input_type, shown_columns):
337
+ shown_columns.append(AutoEvalColumn.params.name)
338
+ return gr.update(visible=(input_type == 'types')), gr.update(visible=(input_type == 'sizes')), shown_columns
339
+
340
 
341
  demo = gr.Blocks(css=custom_css)
342
  with demo:
 
359
  show_label=False,
360
  elem_id="search-bar",
361
  )
362
+ with gr.Box(elem_id="box-filter"):
363
+ filter_type = gr.Dropdown(
364
+ label="⏚ Filter model",
365
+ choices=["types", "sizes"], value="types",
366
+ interactive=True,
367
+ elem_id="filter_type"
368
+ )
369
+ filter_columns = gr.Radio(
370
+ label="⏚ Filter model types",
371
+ show_label=False,
372
+ choices = [
373
+ "all",
374
+ ModelType.PT.to_str(),
375
+ ModelType.FT.to_str(),
376
+ ModelType.IFT.to_str(),
377
+ ModelType.RL.to_str(),
378
+ ],
379
+ value="all",
380
+ elem_id="filter-columns"
381
+ )
382
+ filter_columns_size = gr.Radio(
383
+ label="⏚ Filter model sizes",
384
+ show_label=False,
385
+ choices = [
386
+ "all",
387
+ "< 1B",
388
+ "~3B",
389
+ "~7B",
390
+ "~13B",
391
+ "~35B",
392
+ "60B+"
393
+ ],
394
+ value="all",
395
+ visible=False,
396
+ interactive=True,
397
+ elem_id="filter-columns-size"
398
+ )
399
+
400
  leaderboard_table = gr.components.Dataframe(
401
  value=leaderboard_df[[AutoEvalColumn.model_type_symbol.name, AutoEvalColumn.model.name] + shown_columns.value + [AutoEvalColumn.dummy.name]],
402
  headers=[AutoEvalColumn.model_type_symbol.name, AutoEvalColumn.model.name] + shown_columns.value + [AutoEvalColumn.dummy.name],
 
420
  [hidden_leaderboard_table_for_search, leaderboard_table, search_bar],
421
  leaderboard_table,
422
  )
423
+
424
+ filter_type.change(update_filter_type,inputs=[filter_type, shown_columns],outputs=[filter_columns, filter_columns_size, shown_columns],queue=False).then(select_columns, [hidden_leaderboard_table_for_search, shown_columns], leaderboard_table, queue=False)
425
+ shown_columns.change(select_columns, [hidden_leaderboard_table_for_search, shown_columns], leaderboard_table, queue=False)
426
+ filter_columns.change(filter_items, [hidden_leaderboard_table_for_search, leaderboard_table, filter_columns], leaderboard_table, queue=False)
427
+ filter_columns_size.change(filter_items_size, [hidden_leaderboard_table_for_search, leaderboard_table, filter_columns_size], leaderboard_table, queue=False)
428
  with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
429
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
430
 
 
551
  scheduler = BackgroundScheduler()
552
  scheduler.add_job(restart_space, "interval", seconds=3600)
553
  scheduler.start()
554
+ demo.queue(concurrency_count=40).launch()
src/assets/css_html_js.py CHANGED
@@ -68,6 +68,38 @@ table th:first-child {
68
  #scale-logo .download {
69
  display: none;
70
  }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
71
  """
72
 
73
  get_window_url_params = """
@@ -76,4 +108,4 @@ get_window_url_params = """
76
  url_params = Object.fromEntries(params);
77
  return url_params;
78
  }
79
- """
 
68
  #scale-logo .download {
69
  display: none;
70
  }
71
+ #filter_type{
72
+ border: 0;
73
+ padding-left: 0;
74
+ padding-top: 0;
75
+ }
76
+ #filter_type label {
77
+ display: flex;
78
+ }
79
+ #filter_type label > span{
80
+ margin-top: var(--spacing-lg);
81
+ margin-right: 0.5em;
82
+ }
83
+ #filter_type label > .wrap{
84
+ width: 103px;
85
+ }
86
+ #filter_type label > .wrap .wrap-inner{
87
+ padding: 2px;
88
+ }
89
+ #filter_type label > .wrap .wrap-inner input{
90
+ width: 1px
91
+ }
92
+ #filter-columns{
93
+ border:0;
94
+ padding:0;
95
+ }
96
+ #filter-columns-size{
97
+ border:0;
98
+ padding:0;
99
+ }
100
+ #box-filter > .form{
101
+ border: 0
102
+ }
103
  """
104
 
105
  get_window_url_params = """
 
108
  url_params = Object.fromEntries(params);
109
  return url_params;
110
  }
111
+ """