Gregor Betz commited on
Commit
0f51a5c
1 Parent(s): 9128efb

description

Browse files
Files changed (1) hide show
  1. src/display/about.py +8 -6
src/display/about.py CHANGED
@@ -61,12 +61,7 @@ Unlike these leaderboards, the `/\/` Open CoT Leaderboard assess a model's abili
61
 
62
  d
63
 
64
- | 🤗 Open LLM Leaderboard | `/\/` Open CoT Leaderboard |
65
- |:---|:---|
66
- | Can `model` solve `task`? | Can `model` do CoT to improve in `task`? |
67
- | Measures `task` performance. | Measures ability to reason (about `task`). |
68
- | Metric: absolute accuracy. | Metric: relative accuracy gain. |
69
- | Covers broad spectrum of `tasks`. | Focuses on critical thinking `tasks`. |
70
 
71
 
72
  ## Test dataset selection (`tasks`)
@@ -114,6 +109,13 @@ We're populating the Open CoT Leaderboard step by step. The idea is to grow a di
114
 
115
  """
116
 
 
 
 
 
 
 
 
117
  CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
118
  CITATION_BUTTON_TEXT = r"""
119
  Logikon AI Team. (2024). Open CoT Leaderboard. Retrieved from https://huggingface.co/spaces/logikon/open_cot_leaderboard
 
61
 
62
  d
63
 
64
+
 
 
 
 
 
65
 
66
 
67
  ## Test dataset selection (`tasks`)
 
109
 
110
  """
111
 
112
+ # | 🤗 Open LLM Leaderboard | `/\/` Open CoT Leaderboard |
113
+ # |:---|:---|
114
+ # | Can `model` solve `task`? | Can `model` do CoT to improve in `task`? |
115
+ # | Measures `task` performance. | Measures ability to reason (about `task`). |
116
+ # | Metric: absolute accuracy. | Metric: relative accuracy gain. |
117
+ # | Covers broad spectrum of `tasks`. | Focuses on critical thinking `tasks`. |
118
+
119
  CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
120
  CITATION_BUTTON_TEXT = r"""
121
  Logikon AI Team. (2024). Open CoT Leaderboard. Retrieved from https://huggingface.co/spaces/logikon/open_cot_leaderboard