Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Gregor Betz
commited on
Commit
•
0f51a5c
1
Parent(s):
9128efb
description
Browse files- src/display/about.py +8 -6
src/display/about.py
CHANGED
@@ -61,12 +61,7 @@ Unlike these leaderboards, the `/\/` Open CoT Leaderboard assess a model's abili
|
|
61 |
|
62 |
d
|
63 |
|
64 |
-
|
65 |
-
|:---|:---|
|
66 |
-
| Can `model` solve `task`? | Can `model` do CoT to improve in `task`? |
|
67 |
-
| Measures `task` performance. | Measures ability to reason (about `task`). |
|
68 |
-
| Metric: absolute accuracy. | Metric: relative accuracy gain. |
|
69 |
-
| Covers broad spectrum of `tasks`. | Focuses on critical thinking `tasks`. |
|
70 |
|
71 |
|
72 |
## Test dataset selection (`tasks`)
|
@@ -114,6 +109,13 @@ We're populating the Open CoT Leaderboard step by step. The idea is to grow a di
|
|
114 |
|
115 |
"""
|
116 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
117 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
118 |
CITATION_BUTTON_TEXT = r"""
|
119 |
Logikon AI Team. (2024). Open CoT Leaderboard. Retrieved from https://huggingface.co/spaces/logikon/open_cot_leaderboard
|
|
|
61 |
|
62 |
d
|
63 |
|
64 |
+
|
|
|
|
|
|
|
|
|
|
|
65 |
|
66 |
|
67 |
## Test dataset selection (`tasks`)
|
|
|
109 |
|
110 |
"""
|
111 |
|
112 |
+
# | 🤗 Open LLM Leaderboard | `/\/` Open CoT Leaderboard |
|
113 |
+
# |:---|:---|
|
114 |
+
# | Can `model` solve `task`? | Can `model` do CoT to improve in `task`? |
|
115 |
+
# | Measures `task` performance. | Measures ability to reason (about `task`). |
|
116 |
+
# | Metric: absolute accuracy. | Metric: relative accuracy gain. |
|
117 |
+
# | Covers broad spectrum of `tasks`. | Focuses on critical thinking `tasks`. |
|
118 |
+
|
119 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
120 |
CITATION_BUTTON_TEXT = r"""
|
121 |
Logikon AI Team. (2024). Open CoT Leaderboard. Retrieved from https://huggingface.co/spaces/logikon/open_cot_leaderboard
|