kaikaidai commited on
Commit
c873398
1 Parent(s): 9fb17ef

Update common.py

Browse files
Files changed (1) hide show
  1. common.py +4 -4
common.py CHANGED
@@ -126,19 +126,19 @@ Judge Arena is specifically designed to assess AI models that function as evalua
126
  <br><br>
127
  # FAQ
128
 
129
- -**Isn't this the same as Chatbot Arena?**
130
 
131
  We are big fans of what the LMSYS team have done with Chatbot Arena and fully credit them for the inspiration to develop this. We were looking for a dynamic leaderboard that graded on AI judge capabilities and didn't manage to find one, so we created Judge Arena. This UI is designed especially for evals; to match the format of the model-based eval prompts that you would use in your LLM evaluation / monitoring tool.
132
 
133
- -**Why should I trust this leaderboard?**
134
 
135
  We have listed out our efforts to be fully transparent in the policies above. All of the code for this leaderboard is open-source and can be found on our [Github](https://github.com/atla-ai/judge-arena).
136
 
137
- -**Who funds this effort?**
138
 
139
  Atla currently funds this out of our own pocket. We are looking for API credits (with no strings attached) to support this effort - please get in touch if you or someone you know might be able to help.
140
 
141
- -**What is Atla working on?**
142
 
143
  We are training a general-purpose evaluator that you will soon be able to run in this Judge Arena. Our next step will be to open-source a powerful model that the community can use to run fast and accurate evaluations.
144
  <br><br>
 
126
  <br><br>
127
  # FAQ
128
 
129
+ - **Isn't this the same as Chatbot Arena?**
130
 
131
  We are big fans of what the LMSYS team have done with Chatbot Arena and fully credit them for the inspiration to develop this. We were looking for a dynamic leaderboard that graded on AI judge capabilities and didn't manage to find one, so we created Judge Arena. This UI is designed especially for evals; to match the format of the model-based eval prompts that you would use in your LLM evaluation / monitoring tool.
132
 
133
+ - **Why should I trust this leaderboard?**
134
 
135
  We have listed out our efforts to be fully transparent in the policies above. All of the code for this leaderboard is open-source and can be found on our [Github](https://github.com/atla-ai/judge-arena).
136
 
137
+ - **Who funds this effort?**
138
 
139
  Atla currently funds this out of our own pocket. We are looking for API credits (with no strings attached) to support this effort - please get in touch if you or someone you know might be able to help.
140
 
141
+ - **What is Atla working on?**
142
 
143
  We are training a general-purpose evaluator that you will soon be able to run in this Judge Arena. Our next step will be to open-source a powerful model that the community can use to run fast and accurate evaluations.
144
  <br><br>