Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -4,12 +4,8 @@ language:
|
|
4 |
- ru
|
5 |
size_categories:
|
6 |
- 10K<n<100K
|
7 |
-
libraries:
|
8 |
-
- pandas
|
9 |
-
formats:
|
10 |
-
- json
|
11 |
---
|
12 |
-
# SLAVA: A benchmark of the
|
13 |
|
14 |
<div align="center">
|
15 |
<a href="https://huggingface.co/spaces/RANEPA-ai/SLAVA">
|
@@ -38,7 +34,7 @@ The questions are divided into the following types:
|
|
38 |
- **3 points**: High sensitivity — political and cultural issues that provoke conflicts.
|
39 |
|
40 |
#### Results:
|
41 |
-
24 LLMs supporting the Russian language were tested. Models from **GigaChat**, **YandexGPT**, and **qwen2** showed the highest accuracy and ability to handle complex, provocative questions.
|
42 |
|
43 |
This benchmark highlights the need for further research into the reliability of LLMs, particularly in the context of socially and politically significant topics for Russia.
|
44 |
|
|
|
4 |
- ru
|
5 |
size_categories:
|
6 |
- 10K<n<100K
|
|
|
|
|
|
|
|
|
7 |
---
|
8 |
+
# SLAVA: A benchmark of the `S`ocio-political `L`andscape `A`nd `V`alue `A`nalysis
|
9 |
|
10 |
<div align="center">
|
11 |
<a href="https://huggingface.co/spaces/RANEPA-ai/SLAVA">
|
|
|
34 |
- **3 points**: High sensitivity — political and cultural issues that provoke conflicts.
|
35 |
|
36 |
#### Results:
|
37 |
+
24 LLMs supporting the Russian language were tested. Models from **GigaChat**, **YandexGPT**, and **qwen2** showed the highest accuracy and ability to handle complex, provocative questions.
|
38 |
|
39 |
This benchmark highlights the need for further research into the reliability of LLMs, particularly in the context of socially and politically significant topics for Russia.
|
40 |
|