holylovenia
commited on
Commit
•
8eaaf79
1
Parent(s):
7d44c42
Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -18,9 +18,9 @@ tags:
|
|
18 |
- sentiment-analysis
|
19 |
---
|
20 |
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
|
25 |
|
26 |
## Languages
|
@@ -30,25 +30,25 @@ btk, bew, bug, jav, mad, mak, min, mui, rej, sun
|
|
30 |
## Supported Tasks
|
31 |
|
32 |
Sentiment Analysis
|
33 |
-
|
34 |
## Dataset Usage
|
35 |
### Using `datasets` library
|
36 |
```
|
37 |
-
|
38 |
-
|
39 |
```
|
40 |
### Using `seacrowd` library
|
41 |
```import seacrowd as sc
|
42 |
# Load the dataset using the default config
|
43 |
-
|
44 |
# Check all available subsets (config names) of the dataset
|
45 |
-
|
46 |
# Load the dataset using a specific config
|
47 |
-
|
48 |
```
|
49 |
-
|
50 |
-
|
51 |
-
|
52 |
|
53 |
## Dataset Homepage
|
54 |
|
|
|
18 |
- sentiment-analysis
|
19 |
---
|
20 |
|
21 |
+
Democratizing access to natural language processing (NLP) technology is crucial, especially for underrepresented and extremely low-resource languages. Previous research has focused on developing labeled and unlabeled corpora for these languages through online scraping and document translation. While these methods have proven effective and cost-efficient, we have identified limitations in the resulting corpora, including a lack of lexical diversity and cultural relevance to local communities. To address this gap, we conduct a case study on Indonesian local languages. We compare the effectiveness of online scraping, human translation, and paragraph writing by native speakers in constructing datasets. Our findings demonstrate that datasets generated through paragraph writing by native speakers exhibit superior quality in terms of lexical diversity and cultural content. In addition, we present the NusaWrites benchmark, encompassing 12 underrepresented and extremely low-resource languages spoken by millions of individuals in Indonesia. Our empirical experiment results using existing multilingual large language models conclude the need to extend these models to more underrepresented languages.
|
22 |
+
We introduce a novel high quality human curated corpora, i.e., NusaMenulis, which covers 12 languages spoken in Indonesia. The resource extend the coverage of languages to 5 new languages, i.e., Ambon (abs), Bima (bhp), Makassarese (mak), Palembang / Musi (mui), and Rejang (rej).
|
23 |
+
For the rhetoric mode classification task, we cover 5 rhetoric modes, i.e., narrative, persuasive, argumentative, descriptive, and expository.
|
24 |
|
25 |
|
26 |
## Languages
|
|
|
30 |
## Supported Tasks
|
31 |
|
32 |
Sentiment Analysis
|
33 |
+
|
34 |
## Dataset Usage
|
35 |
### Using `datasets` library
|
36 |
```
|
37 |
+
from datasets import load_dataset
|
38 |
+
dset = datasets.load_dataset("SEACrowd/nusatranslation_senti", trust_remote_code=True)
|
39 |
```
|
40 |
### Using `seacrowd` library
|
41 |
```import seacrowd as sc
|
42 |
# Load the dataset using the default config
|
43 |
+
dset = sc.load_dataset("nusatranslation_senti", schema="seacrowd")
|
44 |
# Check all available subsets (config names) of the dataset
|
45 |
+
print(sc.available_config_names("nusatranslation_senti"))
|
46 |
# Load the dataset using a specific config
|
47 |
+
dset = sc.load_dataset_by_config_name(config_name="<config_name>")
|
48 |
```
|
49 |
+
|
50 |
+
More details on how to load the `seacrowd` library can be found [here](https://github.com/SEACrowd/seacrowd-datahub?tab=readme-ov-file#how-to-use).
|
51 |
+
|
52 |
|
53 |
## Dataset Homepage
|
54 |
|