Update README.md
Browse files
README.md
CHANGED
@@ -7,3 +7,42 @@ widget:
|
|
7 |
example_title: "උදාහරණ 1"
|
8 |
---
|
9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
example_title: "උදාහරණ 1"
|
8 |
---
|
9 |
|
10 |
+
## Model Description
|
11 |
+
|
12 |
+
The Sinhala Story Generation Model is based on fine-tuning the XLM-RoBERTa base model on a dataset of Sinhala language stories. It is designed to generate coherent and contextually relevant Sinhala text based on story beginnings.
|
13 |
+
|
14 |
+
## Intended Use
|
15 |
+
|
16 |
+
The model is intended for generating creative Sinhala stories or text based on initial prompts. It can be used in applications requiring automated generation of Sinhala text, such as chatbots, content generation, or educational tools.
|
17 |
+
|
18 |
+
## Example Use Cases
|
19 |
+
|
20 |
+
- Creative Writing: Generate new story ideas or expand on existing story prompts.
|
21 |
+
- Language Learning: Create exercises or content in Sinhala for language learners.
|
22 |
+
- Content Generation: Automatically generate text for social media posts, blogs, or websites.
|
23 |
+
|
24 |
+
## Limitations and Ethical Considerations
|
25 |
+
|
26 |
+
- The model's output is based on patterns in the training data and may not always generate accurate or contextually appropriate text.
|
27 |
+
- Users are advised to review and refine generated text for accuracy and appropriateness before use in sensitive or critical applications.
|
28 |
+
|
29 |
+
## Model Details
|
30 |
+
|
31 |
+
- Model Architecture: XLM-RoBERTa base
|
32 |
+
- Training Data: Sinhala language stories dataset. Dataset is created using various sources such as social media and web content.
|
33 |
+
- Tokenization: AutoTokenizer from Hugging Face Transformers library
|
34 |
+
- Fine-tuning: Fine-tuned on Sinhala story dataset for text generation task
|
35 |
+
|
36 |
+
## Example Inference
|
37 |
+
|
38 |
+
To use the model for inference via the Hugging Face Inference API, consider the following example Python code:
|
39 |
+
|
40 |
+
```from transformers import pipeline
|
41 |
+
|
42 |
+
model_name = "your-username/model-name"
|
43 |
+
generator = pipeline("text-generation", model=model_name, tokenizer=model_name)
|
44 |
+
|
45 |
+
input_text = "අද සුන්දර දවසක්. හෙට ගැන සිතමින් මම පාර <mask>"
|
46 |
+
output = generator(input_text, max_length=150, num_return_sequences=1)
|
47 |
+
|
48 |
+
print(output[0]['generated_text'])```
|