Text Generation
Transformers
Safetensors
English
falcon_mamba
conversational
Inference Endpoints
Gkunsch commited on
Commit
285408b
1 Parent(s): 6aa14b1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -170,7 +170,7 @@ print(tokenizer.decode(outputs[0]))
170
 
171
  ## Training Data
172
 
173
- Falcon-Mamba has been trained with ~ 5,500 GT mainly coming from [Refined-Web](https://huggingface.co/datasets/tiiuae/falcon-refinedweb), a large volume web-only dataset filtered and deduplicated.
174
  Similar to the others [Falcon](https://huggingface.co/tiiuae/falcon-11B) suite models, Falcon-Mamba has been trained leveraging a multi-stage training strategy to increase the context-length from 2,048 to 8,192.
175
  Moreover, inspired by the concept of Curriculum Learning, we carefully selected data mixtures throughout the training stages, considering both data diversity and complexity.
176
  Note that at inference the context-length is not relevant as the Mamba architecture has no limit on long range dependency.
 
170
 
171
  ## Training Data
172
 
173
+ Falcon-Mamba has been trained with ~ 6,000 GT mainly coming from [Refined-Web](https://huggingface.co/datasets/tiiuae/falcon-refinedweb), a large volume web-only dataset filtered and deduplicated.
174
  Similar to the others [Falcon](https://huggingface.co/tiiuae/falcon-11B) suite models, Falcon-Mamba has been trained leveraging a multi-stage training strategy to increase the context-length from 2,048 to 8,192.
175
  Moreover, inspired by the concept of Curriculum Learning, we carefully selected data mixtures throughout the training stages, considering both data diversity and complexity.
176
  Note that at inference the context-length is not relevant as the Mamba architecture has no limit on long range dependency.