Spaces:
Running
Running
kamranjkhan
commited on
Commit
β’
3f32881
1
Parent(s):
6fdc961
Update README.md
Browse files
README.md
CHANGED
@@ -13,7 +13,7 @@ AWS Inferentia accelerators are designed by AWS to deliver high performance at t
|
|
13 |
|
14 |
AWS Trainium is the second-generation machine learning (ML) accelerator that AWS purpose built for deep learning training of 100B+ parameter models. Each Amazon Elastic Compute Cloud (EC2) Trn1 instance deploys up to 16 AWS Trainium accelerators to deliver a high-performance, low-cost solution for deep learning (DL) training in the cloud. Trainium based EC2 Trn1 instances solve this challenge by delivering faster time to train while offering up to 50% cost-to-train savings over comparable Amazon EC2 instances.
|
15 |
|
16 |
-
**π€ Optimum Neuron**
|
17 |
π€ Optimum Neuron is the interface between the π€ Transformers library and AWS Accelerators including AWS Trainium and AWS Inferentia. It provides a set of tools enabling easy model loading, training and inference on single- and multi-Accelerator settings for different downstream tasks. The list of officially validated models and tasks is available [here](https://huggingface.co/docs/optimum-neuron/package_reference/configuration#supported-architectures).
|
18 |
|
19 |
**Learn More**
|
|
|
13 |
|
14 |
AWS Trainium is the second-generation machine learning (ML) accelerator that AWS purpose built for deep learning training of 100B+ parameter models. Each Amazon Elastic Compute Cloud (EC2) Trn1 instance deploys up to 16 AWS Trainium accelerators to deliver a high-performance, low-cost solution for deep learning (DL) training in the cloud. Trainium based EC2 Trn1 instances solve this challenge by delivering faster time to train while offering up to 50% cost-to-train savings over comparable Amazon EC2 instances.
|
15 |
|
16 |
+
**π€ Optimum Neuron**
|
17 |
π€ Optimum Neuron is the interface between the π€ Transformers library and AWS Accelerators including AWS Trainium and AWS Inferentia. It provides a set of tools enabling easy model loading, training and inference on single- and multi-Accelerator settings for different downstream tasks. The list of officially validated models and tasks is available [here](https://huggingface.co/docs/optimum-neuron/package_reference/configuration#supported-architectures).
|
18 |
|
19 |
**Learn More**
|