-
Mixture-of-Depths: Dynamically allocating compute in transformer-based language models
Paper • 2404.02258 • Published • 104 -
Transformer-Lite: High-efficiency Deployment of Large Language Models on Mobile Phone GPUs
Paper • 2403.20041 • Published • 34 -
ViTAR: Vision Transformer with Any Resolution
Paper • 2403.18361 • Published • 52
Prathamesh Sarang
pratos
AI & ML interests
None yet
Organizations
Collections
2
models
None public yet
datasets
None public yet