Collections
Discover the best community collections!
Collections including paper arxiv:2104.08691
-
Combining Modular Skills in Multitask Learning
Paper • 2202.13914 • Published • 4 -
The Power of Scale for Parameter-Efficient Prompt Tuning
Paper • 2104.08691 • Published • 9 -
Prefix-Tuning: Optimizing Continuous Prompts for Generation
Paper • 2101.00190 • Published • 6 -
GPT Understands, Too
Paper • 2103.10385 • Published • 8
-
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Paper • 2403.03507 • Published • 182 -
Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning
Paper • 2205.05638 • Published • 3 -
The Power of Scale for Parameter-Efficient Prompt Tuning
Paper • 2104.08691 • Published • 9 -
In-Context Learning Demonstration Selection via Influence Analysis
Paper • 2402.11750 • Published • 2
-
Efficient Few-Shot Learning Without Prompts
Paper • 2209.11055 • Published • 3 -
Parameter-Efficient Transfer Learning for NLP
Paper • 1902.00751 • Published • 1 -
GPT Understands, Too
Paper • 2103.10385 • Published • 8 -
The Power of Scale for Parameter-Efficient Prompt Tuning
Paper • 2104.08691 • Published • 9
-
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper • 2402.17764 • Published • 602 -
When Scaling Meets LLM Finetuning: The Effect of Data, Model and Finetuning Method
Paper • 2402.17193 • Published • 23 -
Training-Free Long-Context Scaling of Large Language Models
Paper • 2402.17463 • Published • 19 -
The Power of Scale for Parameter-Efficient Prompt Tuning
Paper • 2104.08691 • Published • 9
-
Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning
Paper • 2303.10512 • Published -
Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning
Paper • 2205.05638 • Published • 3 -
LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention
Paper • 2303.16199 • Published • 4 -
FedPara: Low-Rank Hadamard Product for Communication-Efficient Federated Learning
Paper • 2108.06098 • Published • 2
-
Attention Is All You Need
Paper • 1706.03762 • Published • 44 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 14 -
Universal Language Model Fine-tuning for Text Classification
Paper • 1801.06146 • Published • 6 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 11