HyperTuning: Toward Adapting Large Language Models without Back-propagation Paper • 2211.12485 • Published Nov 22, 2022 • 1
Is Complexity Required for Neural Network Pruning? A Case Study on Global Magnitude Pruning Paper • 2209.14624 • Published Sep 29, 2022 • 1
Backpropagation-free Training of Deep Physical Neural Networks Paper • 2304.11042 • Published Apr 20, 2023 • 1
Lottery Tickets in Evolutionary Optimization: On Sparse Backpropagation-Free Trainability Paper • 2306.00045 • Published May 31, 2023 • 1
ZO-AdaMU Optimizer: Adapting Perturbation by the Momentum and Uncertainty in Zeroth-order Optimization Paper • 2312.15184 • Published Dec 23, 2023 • 1
DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training Paper • 2310.02025 • Published Oct 3, 2023 • 1
AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning Paper • 2406.18060 • Published Jun 26