stereoplegic
's Collections
No backprop
updated
Fine-Tuning Language Models with Just Forward Passes
Paper
•
2305.17333
•
Published
•
2
HyperTuning: Toward Adapting Large Language Models without
Back-propagation
Paper
•
2211.12485
•
Published
•
1
Is Complexity Required for Neural Network Pruning? A Case Study on
Global Magnitude Pruning
Paper
•
2209.14624
•
Published
•
1
Backpropagation-free Training of Deep Physical Neural Networks
Paper
•
2304.11042
•
Published
•
1
Lottery Tickets in Evolutionary Optimization: On Sparse
Backpropagation-Free Trainability
Paper
•
2306.00045
•
Published
•
1
Gradients without Backpropagation
Paper
•
2202.08587
•
Published
•
1
Learning with Local Gradients at the Edge
Paper
•
2208.08503
•
Published
•
1
ZO-AdaMU Optimizer: Adapting Perturbation by the Momentum and
Uncertainty in Zeroth-order Optimization
Paper
•
2312.15184
•
Published
•
1
DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training
Paper
•
2310.02025
•
Published
•
1
AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for
Memory-Efficient Large Language Models Fine-Tuning
Paper
•
2406.18060
•
Published
Zeroth-Order Fine-Tuning of LLMs with Extreme Sparsity
Paper
•
2406.02913
•
Published