AdaGradSelect: A Computationally Efficient and Memory-Optimized Fine-Tuning Method for Large Language Models
By: Yixuan Weng, Minjun Zhu, Qiujie Xie, Qiyao Sun, Zhen Lin, Sifan Liu, Yue Zhang
Published: 2025-12-17
View on arXiv →#cs.AI
Abstract
This paper introduces AdaGradSelect, a novel fine-tuning method for Large Language Models (LLMs) that offers significant computational efficiency and memory optimization. It trains about 12% faster and uses 35% less GPU memory while maintaining performance close to full fine-tuning, outperforming LoRA on certain benchmarks and providing a more effective and resource-efficient alternative to traditional fine-tuning.