AIResearch AIResearch
Back to articles
AI

TS-PEFT: A Smarter Way to Fine-Tune AI Models Saves Time and Resources

In the rapidly evolving world of artificial intelligence, fine-tuning large models like those used in natural language processing and computer vision has become a cornerstone of adapting powerful syst…

AI Research
November 22, 2025
4 min read
TS-PEFT: A Smarter Way to Fine-Tune AI Models Saves Time and Resources

In the rapidly evolving world of artificial intelligence, fine-tuning large models like those used in natural language processing and computer vision has become a cornerstone of adapting powerful systems to specific tasks. However, the computational demands of traditional fine-tuning s are staggering, often requiring extensive resources that limit accessibility and scalability. Enter Parameter-Efficient Fine-Tuning (PEFT), a technique that modifies only a small subset of a model's parameters while keeping the bulk frozen, drastically cutting costs. A groundbreaking study titled 'TS-PEFT: Token-Selective Parameter-Efficient Fine-Tuning with Learnable Threshold Gating' s the status quo by revealing that even PEFT s apply updates indiscriminately to all token positions, introducing unnecessary noise and inefficiency. This research, led by Dabiao Ma and colleagues from Qifu Technology, Inc., proposes a novel approach that could redefine how we optimize AI models, making fine-tuning not just cheaper but smarter by focusing updates only where they matter most.

Traditional PEFT s, such as Low-Rank Adaptation (LoRA) and its variants like DoRA and AdaLoRA, have gained popularity for their ability to reduce trainable parameters by up to 99% while maintaining performance close to full fine-tuning. These techniques work by injecting small, trainable components into a frozen base model, allowing for efficient adaptation to downstream tasks like commonsense reasoning or visual instruction tuning. For instance, LoRA uses low-rank matrices to approximate weight updates, which can be merged back after training without extra inference costs. Despite these advances, the study highlights a critical oversight: once a layer is targeted for PEFT, updates are applied uniformly to every token position in the input sequence. This 'update-all-positions' design ignores the fact that different tokens contribute unequally to task performance, leading to redundant modifications that can harm accuracy and increase computational load, especially when multiple PEFT modules are stacked in complex models.

The core innovation of TS-PEFT lies in its introduction of a token-level selective mechanism that uses a learnable binary gating function to decide which positions receive PEFT updates. Instead of blindly applying modifications, TS-PEFT calculates the relative magnitude of each update compared to the base model's output and compares it to a dynamic threshold. If the update is too small, it's skipped, preserving the original output; if significant, the update is applied. This approach is formalized through a proximal optimization framework that incorporates sparsity regularization, encouraging the model to activate updates on only about 40-60% of tokens. To handle the non-differentiability of the gating function, the researchers developed an approximate gradient combined with Adam-style momentum updates, ensuring stable training. Extensive experiments on models like LLaMA-3-8B and LLaVA-1.5-7B across benchmarks in commonsense reasoning, visual instruction tuning, and natural language understanding demonstrate that TS-PEFT not only matches but often surpasses baseline PEFT s in performance, all while reducing the number of updated positions by nearly half.

Empirical from the study are compelling: on commonsense reasoning tasks using LLaMA-3-8B, TS-PEFT variants like TS-LoRA and TS-DoRA achieved average scores of 88.8 and 75.2, respectively, outperforming their standard counterparts while maintaining sparsity rates around 55-60%. In visual instruction tuning with LLaVA-1.5-7B, improvements of 0.3-0.4 points were consistently observed, and in natural language understanding on the GLUE benchmark, TS-PEFT maintained or slightly enhanced performance without increasing parameter counts. Beyond mere efficiency gains, the token-level sparsity patterns learned by TS-PEFT serve as a robust indicator of module importance, enabling more strategic selection of which parts of a model to fine-tune. For example, selecting the 50% of modules with the lowest sparsity in TS-DoRA led to better performance than using all modules, highlighting how this can guide resource allocation in multi-task environments and potentially accelerate inference in future hardware-optimized systems.

Despite its promise, TS-PEFT has limitations that warrant consideration. The current implementation does not yet provide wall-clock speedups in standard setups, as it relies on logical sparsity that requires specialized kernels for physical latency reductions. Additionally, 's effectiveness varies with the underlying PEFT architecture; for instance, it showed strong compatibility with LoRA-style s but less so with adapter-based approaches in initial tests. The reliance on hyperparameters like the scaling factor and sparsity penalty also introduces tuning complexity, though the study provides guidelines for optimal settings. Looking ahead, these s point to fertile ground for future research, such as integrating TS-PEFT with emerging sparse computation techniques or expanding it to other AI domains like robotics or quantum computing. By addressing the inherent redundancies in fine-tuning, this work not only advances efficiency but also opens doors to more interpretable and scalable AI systems, paving the way for broader adoption in resource-constrained scenarios.

Ma, D., Dai, Z., Xin, Z., Wang, S., Wang, Y., and Fei, H. (2025). TS-PEFT: Token-Selective Parameter-Efficient Fine-Tuning with Learnable Threshold Gating. arXiv preprint arXiv:2511.16147.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn