A new artificial intelligence method called MU-SplitFed enables faster and more efficient machine learning on resource-constrained devices by eliminating the bottleneck of waiting for the slowest participant in distributed systems. This breakthrough addresses a fundamental limitation in federated learning where training speed is typically dictated by the slowest device in the network, significantly hampering scalability and efficiency.
The researchers developed an unbalanced update approach that allows powerful servers to perform multiple training iterations while waiting for slower edge devices to complete their computations. This simple yet effective mechanism decouples training progress from device delays, achieving what the paper describes as "linear speedup" in training rounds. The method combines this unbalanced scheduling with zeroth-order optimization, which reduces computational burdens on edge devices by eliminating the need for complex backpropagation calculations.
The approach works by partitioning neural networks between client devices and servers, with clients handling initial layers and servers processing deeper layers. Unlike traditional methods that require synchronization between all participants, MU-SplitFed enables servers to continue training using available data while waiting for slower devices. The researchers validated their method through experiments on multiple benchmark datasets including CIFAR-10, Fashion-MNIST, CINIC-10, and CIFAR-100, demonstrating consistent performance improvements over baseline methods.
Results showed that with proper parameter selection, the method reduces communication rounds by up to 33% while maintaining or improving accuracy. The paper reports that when using two server iterations per round (τ=2), the method achieved the highest accuracy across all tested datasets, outperforming both vanilla SplitFed and GAS methods. The approach proved particularly effective in scenarios with high device heterogeneity, where computation speeds vary significantly between devices.
This advancement matters because it makes distributed AI training more practical for real-world applications involving diverse hardware, from smartphones to IoT devices. The method's ability to continue training despite slow participants means systems can maintain efficiency even when some devices experience network delays or limited computational resources. The researchers also demonstrated the method's applicability to large language models, showing it can reduce client-side memory usage from 8.02 GB to just 1.05 GB during fine-tuning.
The paper acknowledges that optimal performance requires careful alignment between the model partitioning strategy and the number of server iterations. The researchers found that increasing server iterations beyond an optimal point can actually harm performance if not properly matched with the model architecture. This limitation highlights the need for thoughtful system design when implementing the approach in practical applications.
About the Author
Guilherme A.
Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.
Connect on LinkedIn