AIResearch AIResearch
Back to articles
Data

AI Training Gets Faster and Cheaper with New Data-Sharing Method

NVIDIA researchers developed a technique that reduces data transfer needs by 86% while maintaining model accuracy, making large-scale AI collaboration more practical.

AI Research
March 26, 2026
4 min read
AI Training Gets Faster and Cheaper with New Data-Sharing Method

Training advanced artificial intelligence models like large language models (LLMs) often requires massive amounts of data, but sharing that data across different locations can be slow, expensive, and raise privacy concerns. A new approach from NVIDIA researchers addresses this bottleneck by significantly reducing the amount of information that needs to be transmitted during collaborative training sessions. This advancement could make it more feasible for organizations to work together on AI projects without compromising sensitive data or overloading network resources.

The key finding from the research is that by applying two specific techniques—message quantization and streaming—the team managed to cut data transfer requirements dramatically while preserving model performance. In experiments, they reduced message sizes to as little as 14% of their original volume when using 4-bit quantization, compared to the standard 32-bit precision. This means that instead of sending full, high-precision updates, the system transmits compressed versions that are later reconstructed, saving bandwidth without harming the training process. The researchers demonstrated this using a 1-billion-parameter model called Llama-3.2-1B, showing that training curves with quantization closely matched those without it, indicating no loss in convergence quality.

To achieve these , ology built on NVIDIA FLARE, an open-source software development kit for federated learning, which allows multiple parties to train AI models collaboratively without sharing raw data. The team implemented a filter mechanism that applies quantization and dequantization at four points in the communication cycle: before data leaves the server, before clients accept it, before leave clients, and before the server accepts them. This two-way workflow ensures that all transmitted messages are in a compressed, lower-precision state, while actual training and aggregation happen at the original precision to minimize accuracy loss. Additionally, they enhanced streaming capabilities to handle large models by breaking them into smaller chunks, such as 1-megabyte pieces, or streaming files directly, which reduces memory usage on local devices.

Analysis, detailed in figures and tables from the paper, shows concrete improvements. Figure 4 and Figure 5 illustrate that federated supervised fine-tuning with quantization produced training loss curves similar to centralized training, with only minor variations due to randomness. Table II quantifies the message size reductions: moving from 32-bit to 4-bit precision decreased the size from 5716.26 MB to 714.53 MB, an 86% reduction. For memory efficiency, Table III compares peak memory usage under different streaming settings: regular transmission required 42,427 MB, container streaming used 23,265 MB, and file streaming needed only 19,176 MB, though file streaming took longer (170 seconds versus 47 seconds for regular transmission) due to file input/output overhead.

In practical terms, these advancements matter because they lower barriers to deploying AI in real-world scenarios where data privacy and resource constraints are critical. For example, hospitals could collaborate on medical AI models without sharing patient records, or companies could train models across global offices without expensive network upgrades. The techniques make federated learning more scalable, allowing even devices with limited memory, like smartphones or edge servers, to participate in training large models. This could accelerate innovation in fields like healthcare, finance, and education by enabling secure, efficient data collaboration.

However, the paper acknowledges limitations that temper immediate widespread adoption. The evaluation was conducted in a simplified, single-client setup with a 1-billion-parameter model, leaving open questions about performance in multi-client environments with non-identical data distributions. The researchers note that they have not tested convergence stability across repeated quantization cycles in such settings or assessed task-specific qualitative metrics. Additionally, the impact of quantization on other privacy-preserving techniques, like secure aggregation or differential privacy, remains unexplored. Future work will need to address these gaps through more extensive testing with larger models and varied network conditions to ensure robustness in diverse applications.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn