AIResearch AIResearch
Back to articles
AI

Open-Source AI Simplifies Piano Scores for Inclusive Music Education

A new transformer-based method generates difficulty-controlled sheet music, challenging proprietary systems that limit accessibility.

AI Research
March 26, 2026
4 min read
Open-Source AI Simplifies Piano Scores for Inclusive Music Education

In the realm of music education, artificial intelligence has long promised to democratize learning by making complex compositions accessible to beginners and students with varying skill levels. However, this potential has been stifled by proprietary systems and closed datasets that hinder reproducibility and widen the digital divide, particularly in subjects like music that are often marginalized within STEAM (Science, Technology, Engineering, Arts, and Mathematics) approaches. A new study from researchers at Universitat Pompeu Fabra, Nuremberg University of Music, and Sogang University tackles this issue head-on by introducing an open-source, transformer-based for adjusting the difficulty of piano scores in MusicXML format. This work aims to foster inclusive learning opportunities, allowing educators to adapt repertoire to individual student needs without relying on costly, expert-annotated datasets or closed commercial tools like Yamaha's systems or the Simply Piano app.

At the core of this research is a novel ology that frames score simplification as a conditional generation problem, where the goal is to "translate" existing piano scores from one difficulty level to another while preserving melody, harmony, and style. The researchers developed a transformer-based, decoder-only model pretrained on a large corpus of 136,000 piano scores from the XMander dataset, using Linearized MusicXML (LMX) as an input representation that retains layout and readability—critical for educational use. Unlike previous approaches that rely on MIDI formats lacking visual nuance, LMX converts hierarchical MusicXML into a token sequence compatible with transformers, enabling efficient training and generation. The model is conditioned on melody skyline and harmonic pitch profiles, with a supervised learning setup that uses difficulty-labeled pairs from a synthetic dataset called PianoPairs, generated by mining variations based on pretrained models for difficulty estimation and style similarity.

The experimental demonstrate the effectiveness of this approach, with the proposed filtered mining strategy outperforming random pairing in controlling difficulty adaptation. In the original benchmark, the filtered strategy achieved a 74.4% success rate in generating easier variations for a difficulty gap of two levels or more, compared to 66.6% for random pairing, while reducing the percentage of harder outputs to 8.4% from 8.5%. The mean cosine distance between CLaMP embeddings, which measures stylistic similarity, decreased from 0.305 to 0.299 for gap 2, indicating better preservation of musical coherence. In the extended benchmark, simplification rates reached 78.4% for filtered gap 2, with only 3.0% of variations being harder. Subjective evaluations by 10 expert pianists further validated the model, showing significant preferences for filtered outputs in easiness (44 vs. 16 votes, p-val = 0.0004) and readability (42 vs. 18, p-val = 0.0027), with high pedagogical ratings averaging 4.20 for instructional control on a Likert scale.

Of this research are profound for music education and AI democratization. By releasing all resources—code, dataset, and models—openly, the study s the dominance of proprietary systems and promotes reproducible research in a field often hampered by legal and commercial barriers. This approach supports a Teacher-in-the-Loop framework, enabling educators to personalize learning experiences, enhance student motivation, and bridge the digital divide by making advanced technologies accessible. It also highlights the importance of open-source innovation in fostering inclusive pedagogies, as the system can generate playable and readable scores that adapt to diverse learner needs, from beginners tackling simplified versions of classical pieces to advanced students exploring more challenging arrangements.

Despite its successes, the study acknowledges several limitations that warrant further exploration. primarily focuses on scores with a clear melody-accompaniment structure, which may limit its effectiveness on highly polyphonic or contrapuntal textures, such as those found in complex classical works. Simplification tends to affect accompaniment more than melodic lines, due to the absence of explicit melodic complexity control, and the model does not currently support upward difficulty generation for advanced learners. Additionally, the evaluation relies on synthetic pairs due to a lack of public-aligned datasets, preventing direct benchmarking against human arrangements and introducing potential biases in stylistic preservation. Future work could address these gaps by incorporating more diverse musical genres and refining complexity controls.

In conclusion, this research marks a significant step toward open-source, difficulty-aware music generation, offering a practical tool for educators and a foundation for further innovation. By prioritizing accessibility and reproducibility, it not only advances AI applications in music education but also sets a precedent for ethical technology development in creative domains. As the digital landscape evolves, such initiatives are crucial for ensuring that cutting-edge tools serve broad, inclusive purposes rather than reinforcing existing disparities.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn