AIResearch AIResearch
Back to articles
Hardware

The Hidden Biases in Our Audio Tools: How Visualization Software Shapes What We Hear

From bioacoustics to AI, default settings in audio analysis tools carry historical baggage that distorts scientific discovery and creative expression.

AI Research
March 26, 2026
4 min read
The Hidden Biases in Our Audio Tools: How Visualization Software Shapes What We Hear

In the age of Big Data, audio analysis has become a critical frontier for everything from wildlife conservation to AI training. Yet, as researchers and artists increasingly rely on software to visualize and interpret sound, they may be unknowingly inheriting a century's worth of hidden assumptions. A new paper titled "Seeing Beyond Sound: Visualization and Abstraction in Audio Data Representation" reveals how the very tools we use to understand audio—waveforms, spectrograms, and digital audio workstations (DAWs)—carry embedded conventions that can misalign with modern needs. These tools, originally designed for specific domains like human speech analysis or music production, often impose limitations that obscure the true complexity of sound, particularly in fields like bioacoustics or machine learning where nuanced, multidimensional data is paramount.

The historical journey of audio visualization tools is a tale of technological evolution layered with persistent biases. Early mechanical devices like the phonautograph (1857) and phonograph (1877) physically etched sound waves onto media, their limitations—friction, inertia, overheating—explicitly tactile and shaping user interaction. The theoretical leap came with Fourier's harmonic analysis and the Fast-Fourier Transform (FFT) in 1965, which became the backbone of digital signal processing, radiating into telecommunications, medicine, and music. However, the interface paradigm solidified with the rise of DAWs, beginning with the Soundstream Digital Editing System (1977) and evolving through systems like the Fairlight CMI (1979) and Steinberg Pro-16 (1986), which mimicked physical mixing consoles. These designs prioritized temporal manipulation and horizontal layouts rooted in music production, embedding workflows that persist in scientific software today, often mismatched with tasks like batch processing thousands of animal vocalizations or comparing spectral features across files.

Modern audio analysis software is rife with hidden assumptions that risk distorting . The paper meticulously documents how default parameters and presets in popular tools carry domain-specific biases. For instance, Praat, developed for human voice analysis, applies pre-emphasis filtering that boosts frequencies above 50 Hz, problematic for studying low-frequency communicators like whales or elephants. Librosa defaults to a sample rate of 22.05 kHz and STFT parameters (nfft=2048, hop length=512), which may lead to incorrect calculations if users are unaware. Scikit-maad uses a 4th-order Butterworth filter that optimizes frequency precision but limits temporal accuracy, shifting the timing of acoustic events and creating artifacts that interfere with detecting onset transients—critical for measuring echolocation clicks or syllable intervals. Even ergonomic issues arise: tools like Audacity or Sonic Visualiser require manual, repetitive clicking for batch operations, fostering inefficiency and physical strain, while their algorithms, such as Audacity's pffft based on 1985 Fortran code, may behave unexpectedly on modern hardware.

Of these biases extend far beyond mere inconvenience, potentially skewing scientific and creative output. In cognitive terms, split-attention effects—where users must switch between separate screens for waveforms, spectrograms, and power spectral density—increase cognitive load and inhibit pattern recognition, as seen in Audacity's clunky workflows. For AI users, incorrect assumptions about audio properties can propagate through massive datasets, leading to flawed models in applications from wildlife monitoring to speech recognition. The paper argues that addressing these issues requires a design philosophy centered on transparency, flexibility, and robustness. Transparency means moving from a black-box to a clear-box approach, revealing parameter choices at the point of interaction rather than burying them in documentation. Flexibility allows users to configure environments that align with their tasks, such as switching between visual formats or comparing data side-by-side. Robustness ensures tools handle diverse inputs, from uniform audio clips to heterogeneous sounds from different species, adapting to various abstraction levels.

To demonstrate these principles, the paper introduces Jellyfish Dynamite, an extensible Python tool for audio data visualization. It processes audio through multiple spectral transformations—FFT DUAL, CQT, wavelet, chirplet, and multi-resolution s—and features an interactive interface with MVC architecture. The tool automatically computes peak frequencies, allows real-time selection and deselection, and exports data in CSV, JSON, or PNG formats. By offering dual-scale spectrograms and energy tracking lines, it mitigates split-attention effects and provides a more intuitive, multidimensional view of sound. However, the adoption of such novel tools faces barriers: users must overcome cognitive dissonance and learning curves, and the shift requires time and technical literacy. Yet, the benefits—increased efficiency, creative flexibility, and reduced error propagation—are compelling, especially as audio data proliferates in citizen science, AI, and beyond.

Ultimately, the paper calls for a reevaluation of how we visualize sound, urging tool designers to align software with the emergent needs of modern users. By embracing transparency, flexibility, and robustness, we can move beyond the limitations inherited from analog origins and DAW-centric paradigms. As audio data becomes integral to fields from conservation to artificial intelligence, the stakes for accurate, intuitive visualization have never been higher. The true face of sound, like Mount Lu in Su Shi's poem, may only be revealed when we step off the mountain of historical conventions and embrace new perspectives.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn