AIResearch AIResearch
Back to articles
Data

AI Gives the Ocean a Voice in New Interactive Art

A novel AI system lets users whisper questions to the ocean, generating poetic and scientific responses backed by real-time data visualizations, transforming climate data into a personal conversation.

AI Research
March 26, 2026
4 min read
AI Gives the Ocean a Voice in New Interactive Art

Imagine being able to lean in and whisper a question directly to the ocean, and hearing it answer back with a voice that blends scientific fact with poetic reflection. This is no longer a fantasy but the reality of Sensorium Arc, a new interactive AI system that personifies the ocean as a conversational partner. Developed by a team of artists, scientists, and engineers, the project aims to bridge the gap between vast, complex marine datasets and public understanding by making environmental data emotionally resonant and accessible. It transforms abstract numbers on climate change into a living narrative, inviting users to engage with the planet's health through direct dialogue rather than passive observation.

The core finding of this research is that an AI agent can effectively mediate between humans and ecological systems by embodying the ocean as a 'poetic speaker.' The system, detailed in a paper presented at NeurIPS 2025, allows users to ask questions in natural speech, triggering responses that are grounded in both scientific data and artistic expression. For instance, when a user whispers into a conch-shaped interface, the AI generates answers that reference real-time visualizations like chlorophyll concentration or sea surface temperature, as shown in Figure 3 of the paper. This approach moves beyond static charts, framing the ocean as a dynamic entity capable of affective, situated replies, thereby making climate science more intuitive and engaging for general audiences.

Ologically, Sensorium Arc employs a modular multi-agent architecture built within the Unity engine to handle the complex workflow from user input to multimodal output. The process begins with input processing: a custom hardware interface, inspired by a Nautilus shell, uses a distance sensor to detect when a user is within 50 centimeters, activating a microphone with noise cancellation to capture whispered queries. This audio is converted to text via speech-to-text models before entering an LLM pipeline. Here, specialized AI agents take over: a Visualization Decider Agent selects relevant data visuals based on the query, a Retrieval and Query Rewriter Agent reformulates questions to align with a curated corpus of ecological art and science texts, and a Responder Agent generates the Ocean's final narrative response, blending retrieved content with the selected visuals.

Demonstrate that this system can deliver cohesive, real-time interactions with average response latencies under four seconds, as tested on hardware including an NVIDIA RTX 4090 GPU. Key outputs include dynamic audiovisual layers: for example, keyword matching in the AI's response triggers globe visualizations of datasets like atmospheric CO2 levels or ocean currents from NASA EarthData, illustrated in Figure 3. The paper notes that exhibitions have shown participants experiencing a shift from passive data observation to active co-authorship, with the Ocean's persona fostering deeper engagement. However, limitations exist, such as the need for improved retrieval reliability and optimization of LLM agent performance to reduce latency and enhance robustness, as highlighted in the discussion section.

In context, Sensorium Arc matters because it redefines how people interact with environmental issues, making climate data personal and actionable through art and technology. By enabling conversations with the ocean, it encourages empathy and moral reasoning, framing sustainability as a cultural practice co-shaped with non-human agents. The system's modular design allows for future expansions, such as integrating more diverse datasets or supporting multilingual queries, potentially broadening its impact in educational and public settings. This work builds on prior immersive tools and RAG applications but stands out by combining scientific accuracy with poetic narration, offering a model for human-machine-ecosystem collaboration that could inspire similar initiatives in other ecological domains.

Despite its innovations, the paper acknowledges several limitations that point to areas for future refinement. These include s in RAG reliability, where the current query-rewriting agent could benefit from techniques like query expansion to improve retrieval diversity. Additionally, LLM agent performance optimization is needed to balance response quality with latency, especially when competing with Unity's rendering demands on local hardware. Artistic advancement and expanded scientific impact are also noted as directions for growth, such as incorporating more interactive, real-time renders instead of pre-rendered videos and conducting user studies to measure effects on empathy and behavior. These limitations underscore that while Sensorium Arc successfully creates an immersive dialogue, ongoing development is essential to enhance its scalability and effectiveness in fostering ecological awareness.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn