AIResearch AIResearch
Back to articles
Network

AI Balances Speed and Smarts in Real Conversations

A new hybrid AI system from Baidu keeps digital humans responsive while handling complex tasks in the background, tripling user retention and boosting task completion by over 80% in real-world deployment.

AI Research
March 27, 2026
4 min read
AI Balances Speed and Smarts in Real Conversations

Imagine chatting with a digital assistant that feels instantly responsive, yet can seamlessly plan a vacation or offer medical advice without awkward pauses. This balance between speed and capability has long eluded AI systems, but a new approach from Baidu is showing how to achieve it at scale. The system, called DuCCAE, has been deployed within Baidu Search since June 2025, serving millions of users and demonstrating that it's possible to maintain conversational flow while executing intricate tasks. The key breakthrough lies in decoupling real-time interaction from asynchronous reasoning, allowing the AI to stay engaging without sacrificing functionality.

Researchers found that DuCCAE significantly outperforms existing s by using a dual-track architecture that separates fast responses from slow, complex planning. In evaluations on the Du-Interact dataset, the system achieved an 82.5% dispatch precision for routing user queries correctly and a 72.4% success rate in completing complex tasks, surpassing even large models like Llama-3.3-70B, which scored 61.1% in success rate. More impressively, deployment metrics revealed a tripling of Day-7 user retention to 34.2% and a surge in complex task completion rate to 65.2%, up from 35.5% in earlier versions. These indicate that the hybrid design not only improves technical performance but also enhances user engagement in practical settings.

Ology behind DuCCAE involves five integrated subsystems: Info, Conversation, Collaboration, Augmentation, and Evolution. The system starts by processing multimodal inputs like speech and video through the Info System, converting them into text to reduce latency. For instance, video streams are summarized into captions by a lightweight vision-language model, cutting visual perception latency from 2,100 milliseconds to 480 milliseconds. The Conversation System then acts as a gatekeeper, classifying user requests into three tiers—lightweight chats, tool-based queries, and complex domain tasks—and routing them accordingly. Simple queries go to a Fast Track for immediate response within 500 milliseconds, while complex ones are offloaded to a Slow Track handled by the Collaboration System, which uses multi-agent planning and tool execution without blocking the conversation.

Analysis of , as detailed in Table 2 of the paper, shows that DuCCAE's evolved versions consistently outperform baselines across key metrics. The V3 iteration, trained on 50,000 interaction episodes through supervised fine-tuning and reinforcement learning, achieved a 71.1% response fidelity and high scores in persona consistency (4.1 out of 5) and empathy (4.3 out of 5). In contrast, large zero-shot models like Llama-3.3-70B, while powerful, suffered from high latency of up to 5,800 milliseconds, making them unsuitable for real-time use. The system's efficiency is further highlighted by its average latency of 1,880 milliseconds, which remains low despite handling complex tasks. Additionally, Figure 4 illustrates how user stickiness and interaction depth improved over iterations, with average session turns increasing from 4.2 to 12.5, indicating deeper engagement.

Of this work extend beyond technical benchmarks to real-world applications where seamless AI interaction is critical. By maintaining conversational continuity, DuCCAE enables more natural and trustworthy digital assistants in domains like customer support, healthcare, and education. The system's ability to integrate asynchronous back into live dialogue means users can receive timely updates without disruption, enhancing overall experience. This approach could set a new standard for industrial AI deployment, balancing responsiveness with robust task execution in high-traffic environments.

Despite its successes, the paper acknowledges several limitations that point to future research directions. One key issue is the trade-off between perception granularity and latency; converting video to text may lose subtle visual cues like micro-expressions, potentially limiting empathetic resonance. The system also incurs significant GPU memory overhead due to its dual-track architecture, raising inference costs. Furthermore, DuCCAE currently relies on Baidu's proprietary infrastructure, including ERNIE foundation models, which may hinder reproducibility for the broader research community. Safety concerns in autonomous tool execution, such as agentic misalignment in high-stakes tasks, remain an area for improvement, with plans to integrate critic-in-the-loop modules for better validation.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn