DUBLIN–(BUSINESS WIRE)–The “China Automotive Multimodal Interaction Development Research Report, 2025” report has been added to ResearchAndMarkets.com’s offeringDUBLIN–(BUSINESS WIRE)–The “China Automotive Multimodal Interaction Development Research Report, 2025” report has been added to ResearchAndMarkets.com’s offering

China Automotive Multimodal Interaction Development Research Report 2025: Closed-Loop Evolution of Multimodal Interaction – Progressive Evolution of L1~L4 Intelligent Cockpits – ResearchAndMarkets.com

2026/01/26 23:36
7 min read

DUBLIN–(BUSINESS WIRE)–The “China Automotive Multimodal Interaction Development Research Report, 2025” report has been added to ResearchAndMarkets.com’s offering.

Research on Automotive Multimodal Interaction: The Interaction Evolution of L1~L4 Cockpits.

This report comprehensively sorts out the installation of Interaction Modalities in automotive cockpits, multimodal interaction patents, mainstream cockpit interaction modes, application of interaction modes in key vehicle models launched in 2025, cockpit interaction solutions of automakers/suppliers, and integration trends of multimodal interaction.

I. Closed-Loop Evolution of Multimodal Interaction: Progressive Evolution of L1~L4 Intelligent Cockpits

According to the “White Paper on Automotive Intelligent Cockpit Levels and Comprehensive Evaluation” jointly released by the China Society of Automotive Engineers (China-SAE), five levels of intelligent cockpits are defined: L0-L4.

As a key driver for cockpit intelligence, multimodal interaction capability relies on the collaboration of AI large models and multiple hardware to achieve the fusion processing of multi-source interaction data. On this basis, it accurately understands the intentions of drivers and passengers and provides scenario-based feedback, ultimately achieving natural, safe, and personalized human-machine interaction. Currently, the automotive intelligent cockpit industry is generally in the L2 stage, with some leading manufacturers exploring and moving towards the L3.

The core feature of L2 intelligent cockpits is “strong perception, weak cognition”. In the L2 stage, the multimodal interaction function of cockpits achieves signal-level fusion. Based on multimodal large model technology, it can “understand users’ ambiguous intentions” and “simultaneously process multiple commands” to execute users’ immediate and explicit commands. At present, most mass-produced intelligent cockpits can enable this.

In the case of Li i6, it is equipped with MindGPT-4o, the latest multimodal model which boasts understanding and response capabilities with ultra-long memory and ultra-low latency, and features more natural language generation. It supports multimodal “see and speak” (voice + vision fusion search: allowing illiterate children to select the cartoons they want to watch by describing the content on the video cover); multimodal referential interaction (voice + gesture: Voice reference to objects: while issuing commands, extend the index finger: pointing left can control the window and complete vehicle control. Voice reference to personnel: passengers in the same row can achieve voice control over designated personnel through gesture and voice coordination, e.g., pointing right and saying “Turn on the seat heating for him”).

The core feature of L3 intelligent cockpits is “strong perception, strong cognition”. In the L3 stage, the multimodal interaction function of cockpits achieves cognitive-level fusion. Relying on large model capabilities, the cockpit system can comprehensively understand the complete current scenario and actively initiate reasonable services or suggestions without the user issuing explicit commands.

The core feature of L4 intelligent cockpits is “full-domain cognition and autonomous evolution”, creating a “full-domain intelligent manager” for users. In the L4 stage, the application of intelligent cockpits will go far beyond the tool attribute and become a “digital twin partner” that can predict users’ unspoken needs, have shared memories, and dispatch all resources for users. Its core experience is: before the user clearly perceives or expresses the need, the system has completed prediction and planning and entered the execution state.

II. Multimodal AI Agent: Understand What You Need and Predict What You Think

AI Agent can be regarded as the core execution unit and key technical architecture for the specific implementation of functions in the evolution of intelligent cockpits from L2 to L4. By integrating voice, vision, touch and situational information, AI Agent can not only “understand” commands, but also “see” the environment and “perceive” the state, thereby integrating the original discrete cockpit functions into a coherent, active and personalized service process.

Agent applications under L2 can be regarded as “enhanced command execution”, which is the ultimate extension of L2 cockpit interaction capabilities. Based on large model technology, the cockpit system decomposes a user’s complex command into multiple steps and then calls different Agent tools to execute them.

In the next level of intelligent cockpits, Agent applications will change from “you say, I do” to “I watch, I guess, I suggest, let’s do it together”. Users do not need to issue any explicit commands. They just sigh and rub their temples, and the system can comprehensively judge data from “camera” (tired micro-expressions), “biological sensors” (heart rate changes), “navigation data” (continuous driving for 2 hours), and “time” (3 pm (afternoon sleepiness period)) via the large model to know that “the user is in the tired period of long-distance driving and has the need to rest and refresh”.

Based on this, the system will take the initiative to initiate interaction: “You seem to need a rest. There is a service zone* kilometers ahead with your favorite ** coffee. Do you need me to turn on the navigation? At the same time, I can play refreshing music for you.” After the user agrees, the system then calls navigation, entertainment and other Agent tools.

Key Topics Covered:

1 Overview of Multimodal Interaction in Automotive Cockpits
1.1 Development Stages of Intelligent Cockpits

1.2 Definition of Multimodal Interaction

1.3 Development System of Multimodal Interaction

1.4 Introduction to Core Interaction Modality Technologies: Haptic Interaction

1.5 Application Scenarios of Large Models in Intelligent Cockpits

1.6 Vehicle-Human Interaction Functions Based on Multimodal AI Large Models

1.7 Industry Chain of Multimodal Interaction

1.8 Industry Chain of Multimodal AI Large Models

1.9 Policy Environment for Multimodal Interaction

1.10 Installation of Interaction Modalities in Cockpits

2 Summary of Patents Related to Automotive Multimodal Interaction
2.1 Summary of Patents Related to Haptic Interaction

2.2 Summary of Patents Related to Auditory Interaction

2.3 Summary of Patents Related to Visual Interaction

2.4 Summary of Patents Related to Olfactory Interaction

2.5 Summary of Patents Related to Other Featured Interaction Modalities

3 Multimodal Interaction Cockpit Solutions of OEMs
3.1 BYD

3.2 SAIC IM Motors

3.3 FAW Hongqi

3.4 Geely

3.5 Great Wall Motor

3.6 Chery

3.7 Changan

3.8 Voyah

3.9 Li Auto

3.10 NIO

3.11 Leapmotor

3.12 Xpeng

3.13 Xiaomi

3.14 BMW

4 Multimodal Cockpit Solutions of Suppliers
4.1 Desay SV

4.2 Joyson Electronics

4.3 SenseTime

4.4 iFLYTEK

4.5 Thundersoft

4.6 Huawei

4.7 Baidu

4.8 Banma Zhixing

5 Application Cases of Multimodal Interaction Solutions for Typical Vehicle Models
5.1 Summary of Application Cases of Multimodal Interaction Solutions for Typical Vehicle Models

5.2 All-New IM L6

5.2.1 Panoramic Summary of Multimodal Interaction Functions

5.2.2 Analysis of Featured Modal Interaction Capabilities

5.3 Fangchengbao Bao 8

5.4 Hongqi Jinkuihua Guoya

5.5 Denza N9

5.6 Zeekr 9X

5.7 Geely Galaxy A7

5.8 Leapmotor B10

5.9 Li i6

5.10 Xpeng G7

5.11 Xiaomi YU7

5.12 MAEXTRO S800

5.13 2025 AITO M9

5.14 All-New BMW X3 M50

5.15 2026 Audi E5 Sportback

5.16 All-New Mercedes-Benz Electric CLA

6 Summary and Development Trends of Multimodal Interaction
6.1 Summary of Large Model Configuration Parameters of OEMs

6.2 Trend 1: Evolution of Multimodal Interaction Based on AI Large Models}

6.3 Trend 2: Cockpit Scenario Application Cases

6.4 Trend 3 (Voice Interaction)

6.5 Trend 4 (Visual Interaction)

For more information about this report visit https://www.researchandmarkets.com/r/layqmj

About ResearchAndMarkets.com
ResearchAndMarkets.com is the world’s leading source for international market research reports and market data. We provide you with the latest data on international and regional markets, key industries, the top companies, new products and the latest trends.

Contacts

ResearchAndMarkets.com

Laura Wood, Senior Press Manager

press@researchandmarkets.com

For E.S.T Office Hours Call 1-917-300-0470

For U.S./ CAN Toll Free Call 1-800-526-8630

For GMT Office Hours Call +353-1-416-8900

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Microsoft Corp. $MSFT blue box area offers a buying opportunity

Microsoft Corp. $MSFT blue box area offers a buying opportunity

The post Microsoft Corp. $MSFT blue box area offers a buying opportunity appeared on BitcoinEthereumNews.com. In today’s article, we’ll examine the recent performance of Microsoft Corp. ($MSFT) through the lens of Elliott Wave Theory. We’ll review how the rally from the April 07, 2025 low unfolded as a 5-wave impulse followed by a 3-swing correction (ABC) and discuss our forecast for the next move. Let’s dive into the structure and expectations for this stock. Five wave impulse structure + ABC + WXY correction $MSFT 8H Elliott Wave chart 9.04.2025 In the 8-hour Elliott Wave count from Sep 04, 2025, we saw that $MSFT completed a 5-wave impulsive cycle at red III. As expected, this initial wave prompted a pullback. We anticipated this pullback to unfold in 3 swings and find buyers in the equal legs area between $497.02 and $471.06 This setup aligns with a typical Elliott Wave correction pattern (ABC), in which the market pauses briefly before resuming its primary trend. $MSFT 8H Elliott Wave chart 7.14.2025 The update, 10 days later, shows the stock finding support from the equal legs area as predicted allowing traders to get risk free. The stock is expected to bounce towards 525 – 532 before deciding if the bounce is a connector or the next leg higher. A break into new ATHs will confirm the latter and can see it trade higher towards 570 – 593 area. Until then, traders should get risk free and protect their capital in case of a WXY double correction. Conclusion In conclusion, our Elliott Wave analysis of Microsoft Corp. ($MSFT) suggested that it remains supported against April 07, 2025 lows and bounce from the blue box area. In the meantime, keep an eye out for any corrective pullbacks that may offer entry opportunities. By applying Elliott Wave Theory, traders can better anticipate the structure of upcoming moves and enhance risk management in volatile markets. Source: https://www.fxstreet.com/news/microsoft-corp-msft-blue-box-area-offers-a-buying-opportunity-202509171323
Share
BitcoinEthereumNews2025/09/18 03:50
XRP Buyers Defend Most Major 200-Week Price Average: Can It Be Bottom of 2026?

XRP Buyers Defend Most Major 200-Week Price Average: Can It Be Bottom of 2026?

The post XRP Buyers Defend Most Major 200-Week Price Average: Can It Be Bottom of 2026? appeared on BitcoinEthereumNews.com. XRP has returned to its 200-week moving
Share
BitcoinEthereumNews2026/02/08 19:49
Expert Tags Ethereum’s ERC-8004 Mainnet Launch An “iPhone Moment”, Here’s What It Means

Expert Tags Ethereum’s ERC-8004 Mainnet Launch An “iPhone Moment”, Here’s What It Means

Market analyst says Ethereum is having an “iPhone moment” as it approaches the ERC-8004 mainnet launch.
Share
Coinstats2026/02/08 19:56