China Automotive Gesture Interaction Development Research Report,2022-2023
  • Mar.2023
  • Hard Copy
  • USD $3,800
  • Pages:186
  • Single User License
    (PDF Unprintable)       
  • USD $3,600
  • Code: LYX001
  • Enterprise-wide License
    (PDF Printable & Editable)       
  • USD $5,400
  • Hard Copy + Single User License
  • USD $4,000
      

Vehicle gesture interaction research: in 2022, the installations rocketed by 315.6% year on year.

China Automotive Gesture Interaction Development Research Report, 2022-2023 released by ResearchInChina analyzes and studies four aspects: gesture interaction technology, benchmarking vehicle gesture interaction solutions, gesture interaction industry chain, and gesture interaction solution providers.

1. In 2022, the installations of vehicle gesture recognition functions soared by 315.6% on an annual basis.

Accompanied by iterative upgrade of intelligent cockpit technology, cockpit services are evolving from passive intelligence to active intelligence, and the human-computer interaction mode is also shifting from single-modal to multi-modal interaction. In this trend, vehicle gesture interaction functions enjoy a boom. In 2022, gesture recognition (standard configuration) was installed in 427,000 passenger cars in China, a year-on-year spurt of 315.6%, with installation rate up to 2.1%, 1.6 percentage points higher than 2021.

手势交互 1_副本.png

By brand, in 2022 Changan Automobile boasted the highest gesture recognition installation rate, up to 33.0%, 13.1 percentage points higher than 2021. In terms of models, in 2022, Changan Automobile had a total of 6 models (e.g., UNI-V, CS75 and UNI-K) equipped with gesture recognition as a standard configuration, 5 models more than in 2021.

手势交互 2_副本.png

The gesture recognition feature of Changan UNI-K adopts a 3D ToF solution, enabling such functions as song switch and navigation activation. The specific gestures are: swipe the palm horizontally to left/right for playing the previous/next song; make a finger heart for navigating back home; thumb up for navigating to the workplace.

手势交互 3_副本.png

2. The control scope of gesture recognition is extending from software to hardware, and from the inside to the outside of cars.

As gesture interaction technology gains popularity and finds application in ever more scenarios, vehicle gesture interaction also springs up. At present, automakers are working hard on layout of cockpit interaction functions. Gesture controlled functions have increased from initially in-vehicle infotainment system features (e.g., phone call, media volume and navigation), to body hardware and safety systems (e.g., windows/sunroof/sunshades, doors, and driving).

In addition, manufacturers also make efforts to develop exterior gesture control technology. One example is WEY Mocha that has allowed for gesture control over ignition, forward/backward movement, stop, and flameout outside the car. In the future, gesture recognition will no longer be limited to occupants, and will gradually cover actions of passers-by outside the car, for instance, recognizing command gestures of traffic police on road or gestures made by cyclists around the car.

3. Six gesture recognition technology routes.

From the perspective of technology route, gesture recognition technologies are led by 3D camera based structured light, ToF, stereo imaging, radar-based millimeter wave, ultrasonics, and bioelectricity-based myoelectricity.

手势交互 4_副本.png

In current stage, 3D camera based gesture sensing prevails among vehicle gesture recognition technology routes. The technology route consists of 3D camera and control unit. Composed of camera, infrared LED and sensor, the 3D camera is used to capture hand movements, and then recognize the type of gestures via corresponding image processing algorithms and issue relevant instructions. The 3D camera based technology route can be subdivided into structured light, ToF and stereo vision.

1. Structured light technology refers to a solution where the light with coded information is projected onto the human body, the infrared sensor collects the reflected structural pattern, and finally the processor builds a 3D model. With benefits of mature hardware, high recognition accuracy and high resolution, it is applicable to close-range scenarios within 10 meters. The gesture recognition carried by Neta S rolled out in July 2022 is a structured light solution.

The in-cabin gesture recognition sensor of Neta S is located above the interior rearview mirror. It can recognize 6 gestures, including: swipe the palm back and forth to adjust the light transmittance of the sunroof; make the "shh" sign for the silent mode; rotate a finger clockwise/counterclockwise to adjust the volume; move the palm to left and right to switch audio and video programs; make the "V" sign to take selfies in the car; thumb up to save favorite programs.

手势交互 5_副本.png

2. ToF technology, namely, time-of-flight (ToF) based ranging, is enabled with 3D images constructed by the underlying photosensitive elements. It can obtain effective and real-time depth information within 5 meters. With the applicability to a wider range of scenarios, it acquires effective depth of field information regardless of whether the ambient light is strong (e.g., sunlight) or weak. The gesture recognition solutions installed in production models like BMW iX, Li Auto L9 and new ARCFOX αS HI Edition are all ToF solutions.

The in-cabin gesture recognition sensor of BMW iX lies at the dome light above the center console screen. It can recognize 8 gestures, including:
① swipe hand to left and right to reject phone call/close the pop-up;
② point the index finger back and forth to answer phone call/confirm the pop-up;
③ rotate the index finger clockwise to turn the volume up or zoom in on the navigation map;
④ rotate the index finger counterclockwise to turn the volume down or zoom out on the navigation map;
⑤move a fist with thumb extended to left right back and forth to play the previous/next song;
⑥ point the index and middle fingers extended into the display to perform individually assignable gesture;
⑦ stretch out all five fingers, make a fist and then stretch out all five fingers again to perform individually assignable gesture;
⑧ bring thumb and index finger together and swipe the hand to the right or left for a view around the car (requiring the car to pack the automated parking assist system PLUS).  

手势交互 6_副本.png

To ensure gesture recognition and control by occupants, Li Auto L9 has gesture recognition sensors installed in the fore cabin and rear cabin. The fore cabin sensor is located above the interior rearview mirror, and the rear one lies above the rear entertainment screen.

The fore cabin sensor can recognize 2 gestures, including:
①point towards windows/sunroof/sunshades to control (combined with voice interaction capability);
②make a fist and hold, and swipe up and down on the play page to adjust the volume.   

The rear cabin sensor can recognize 7 gestures, including: ①stretch out all five fingers and place the inner side of elbow on the armrest for 2 seconds to activate the gesture control function; ②stretch out all five fingers and swipe the hand down to turn on the screen; ③stretch out all five fingers and swipe the hand to move the cursor; ④stretch out all five fingers and make a fist to spot the icon; ⑤stretch out all five fingers, make a fist and hold, and move the hand to share the content on the rear entertainment screen to the front display; ⑥stretch out all five fingers, make a fist and hold, and swipe the hand on the play page to left and right to adjust the play progress; ⑦stretch out all five fingers and swipe the hand up to exit the current content. 

手势交互 7_副本.png

3. Stereo imaging technology based on the parallax principle is enabled with 3D geometric information of objects that is obtained from multiple images. This technology is a cost-effective solution posing low requirements for hardware and needing no additional special device. The gesture recognition solution carried by the Mercedes-Benz EQS launched in May 2022 is a stereo imaging solution.

The in-cabin gesture recognition sensor of Mercedes-Benz EQS is located at the reading light on the roof, and can recognize 3 gestures, including:
① make the "V" sign to call up favorites;
②swipe hand back and forth under the interior rearview mirror to control the sunroof;
③swipe hand toward the inside of the car to automatically close doors (requiring optional four-door electric switches). 

手势交互 8_副本.png


Currently gesture recognition technologies such as radar-based millimeter wave, ultrasonics, and bioelectricity-based myoelectricity have yet to be used widely in in-cabin gesture recognition functions. Compared with conventional vision-based gesture recognition, these technologies still have some limitations and pose challenges.  

1. Radar is a radio wave sensor that enables accurate detection of the position and movements of hands even in the presence of obstacles.

In 2020, Ainstein, the American subsidiary of Muniu Technology, together with ADAC Automotive established a joint venture brand - RADAC. At the CES 2020, Ainstein introduced a radar–based vehicle gesture recognition solution. The gesture recognition sensor in this solution lies on the top of the tailgate, allowing users to open the door by swiping hand to left and right.

手势交互 9_副本.png

2. Ultrasonic radar. In February 2020, DS showcased the Aero Sport Lounge concept car at the Geneva International Motor Show. Integrating Leap Motion and Ultrahaptics technologies, this car can easily recognize and understand every gesture made by occupants, and give haptic feedback to them through the stereo ultrasonic waves emitted by the micro-speaker.

The in-cabin gesture recognition and ultrasonic feedback sensor of DS Aero Sport Lounge is located at the center armrest of the car, and can recognize 5 gestures, including:
①adjust in-cabin temperature and blowing velocity;
②adjust tracks and volume;
③process navigation/map, including new route settings;
④answer/reject phone calls;
⑤switch menu functions.

手势交互 10_副本.png

3. Bioelectricity refers to the electric signals generated by human muscular movements. Bioelectric sensors can recognize gestures and movements by measuring these signals. At present, the bioelectricity-based myoelectric gesture recognition technology is more used to control external devices and interaction interfaces, such as prosthetics, virtual reality and gaming devices. Thalmic Labs, a Canadian company dedicated to developing smart gesture control products, introduced the first wearable device, the MYO armband, which uses myoelectricity technology. The eight myoelectric sensors embedded in the armband record the electric signals of arm muscles, and recognize different gestures by analyzing these signals. In the actual application, users can control drones, computers, smartphones and other electronic devices through the Bluetooth connect of MYO. There are no vehicle use cases at present.

手势交互 11_副本.png

1 Overview of Gesture Interaction
1.1 Introduction to Gesture Interaction
1.2 Key Features of Gesture Interaction
1.3 Application Scenarios of Gesture Interaction
1.3.1 Mobile Devices
1.3.2 Smart Wearables
1.3.3 Smart Home
1.3.4 Outdoor/Indoor Experience Areas
1.3.5 Automobiles
1.4 Development Route of Automotive Gesture Interaction 
1.4.1 Development Route of Gesture Interaction and Intelligent Cockpit 
1.4.2 Development Route of Gesture Interaction and Multimodal Interaction
1.4.3 Installation History of Automotive Gesture Interaction
1.4.4 Development History and Trends of Intelligent Cockpit Interaction Scenarios (1)
1.4.5 Development History and Trends of Intelligent Cockpit Interaction Scenarios (2)
1.4.6 Development History and Trends of Intelligent Cockpit Interaction Scenarios (3) 
1.4.7 Development Trends of Gesture Interaction (1)
1.4.8 Development Trends of Gesture Interaction (2)
1.4.9 Development Trends of Gesture Interaction (3)

2 Gesture Interaction Industry Chain
2.1 Types of Gesture Interaction Technology
2.2 Development History of Gesture Interaction Technology by Type
2.3 Gesture Interaction Industry Chain
2.4 Gesture Interaction Algorithm
2.5 Gesture Interaction Solution Providers
2.6 Trends of Gesture Interaction Patent Filings
2.7 TOP10 Companies by Number of Gesture Interaction Patents

3 Gesture Interaction Solutions for Benchmarking Models
3.1 Installation of Vehicle Gesture Recognition 
3.1.1 Installations & Installation Rate 
3.1.2 Ranking of Brands
3.1.3 Ranking of Vehicle Models
3.1.4 Price Features 
3.1.5 Installation of Gesture Function in New Models in 2022
3.1.6 Price Features of New Models in 2022
3.2 Cockpit Interaction Modes and Gesture Interaction Functions of Major Models
3.3 Gesture Interaction Solutions for Benchmarking Models
3.3.1 BMW
3.3.2 Li Auto
3.3.3 ARCFOX 
3.3.4 Mercedes-Benz
3.3.5 Neta
3.3.6 RADAC & DS & BYTON

4 Gesture Interaction Solution Providers
4.1 Cipia Vision (Eyesight Technologies)
4.1.1 Profile
4.1.2 Gesture Interaction Products 
4.1.3 In-cabin Interaction Solutions
4.1.4 In-cabin Solutions 
4.2 Ultraleap (Leap Motion)
4.2.1 Profile
4.2.2 Hand Tracking Technology
4.2.3 Hand Tracking Hardware
4.2.4 Gesture Tracking Software
4.2.5 Mid-air Haptic Technology
4.2.6 Haptic Feedback Kit
4.2.7 Technical Reference Platform
4.2.8 Application of Products: Gesture Solution Based on AR Helmet
4.2.8 Application of Products: Outdoor Gesture Solution
4.2.8 Application of Products: Vehicle Gesture Solution
4.2.9 Development Route
4.3 Aptiv
4.3.1 Profile
4.3.2 Solutions
4.3.3 In-cabin Sensing Platform
4.3.4 Gesture Recognition Technology
4.4 Cerence Inc.
4.4.1 Profile 
4.4.2 Cockpit Interaction Solutions
4.4.3 Core Technologies
4.4.4 Human-Computer Interaction: Eyesight + Gesture/Voice
4.4.4 Human-Computer Interaction: Gesture
4.4.5 Development Plan
4.5 Melexis
4.5.1 Profile
4.5.2 ToF Sensor Chip
4.5.3 Gesture Interaction
4.6 SenseTime
4.6.1 Profile
4.6.2 Core Technologies of Smart Cars
4.6.3 Smart Car Solutions 
4.6.4 SenseAuto Cabin
4.6.5 In-cabin Interaction Technology
4.7 uSens
4.7.1 Profile
4.7.2 Core Technologies
4.7.3 Product Layout
4.7.4 Gesture Recognition System Solutions
4.7.5 Strategic Planning
4.7.6 Development Route of Interaction
4.8 Geefish Technology
4.8.1 Profile
4.8.2 ToF Modules for Gesture Recognition 
4.8.3 In-depth Human Vehicle Interaction Solution
4.8.4 Gesture Recognition Technology
4.8.5 Exploration Route
4.9 iGentAI 
4.9.1 Profile
4.9.2 IVI Solutions

5 Appendix: Summary of Automotive Gesture Interaction Patents
5.1 Summary of Gesture Interaction Patents of OEMs
5.2 Summary of Gesture Interaction Patents of Suppliers
5.3 Summary of Gesture Interaction Patents of Research Institutes
5.4 Summary of Gesture Interaction Patents of Colleges and Universities 
5.5 Summary of Gesture Interaction Patents of Individuals 
 

Intelligent Cockpit Domain Controller and SoC Market Analysis Report, 2023Q2

Cockpit domain controller and chip in 2023Q2: by intelligent cockpit level, L1 surged by 105% on a like-on-like basis, and L2 soared by 171%.On May 17, 2023, the “White Paper on Automotive Intelligent...

Intelligent Vehicle E/E Architecture Research Report, 2023

E/E Architecture Research: How will the zonal EEA evolve and materialize from the perspective of supply chain deployment?Through the lens of development trends, automotive EEA (Electronic/electrical A...

China Passenger Car Brake-by-wire Industry Report, 2023

Passenger car brake-by-wire research: One-box solution takes an over 50% share. China Passenger Car Brake-by-wire Industry Report, 2023 released by ResearchInChina combs through and summarizes passe...

Smart Car OTA Industry Report, 2023

Vehicle OTA Research: OTA functions tend to cover a full life cycle and feature SOA and central supercomputing.In the trend for software-defined vehicles, OTA installations are surging, and software i...

Intelligent Vehicle Multi-Domain Computing Industry Report, 2023

Multi-domain computing research: in the coming first year of cross-domain fusion, major suppliers will quicken their pace of launching new solutions. As vehicle intelligence develops, electrical/ele...

Automotive Head-up Display (HUD) Industry Report, 2023

Automotive HUD research: in the "technology battle" in AR-HUD, who will be the champion of mass production? Automotive head-up display (HUD) works on the optical principle for real-time display of s...

Automotive Cloud Service Platform Industry Report, 2023

Research on Automotive Cloud Services: As Dedicated Automotive Cloud Platforms Are Launched, the Market Enters A Phase of Differentiated Competition  1. The exponentially increasing amount of v...

Global and China Automotive Gateway Industry Report, 2023

Automotive gateway research: integrated gateways have become an important trend in zonal architecture. Automotive gateway is a core component in the automotive electronic/electrical architecture. As ...

In-vehicle Communication and Network Interface Chip Industry Report, 2023

In-vehicle communication chip research: automotive Ethernet is evolving towards high bandwidth and multiple ports, and the related chip market is growing rapidly. By communication connection form, au...

China Autonomous Driving Data Closed Loop Research Report, 2023

Data closed loop research: in the stage of Autonomous Driving 3.0, work hard on end-to-end development to control data. At present, autonomous driving has entered the stage 3.0. Differing from the s...

ADAS and Autonomous Driving Tier 1 Research Report, 2023 - Foreign Companies

Research on foreign ADAS Tier 1 suppliers: 4D radar starts volume production, and CMS becomes a new battlefield. 1. Global Tier 1 suppliers boast complete ADAS/AD product matrix, and make continuous...

China Passenger Car Driving-parking Integrated Solution Industry Report, 2023

Research on driving-parking integration: with the declining share of the self-development model, suppliers' solutions blossom. Local suppliers lead the driving-parking integration market. The statis...

Passenger Car Cockpit Entertainment Research Report, 2023

Cockpit entertainment research: vehicle games will be the next hotspot. The Passenger Car Cockpit Entertainment Research Report, 2023 released by ResearchInChina combs through the cockpit entertainme...

Smart Road - Roadside Perception Industry Report, 2023

Roadside perception research: evolution to integration, high performance and cost control.In June 2023, at a regular policy briefing of the State Council the Ministry of Industry and Information Techn...

China Passenger Car ADAS Domain Controller,Master Chip Market Data and Supplier Research Report, 2023Q1

Quarterly Report on ADAS Domain Controllers: L2+ and above ADAS Domain Controller Master Chip Market Structure    This report highlights the passenger car L2+ and above (including L2+, ...

Automotive Cockpit Domain Controller Research Report, 2023

Research on cockpit domain controllers: various forms of products are mass-produced and mounted on vehicles, and product iteration speeds up. Both quality and quantity have been improved, and the it...

Chinese Passenger Car OEMs’ Overseas Layout Research Report, 2023

OEMs’ overseas layout research: automobile exports are expected to hit 7.18 million units in 2025.  1. China’s automobile export market bucked the trend. During 2021-2022, the global economy ...

Global and Chinese Automakers’ Modular Platform and Technology Planning Research Report, 2023

Research on modular platforms: explore intelligent evolution strategy of automakers after modular platforms become widespread. By analyzing the planning of international automakers, Chinese conventi...

2005- www.researchinchina.com All Rights Reserved 京ICP备05069564号-1 京公网安备1101054484号