Hostname: page-component-77f85d65b8-5ngxj Total loading time: 0 Render date: 2026-03-30T01:06:43.946Z Has data issue: false hasContentIssue false

Interpretable hand gesture recognition: a rule-based framework for radar-based gesture onset detection and classification

Published online by Cambridge University Press:  04 November 2025

Sarah Seifi*
Affiliation:
Technical University of Munich, Munich, Germany Infineon Technologies AG, Neubiberg, Germany
Julius Ott
Affiliation:
Technical University of Munich, Munich, Germany Infineon Technologies AG, Neubiberg, Germany
Cecilia Carbonelli
Affiliation:
Infineon Technologies AG, Neubiberg, Germany
Lorenzo Servadei
Affiliation:
Technical University of Munich, Munich, Germany
Robert Wille
Affiliation:
Technical University of Munich, Munich, Germany Software Competence Center Hagenberg GmbH (SCCH), Hagenberg, Austria
*
Corresponding author: Sarah Seifi; Email: sarah.seifi@tum.de
Rights & Permissions [Opens in a new window]

Abstract

Hand gesture recognition (HGR) has gained significant attention in human-computer interaction, enabling touchless control in various domains, such as virtual reality, automotive systems, and healthcare. While deep learning approaches achieve high accuracy in gesture classification, their lack of interpretability hinders transparency and user trust in critical applications. To address this, we extend MIRA, an interpretable rule-based HGR system, with a novel gesture onset detection method that autonomously identifies the start of a gesture before classification. Our onset detection approach achieves 90.13% accuracy on average, demonstrating its robustness across users. By integrating signal processing techniques, MIRA enhances interpretability while maintaining real-time adaptability to dynamic environments. Additionally, we introduce a background class, enabling the system to differentiate between gesture and non-gesture frames and expand the dataset with new users and recordings to improve generalization. We further analyze how feature diversity affects performance, showing that low diversity can suppress personalization due to early misclassifications. Using a foundational and personalized rule framework, our approach correctly classifies up to 94.9% of gestures, reinforcing the impact of personalization in rule-based systems. These findings demonstrate that MIRA is a robust and interpretable alternative to deep learning models, ensuring transparent decision-making for real-world radar-based gesture recognition.

Information

Type
Research Paper
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NC
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial licence (http://creativecommons.org/licenses/by-nc/4.0), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original article is properly cited. The written permission of Cambridge University Press must be obtained prior to any commercial use.
Copyright
© The Author(s), 2025. Published by Cambridge University Press in association with The European Microwave Association.
Figure 0

Figure 1. Radar signal preprocessing. a) Using Infineon’s XENSIV™ BGT60TR13C $60\,$GHz radar, the raw gesture data are collected in the format $[\text{frames} \times \text{receive channels} \times \text{chirps} \times \text{samples}]$ (depicted only for one frame). b) The range profile is generated, where a local peak search is performed to find the range bin of the hand (marked in blue). c) In total, five features are extracted, bringing the data into the format $[\text{frames} \times \text{features}]$ (depicted only for one frame). d) Frame-based labeling is performed, where the area around $F_{\text{gesture}}$, i.e., the bin with the closest distance to the radar, is labeled as the gesture. All remaining frames are labeled as Background.

Figure 1

Table 1. Illustrative example of foundational and personalized rules. The default-else rule (crossed-out rule) is replaced with personalized rules

Figure 2

Figure 2. Visualization of the dynamic gesture accuracy metric for gesture classification. The blue box represents the true gesture, the green box denotes the tolerance window, and the red box indicates the predicted gesture. A correctly classified gesture falls within the tolerance window and satisfies the duration requirement, whereas a misclassified gesture does not. This metric accommodates slight offsets, providing a more practical evaluation of model performance.

Figure 3

Table 2. MIRA classification parameters

Figure 4

Table 3. Configuration of experimental settings

Figure 5

Figure 3. Gesture onset detection. (a) Range spectrogram and (b) Doppler spectrogram of a recording with a duration of 100 frames. The yellow overlay indicates the gesture duration. (c) Frame energy over time showing the energy (blue dotted line) and filtered (smoothed) energy (black solid line). The red marker denotes the detected gesture onset using the filtered signal, while peaks below the red-dashed threshold line are not considered. The green marker denotes a suboptimal onset detected in the unfiltered signal. The purple-shaded region represents the ten-frame interval predicted as a gesture based on the detected onset.

Figure 6

Table 4. Dynamic gesture onset detection accuracy for filtered and unfiltered frame energy signals across twelve users

Figure 7

Figure 4. Impact of feature distribution mismatch on classification accuracy. (a) Confusion matrix for user$_{12}$ showing high misclassification as Push in configuration$_1$. (b) Range vs. Doppler plot showing misclassification region based on the first rule. (c) Range feature distribution comparison for Push class between training and user$_{12}$.

Figure 8

Table 5. Feature variability and classification accuracy across experimental configurations

Figure 9

Figure 5. Recursive ablation heatmap showing the accuracy drop per feature per round. Top: Foundational rules. Bottom: Personalized rules. White cells indicate removed features.

Figure 10

Table 6. Impact of calibration size on personalized accuracy for configuration$_1$ and configuration$_6$