Microsoft and Georgia Tech Researchers Unveil TongueTap: Cutting-Edge Multimodal Tongue Gesture Recognition for Head-Worn Devices

TL;DR:

  • TongueTap technology enables tongue gesture recognition for head-worn devices, offering hands-free interaction.
  • Microsoft Research and Georgia Tech collaborated to create TongueTap by combining sensors from off-the-shelf headsets.
  • Data synchronization was achieved using the Lab Streaming Layer (LSL).
  • Preprocessing involved low-pass filtering, Independent Component Analysis (ICA), and Principal Component Analysis (PCA).
  • Gesture recognition employed a Support Vector Machine (SVM) with high accuracy.
  • IMU sensors, particularly from the Muse headset, proved most effective for tongue gesture classification.
  • TongueTap has potential applications in augmented reality (AR) interfaces.
  • Further research will explore multi-organ interactions in AR headsets.

Main AI News:

In the dynamic realm of wearable technology, the quest for seamless, hands-free interaction has yielded remarkable breakthroughs. Enter TongueTap, an innovation that harmonizes multiple data streams to enable tongue gesture recognition for the control of head-worn devices—a promising stride forward. This ingenious method empowers users to engage silently, liberating them from the constraints of using their hands or eyes and obviating the need for specialized interfaces typically located inside or near the mouth.

Collaborating with Microsoft Research in Redmond, Washington, USA, the researchers at the Georgia Institute of Technology have unveiled the TongueTap—a groundbreaking tongue gesture interface. This innovation amalgamates sensors from two readily available off-the-shelf headsets, both equipped with IMUs and photoplethysmography (PPG) sensors. Notably, one of the headsets incorporates EEG (electroencephalography), eye tracking, and head tracking sensors. The fusion of data from these two headsets, namely the Muse 2 and Reverb G2 OE devices, is orchestrated through the Lab Streaming Layer (LSL), a time synchronization system commonly employed in multimodal brain-computer interfaces.

The research team meticulously preprocesses the data pipeline. Employing a 128Hz low-pass filter using SciPy, they subject the EEG signals to Independent Component Analysis (ICA). Simultaneously, Principal Component Analysis (PCA) is applied to the other sensors, with each sensor treated independently. For gesture recognition, the researchers harness a Support Vector Machine (SVM) in Scikit-Learn, leveraging a radial basis function (RBF) kernel with hyperparameters C=100 and gamma=1 for binary classification. This discerns whether a given data window contains a gesture or falls into the non-gesture category.

To evaluate tongue gesture recognition, an extensive dataset is collected from 16 participants. One of the most intriguing findings of this study revolves around the effectiveness of various sensors in classifying tongue gestures. Remarkably, the IMU on the Muse headset emerges as the most potent sensor, boasting an 80% classification accuracy when used in isolation. However, the true efficacy lies in multimodal combinations, with a diverse array of PPG sensors achieving an impressive 94% accuracy.

Building on sensors with the highest accuracy, the researchers establish that the IMU positioned behind the ear offers a cost-effective means of detecting tongue gestures. Its strategic placement allows for seamless integration with previous mouth-sensing approaches. A pivotal milestone in the journey toward tangible product applications for tongue gestures is the development of a dependable, user-independent classification model. Achieving real-world viability necessitates a more ecologically valid study design, encompassing multiple sessions and mobility across diverse environments.

TongueTap signifies a significant leap toward achieving smooth and intuitive interaction with wearable devices. Its capacity to discern and categorize tongue gestures using commercially available technology lays the foundation for a future where precise, user-friendly control of head-worn devices becomes a reality. Among the most promising applications for tongue interactions is their utilization in governing augmented reality (AR) interfaces. The researchers intend to delve deeper into this multi-organ interaction paradigm, conducting experiments within AR headsets and drawing comparisons with other gaze-based interactions.

Conclusion:

The development of TongueTap marks a significant stride in the wearable technology market, offering a hands-free interaction solution that can revolutionize user experiences. With the potential to enhance control of head-worn devices and facilitate augmented reality interfaces, TongueTap paves the way for new market opportunities and a more intuitive user interaction paradigm.

Source