EchoSpeech: Artificial Intelligence-Equipped Eyeglasses Can Read Silent Speech

Apr 10, 2023 by News Staff

EchoSpeech, developed by Cornell University researchers, uses speakers and microphones mounted on a glass-frame and emits inaudible sound waves towards the skin. By analyzing echoes from multiple paths, EchoSpeech captures subtle skin deformations caused by silent utterances and uses them to infer silent speech.

EchoSpeech requires just a few minutes of user training data before it will recognize commands and can be run on a smartphone. Image credit: Zhang et al., doi: 10.1145/3544548.3580801.

EchoSpeech requires just a few minutes of user training data before it will recognize commands and can be run on a smartphone. Image credit: Zhang et al., doi: 10.1145/3544548.3580801.

“For people who cannot vocalize sound, this silent speech technology could be an excellent input for a voice synthesizer. It could give patients their voices back,” said Ruidong Zhang, a doctoral student at Cornell University.

In its present form, EchoSpeech could be used to communicate with others via smartphone in places where speech is inconvenient or inappropriate, like a noisy restaurant or quiet library.

The silent speech interface can also be paired with a stylus and used with design software like CAD, all but eliminating the need for a keyboard and a mouse.

Outfitted with a pair of microphones and speakers smaller than pencil erasers, the EchoSpeech glasses become a wearable AI-powered sonar system, sending and receiving soundwaves across the face and sensing mouth movements.

A deep learning algorithm then analyzes these echo profiles in real time, with about 95% accuracy.

“We’re moving sonar onto the body,” said Dr. Cheng Zhang, also from Cornell University.

“We’re very excited about this system, because it really pushes the field forward on performance and privacy.”

“It’s small, low-power and privacy-sensitive, which are all important features for deploying new, wearable technologies in the real world.”

“Most technology in silent-speech recognition is limited to a select set of predetermined commands and requires the user to face or wear a camera, which is neither practical nor feasible.”

“There also are major privacy concerns involving wearable cameras — for both the user and those with whom the user interacts.”

Acoustic-sensing technology like EchoSpeech removes the need for wearable video cameras.

“And because audio data is much smaller than image or video data, it requires less bandwidth to process and can be relayed to a smartphone via Bluetooth in real time,” said Cornell University’s Professor François Guimbretière.

“And because the data is processed locally on your smartphone instead of uploaded to the cloud, privacy-sensitive information never leaves your control.”

The researchers will present their technology this month at the Association for Computing Machinery Conference on Human Factors in Computing Systems (CHI’ 23) in Hamburg, Germany.

_____

Ruidong Zhang et al. 2023. EchoSpeech: Continuous Silent Speech Recognition on Minimally-obtrusive Eyewear Powered by Acoustic Sensing. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (CHI ’23); doi: 10.1145/3544548.3580801

Share This Page