This website uses cookies primarily for visitor analytics. Certain pages will ask you to fill in contact details to receive additional information. On these pages you have the option of having the site log your details for future visits. Indicating you want the site to remember your details will place a cookie on your device. To view our full cookie policy, please click here. You can also view it at any time by going to our Contact Us page.

These innovative hearing aids can read lips through masks

07 September 2022

An international team of engineers and computing scientists have developed a new technology, combining radio frequency with AI, to read lips with remarkable accuracy, even through face masks.

The system, when integrated with conventional hearing aid technology, could help tackle the ‘cocktail party effect’, a common shortcoming of traditional hearing aids.
Currently, hearing aids assist hearing-impaired people by amplifying all ambient sounds around them, which can be helpful in many aspects of everyday life.
However, in noisy situations such as cocktail parties, hearing aids’ broad spectrum of amplification can make it difficult for users to focus on specific sounds, like conversation with a particular person.
One potential solution to the cocktail party effect is to make ‘smart’ hearing aids, which combine conventional audio amplification with a second device to collect additional data for improved performance.
While other researchers have had success in using cameras to aid with lip reading, collecting video footage of people without their explicit consent raises concerns for individual privacy. Cameras are also unable to read lips through masks, an everyday challenge for people who wear face coverings for cultural or religious purposes and a broader issue in the age of COVID-19.
In a new paper published today in the journal Nature Communications, the University of Glasgow-led team outline how they set out to harness cutting-edge sensing technology to read lips. Their system preserves privacy by collecting only radio-frequency data, with no accompanying video footage. 
To develop the system, the researchers asked male and female volunteers to repeat the five vowel sounds (A, E, I, O, and U) first while unmasked and then while wearing a surgical mask.
As the volunteers repeated the vowel sounds, their faces were scanned using radio-frequency signals from both a dedicated radar sensor and a Wi-Fi transmitter. Their faces were also scanned while their lips remained still.
Then, the 3,600 samples of data collected during the scans was used to ‘teach’ machine learning and deep learning algorithms how to recognise the characteristic lip and mouth movements associated with each vowel sound.
Because the radio-frequency signals can easily pass through the volunteers’ masks, the algorithms could also learn to read masked users’ vowel formation.
The system proved to be capable of correctly reading the volunteers’ lips most of the time. Wifi data was correctly interpreted by the learning algorithms up to 95 percent of the time for unmasked lips, and 80 percent for masked. 
Meanwhile, the radar data was interpreted correctly up to 91 percent without a mask, and 83 percent of the time with a mask.
Dr Qammer Abbasi, of the University of Glasgow’s James Watt School of Engineering, is the paper’s lead author. He said: “Around five percent of the world’s population – about 430 million people – have some kind of hearing impairment.
“Hearing aids have provided transformative benefits for many hearing-impaired people. A new generation of technology which collects a wide spectrum of data to augment and enhance the amplification of sound could be another major step in improving hearing-impaired people’s quality of life.
“With this research, we have shown that radio-frequency signals can be used to accurately read vowel sounds on people’s lips, even when their mouths are covered. While the results of lip-reading with radar signals are slightly more accurate, the Wi-Fi signals also demonstrated impressive accuracy.
“Given the ubiquity and affordability of Wi-Fi technologies, the results are highly encouraging which suggests that this technique has value both as a standalone technology and as a component in future multimodal hearing aids.” 
Professor Muhammad Imran, Head of the University of Glasgow’s Communications, Sensing and Imaging research group and a co-author of the paper, added: “This technology is an outcome from two research projects funded by the Engineering and Physical Sciences Research Council (EPSRC), called COG-MHEAR and QUEST.
“Both aim to find new methods of creating the next generation of healthcare devices, and this development will play a major role in supporting that goal.”

More information...

Print this page | E-mail this page