Schedule Your Appointment Today

12780 W North Avenue, Suite B Brookfield, WI 53005

(262) 784-9300

Advanced Hearing Solutions of Wisconsin, S.C. BBB Business Review
Hearing Test

A hearing evaluation is the first step in determining if hearing loss affects you

Our diagnostic hearing evaluation consists of a variety of tests to determine the unique aspects of your hearing loss.

Learn More
Earplugs and Earmolds

We offer custom hearing protection for your professional and recreational needs

We offer custom fitted earmolds and earplugs for maximum protection against sudden and prolonged noise exposure.

Learn More

We're here to help you experience life more fully

Brain-Computer Interface Could Improve Hearing Aids

Researchers are working on the early stages of a brain-computer interface to help program hearing aids

Researchers are working on the early stages of a brain-computer interface (BCI) that can tell who you're listening to in a room full of noise and other people talking. In the future, the technology could be incorporated into hearing aids as tiny beam-formers that point in the direction of interest so that they tune into certain conversations, sounds, or voices that an individual is paying attention to, while tuning out unwanted background noise.

The researchers, Carlos da Silva Souto and coauthors at the University of Oldenburg's Department of Medical Physics and Cluster of Excellence Hearing4all, have published a paper on a proof-of-concept auditory BCI in a recent issue of Biomedical Physics and Engineering Express.

So far, most BCI research has focused primarily on visual stimuli, which currently outperforms systems that use auditory stimuli, possibly because of the larger cortical surface of the visual system compared to the auditory cortex. However, for individuals who are visually impaired or completely paralyzed, auditory-based BCI could offer a potential alternative.

In the new study, 12 volunteers listened to two recorded voices (one male, one female) speaking repeated syllables, and were asked to pay attention to just one of the voices. In early sessions, electroencephalogram (EEG) data on the electrical activity of the brain was used to train the participants. In later sessions, the system was tested on how accurately it classified the EEG data to determine which voice the volunteer was paying attention to.

At first, the system classified the data with about 60 percent accuracy—above the level of chance, but not good enough for practical use. By increasing the analysis time from 20 seconds to 2 minutes, the researchers could improve the accuracy to nearly 80 percent. Although this analysis window is too long for real-life situations, and the accuracy can be further improved, the results are among the first to show that it is possible to classify EEG data evoked from spoken syllables.

In the future, the researchers plan to optimize the classification methods to further improve the system's performance.