Senior student helps develop real-time spatial audio technology

2/15/2013 Megan Kelly

The advent of Skype and other video chat applications enabled people to see friends and family in real-time while virtually communicating.

Written by Megan Kelly

The advent of Skype and other video chat applications enabled people to see friends and family in real-time while virtually communicating.

Senior ECE student Ryan Rogowski and CSL researcher Doug Jones are taking video chat to the next level, developing 3D audio systems that can process real-time spatial audio.

In other words, if you’re video chatting with a friend in New York while moving around your Champaign room, your friend will be able to hear your voice travel as though he was in the room with you.

These systems have the potential to not only enhance video chat applications but may transform hearing-aid capabilities as well.

“It’s really cool technology,” Rogowski said. “I kept thinking, ‘How will this be possible?’ while working on it, but we did it.”

Last fall, Rogowski heard that Jones, a professor of electrical and computer engineering, was working on 3D audio systems with real-time spatial audio. Interested, he asked Jones to be his senior thesis adviser and the collaboration got underway. ECE graduate student Nam Nguyen and ECE technician Mark Smart assisted.

Rogowski and Jones record 3D audio using an array of four small microphones, an idea that differs from previous 3D recordings.

“In the past, recording 3D sounds required many expensive microphones, which were implemented on a variety of different systems, whether it be a video camera or hearing aid,” Rogowski said. “We took the 3D sound, recorded it and reproduced it in a stereo headset.”

Rogowski added that they created surround sound in the headset by approximating “head-related transfer functions.” These functions describe the interaction among the head, inner ear and pinna (ear-flaps) to derive audio information. Head-related transfer functions detect how sound waves’ input is filtered and interpreted before it reaches the eardrum and inner ear.

“It turns out by using head-related transfer functions, you can create the change in the sounds that your brain associates with directions,” Rogowski said. “We can manipulate sound so it seems like you’re hearing it in surround sound and not just from (the headset surrounding) your ears.”

In addition, Rogowski and Jones made the system record in real-time, so instead of recording and playing it back, they recorded and listened to it simultaneously.

“At the same time we’re recording in one room, we’d hook (the system) up in a different room and hear the sound move around there,” Rogowski said. “It seemed as if the people we were recording were in the same room as us.”

Rogowski said the systems could be used in video chat and hearing aid applications, among others.

“Currently, if you use a hearing aid, you may only hear background noise without spatial direction,” Rogowski explained. “For example if someone was walking by, a person with a hearing aid might not be able to tell which direction they were coming from unless they were watching the person. This technology could change all that.”

Rogowski said the next step is to find interested businesses and expand on what they’ve already accomplished.

“We’re demonstrating the project to several visiting companies,” Rogowski said. “There’s a little more work to be done in improving sound localization, like how well you can differentiate where the sound comes from.”

After graduating from the College of Engineering this month, Rogowski will study Mandarin in China for a year on a Boren Scholarship. He plans to attend graduate school and build upon his current research.


Share this story

This story was published February 15, 2013.