We are re-imagining audio technology to empower people to be creative without the tools getting in the way. We are working on dramatically simplifying the creation of audio content, while still maintaining high production value. Our work spans a number of audio research areas including speech enhancement, music information retrieval, speech and music synthesis, computational acoustics, spatial audio, and audio event detection. We also work on problems at the intersection of audio with video, augmented reality, and natural language processing. To advance all of these research areas, we develop new machine learning algorithms, novel signal processing algorithms, and new human computer interaction paradigms.

Meet some of our researchersView More

Juan-Pablo Caceres

Research Engineer

Zeyu Jin

Research Scientist

Paris Smaragdis

Senior Research Scientist

View our latest publicationsView More

MakeItTalk: Speaker-Aware Talking Head Animation

Zhou, Y., Li, D., Shechtman, E., Echevarria, J., Han, X., Kalogerakis, E. (Nov. 30, 2020)

SIGGRAPH Asia 2020

A Differentiable Perceptual Audio Metric Learned from Just Noticeable Differences

Manocha, P., Finkelstein, A., Zhang, R., Bryan, N., Mysore, G., Jin, Z. (Oct. 26, 2020)

Interspeech 2020

Controllable Neural Prosody Synthesis

Morrison, M., Jin, Z., Salamon, J., Bryan, N., Mysore, G. (Oct. 26, 2020)

Interspeech 2020

Project SoundSeek

SoundSeek, an experimental technology, allows users to find any sound in an audio track quickly and easily. All the user has to do is select one or more examples of the target sound, and SoundSeek will find everywhere else in the recording where a similar sound occurs, using a few-shot deep learning model.

View our latest newsView All News

Join us!

We are looking for researchers, engineers, and interns to take our technologies to the next level. We're recruiting, and we would love to hear from you!