Martes 7 de mayo 10:00hs, Salón Azul (piso 5, salón 502) – Facultad de Ingeniería, J. Herrera y Reissig 565
El Instituto de Ingeniería Eléctrica de la Facultad de Ingeniería de la Universidad de la República invita a la siguiente charla : “Robust sound recognition in acoustic sensor networks” por Justin Salamon
The combination of remote acoustic sensors with automatic sound recognition represents a powerful emerging technology for studying both natural and urban environments. At NYU we’ve been working on two projects whose aim is to develop and leverage this technology : the Sounds of New York City (SONYC) project is using acoustic sensors to understand noise patterns across NYC to improve noise mitigation efforts, and the BirdVox project is using them for the purpose of tracking bird migration patterns in collaboration with the Cornell Lab of Ornithology. Acoustic sensors present both unique opportunities and unique challenges when it comes to developing machine listening algorithms for automatic sound event detection: they facilitate the collection of large quantities of audio data, but the data is unlabeled, constraining our ability to leverage supervised machine learning algorithms. Training generalizable models becomes particularly challenging when training data come from a limited set of sensor locations (and times), and yet our models must generalize to unseen natural and urban environments with unknown and sometimes surprising confounding factors. In this talk I will present our work towards tackling these challenges along several different lines with neural network architectures, including novel pooling layers that allow us to better leverage weakly labeled training data, self-supervised audio embeddings that allow us to train high-accuracy models with a limited amount of labeled data, and context-adaptive networks that improve the robustness of our models to heterogeneous acoustic environments.
Justin Salamon, PhD :
Justin Salamon is a research scientist and member of the Audio Research Group at Adobe Research in San Francisco. Previously he was a senior research scientist at the Music and Audio Research Laboratory and Center for Urban Science and Progress of New York University. His research focuses on the application of machine learning and signal processing to audio and video signals, with applications in machine listening, music information retrieval, bioacoustics, environmental sound analysis and open source software & data. He holds a B.A. in Computer Science from the University of Cambridge (UK), completed his M.Sc. and Ph.D. in Computer Science with the Music Technology Group of Pompeu Fabra University (Spain), and was a visiting researcher at IRCAM (France). Please visit his personal website for a complete list of publications, research topics, updates and code/data releases.