Cadenza logo
The 1st Cadenza Workshop on Machine Learning Challenges to Improve Music for People with Hearing Loss (Cadenza-2023)
EPSRC logo

Programme

Cadenza-2023 will be a one-day online workshop.

Programme will be available soon

GMT    US East  Taipei 
12:00 07:00 20:00 Welcome
12:10 07:10 20:10 Keynote 1
12:40 07:40 20:40 Cadenza 1 - Overview / Results
13:10 08:10 21:10 Break
13:20 08:20 21:20 Challenge papers
14:20 09:20 22:20 Break
14:30 09:30 22:30 Cadenza Team Submissions
15:00 10:00 23:00 Invited Talks
15:55 10:55 23:55 Future Challenges
16:10 11:10 24:100 Close

Keynote 1

Session chair: Trevor Cox
12:10-12:40 Intelligent mixing for the hearing impaired Joshua Reiss (Queen Mary University of London)
Joshua Reiss

Joshua Reiss
Centre for Digital Music at Queen Mary University of London, UK

Intelligent mixing for the hearing impaired

Intelligent mixing for the hearing impaired

Abstract

Hearing aids are limited in how they can address the enjoyment of produced audio content, partly due to the fact that they can only apply processing on the mixed content received by the listener. However, music and broadcast content may be distributed as multitracks, where individual sources or groups of sources are stored separately, allowing for them to be remixed for different situations. In particular, they may be mixed for a given listener hearing profile. Furthermore, hearing loss simulators allow one to establish the profile of hearing impaired listeners. In this talk, we will discuss recent attempts to automatically re-mix multitrack music content such that the new content, when passed through a hearing loss simulator, will sound as close as possible to how the original mix sounds to normal hearing listeners. The focus here is not on the de-mixing or presence of noise aspects of recent Cadenza challenges, but the potential scenarios that arise when high quality multitracks can be accessed. We will also discuss the challenges in this area of research, and the opportunities for delivering far superior and customised audio content for listeners with hearing impairments.

Bio

Joshua Reiss is Professor of Audio Engineering with the Centre for Digital Music at Queen Mary University of London. He has published more than 200 scientific papers (including over 50 in premier journals and 7 best paper awards) and co-authored three books. His research has been featured in dozens of original articles and interviews on TV, radio, and in the press. He is a Fellow and Past President of the Audio Engineering Society, and chair of their Publications Policy Committee. He co-founded the highly successful spin-out company, LandR, and recently co-founded Waveshaper AI, RoEx and Nemisindo, also based on his team’s research. He maintains a popular blog, YouTube channel, and Twitter feed for scientific education and dissemination of research activities.

Cadenza Challenge papers

Session chair: Jennifer Firth
12:40-13:10 The First Cadenza Signal Processing Challenge [Recording] [Slides]
Trevor J. Cox4, Michael A. Akeroyd1, Scott Bannister2, Jon Barker3, Bruno Fazenda4, Jennifer Firth1, Simone Graetzer4, Alinka Greasley2, Gerardo Roa Dabike4, Rebecca Vos4, William Whitmer1 (1University of Nottingham; 2University of Leeds; 3University of Sheffield; 4University of Salford)
Session chair: Scott Bannister
13:20-13:35 Better Music Demixing with the sliCQ Transform [Recording] [Slides] [Paper]
Sevag Hanssian (Montréal, Canada)
13:35-13:50 Advancing Auditory Enhancement: Integration of Demucs and Advanced Remixing Techniques in The Cadenza Challenge 2023 [Recording] [Slides] [Paper]
Ritu Munshi, Anulan Wu (University of Leeds, Leeds, UK)
13:50-14:05 The Cadenza Challenge – Task 1 - Sleeter [Recording] [Slides] [Paper]
Debangshu Sarkar, Arundhuti Mukherjee (University of Leeds, Leeds, UK)
14:05-14:20 Audio Mixing and Stem Proportions Adjustment [Recording] [Slides] [Paper]
Epri Wahyu Pratiwi, Yu Tsao, Stefano Rini (Academia Sinica, Taiwan)
Session chair: Jennifer Firth
14:30-15:00 Cadenza Team Submissions [Recording] [Slides]
Gerardo Roa Dabike4, Michael A. Akeroyd1, Scott Bannister2, Jon Barker3, Trevor J. Cox4, Bruno Fazenda4, Jennifer Firth1, Simone Graetzer4, Alinka Greasley2, Rebecca Vos4, William Whitmer1 (1University of Nottingham; 2University of Leeds; 3University of Sheffield; 4University of Salford)

Invited talks

Session chair: Rebecca Vos
15:00-15:15 Who, Where, When, Why, What and How of listening to music via hearing aids
Katherine Phipps
15:15-15:35 The Development and Application of the New Musical Scene Analysis Test [Slides]
Robin Hake (Universität Oldenburg, Germany)
15:35-15:55 Mixing preferences of hearing-impaired listeners [Recording]
Aravindan Benjamin (Universität Oldenburg, Germany)
Katherine Phipps

Katherine Phipps

Who, Where, When, Why, What and How of listening to music via hearing aids

Who, Where, When, Why, What and How of listening to music via hearing aids

Bio

Katherine became severely – profoundly deafened in her late teens, and was the first person in Scotland to get digital hearing aids on the NHS. She has over 25 years of experience working in voluntary and public sector organisations, mainly in the field of deafness and disability. She was a director at RNID, Vice Chair at United Kingdom Council on Deafness (UKCoD), and a board member of the National Register for Communication Professionals for Deaf and Deafblind People (NRCPD). She has an MSc in Inclusive Environments: Design & Management and is a qualified Access Consultant.

Robin Hake

Robin Hake
Universität Oldenburg, Germany

The Development and Application of the New Musical Scene Analysis Test

The Development and Application of the New Musical Scene Analysis Test

Abstract

Auditory Scene Analysis (ASA) is a fundamental process in music perception, enabling listeners to organize complex acoustic environments into coherent auditory streams. This talk first introduces the newly developed Musical Scene Analysis (MSA) test, an adaptive tool designed to assess musical ASA abilities for a diverse audience, including both normal-hearing and hearing-impaired individuals. The MSA test is distinguished by its adaptability and the use of ecologically valid, realistic stimuli, which render it particularly effective at modeling individual differences in musical ASA abilities. The second part of this talk will delve into these individual differences, examining how factors such as age, musical sophistication, hearing impairment, and cognitive functions can influence ASA abilities.

Bio

Currently, as a PhD student at the Music Perception and Processing Lab under Dr. Kai Siedenburg (University of Oldenburg), Robin Hake is engaged in research on Auditory Scene Analysis (ASA) in music. Specifically, he focuses on modelling individual differences in ASA within a musical context for normal-hearing and hearing-impaired individuals. With a background in psychology (University of Groningen), Robin’s academic journey has been marked by a keen interest in understanding the processes that drive music perception.

Aravindan Benjamin

Aravindan Benjamin
Universität Oldenburg, Germany

Mixing preferences of hearing-impaired listeners

Mixing preferences of hearing-impaired listeners

Abstract

Multi-track mixing is an essential practice in modern music production. Several studies have investigated mixing preferences among normal hearing (NH) participants but relatively fewer studies consider hearing impaired (HI) participants. In this study, the subjective mixing preferences of NH and HI listeners was investigated with respect to lead-to-accompaniment level ratios and the low-high frequency spectral energy balance. Furthermore, preferences of transformed equalization were assessed, achieved by linearly extrapolating between the power spectrum of individual tracks and a reference spectrum. It was evident that increasing levels of hearing loss among the participants brought about higher preferences in enhanced vocal levels, larger weightings at higher frequencies (> 1 kHz), and transformed equalization yielding spectrally sparser mixes. To assess if spectrally sparser mixes would also increase mix transparency as measured with a task of musical scene analysis, an independent study was conducted. Results suggest that modifying mixes can make them more accessible to HI and also that sparser mixes may mitigate the depreciation of MSA performance brought on by hearing loss.

Bio

Aravindan Joseph Benjamin, an audio engineer by specialty, obtained a Bachelor of engineering degree in electronic and communication engineering at the Northumbria University in the United Kingdom and a Master of Science degree in electronic media technology at the Technical University of Ilmenau in Germany. His areas of expertise include: music signal processing, psychoacoustics, head-tracking-based dynamic binaural auralization of virtual room acoustics, algorithms for the simulation of virtual room acoustics, audio coding and compression, and electroacoustics. He is currently pursuing his doctorate in music processing for hearing impaired listeners at the music perception and processing laboratory at the Carl von Ossietzky University in Oldenburg. His research areas include pre-processing techniques in multi-track music for listeners with moderate hearing loss and the influence of frequency domain sparsity on music scene analysis abilities among such liseners.

Cadenza Challenge plans

Session chair: Alinka Greasley
16:10-16:30 CAD2 plans and discussion [Slides]
Trevor J. Cox4, Michael A. Akeroyd1, Scott Bannister2, Jon Barker3, Bruno Fazenda4, Jennifer Firth1, Simone Graetzer4, Alinka Greasley2, Gerardo Roa Dabike4, Rebecca Vos4, William Whitmer1 (1University of Nottingham; 2University of Leeds; 3University of Sheffield; 4University of Salford)