Download DataRES and PyRES: A Room Dataset and a Python Library for Reverberation Enhancement System Development, Evaluation, and Simulation Reverberation is crucial in the acoustical design of physical
spaces, especially halls for live music performances. Reverberation Enhancement Systems (RESs) are active acoustic systems that
can control the reverberation properties of physical spaces, allowing them to adapt to specific acoustical needs. The performance of
RESs strongly depends on the properties of the physical room and
the architecture of the Digital Signal Processor (DSP). However,
room-impulse-response (RIR) measurements and the DSP code
from previous studies on RESs have never been made open access, leading to non-reproducible results. In this study, we present
DataRES and PyRES—a RIR dataset and a Python library to increase the reproducibility of studies on RESs. The dataset contains RIRs measured in RES research and development rooms and
professional music venues. The library offers classes and functionality for the development, evaluation, and simulation of RESs.
The implemented DSP architectures are made differentiable, allowing their components to be trained in a machine-learning-like
pipeline. The replication of previous studies by the authors shows
that PyRES can become a useful tool in future research on RESs.
Download A Wavelet-Based Method for the Estimation of Clarity of Attack Parameters in Non-Percussive Instruments From the exploration of databases of instrument sounds to the selfassisted practice of musical instruments, methods for automatically
and objectively assessing the quality of musical tones are in high
demand. In this paper, we develop a new algorithm for estimating
the duration of the attack, with particular attention to wind and
bowed string instruments. In fact, for these instruments, the quality
of the tones is highly influenced by the attack clarity, for which,
together with pitch stability, the attack duration is an indicator often
used by teachers by ear. Since the direct estimation of the attack
duration from sounds is made difficult by the initial preponderance of the excitation noise, we propose a more robust approach
based on the separation of the ensemble of the harmonics from the
excitation noise, which is obtained by means of an improved pitchsynchronous wavelet transform. We also define a new parameter,
the noise ducking time, which is relevant for detecting the extent of
the noise component in the attack. In addition to the exploration of
available sound databases, for testing our algorithm, we created an
annotated data set in which several problematic sounds are included.
Moreover, to check the consistency and robustness of our duration
estimates, we applied our algorithm to sets of synthetic sounds with
noisy attacks of programmable duration.
Download DiffVox: A Differentiable Model for Capturing and Analysing Vocal Effects Distributions This study introduces a novel and interpretable model, DiffVox,
for matching vocal effects in music production. DiffVox, short
for “Differentiable Vocal Fx", integrates parametric equalisation,
dynamic range control, delay, and reverb with efficient differentiable implementations to enable gradient-based optimisation for
parameter estimation. Vocal presets are retrieved from two datasets,
comprising 70 tracks from MedleyDB and 365 tracks from a private collection. Analysis of parameter correlations reveals strong
relationships between effects and parameters, such as the highpass and low-shelf filters often working together to shape the low
end, and the delay time correlating with the intensity of the delayed signals. Principal component analysis reveals connections to
McAdams’ timbre dimensions, where the most crucial component
modulates the perceived spaciousness while the secondary components influence spectral brightness. Statistical testing confirms
the non-Gaussian nature of the parameter distribution, highlighting
the complexity of the vocal effects space. These initial findings on
the parameter distributions set the foundation for future research
in vocal effects modelling and automatic mixing.