Download Acoustical Simulations of the Human Vocal Tract Using the 1D and 2D Digital Waveguide Software Model This paper details software under development that uses the digital waveguide physical model to represent the sound creation mechanism and environment associated with the production of speech, specifically the human vocal tract. Focus is directed towards a comparison between the existing 1D waveguide method, on which several studies have already been conducted, and the developing 2D waveguide mesh method. The construction of the two models and the application of the tract geometry is examined, in addition, the inclusion of dynamic articulatory variations to increase the ability of such systems to create natural sounding speech is discussed. Results obtained from each suggest that the 2D model is capable of producing similarly accurate vowel spectra to that already accomplished with the 1D version, although speech-like sounds created with the 2D mesh appear to exhibit greater realism.
Download Hybrid room impulse response synthesis in digital waveguide mesh based room acoustics simulation The digital waveguide mesh (DWM) and related finite difference time domain techniques offer significant promise for room acoustics simulation problems. However high resolution 3-D DWMs of large spaces remain beyond the capabilities of current desktop based computers, due to prohibitively long run-times and large memory requirements. This paper examines how hybrid room impulse response synthesis might be used to better enable virtual environment simulation through the use of otherwise computationally expensive DWM models. This is facilitated through the introduction of the RenderAIR virtual environment simulation system and comparison with both real-world measurements and more established modelling techniques. Results demonstrate good performance against acoustic benchmarks and significant computational savings when a 2-D DWM is used as part of an appropriate hybridization strategy.
Download RoomWeaver: A Digital Waveguide Mesh Based Room Acoustics Research Tool RoomWeaver is a Digital Waveguide Mesh (DWM) based Integrated Development Environment (IDE) style research tool, similar in appearance and functionality to other current acoustics software. The premise of RoomWeaver is to ease the development and application of DWM models for virtual acoustic spaces. This paper demonstrates the basic functionality of RoomWeaver’s 3D modelling and Room Impulse Response (RIR) generation capabilities. A case study is presented to show how new DWM types can be quickly developed and easily tested using RoomWeaver’s built in plug-in architecture through the implementation of a hybrid-type mesh. This hybrid mesh is comprised of efficient, yet geometrically inflexible, finite difference DWM elements and the geometrically versatile, but slow, wave-based DWM elements. The two types of DWM are interfaced using a KW-pipe and this hybrid model exhibits a significant increase in execution speed and a smaller memory footprint than standard wave-based DWM models and allows nontrivial geometries to be successfully modelled.
Download Audio-Tactile Glove This paper introduces the Audio-Tactile Glove, an experimental tool for the analysis of vibrotactile feedback in instrument design. Vibrotactile feedback provides essential information in the operation of acoustic instruments. The Audio-Tactile Glove is designed as a research tool for the investigation of the various techniques used to apply this theory to digital interfaces. The user receives vibrations via actuators distributed throughout the glove, located so as not to interrupt the physical contact required between user and interface. Using this actuator array, researchers will be able to independently apply vibrotactile information to six stimulation points across each hand exploiting the broad frequency range of the device, with specific sensitivity within the haptic frequency range of the hand. It is proposed that researchers considering the inclusion of vibrotactile feedback in existing devices can utilize this device without altering their initial designs.
Download Articulatory vocal tract synthesis in Supercollider The APEX system [1] enables vocal tract articulation using a reduced set of user controllable parameters by means of Principal Component Analysis of X-ray tract data. From these articulatory profiles it is then possible to calculate cross-sectional area function data that can be used as input to a number of articulatory based speech synthesis algorithms. In this paper the Kelly-Lochbaum 1-D digital waveguide vocal tract is used, and both APEX control and synthesis engine have been implemented and tested in SuperCollider. Accurate formant synthesis and real-time control are demonstrated, although for multi-parameter speech-like articulation a more direct mapping from tract-to-synthesizer tube sections is needed. SuperCollider provides an excellent framework for the further exploration of this work.
Download Measuring Diffusion in a 2D Digital Waveguide Mesh The digital waveguide mesh is a method by which the propagation of sound waves in an acoustic system can be simulated. An important consideration in modelling such systems is the accurate modelling of reflection characteristics at boundaries and surfaces. A significant property of an acoustic boundary is its diffusivity. In fact partially diffuse sound reflections are observed at most real acoustic surfaces and so this is an important consideration when implementing a digital waveguide mesh model. This paper presents a method for modelling diffusion that offers a high degree of control. The model is implemented with varying amounts of diffusivity, and a method for measuring its diffusive properties is outlined. Results for the model are presented and a method to calculate the diffusion coefficient is described.
Download Real-Time Dynamic Image-Source Implementation For Auralisation This paper describes a software package for auralisation in interactive virtual reality environments. Its purpose is to reproduce, in real time, the 3D soundfield within a virtual room where listener and sound sources can be moved freely. Output sound is presented binaurally using headphones. Auralisation is based on geometric acoustic models combined with head-related transfer functions (HRTFs): the direct sound and reflections from each source are computed dynamically by the image-source method. Directional cues are obtained by filtering these incoming sounds by the HRTFs corresponding to their propagation directions relative to the listener, computed on the basis of the information provided by a head-tracking device. Two interactive real-time applications were developed to demonstrate the operation of this software package. Both provide a visual representation of listener (position and head orientation) and sources (including image sources). One focusses on the auralisation-visualisation synchrony and the other on the dynamic calculation of reflection paths. Computational performance results of the auralisation system are presented.
Download Spatial audio quality and user preference of listening systems in video games Spatial audio playback solutions provide video game players with ways to experience more immersive and engaging video game content. This paper aims to find whether listening systems that are able to more accurately convey spatial information are preferred by video game players, and to what extent this is true for different loudspeaker configurations whilst engaged in video game play. Results do suggest that a listening system with high perceived spatial quality is more preferred.
Download Scalable Spectral Reflections In Conic Sections The object of this project is to present a novel digital audio effect based on a real-time, windowed block-based FFT and inverse FFT. The effect is achieved by mirroring the spectrum, producing a sound effect ranging from a purer rendition of the original, through a rougher one, to a sound unrecognisable from the original. A mirror taking the shape of a conic section is constructed between certain partials, and the modified spectrum is created by reflecting the original spectrum in this mirror. The user can select the type and continuously vary the amount of curvature, typically ‘roughening’ the input sound quite gratifyingly. We demonstrate the system with live real-time audio via microphone.
Download A Deep Learning Approach to the Prediction of Time-Frequency Spatial Parameters for Use in Stereo Upmixing This paper presents a deep learning approach to parametric timefrequency parameter prediction for use within stereo upmixing algorithms. The approach presented uses a Multi-Channel U-Net with Residual connections (MuCh-Res-U-Net) trained on a novel dataset of stereo and parametric time-frequency spatial audio data to predict time-frequency spatial parameters from a stereo input signal for positions on a 50-point Lebedev quadrature sampled sphere. An example upmix pipeline is then proposed which utilises the predicted time-frequency spatial parameters to both extract and remap stereo signal components to target spherical harmonic components to facilitate the generation of a full spherical representation of the upmixed sound field.