Recently, at the Audio Mostly 2017 conference, my work with Rod Selfridge and Josh Reiss was published on Propellor Sound Synthesis. I was both published at the conference, on the conference organising committee, as a the webmaster and a member of the music team.More information is available here on the Intelligent Sound Engineering Blog, and an example of the propellor synthesis is available on youtube.
It has been quite a while since I have posted, but I hope to resolve that shortly with a number of academic papers being published this summer,
In the meantime, there is some discussion over the use of sound effects in port production, and the fundamental fact that many things you hear as part of a soundscape are not the original recorded sound – this is the one of the fundamental justifications for my PhD and this is very well explained in this TED Talk:
The 61th International Conference of the Audio Engineering Society on Audio for Games took place in London from 10 to 12 February. This is the fifth edition of the Audio for Games conference which features a mixture of invited talks and academic paper sessions. Traditionally a biennial event, by popular demand the conference was organised in 2016 again following a very successful 4th edition in 2015.
Christian Heinrichs presented work from his doctoral research with Andrew McPherson, discussing Digital Foley and introducing FoleyDesigner, which allows for effectively using human gestures to control sound effects models.
I presented a paper in the Synthesis and Sound Design paper session, on weapon sound synthesis and my colleague William Wilkinson presented work on mammalian growls, both of which can be found in the conference proceedings.
Furthermore, Xavier Serra and Frederic Font presented the Audio Commons project and how the creative industries could benefit from and get access to content with liberal licenses.
Along with presenting work at this conference, I was also involved as the technical coordinator and webmaster for the Audio for Games community.
More information about the conference can be found on the conference website.
Day three of the Digital Audio Effects Conference (DAFx15) began with an excellent introduction and summary of Wave Digital filters and Digital Wave Guides by Kurt Werner and Julius O. Smith from CCRMA, in which the current state of the art in physical modelling no nonlinearities was presented and some potential avenues for future exploration was discussed. Following on from this work was discussed
- identification of metrical structure of music, by Elio from C4DM
- research on whether computer games noticeably prefer spacial audio, from York University
- Discussion and evaluation of feature extraction toolboxes, when to use different feature extraction tools, and how we can develop them in the future, by Dave from C4DM
- Work on vocal tract modelling from York, PPCU Budapest and KTH Sweden.
Day two of DAFx conference in Trondheim NTNU opened with Marije Baalmans keynote on the range of hardware and software audio effects and synthesisers are available to artists, and how different artists utilise these effects. This talk was focused primarily on small embedded systems that artists use, such as Arduino, Beaglebone Black and Raspberry Pi. Later in the day, some excellent work including:
- Granular Synthesis was presented by Sadjad Siddiq from Square Enix,
- A collaboration on synthesising Percussive Drilling Sounds, between IRCAM and HUT,
- Using a modal reverberator structure to modify samples from CCRMA
- Work on intelligent multitrack audio subgrouping by Dave Ronan and Dave Moffat from the Center for Digital Music, Queen Mary University London
He discussed how the crowd funding sources, how to budget for small start up projects. The importance of open source, both in terms of software and hardware was discussed at length, and is a vital aspect of what the OWL team set out to do.
The OWL is a custom build programmable guitar effects pedal that allows anyone to write their own effect pedal and load it onto the standalone program. Effects can be written in C++, Faust or even Pure Data (PD). There is also a wrapper that allows users to run their patches as a VST or AU within a Digital Audio Workstation and in the future, it will also be possible to run patches in the browser. Recently a modular synthesiser version of the Owl has also been released.
Today, 28th August 2015, C4DM presented a one day workshop entitled Listening In The Wild, organised by Dan Stowell, Bob Sturm and Emmanouil Benetos.
The morning session presented a range of research including sound event detection using NMF and DTW techniques, understanding detectability variations of species and habitats, animal vocalisation synthesis through probabilistic models.
The post lunch session saw discussion on vocal modelling and analysis working towards understanding how animals produce their given associated sounds. Following this there was further discussion on NMF followed by work on using bird songs as part of a musical composition.
The poster session included work on auditory scene analysis, bird population vocalisation variations, CHiME: a sound source recognition dataset, technology assisted animal population size measures, bird identification through the use of identity vectors, and DTW for bird song dissimilarity.
Further information on the presenters and posters is avaliable here
Over the past month, I have been working closely with visiting researcher Luca Turchet [http://www.lucaturchet.it/].
We have been working on perceptual evaluation of synthesised footstep sounds. Within the experiment that we ran, participants put on shoes with sensors mounted in them. The sound of different floor surfaces and shoe types is then synthesised, through quality noise blocking headphones, and the participants are then asked to shape the spectral content with the aid of some very basic audio filters.
The intended outcome is to identify the extent to which different participants will vary the spectral characteristics of their footsteps.
Further updates on this research to follow.