Don't miss our weekly PhD newsletter | Sign up now Don't miss our weekly PhD newsletter | Sign up now

This project is no longer listed on FindAPhD.com and may not be available.

Click here to search FindAPhD.com for PhD studentship opportunities
  Dr Rod Selfridge, Dr Callum Goddard  Applications accepted all year round  Self-Funded PhD Students Only

About the Project

This aim of this research is to improve the audio quality of synthesized procedural audio sound effects, optimising parameters through the use of differential digital signal processing (DDSP) techniques. Physically inspired synthesis techniques often used for procedural audio sound effects [1], where basic knowledge of the sound producing process and behaviour modelling are integrated within the synthesis process. Previous research has incorporated deeper knowledge of the physical processes to improve the quality of the sounds synthesised, but it is still possible for listeners to identify synthesised sounds when compared to the recorded samples [2].

DDSP covers a number of techniques where signal processors are integrated within neural networks [3]. Through backpropagation of loss functions, the signal processors can be optimised for specific synthesis models.

One drawback of physically inspired procedural models is that potential critical aspects of the physical process as well as the behaviour model that controls the sound synthesis process can be missed. By training the parameters of the synthesis models using DDSP, based on pre-recorded samples, it should be possible to capture missing elements of the models, (behaviour etc), and apply these to new synthesis models. Similar separation of the sounds generated by a musical instrument has been carried out in [4] where the performance data is preserved while the timbre.

The use of DDSP and neural networks for the purposes of sound effects is an ongoing area of research. DDSP has more recently be used to generate sound effects [5] or inspired vocalisation synthesis techniques [6], and different neural synthesis approaches to foley have also been explored [7, 8, 9, 10]. This research looks to build on this body of research, using DDSP to control new physically inspired sound effect models, to improve behaviour and plausibility, and ultimately the quality of synthesised sound effects.

Academic qualifications

A first-class honours degree, or a distinction at master level, or equivalent achievements in in Computer Science, Sound and Music Computing, Music Technology, Sound Design, Acoustics or Artificial Intelligence. 

English language requirement

If your first language is not English, comply with the University requirements for research degree programmes in terms of English language.

Application process

Prospective applicants are encouraged to contact the supervisor, Dr. Rod Selfridge ([Email Address Removed]) to discuss the content of the project and the fit with their qualifications and skills before preparing an application. 

Contact details

Should you need more information, please email [Email Address Removed].

The application must include: 

Research project outline of 2 pages (list of references excluded). The outline may provide details about

  • Background and motivation, explaining the importance of the project, should be supported also by relevant literature. You can also discuss the applications you expect for the project results.
  • Research questions or
  • Methodology: types of data to be used, approach to data collection, and data analysis methods.
  • List of references

The outline must be created solely by the applicant. Supervisors can only offer general discussions about the project idea without providing any additional support.

  • Statement no longer than 1 page describing your motivations and fit with the project.
  • Recent and complete curriculum vitae. The curriculum must include a declaration regarding the English language qualifications of the candidate.
  • Supporting documents will have to be submitted by successful candidates.
  • Two academic references (but if you have been out of education for more than three years, you may submit one academic and one professional reference), on the form can be downloaded here.

Applications can be submitted here.

Download a copy of the project details here.

Computer Science (8)

References

[1] Farnell, A. (2010). Designing sound. Mit Press.
[2] Selfridge, R., Moffat, D., Avital, E. J., & Reiss, J. D. (2018). Creating real-time aeroacoustic sound effects using physically informed models. Journal of the
Audio Engineering Society, 66(7/8), 594-607.
[3] Hayes, B., Shier, J., Fazekas, G., McPherson, A., & Saitis, C. (2023). A Review of Differentiable Digital Signal Processing for Music & Speech Synthesis. arXiv preprint arXiv:2308.15422
[4] Dai, S., Zhang, Z., & Xia, G. G. (2018). Music style transfer: A position paper. arXiv preprint arXiv:1803.06841
[5] Barahona-Ríos, A., & Collins, T. (2023). NoiseBandNet: Controllable TimeVarying Neural Synthesis of Sound Effects Using Filterbanks. arXiv preprint
arXiv:2307.08007.
[6] Hagiwara, M., Cusimano, M., & Liu, J. Y. (2022). Modeling Animal Vocalizations through Synthesizers. arXiv preprint arXiv:2210.10857.
[7] Andreu, S., & Aylagas, M. V. (2022, October). Neural synthesis of sound effects using flow-based deep generative models. In Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment (Vol. 18, No. 1, pp. 2-9).
[8] Comunità, M., Phan, H., & Reiss, J. D. (2021). Neural synthesis of footsteps sound effects with generative adversarial networks. arXiv preprint arXiv:2110.09605
[9] Chung, Y., Lee, J., & Nam, J. (2023). Foley sound synthesis in waveform domain with diffusion model. Tech. Rep., June. [10]Liu, Y., & Jin, C. (2023). Conditional Sound Effects Generation with Regularized WGAN
Search Suggestions
Search suggestions

Based on your current searches we recommend the following search filters.

 About the Project