Acoustic event detection in polyphonic environments using artificial neural networks

dc.contributor.advisorSooriyaarachchi S
dc.contributor.authorMihiranga JPM
dc.date.accept2021
dc.date.accessioned2021
dc.date.available2021
dc.date.issued2021
dc.description.abstractOur environment is a mixture of hundreds of sounds that are emitted by different sound sources. These sounds are overlapped in both time and frequency domains in an unstructured manner composing a polyphonic environment. Identification of acoustic events in a polyphonic environment has become an emerging topic with many applications such as surveillance, context-aware computing, automatic audio indexing, health care monitoring and bioacoustics monitoring. Polyphonic acoustic event detection is a challenging task aimed at detecting the presence of multiple sound events that are overlapped at a particular time instance and labeling. It requires a large amount of training data with a complex machine learning architecture thus making it a highly resource-consuming task. Hence, the accuracy of this research area is still not at a satisfactory level. This study presents a neural networks-based classifier architecture with data augmentation and post-processing methods to improve accuracy. Two neural network architectures as a multi-label and combined single label are implemented and compared in the study. Previous literature reveals that Mel frequency cepstral coefficients and log Mel-band energies are the widely used features in the state of the art research in the area. Different data augmentation methods were used to ensure that the neural networks are trained for even the slight variations of the environmental sounds. A novel binarization method based on the signal energy is proposed to calculate the threshold value for binarizing the source presence predictions. Finally, the median filter based post processing was implemented to smoothen the detection results. The experimental results show that the proposed binarizing method improved the detection accuracy and recorded a maximum of 62.5% combined with the data augmentation and post-processing.en_US
dc.identifier.accnoH4864en_US
dc.identifier.citationMihiranga, J.P.M. (2021). Acoustic event detection in polyphonic environments using artificial neural networks [Master's theses, University of Moratuwa]. Institutional Repository University of Moratuwa. http://dl.lib.uom.lk/handle/123/22278
dc.identifier.degreeMSc in Computer Science & Engineering By researchen_US
dc.identifier.departmentDepartment of Computer Science & Engineeringen_US
dc.identifier.facultyEngineeringen_US
dc.identifier.urihttp://dl.lib.uom.lk/handle/123/22278
dc.language.isoenen_US
dc.subjectPOLYPHONIC ACOUSTIC EVENT DETECTIONen_US
dc.subjectDYNAMIC THRESHOLD BINARIZATIONen_US
dc.subjectDEEP NEURAL NETWORKSen_US
dc.subjectCOMPUTER SCIENCE & ENGINEERING - Dissertationen_US
dc.subjectCOMPUTER SCIENCE- Dissertationen_US
dc.titleAcoustic event detection in polyphonic environments using artificial neural networksen_US
dc.typeThesis-Abstracten_US

Files

Original bundle

Now showing 1 - 3 of 3
Loading...
Thumbnail Image
Name:
TH4864-1.pdf
Size:
313.19 KB
Format:
Adobe Portable Document Format
Description:
Pre-Text
Loading...
Thumbnail Image
Name:
TH4864-2.pdf
Size:
98.63 KB
Format:
Adobe Portable Document Format
Description:
Post- Text
Loading...
Thumbnail Image
Name:
TH4864.pdf
Size:
3.34 MB
Format:
Adobe Portable Document Format
Description:
Full theses

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: