TY - GEN
T1 - Acoustic Scene Classification using Fusion of Features and Random Forest Classifier
AU - Sachdeva, Shantanu
AU - Mulimani, Manjunath
N1 - Publisher Copyright:
© 2022 Bharati Vidyapeeth, New Delhi.
PY - 2022
Y1 - 2022
N2 - This paper proposes a model for the task of Acoustic Scene Classification. The proposed model utilizes convolutional neural networks and a random forest classifier to predict the class of the audio clips. The features used by the proposed model are log-Mel, Mel-frequency cepstral coefficient, and Gammatone cepstral coefficient spectrograms. Each spectrogram is processed using a convolutional neural network and combined into a single vector. The processed feature vectors are classified into one of the acoustic scenes using the random forest classifier. The proposed model is evaluated on Tampere University of Technology Urban Acoustic Scenes 2018 and the Tampere University Urban Acoustic Scenes 2019 development datasets. The performance of the proposed model is compared with the Detection and Classification of Acoustic Scenes and Events 2018 and 2019 challenge baseline models to show its efficacy. The proposed model has an accuracy of 68.1% and 67.1% for the two datasets, respectively.
AB - This paper proposes a model for the task of Acoustic Scene Classification. The proposed model utilizes convolutional neural networks and a random forest classifier to predict the class of the audio clips. The features used by the proposed model are log-Mel, Mel-frequency cepstral coefficient, and Gammatone cepstral coefficient spectrograms. Each spectrogram is processed using a convolutional neural network and combined into a single vector. The processed feature vectors are classified into one of the acoustic scenes using the random forest classifier. The proposed model is evaluated on Tampere University of Technology Urban Acoustic Scenes 2018 and the Tampere University Urban Acoustic Scenes 2019 development datasets. The performance of the proposed model is compared with the Detection and Classification of Acoustic Scenes and Events 2018 and 2019 challenge baseline models to show its efficacy. The proposed model has an accuracy of 68.1% and 67.1% for the two datasets, respectively.
UR - http://www.scopus.com/inward/record.url?scp=85130053681&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85130053681&partnerID=8YFLogxK
U2 - 10.23919/INDIACom54597.2022.9763271
DO - 10.23919/INDIACom54597.2022.9763271
M3 - Conference contribution
AN - SCOPUS:85130053681
T3 - Proceedings of the 2022 9th International Conference on Computing for Sustainable Global Development, INDIACom 2022
SP - 654
EP - 658
BT - Proceedings of the 2022 9th International Conference on Computing for Sustainable Global Development, INDIACom 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 9th International Conference on Computing for Sustainable Global Development, INDIACom 2022
Y2 - 23 March 2022 through 25 March 2022
ER -