Acoustic Classification of Mosquitoes using Convolutional Neural Networks Combined with Activity Circadian Rhythm Information.
DOI:
https://doi.org/10.9781/ijimai.2021.08.009Keywords:
Artificial Intelligence, Convolutional Neural Network (CNN), Mosquitoes Classification, A Priori Probability, Bayes’ RuleAbstract
Many researchers have used sound sensors to record audio data from insects, and used these data as inputs of machine learning algorithms to classify insect species. In image classification, the convolutional neural network (CNN), a well-known deep learning algorithm, achieves better performance than any other machine learning algorithm. This performance is affected by the characteristics of the convolution filter (ConvFilter) learned inside the network. Furthermore, CNN performs well in sound classification. Unlike image classification, however, there is little research on suitable ConvFilters for sound classification. Therefore, we compare the performances of three convolution filters, 1D-ConvFilter, 3×1 2D-ConvFilter, and 3×3 2D-ConvFilter, in two different network configurations, when classifying mosquitoes using audio data. In insect sound classification, most machine learning researchers use only audio data as input. However, a classification model, which combines other information such as activity circadian rhythm, should intuitively yield improved classification results. To utilize such relevant additional information, we propose a method that defines this information as a priori probabilities and combines them with CNN outputs. Of the networks, VGG13 with 3×3 2D-ConvFilter showed the best performance in classifying mosquito species, with an accuracy of 80.8%. Moreover, adding activity circadian rhythm information to the networks showed an average performance improvement of 5.5%. The VGG13 network with 1D-ConvFilter achieved the highest accuracy of 85.7% with the additional activity circadian rhythm information.
Downloads
References
World Health Organization: Mosquito-Borne Diseases. Available online: http://www.who.int/neglected_diseases/vector_ecology/mosquitoborne-diseases/en/ (accessed on 1 April 2017).
K. Okayasu, K. Yoshida, M. Fuchida, and A. Nakamura, “Vision-Based Classification of Mosquito Species: Comparison of Conventional and Deep Learning Methods,” Applied Sciences, vol. 10, no. 1, pp. 3935, 2019, doi: 10.3390/app9183935.
J. Park, D. I. Kim, B. Choi, W. Kang, and H. W. Kwon, “Classification and Morphological Analysis of Vector Mosquitoes using Deep Convolutional Neural Networks,” Scientific Reports, vol. 10, no. 1, pp. 1-12, 2020, doi: 10.1038/s41598-020-57875-1.
G. E. Batista, E. J. Keogh, A. Mafra-Neto, and E. Rowton, “SIGKDD demo: sensors and software to allow computational entomology, an emerging application of data mining,” in Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining, 2011, pp. 761-764, doi: 10.1145/2020408.2020530.
Y. Chen, A. Why, G. E. Batista, A. Mafra-Neto, and E. J. Keogh, “Flying insect classification with inexpensive sensors,” Journal of insect behavior, vol. 27, no.5, pp. 657-677, 2014, doi: 10.1007/s10905-014-9454-4.
J. J. Noda, C. M. Travieso-González, D. Sánchez-Rodríguez, and J. B. Alonso-Hernández, “Acoustic Classification of Singing Insects Based on MFCC/LFCC Fusion,” Applied Sciences, vol. 9, no. 19, pp. 4097, 2019, doi: 10.3390/app9194097.
Z. Le-Qing, “Insect sound recognition based on MFCC and PNN,” in 2011 International Conference on Multimedia and Signal Processing, 2011, pp. 42-46, doi: 10.1109/CMSP.2011.100.
N. Saleem, and T. G. Tareen, “Spectral Restoration based speech enhancement for robust speaker identification,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 5, no. 1, pp. 34-39, 2018, doi: 10.9781/ijimai.2018.01.002.
N. Saleem, and M. I. Khattak, “Deep Neural Networks for Speech Enhancement in Complex-Noisy Environments,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 6, no. 1, pp. 84-90, 2020, doi: 10.9781/ijimai.2019.06.001.
X. Fan, H. Feng, and M. Yuan, “PCA based on mutual information for acoustic environment classification,” In 2012 International Conference on Audio, Language and Image Processing, 2012, pp. 270-275, doi: 10.1109/ICALIP.2012.6376624.
S. Ghosh, E. Laksana, L. P. Morency, and S. Scherer, “Learning representations of affect from speech,” 2015, arXiv:1511.04747.
H. Lim, J. Park, and Y. Han, “Rare sound event detection using 1D convolutional recurrent neural networks,” in Proceedings of the Detection and Classification of Acoustic Scenes and Events 2017 Workshop, 2017, pp. 80-84.
J. Naranjo-Alcazar, S. Perez-Castanos, P. Zuccarello, and M. Cobos, “DCASE 2019: CNN depth analysis with different channel inputs for Acoustic Scene Classification,” 2019, arXiv:1906.04591.
K. Simonyan, and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” 2014, arXiv:1409.1556.
C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, “Going deeper with convolutions,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 1-9, doi: 10.1109/CVPR.2015.7298594.
K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770-778, doi: 10.1109/CVPR.2016.90.
S. Hershey, S. Chaudhuri, D. P. Ellis, J. F. Gemmeke, A. Jansen, R. C. Moore, M. Plakal, D. Platt, R. A. Saurous, B. Seybold, M. Slaney, R. J. Weiss and K. Wilson, “CNN architectures for large-scale audio classification,” in 2017 IEEE international conference on acoustics, speech and signal processing, 2017, pp. 131-135, doi: 10.1109/ICASSP.2017.7952132.
Y. Wu, F. Yang, Y. Liu, X. Zha, and S. Yuan. “A comparison of 1-D and 2-D deep convolutional neural networks in ECG classification,” 2018, arXiv:1810.07088.
T. Grill, and J. Schlüter, “Two convolutional neural networks for bird detection in audio signals,” in 2017 25th European Signal Processing Conference (EUSIPCO), 2017, pp. 1764-1768, doi: 10.23919/EUSIPCO.2017.8081512.
J. Sharma, O. C. Granmo, and M. Goodwin, “Environment Sound Classification using Multiple Feature Channels and Attention based Deep Convolutional Neural Network,” 2019, arXiv:1908.11219.
M. Saerens, P. Latinne, and C. Decaestecker, “Adjusting the outputs of a classifier to new a priori probabilities: a simple procedure,” Neural computation, vol. 14, no. 1, pp. 21-41, 2002, doi: 10.1162/089976602753284446.
P. Latinne, M. Saerens, and C. Decaestecker, “Adjusting the outputs of a classifier to new a priori probabilities may significantly improve classification accuracy: evidence from a multi-class problem in remote sensing,” in Proceedings of the Eighteenth International Conference on Machine Learning, 2001, Vol. 1, pp. 298-305.
Y. Chen, A. Why, G. E. Batista, A. Mafra-Neto, and E. J. Keogh. “Flying Insect Classification with Inexpensive Sensors.” Distributed by Y. Chen. https://sites.google.com/site/insectclassification/
X. Glorot, and Y. Bengio, “Understanding the difficulty of training deep feedforward neural networks,” in Proceedings of the thirteenth international conference on artificial intelligence and statistics, 2010, pp. 249-256.
D. P. Kingma, and J. Ba, “Adam: A method for stochastic optimization,” 2014, arXiv:1412.6980.
Downloads
Published
-
Abstract172
-
PDF17






