VGG16-based deep learning architectures for classification of lung sounds into normal, crackles, and wheezes using Gammatonegrams

Breathing sounds are a rich source of information that can assist doctors in diagnosing pulmonary diseases in a non-invasive manner. Several algorithms can be developed based on these sounds to create an automatic classification system for lung diseases. To implement these systems, researchers...

Full description

Saved in:
Bibliographic Details
Main Authors: Zakaria, Neili, Sundaraj, Kenneth
Format: Conference or Workshop Item
Language:en
Published: 2023
Online Access:http://eprints.utem.edu.my/id/eprint/28048/1/VGG16-based%20deep%20learning%20architectures%20for%20classification%20of%20lung%20sounds%20into%20normal%2C%20crackles%2C%20and%20wheezes%20using%20gammatonegrams.pdf
http://eprints.utem.edu.my/id/eprint/28048/
https://www.researchgate.net/publication/377242904_Advancing_Educational_Practices_Implementation_and_Impact_of_Virtual_Reality_in_Islamic_Religious_Education
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Breathing sounds are a rich source of information that can assist doctors in diagnosing pulmonary diseases in a non-invasive manner. Several algorithms can be developed based on these sounds to create an automatic classification system for lung diseases. To implement these systems, researchers traditionally follow two main steps: feature extraction and pattern classification. In recent years, deep neural networks have gained attention in the field of breathing sound classification as they have proven effective for training large datasets. In this study, we conducted a comparison of two versions of the VGG16-based deep learning model for breathing sound classification using Gammatonegrams as input. We implemented two extensions of the VGG16 model - one executed from scratch and the other based on a pretrained VGG16 model using transfer learning. We processed digital recordings of cycle-based breathing sounds to obtain Gammatonegrams images, which were then fed as input to the VGG16 network. In addition, we performed data augmentation in our experiments using audio cycles from the ICBHI database to evaluate the performance of the proposed method. The classification results were obtained using the Google Collaboratory platform.