Show simple item record

dc.contributor.advisorHuq, Aminul
dc.contributor.advisorRahman, Rafeed
dc.contributor.authorHossain, Md. Sakib
dc.contributor.authorIslam, Syed Tamzidul
dc.contributor.authorMazumder, Sujat
dc.contributor.authorJoy, Ali Imran
dc.contributor.authorSakib, Md. Sadman
dc.date.accessioned2023-08-20T06:02:54Z
dc.date.available2023-08-20T06:02:54Z
dc.date.copyright2023
dc.date.issued2023-03
dc.identifier.otherID 18101201
dc.identifier.otherID 22241133
dc.identifier.otherID 18101300
dc.identifier.otherID 18301179
dc.identifier.otherID 18301061
dc.identifier.urihttp://hdl.handle.net/10361/19457
dc.descriptionThis thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science, 2023.en_US
dc.descriptionCataloged from PDF version of thesis.
dc.descriptionIncludes bibliographical references (pages 37-38).
dc.description.abstractIn our day-to-day life there are lots of sounds that we are processing. To process these sounds our brain absorb sound signals and provide us informative knowledge. For human being this is not possible to extract every sounds properly so that, there are lots of equipment which helps us to extract essential information from an audio source. Around the year lots of model came to help thorough extract informations using various algorithms. Also, some models are Convolutional Neural Network (CNN), Region-Convolutional Neural Network (R-CNN), Artificial Neural Network (ANN), VGG16, ResNet50 and Numerous machine learning algorithms have been utilized to effectively categorize audio, and these methods have recently demonstrated encouraging results in separating spectrotemporal images from various sound classifications. The study purpose of this research was to analyze which feature extraction method shows maximum result using Convolutional Neural Network (CNN), VGG16 and ResNet50. In the proposed model, MFCC feature extraction method are taken from the dataset and trained using a multiple layer-based con volution neural network. In the experimental assessment, a sound dataset consisting of 105829 audio clips separated up into multiple groups of important sounds during study used to develop the models. Additionally, we evaluated the models’ validity which reach an accuracy of 94.53% on Speech Command dataset.en_US
dc.description.statementofresponsibilityMd. Sakib Hossain
dc.description.statementofresponsibilitySyed Tamzidul Islam
dc.description.statementofresponsibilitySujat Mazumder
dc.description.statementofresponsibilityAli Imran Joy
dc.description.statementofresponsibilityMd. Sadman Sakib
dc.format.extent49 pages
dc.language.isoenen_US
dc.publisherBrac Universityen_US
dc.rightsBrac University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission.
dc.subjectSound classificationen_US
dc.subjectSpectrogramsen_US
dc.subjectSpeech commanden_US
dc.subjectCNNen_US
dc.subjectResNet50en_US
dc.subject.lcshNeural networks (Computer science)
dc.titleSpeech command classification based on deep neural networksen_US
dc.typeThesisen_US
dc.contributor.departmentDepartment of Computer Science and Engineering, Brac University
dc.description.degreeB. Computer Science


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record