Show simple item record

dc.contributor.advisorFeroz, Farhan
dc.contributor.advisorMostakim, Moin
dc.contributor.authorLabiba, Mansura Rahman
dc.contributor.authorJahura, Fatema Tuj
dc.contributor.authorAlam, Sadia
dc.contributor.authorBinte Morshed, Tasfia
dc.contributor.authorRahman, Wasey
dc.date.accessioned2023-08-09T06:28:45Z
dc.date.available2023-08-09T06:28:45Z
dc.date.copyright2023
dc.date.issued2023-01
dc.identifier.otherID: 20201227
dc.identifier.otherID: 18101181
dc.identifier.otherID: 18301200
dc.identifier.otherID: 18101173
dc.identifier.otherID: 18101178
dc.identifier.urihttp://hdl.handle.net/10361/19369
dc.descriptionThis thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science and Engineering, 2023.en_US
dc.descriptionCataloged from PDF version of thesis.
dc.descriptionIncludes bibliographical references (pages 27-29).
dc.description.abstractThe development of the Internet of Things and voice-based multimedia apps has allowed for the association and capture of several aspects of human behavior through the use of big data, which consists of trends and patterns. In the emotion of human speech, there is a latent representation of numerous aspects that are expressed. By mining audio-based data, it has been prioritized to extract sentiment from human speech. This capacity to recognize and categorize human emotion will be crucial for developing the next generation of AI. The machine will then begin to connect with human desires as a result. The audio-based data, such as voice emotion recognition, has not been able to produce results as accurate as those of text-based emotion recognition in terms of performance. For acoustic modal data, this study presents a combined strategy of feature extraction and data encoding with one hot vector embedding. When real-time data is available, LSTM has even employed an RNN based model to forecast the emotion that captures the human voice’s tone and signifies it. When predicting categorical emotion, the model has been assessed and shown to perform better than the other models by about 10%. The model has been tested against two benchmark datasets, RAVDESS and TESS, which contain voice actors’ renditions of eight different emotions. This model beat other cutting-edge models, achieving approximately 80% accuracy for weighted data and approximately 85% accuracy for unweighted data.en_US
dc.description.statementofresponsibilityMansura Rahman Labiba
dc.description.statementofresponsibilityFatema Tuj Jahura
dc.description.statementofresponsibilitySadia Alam
dc.description.statementofresponsibilityTasfia Binte Morshed
dc.description.statementofresponsibilityWasey Rahman
dc.format.extent29 pages
dc.language.isoenen_US
dc.publisherBrac Universityen_US
dc.rightsBrac University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission.
dc.subjectMachine learningen_US
dc.subjectSpeech emotion recognitionen_US
dc.subjectPredictionen_US
dc.subjectRNNen_US
dc.subjectLSTMen_US
dc.subjectReal-time predictionen_US
dc.subject.lcshHuman-computer interaction.
dc.subject.lcshArtificial intelligence.
dc.titleRED-LSTM: Real time emotion detection using LSTMen_US
dc.typeThesisen_US
dc.contributor.departmentDepartment of Computer Science and Engineering, Brac University
dc.description.degreeB. Computer Science and Engineering


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record