dc.contributor.advisor | Feroz, Farhan | |
dc.contributor.advisor | Mostakim, Moin | |
dc.contributor.author | Labiba, Mansura Rahman | |
dc.contributor.author | Jahura, Fatema Tuj | |
dc.contributor.author | Alam, Sadia | |
dc.contributor.author | Binte Morshed, Tasfia | |
dc.contributor.author | Rahman, Wasey | |
dc.date.accessioned | 2023-08-09T06:28:45Z | |
dc.date.available | 2023-08-09T06:28:45Z | |
dc.date.copyright | 2023 | |
dc.date.issued | 2023-01 | |
dc.identifier.other | ID: 20201227 | |
dc.identifier.other | ID: 18101181 | |
dc.identifier.other | ID: 18301200 | |
dc.identifier.other | ID: 18101173 | |
dc.identifier.other | ID: 18101178 | |
dc.identifier.uri | http://hdl.handle.net/10361/19369 | |
dc.description | This thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science and Engineering, 2023. | en_US |
dc.description | Cataloged from PDF version of thesis. | |
dc.description | Includes bibliographical references (pages 27-29). | |
dc.description.abstract | The development of the Internet of Things and voice-based multimedia apps has
allowed for the association and capture of several aspects of human behavior through
the use of big data, which consists of trends and patterns. In the emotion of human
speech, there is a latent representation of numerous aspects that are expressed. By
mining audio-based data, it has been prioritized to extract sentiment from human
speech. This capacity to recognize and categorize human emotion will be crucial for
developing the next generation of AI. The machine will then begin to connect with
human desires as a result. The audio-based data, such as voice emotion recognition,
has not been able to produce results as accurate as those of text-based emotion
recognition in terms of performance. For acoustic modal data, this study presents
a combined strategy of feature extraction and data encoding with one hot vector
embedding. When real-time data is available, LSTM has even employed an RNN based model to forecast the emotion that captures the human voice’s tone and
signifies it. When predicting categorical emotion, the model has been assessed and
shown to perform better than the other models by about 10%. The model has been
tested against two benchmark datasets, RAVDESS and TESS, which contain voice
actors’ renditions of eight different emotions. This model beat other cutting-edge
models, achieving approximately 80% accuracy for weighted data and approximately
85% accuracy for unweighted data. | en_US |
dc.description.statementofresponsibility | Mansura Rahman Labiba | |
dc.description.statementofresponsibility | Fatema Tuj Jahura | |
dc.description.statementofresponsibility | Sadia Alam | |
dc.description.statementofresponsibility | Tasfia Binte Morshed | |
dc.description.statementofresponsibility | Wasey Rahman | |
dc.format.extent | 29 pages | |
dc.language.iso | en | en_US |
dc.publisher | Brac University | en_US |
dc.rights | Brac University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. | |
dc.subject | Machine learning | en_US |
dc.subject | Speech emotion recognition | en_US |
dc.subject | Prediction | en_US |
dc.subject | RNN | en_US |
dc.subject | LSTM | en_US |
dc.subject | Real-time prediction | en_US |
dc.subject.lcsh | Human-computer interaction. | |
dc.subject.lcsh | Artificial intelligence. | |
dc.title | RED-LSTM: Real time emotion detection using LSTM | en_US |
dc.type | Thesis | en_US |
dc.contributor.department | Department of Computer Science and Engineering, Brac University | |
dc.description.degree | B. Computer Science and Engineering | |