Show simple item record

dc.contributor.advisorAlam, Md. Ashraful
dc.contributor.advisorAlam, Md. Golam Rabiul
dc.contributor.authorDas, Saurav
dc.contributor.authorBiswas, Shammo
dc.contributor.authorFahim, Taimoor
dc.contributor.authorSanjan, M.A.B. Siddique
dc.contributor.authorTarannum, Tasnia Alam
dc.date.accessioned2024-10-17T05:33:21Z
dc.date.available2024-10-17T05:33:21Z
dc.date.copyright©2024
dc.date.issued2024-05
dc.identifier.otherID 20101100
dc.identifier.otherID 20101359
dc.identifier.otherID 23241093
dc.identifier.otherID 19201068
dc.identifier.otherID 20301179
dc.identifier.urihttp://hdl.handle.net/10361/24342
dc.descriptionThis thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science, 2024.en_US
dc.descriptionCataloged from PDF version of thesis.
dc.descriptionIncludes bibliographical references (pages 52-55).
dc.description.abstractVideo understanding and description have an important role to play in the field of computer vision and natural language processing. The capacity of automatically generating natural language descriptions for video content has many real-world applications, for example, quoting accessibility tools up to multimedia retrieval systems. Although understanding and describing video content in natural language is a challenging job, it is more so in resource-constrained languages like Bangla. This study investigates the integration of a feature fusion method and the attention-based encoder-decoder framework to improve comprehension of videos and to generate accurate captions for single-action video clips in Bangla. We propose a novel model based on multimodal fusion by combining visual features from video frames and motion information derived from optical flow. The adopted multimodal representations are then fed into an attention-based encoder-decoder architecture aiming to generate descriptive captions in the Bangla language. To facilitate our research, we collected and annotated a new dataset comprising single-action videos sourced from various online platforms. Extensive experiments are conducted on this newly created Bangla single-action videos dataset, with the models evaluated using standard metrics like BLEU, METEOR, and CIDEr. Among the models tested, including architectural variations, the GRU-Gaussian Attention model achieves the best performance, generating captions closest to the ground truth. As this is a new dataset with no previous benchmarks, the proposed approach establishes a strong baseline for Bangla video captioning, achieving a BLEU score of 0.53 and a CIDEr score of 0.492. Additionally, we analyze the attention mechanisms to interpret the learned representations, providing insights into the model’s behavior and decision-making process. This work on developing solutions for under-resourced languages paves the way for enhanced video comprehension with potential applications in human-computer interaction, accessibility, and multimedia retrieval.en_US
dc.description.statementofresponsibilitySaurav Das
dc.description.statementofresponsibilityShammo Biswas
dc.description.statementofresponsibilityTaimoor Fahim
dc.description.statementofresponsibilityM.A.B. Siddique Sanjan
dc.description.statementofresponsibilityTasnia Alam Tarannum
dc.format.extent65 pages
dc.language.isoenen_US
dc.publisherBrac Universityen_US
dc.rightsBrac University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission.
dc.subjectVideo captioningen_US
dc.subjectBangla languageen_US
dc.subjectVideo processingen_US
dc.subjectNatural language processingen_US
dc.subjectFeature fusionen_US
dc.subjectEncoder-decoder frameworken_US
dc.subjectMultimodal fusionen_US
dc.subjectGRU-Gaussian attention modelen_US
dc.subjectCIDEr scoreen_US
dc.subject.lcshNatural language processing (Computer science).
dc.subject.lcshNeural networks (Computer science).
dc.titleEnhancing Bangla video comprehension through multimodal feature integration and attention-based encoder-decoder captioning models for single-action videosen_US
dc.typeThesisen_US
dc.contributor.departmentDepartment of Computer Science and Engineering, Brac University
dc.description.degreeB.Sc. in Computer Science


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record