Show simple item record

dc.contributor.advisorSadeque, Farig Yousuf
dc.contributor.authorRodoshi, Mashiat Hasin
dc.contributor.authorAhmed, Moin Uddin
dc.contributor.authorAshraf, Md. Sobhan
dc.contributor.authorMim, Md. Galib Hasan
dc.contributor.authorKhanam, Ashfia
dc.date.accessioned2023-12-20T05:06:39Z
dc.date.available2023-12-20T05:06:39Z
dc.date.copyright2023
dc.date.issued2023-01
dc.identifier.otherID 19201089
dc.identifier.otherID 19301095
dc.identifier.otherID 19301046
dc.identifier.otherID 19301094
dc.identifier.otherID 18301231
dc.identifier.urihttp://hdl.handle.net/10361/22012
dc.descriptionThis thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science, 2023.en_US
dc.descriptionCataloged from PDF version of thesis.
dc.descriptionIncludes bibliographical references (pages 37-38).
dc.description.abstractExperiencing an image on-screen is a privilege that we often seem not to care about. A visually impaired person does not have that luxury. A system that can automatically produce closed captions of an image can thus help visually impaired people experience what’s appearing on a digital screen. Research in this area has been in the forefront of multimodal machine learning for quite some time; but while a plethora of languages has benefited from all that research, Bangla has been left behind. For our thesis, we would like to build a Bangla Caption Generator using multimodal learning with high accuracy which automatically produces closed captioning in Bangla for digital images. The generator will be able to identify different objects in the image, relations among the objects and the actions happening in the image using neural networks. Combining the information collected, it may construct an information-rich, descriptive caption for the image. These captions can be later read aloud so that visually impaired people can get an idea about what is happening around them. This thesis aims to achieve further improvement upon the existing image caption generator in Bangla so that it can greatly help to improve the lives of visually impaired people as well as advance this research towards the state of the art. We have used the Flickr8k and Flickr30k datasets containing 8091 and 31783 images respectively and there are five Bangla captions for each image. We have used the VGG16, VGG19, ResNet50, InceptionV3 and EfficientNetB3 CNN architectures for feature extraction. Our best model has achieved a BLEU-1, BLEU-2, BLEU-3 and BLEU-4 score of 0.553197, 0.341976, 0.234436 and 0.113089 respectively.en_US
dc.description.statementofresponsibilityMashiat Hasin Rodoshi
dc.description.statementofresponsibilityMoin Uddin Ahmed
dc.description.statementofresponsibilityMd. Sobhan Ashraf
dc.description.statementofresponsibilityMd. Galib Hasan Mim
dc.description.statementofresponsibilityAshfia Khanam
dc.format.extent38 pages
dc.language.isoenen_US
dc.publisherBrac Universityen_US
dc.rightsBrac University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission.
dc.subjectImage captioningen_US
dc.subjectCNNen_US
dc.subjectLSTMen_US
dc.subjectRNNen_US
dc.subjectDeep learningen_US
dc.subjectBanglaen_US
dc.subjectNatural language processingen_US
dc.subject.lcshMachine learning
dc.subject.lcshNeural networks (Computer science)
dc.subject.lcshNatural language processing (Computer science)
dc.subject.lcshCognitive learning theory
dc.titleAutomated image caption generator in Bangla using multimodal learningen_US
dc.typeThesisen_US
dc.contributor.departmentDepartment of Computer Science and Engineering, Brac University
dc.description.degreeB.Sc. in Computer Science and Engineering


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record