Show simple item record

dc.contributor.advisorAlam, Md. Golam Rabiul
dc.contributor.advisorReza, Md. Tanzim
dc.contributor.authorAmin, Mahzabin Yasmin Binte
dc.contributor.authorShammo, Weney Hasan
dc.contributor.authorSayed, Jawad Bin
dc.contributor.authorHossain, MD Junaied
dc.date.accessioned2023-12-31T05:38:13Z
dc.date.available2023-12-31T05:38:13Z
dc.date.copyright2023
dc.date.issued2023-05
dc.identifier.otherID 18101479
dc.identifier.otherID 19101601
dc.identifier.otherID 21341025
dc.identifier.otherID 20101204
dc.identifier.urihttp://hdl.handle.net/10361/22042
dc.descriptionThis thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science, 2023.en_US
dc.descriptionCataloged from PDF version of thesis.
dc.descriptionIncludes bibliographical references (pages 36-37).
dc.description.abstractThe intent of this paper is to make the process of interpreting and understanding information within ultrasound pictures simpler and quicker by addressing the lack of techniques for automatically deciphering medical images. In order to do so, we propose a method of ultrasound image caption generation using AI that highlights the potential Machine Translation has in translating medical images to textual notations. The model needs to be trained on an ultrasound image dataset of the abdominal region including the uterus, myometrium, endometrium and cervix, a field of the medical sector that remains inadequately addressed. Two pre-trained CNN models, namely, VGG16 and Inception v3 have been used to extract features from the ultrasound images. Subsequently, the encoder-decoder model takes in two types of inputs, one for each of its layers. The two kinds of inputs are the text sequence and the image features. Both Vanilla LSTM and Bi-directional LSTM have been used to build the language generation model. The embedding layer along with the LSTM layer will process the text input. At last, the output from the two layers stated above will be merged.en_US
dc.description.statementofresponsibilityMahzabin Yasmin Binte Amin
dc.description.statementofresponsibilityWeney Hasan Shammo
dc.description.statementofresponsibilityJawad Bin Sayed
dc.description.statementofresponsibilityMD Junaied Hossain
dc.format.extent49 pages
dc.language.isoenen_US
dc.publisherBrac Universityen_US
dc.rightsBrac University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission.
dc.subjectUltrasound imageen_US
dc.subjectImage captioningen_US
dc.subjectMedical image captioningen_US
dc.subjectConvolutional Neural Networken_US
dc.subjectLSTMen_US
dc.subject.lcshImaging systems in medicine
dc.subject.lcshDiagnostic ultrasonic imaging
dc.subject.lcshNeural networks (Computer science)
dc.titleA comparative analysis of the different CNN-LSTM model caption generation of medical imagesen_US
dc.typeThesisen_US
dc.contributor.departmentDepartment of Computer Science and Engineering, Brac University
dc.description.degreeB.Sc. in Computer Science


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record