Show simple item record

dc.contributor.advisorSadeque, Farig Yousuf
dc.contributor.authorTaki, S.M. Abrar Mustakim
dc.contributor.authorKar, Showmick
dc.contributor.authorNiloy, Soumik Deb
dc.contributor.authorRakib, Mazharul Islam
dc.contributor.authorBiswas, Abdullah Al Nahid
dc.date.accessioned2024-05-07T08:58:35Z
dc.date.available2024-05-07T08:58:35Z
dc.date.copyright©2024
dc.date.issued2024-01
dc.identifier.otherID: 20301125
dc.identifier.otherID: 20301177
dc.identifier.otherID: 20301207
dc.identifier.otherID: 20101408
dc.identifier.otherID: 20301024
dc.identifier.urihttp://hdl.handle.net/10361/22762
dc.descriptionThis thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science, 2024.en_US
dc.descriptionCataloged from PDF version of thesis.
dc.descriptionIncludes bibliographical references (pages 78-83).
dc.description.abstractIn recent years, Large Language Models(LLM) have shown excellent performance in a variety of Natural Language Processing tasks. However, they often produce hallucinated content. Contents that are seemingly correct and make sense linguistically, but are factually incorrect. Since researchers have started working on LLM hallucinations very recently, the problem of mitigating hallucination and understanding which factors play a role in correcting hallucinated content is relatively new. In this paper, we modified a multi-step pipeline called ’Chain of Verification’ that reduces hallucination in Large Language Models by itself without having to feed in external resources. This method is particularly useful for reasoning and reading comprehension types of language tasks. In addition, we extracted the decoder layers of an large language model Mistral 7B to interpret and analyze how the correction was done under the hood. A custom attention weight pruning method was used to prune the defective layers and after pruning, the LLM model passed 3/4 test cases to give proper and correct output results.en_US
dc.description.statementofresponsibilityS.M. Abrar Mustakim Taki
dc.description.statementofresponsibilityShowmick Kar
dc.description.statementofresponsibilitySoumik Deb Niloy
dc.description.statementofresponsibilityMazharul Islam Rakib
dc.description.statementofresponsibilityAbdullah Al Nahid Biswas
dc.format.extent84 pages
dc.language.isoenen_US
dc.publisherBrac Universityen_US
dc.rightsBrac University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission.
dc.subjectMistral 7B AIen_US
dc.subjectLarge language modelen_US
dc.subjectSelf attentionen_US
dc.subjectBlack-BoxNLPen_US
dc.subject.lcshNeural networks (Computer science)
dc.subject.lcshArtificial intelligence
dc.titleMitigation of hallucination and interpretations of self attention of Mistral 7B AI to analyze and visualize context understanding ability of large language modelsen_US
dc.typeThesisen_US
dc.contributor.departmentDepartment of Computer Science and Engineering, Brac University
dc.description.degreeB.Sc. in Computer Science


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record