dc.contributor.advisor | Sadeque, Farig Yousuf | |
dc.contributor.author | Taki, S.M. Abrar Mustakim | |
dc.contributor.author | Kar, Showmick | |
dc.contributor.author | Niloy, Soumik Deb | |
dc.contributor.author | Rakib, Mazharul Islam | |
dc.contributor.author | Biswas, Abdullah Al Nahid | |
dc.date.accessioned | 2024-05-07T08:58:35Z | |
dc.date.available | 2024-05-07T08:58:35Z | |
dc.date.copyright | ©2024 | |
dc.date.issued | 2024-01 | |
dc.identifier.other | ID: 20301125 | |
dc.identifier.other | ID: 20301177 | |
dc.identifier.other | ID: 20301207 | |
dc.identifier.other | ID: 20101408 | |
dc.identifier.other | ID: 20301024 | |
dc.identifier.uri | http://hdl.handle.net/10361/22762 | |
dc.description | This thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science, 2024. | en_US |
dc.description | Cataloged from PDF version of thesis. | |
dc.description | Includes bibliographical references (pages 78-83). | |
dc.description.abstract | In recent years, Large Language Models(LLM) have shown excellent performance in
a variety of Natural Language Processing tasks. However, they often produce hallucinated
content. Contents that are seemingly correct and make sense linguistically,
but are factually incorrect. Since researchers have started working on LLM hallucinations
very recently, the problem of mitigating hallucination and understanding
which factors play a role in correcting hallucinated content is relatively new. In this
paper, we modified a multi-step pipeline called ’Chain of Verification’ that reduces
hallucination in Large Language Models by itself without having to feed in external
resources. This method is particularly useful for reasoning and reading comprehension
types of language tasks. In addition, we extracted the decoder layers of an
large language model Mistral 7B to interpret and analyze how the correction was
done under the hood. A custom attention weight pruning method was used to prune
the defective layers and after pruning, the LLM model passed 3/4 test cases to give
proper and correct output results. | en_US |
dc.description.statementofresponsibility | S.M. Abrar Mustakim Taki | |
dc.description.statementofresponsibility | Showmick Kar | |
dc.description.statementofresponsibility | Soumik Deb Niloy | |
dc.description.statementofresponsibility | Mazharul Islam Rakib | |
dc.description.statementofresponsibility | Abdullah Al Nahid Biswas | |
dc.format.extent | 84 pages | |
dc.language.iso | en | en_US |
dc.publisher | Brac University | en_US |
dc.rights | Brac University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. | |
dc.subject | Mistral 7B AI | en_US |
dc.subject | Large language model | en_US |
dc.subject | Self attention | en_US |
dc.subject | Black-BoxNLP | en_US |
dc.subject.lcsh | Neural networks (Computer science) | |
dc.subject.lcsh | Artificial intelligence | |
dc.title | Mitigation of hallucination and interpretations of self attention of Mistral 7B AI to analyze and visualize context understanding ability of large language models | en_US |
dc.type | Thesis | en_US |
dc.contributor.department | Department of Computer Science and Engineering, Brac University | |
dc.description.degree | B.Sc. in Computer Science | |