Show simple item record

dc.contributor.advisorAlam, Golam Rabiul
dc.contributor.authorSusmit, Tahsin Ashrafee
dc.contributor.authorMehejabin, Maliha
dc.contributor.authorHasan, Isratul
dc.contributor.authorKausar, Azmain Ibn
dc.contributor.authorAkbar, Suraiya Binte
dc.date.accessioned2025-02-13T06:24:57Z
dc.date.available2025-02-13T06:24:57Z
dc.date.copyright2024
dc.date.issued2024-10
dc.identifier.otherID 20301088
dc.identifier.otherID 20301264
dc.identifier.otherID 20301072
dc.identifier.otherID 20301144
dc.identifier.urihttp://hdl.handle.net/10361/25395
dc.descriptionThis thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science, 2024.en_US
dc.descriptionCataloged from PDF version of thesis.
dc.descriptionIncludes bibliographical references (pages 70-72).
dc.description.abstractThe purpose of our study is to create new technology that will provide a revolutionary navigation system with significant improvement of mobility and independence for visually impaired people. We utilize YOLOv11 and Faster R-CNN to detect an object which is used in combination with Llama 3.2–3B Instruct for context-aware navigation by providing helpful guidance of our current essential location. Our paper tackles the failure points in today’s technologies with lack of flexibility for dynamic and unfamiliar environments, unreliable performance under changes in lighting conditions and inefficient obstacle detection. By training these models together and selecting the one with the highest confidence score, we enhance spatial awareness, identifying obstacles in key areas like the left, right, or center. This approach, complemented by personalized navigation instructions, ensures improved decisionmaking and safety in real-world scenarios. Using advanced locational technologies available today and imagining those of tomorrow, we aspire to render current navigation methods obsolete by fostering more efficient, real-time and autonomous tools for visually impaired people as they become part of the familiar or unfamiliar environments. After fine-tuning the Llama 3.2-3B-Instruct model, BLEU-4 increased from 0.0442 to 0.1175, and ROUGE-L improved from 0.2102 to 0.3204, indicating enhanced text generation fluency and coherence.en_US
dc.description.statementofresponsibilityTahsin Ashrafee Susmit
dc.description.statementofresponsibilityMaliha Mehejabin
dc.description.statementofresponsibilityIsratul Hasan
dc.description.statementofresponsibilityAzmain Ibn Kausar
dc.description.statementofresponsibilitySuraiya Binte Akbar
dc.format.extent72 pages
dc.language.isoenen_US
dc.publisherBRAC Universityen_US
dc.rightsBRAC University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission.
dc.subjectYOLOv11en_US
dc.subjectFaster R-CNNen_US
dc.subjectLlama 3.2-3B Instructen_US
dc.subjectObject detectionen_US
dc.subjectNavigation systemen_US
dc.subjectVisually impaireden_US
dc.subject.lcshVisually impaired persons--Navigation.
dc.subject.lcshComputer vision.
dc.subject.lcshObject recognition--Computer algorithms.
dc.titlePathway to perception: a smart navigation approach for visually impaired individuals leveraging YOLO, faster R-CNN, and LLaMAen_US
dc.typeThesisen_US
dc.contributor.departmentDepartment of Computer Science and Engineering, BRAC University
dc.description.degreeB.Sc. in Computer Science


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record