dc.contributor.advisor | Alam, Golam Rabiul | |
dc.contributor.author | Susmit, Tahsin Ashrafee | |
dc.contributor.author | Mehejabin, Maliha | |
dc.contributor.author | Hasan, Isratul | |
dc.contributor.author | Kausar, Azmain Ibn | |
dc.contributor.author | Akbar, Suraiya Binte | |
dc.date.accessioned | 2025-02-13T06:24:57Z | |
dc.date.available | 2025-02-13T06:24:57Z | |
dc.date.copyright | 2024 | |
dc.date.issued | 2024-10 | |
dc.identifier.other | ID 20301088 | |
dc.identifier.other | ID 20301264 | |
dc.identifier.other | ID 20301072 | |
dc.identifier.other | ID 20301144 | |
dc.identifier.uri | http://hdl.handle.net/10361/25395 | |
dc.description | This thesis is submitted in partial fulfillment of the requirements for the degree of Bachelor of Science in Computer Science, 2024. | en_US |
dc.description | Cataloged from PDF version of thesis. | |
dc.description | Includes bibliographical references (pages 70-72). | |
dc.description.abstract | The purpose of our study is to create new technology that will provide a revolutionary
navigation system with significant improvement of mobility and independence
for visually impaired people. We utilize YOLOv11 and Faster R-CNN to detect an
object which is used in combination with Llama 3.2–3B Instruct for context-aware
navigation by providing helpful guidance of our current essential location. Our paper
tackles the failure points in today’s technologies with lack of flexibility for dynamic
and unfamiliar environments, unreliable performance under changes in lighting conditions
and inefficient obstacle detection. By training these models together and
selecting the one with the highest confidence score, we enhance spatial awareness,
identifying obstacles in key areas like the left, right, or center. This approach,
complemented by personalized navigation instructions, ensures improved decisionmaking
and safety in real-world scenarios. Using advanced locational technologies
available today and imagining those of tomorrow, we aspire to render current navigation
methods obsolete by fostering more efficient, real-time and autonomous tools
for visually impaired people as they become part of the familiar or unfamiliar environments.
After fine-tuning the Llama 3.2-3B-Instruct model, BLEU-4 increased
from 0.0442 to 0.1175, and ROUGE-L improved from 0.2102 to 0.3204, indicating
enhanced text generation fluency and coherence. | en_US |
dc.description.statementofresponsibility | Tahsin Ashrafee Susmit | |
dc.description.statementofresponsibility | Maliha Mehejabin | |
dc.description.statementofresponsibility | Isratul Hasan | |
dc.description.statementofresponsibility | Azmain Ibn Kausar | |
dc.description.statementofresponsibility | Suraiya Binte Akbar | |
dc.format.extent | 72 pages | |
dc.language.iso | en | en_US |
dc.publisher | BRAC University | en_US |
dc.rights | BRAC University theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. | |
dc.subject | YOLOv11 | en_US |
dc.subject | Faster R-CNN | en_US |
dc.subject | Llama 3.2-3B Instruct | en_US |
dc.subject | Object detection | en_US |
dc.subject | Navigation system | en_US |
dc.subject | Visually impaired | en_US |
dc.subject.lcsh | Visually impaired persons--Navigation. | |
dc.subject.lcsh | Computer vision. | |
dc.subject.lcsh | Object recognition--Computer algorithms. | |
dc.title | Pathway to perception: a smart navigation approach for visually impaired individuals leveraging YOLO, faster R-CNN, and LLaMA | en_US |
dc.type | Thesis | en_US |
dc.contributor.department | Department of Computer Science and Engineering, BRAC University | |
dc.description.degree | B.Sc. in Computer Science | |