Automatic question & answer generation using generative Large Language Model (LLM)
Abstract
In the realm of education, student evaluation holds equal significance as imparting
knowledge. To be evaluated, students usually need to go through text-based academic
assessment methods. Instructors need to make diverse sets of questions that
need to be fair for all students to prove their adequacy over a particular topic. This
can prove to be quite challenging as they may need to manually go through several
different lecture materials. Our objective is to make this whole process much easier
by implementing Automatic Question Answer Generation (AQAG), using fine-tuned
generative LLM. For tailoring the instructor’s preferred question style (MCQ, conceptual,
or factual questions), prompt Engineering (PE) is being utilized. In this
research, we propose to leverage unsupervised learning methods in NLP, primarily
focusing on the English language. This approach empowers the base Meta-Llama
2-7B model to integrate RACE dataset as training data for the fine-tuning process.
Creating a customized model that will offer efficient solutions for educators, instructors,
and individuals engaged in text-based evaluations. A reliable and efficient tool
for generating questions and answers can free up valuable time and resources, thus
streamlining their evaluation processes.