Customizing large language models for ncert textbooks and development of interface which works offline

Research Article
Dr. Anusha Preetham , P Srihith Reddy and * Shamitha Gaddam
DOI: 
http://dx.doi.org/10.24327/ijrsr.20241504.0873
Subject: 
Other
KeyWords: 
Offline LLMs, Generative AI, Model trained on NCERT Dataset
Abstract: 

Offline Language Models (LLMs) as a solution designed for pupils with restricted internet access. Our approach combines the National Council of Educational Research and Training (NCERT) dataset for foundational knowledge and straightforward answers with the Question Answering Dataset to encourage critical thinking. The model training prioritizes simplicity for NCERT-based content, delivering concise responses to straightforward queries. Simultaneously, question answer dataset integration prompts the model to provide nuanced responses; fostering out-of-the-box thinking. This paper details the methodology, encompassing pre-processing, model architecture, and evaluation metrics. Preliminary results indicate the model's effectiveness in delivering both fundamental and advanced conceptual understanding. The proposed Offline LLMs strive for inclusivity, ensuring all students benefit from tailored language models, irrespective of internet connectivity. This research lays the foundation for a more equitable education system by making sophisticated learning tools accessible to all.