학술논문

Automated Multi-Choice Question Answering System using Natural Language Processing
Document Type
Conference
Source
2024 3rd International Conference for Innovation in Technology (INOCON) Innovation in Technology (INOCON), 2024 3rd International Conference for. :1-6 Mar, 2024
Subject
Bioengineering
Communication, Networking and Broadcast Technologies
Components, Circuits, Devices and Systems
Computing and Processing
Engineered Materials, Dielectrics and Plasmas
Engineering Profession
Fields, Waves and Electromagnetics
General Topics for Engineers
Nuclear Engineering
Photonics and Electrooptics
Power, Energy and Industry Applications
Robotics and Control Systems
Signal Processing and Analysis
Transportation
Training
Technological innovation
Computational modeling
Transfer learning
Graphics processing units
Portable document format
Transformers
UnifiedQA
BERT
RoBERTa
XLNET
Transfer Learning
Discriminators
and Natural Language Processing
Language
Abstract
Machine reading comprehension not only involves reading unstructured text but also understanding it. Compared to questions of a simpler kind, this job calls for more sophisticated reading comprehension, such as logical reasoning, summarizing, and computation. They demand the intelligent system's ability to understand human language, making it an essential tool for testing its capabilities. We used unstructured corpora such as ARC corpora and Aristo mini corpora as contexts to answer multiple choice questions, which are indexed using ElasticSearch API. To answer multiple choice questions, we assigned questions from subjects such as Chemistry, and Biology; The goal is to target science-related questions instead of general questions, as they are clearly more difficult to answer. We have developed an automated multiple choice question answering template that is deployed on localhost using the Flask framework; the model accepts a question followed by a series of answer choices, and there is a PDF file upload function where the user can also upload a PDF file containing a question or a set of questions followed by a pair of candidate answers options. Searching the corpus for the appropriate context for each question is an important part of answering the question, using discriminators such as the answer discriminator and the document discriminator, because it was found that the TF-IDF score alone does not always provide the result. The best possible documents to answer the question in hand. UnifiedQA was used for modeling and a model without discriminators was also made to compare the results, and it was found that with discriminators, better performance was achieved at the expense of computational speed. A current increase of up to 7% was observed, which is huge, so better context has been shown to improve model performance and accuracy.