INTEGRATING FIRST-ORDER LOGIC FOR READING COMPREHENSION MODEL ON RECLOR DATASET

ReClor is a Machine Reading Comprehension dataset that requires logical reasoning. Studies have shown that pre-trained language models are still weak in reasoning tasks. Several studies have been conducted to address this shortcoming. However, among the existing approaches tend to improve the lan...

Full description

Saved in:
Bibliographic Details
Main Author: Nuraeni, Wulantika
Format: Theses
Language:Indonesia
Online Access:https://digilib.itb.ac.id/gdl/view/85304
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Institut Teknologi Bandung
Language: Indonesia
Description
Summary:ReClor is a Machine Reading Comprehension dataset that requires logical reasoning. Studies have shown that pre-trained language models are still weak in reasoning tasks. Several studies have been conducted to address this shortcoming. However, among the existing approaches tend to improve the language model’s capabilities through statistical features rather than performing actual reasoning. Logical reasoning systems typically involve logical symbols and strict inference rules. Some studies have attempted to combine the strenght of statistical learning and logical reasosning using hybrid approach between pre-trained language models and logic programming. Given the complexity of sentences and the diversity of question types in the ReClor data, this hybrid approach might be inadequate. This study proposes an approach using Graph Neural Networks, which have the ability to learn structured data and perform reasoning using First-Order Logic, which is more expressive in representing relationships between objects. The Graph Neural Network model then combined with a language model using joint loss training to enhance the language model’s performance on the Machine Reading Comprehension task. Study results show that the Graph Neural Network model trained on First-Order Logic data can process logical symbols and learn resolution path, but its performance heavily depends on the accuracy of the semantic parser that converts ReClor’s text data into first-order logic. Additionally, the combination of loss values from the graph model and language model effectively improved RoBERTa’s performance but did not significantly enhance BERT’s performance. This outcome is due to the next-sentence prediction task during BERT’s pre-training, which may limit the performace in downstream tasks, unlike RoBERTa, which removes this task during pre-trainig and achieves accuracy improvements using proposed method.