Recommendation of reviewers based on text analysis and machine learning : part b

This project aims to develop an automated system for recommendation of theses reviewers using Natural Language Processing (NLP) tools and state-of-the-art language models proposed in recent years. The review of theses or dissertations is crucial to access the research outcomes of Doctor of Philosoph...

Full description

Saved in:
Bibliographic Details
Main Author: Guo, Zechuan
Other Authors: Lihui CHEN
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2020
Subjects:
Online Access:https://hdl.handle.net/10356/139869
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:This project aims to develop an automated system for recommendation of theses reviewers using Natural Language Processing (NLP) tools and state-of-the-art language models proposed in recent years. The review of theses or dissertations is crucial to access the research outcomes of Doctor of Philosophy (PhD) students. However, allocation of reviewers can often be a challenging task due to the scope of the project, expertise requirement and availability. In addition, some key details may possibly be overlooked when matching a thesis to reviewers. This will result in a mismatch of the research area between the PhD students and certain reviewers, which can affect the accuracy of the final assessment. Thus, there is a need to develop a system to recommend reviewers based on the similarity of their research fields. This project is based on the concept of semantic text matching, which measures the semantic similarity between source and target text documents. Starting from this idea, various word embedding techniques and deep learning models for comparing document semantics were explored and experimented on the dataset which consists of information pertaining to the research topics of PhD students and reviewers. The results of the Siamese Network were used as implementation benchmarks for the dataset. The performance of the other models was compared against the benchmark, in terms of four evaluation measures. Subsequently, ensemble learning and genetic algorithms were incorporated into the Siamese Network. The resulting model outperformed previous Siamese Networks. This significant improvement highlights the importance of considering various learning and optimization algorithms during the modelling process. Besides, careful tuning of hyperparameters is essential to achieve high-performance and robust language representation models. Finally, Transformer-based language representation models, BERT and ALBERT were implemented and tweaked appropriately to suit the dataset. These deep bidirectional architecture outperformed all previous models and achieved state-of-the-art results for the dataset.