Clause complexing in research-article abstracts: comparing human- and AI-generated texts

The ability of chatbots to produce plausible, human-like responses raises questions about the extent of their similarity with original texts. Using a modified version of Halliday’s clause-complexing framework, this study compared 50 abstracts of scientific research articles from Nature with generate...

Full description

Saved in:
Bibliographic Details
Main Author: Leong, Alvin Ping
Other Authors: School of Humanities
Format: Article
Language:English
Published: 2024
Subjects:
Online Access:https://hdl.handle.net/10356/172988
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:The ability of chatbots to produce plausible, human-like responses raises questions about the extent of their similarity with original texts. Using a modified version of Halliday’s clause-complexing framework, this study compared 50 abstracts of scientific research articles from Nature with generated versions produced by Bard, ChatGPT, and Poe Assistant. None of the chatbots matched the original abstracts in all categories. The only chatbot that came closest was ChatGPT, but differences in the use of finite adverbial clauses and –ing elaborating clauses were detected. Incorporating distinct grammatical features in the algorithms of AI-detection tools is crucially needed to enhance the reliability of their results. A genre-based approach to detecting AI-generated content is recommended.