Unlocking markets: A multilingual benchmark to cross-market question answering
Users post numerous product-related questions on e-commerce platforms, affecting their purchase decisions. Product-related question answering (PQA) entails utilizing product-related resources to provide precise responses to users. Wepropose a novel task of Multilingual Crossmarket Product-based Ques...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9539 https://ink.library.smu.edu.sg/context/sis_research/article/10539/viewcontent/21f2d042_e572_43ec_9332_665f06331a15.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-10539 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-105392024-11-15T07:25:40Z Unlocking markets: A multilingual benchmark to cross-market question answering YUAN, Yifei DENG, Yang SOGAARD, Anders ALLIANNEJADI, Mohammad Users post numerous product-related questions on e-commerce platforms, affecting their purchase decisions. Product-related question answering (PQA) entails utilizing product-related resources to provide precise responses to users. Wepropose a novel task of Multilingual Crossmarket Product-based Question Answering (MCPQA) and define the task as providing answers to product-related questions in a main marketplace by utilizing information from another resource-rich auxiliary marketplace in a multilingual context. We introduce a largescale dataset comprising over 7 million questions from 17 marketplaces across 11 languages. We then perform automatic translation on the Electronics category of our dataset, naming it as McMarket. We focus on two subtasks: review-based answer generation and productrelated question ranking. For each subtask, we label a subset of McMarket using an LLM and further evaluate the quality of the annotations via human assessment. We then conduct experiments to benchmark our dataset, using models ranging from traditional lexical models to LLMsin both single-market and cross-market scenarios across McMarket and the corresponding LLM subset. Results show that incorporating cross-market information significantly enhances performance in both tasks. 2024-11-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9539 https://ink.library.smu.edu.sg/context/sis_research/article/10539/viewcontent/21f2d042_e572_43ec_9332_665f06331a15.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Large-scale dataset E-commerce platforms Large Language Models (LLMs) Databases and Information Systems |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Large-scale dataset E-commerce platforms Large Language Models (LLMs) Databases and Information Systems |
spellingShingle |
Large-scale dataset E-commerce platforms Large Language Models (LLMs) Databases and Information Systems YUAN, Yifei DENG, Yang SOGAARD, Anders ALLIANNEJADI, Mohammad Unlocking markets: A multilingual benchmark to cross-market question answering |
description |
Users post numerous product-related questions on e-commerce platforms, affecting their purchase decisions. Product-related question answering (PQA) entails utilizing product-related resources to provide precise responses to users. Wepropose a novel task of Multilingual Crossmarket Product-based Question Answering (MCPQA) and define the task as providing answers to product-related questions in a main marketplace by utilizing information from another resource-rich auxiliary marketplace in a multilingual context. We introduce a largescale dataset comprising over 7 million questions from 17 marketplaces across 11 languages. We then perform automatic translation on the Electronics category of our dataset, naming it as McMarket. We focus on two subtasks: review-based answer generation and productrelated question ranking. For each subtask, we label a subset of McMarket using an LLM and further evaluate the quality of the annotations via human assessment. We then conduct experiments to benchmark our dataset, using models ranging from traditional lexical models to LLMsin both single-market and cross-market scenarios across McMarket and the corresponding LLM subset. Results show that incorporating cross-market information significantly enhances performance in both tasks. |
format |
text |
author |
YUAN, Yifei DENG, Yang SOGAARD, Anders ALLIANNEJADI, Mohammad |
author_facet |
YUAN, Yifei DENG, Yang SOGAARD, Anders ALLIANNEJADI, Mohammad |
author_sort |
YUAN, Yifei |
title |
Unlocking markets: A multilingual benchmark to cross-market question answering |
title_short |
Unlocking markets: A multilingual benchmark to cross-market question answering |
title_full |
Unlocking markets: A multilingual benchmark to cross-market question answering |
title_fullStr |
Unlocking markets: A multilingual benchmark to cross-market question answering |
title_full_unstemmed |
Unlocking markets: A multilingual benchmark to cross-market question answering |
title_sort |
unlocking markets: a multilingual benchmark to cross-market question answering |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2024 |
url |
https://ink.library.smu.edu.sg/sis_research/9539 https://ink.library.smu.edu.sg/context/sis_research/article/10539/viewcontent/21f2d042_e572_43ec_9332_665f06331a15.pdf |
_version_ |
1816859126726131712 |