Collaborative cross-modal fusion with Large Language Model for recommendation

Despite the success of conventional collaborative filtering (CF) approaches for recommendation systems, they exhibit limitations in leveraging semantic knowledge within the textual attributes of users and items. Recent focus on the application of large language models for recommendation (LLM4Rec) ha...

Full description

Saved in:
Bibliographic Details
Main Authors: LIU, Zhongzhou, ZHANG, Hao, DONG, Kuicai, FANG, Yuan
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2024
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/9751
https://ink.library.smu.edu.sg/context/sis_research/article/10751/viewcontent/CIKM24_CCFLLM.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-10751
record_format dspace
spelling sg-smu-ink.sis_research-107512024-12-16T03:21:00Z Collaborative cross-modal fusion with Large Language Model for recommendation LIU, Zhongzhou ZHANG, Hao DONG, Kuicai FANG, Yuan Despite the success of conventional collaborative filtering (CF) approaches for recommendation systems, they exhibit limitations in leveraging semantic knowledge within the textual attributes of users and items. Recent focus on the application of large language models for recommendation (LLM4Rec) has highlighted their capability for effective semantic knowledge capture. However, these methods often overlook the collaborative signals in user behaviors. Some simply instruct-tune a language model, while others directly inject the embeddings of a CF-based model, lacking a synergistic fusion of different modalities. To address these issues, we propose a framework of Collaborative Cross-modal Fusion with Large Language Models, termed CCF-LLM, for recommendation. In this framework, we translate the user-item interactions into a hybrid prompt to encode both semantic knowledge and collaborative signals, and then employ an attentive cross-modal fusion strategy to effectively fuse latent embeddings of both modalities. Extensive experiments demonstrate that CCF-LLM outperforms existing methods by effectively utilizing semantic and collaborative signals in the LLM4Rec context. 2024-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9751 info:doi/10.1145/3627673.3679596 https://ink.library.smu.edu.sg/context/sis_research/article/10751/viewcontent/CIKM24_CCFLLM.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Large Language Models Recommendation systems Cross-modal Collaborative filtering Artificial Intelligence and Robotics Computer Sciences
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Large Language Models
Recommendation systems
Cross-modal
Collaborative filtering
Artificial Intelligence and Robotics
Computer Sciences
spellingShingle Large Language Models
Recommendation systems
Cross-modal
Collaborative filtering
Artificial Intelligence and Robotics
Computer Sciences
LIU, Zhongzhou
ZHANG, Hao
DONG, Kuicai
FANG, Yuan
Collaborative cross-modal fusion with Large Language Model for recommendation
description Despite the success of conventional collaborative filtering (CF) approaches for recommendation systems, they exhibit limitations in leveraging semantic knowledge within the textual attributes of users and items. Recent focus on the application of large language models for recommendation (LLM4Rec) has highlighted their capability for effective semantic knowledge capture. However, these methods often overlook the collaborative signals in user behaviors. Some simply instruct-tune a language model, while others directly inject the embeddings of a CF-based model, lacking a synergistic fusion of different modalities. To address these issues, we propose a framework of Collaborative Cross-modal Fusion with Large Language Models, termed CCF-LLM, for recommendation. In this framework, we translate the user-item interactions into a hybrid prompt to encode both semantic knowledge and collaborative signals, and then employ an attentive cross-modal fusion strategy to effectively fuse latent embeddings of both modalities. Extensive experiments demonstrate that CCF-LLM outperforms existing methods by effectively utilizing semantic and collaborative signals in the LLM4Rec context.
format text
author LIU, Zhongzhou
ZHANG, Hao
DONG, Kuicai
FANG, Yuan
author_facet LIU, Zhongzhou
ZHANG, Hao
DONG, Kuicai
FANG, Yuan
author_sort LIU, Zhongzhou
title Collaborative cross-modal fusion with Large Language Model for recommendation
title_short Collaborative cross-modal fusion with Large Language Model for recommendation
title_full Collaborative cross-modal fusion with Large Language Model for recommendation
title_fullStr Collaborative cross-modal fusion with Large Language Model for recommendation
title_full_unstemmed Collaborative cross-modal fusion with Large Language Model for recommendation
title_sort collaborative cross-modal fusion with large language model for recommendation
publisher Institutional Knowledge at Singapore Management University
publishDate 2024
url https://ink.library.smu.edu.sg/sis_research/9751
https://ink.library.smu.edu.sg/context/sis_research/article/10751/viewcontent/CIKM24_CCFLLM.pdf
_version_ 1819113127848968192