Collaborative cross-modal fusion with Large Language Model for recommendation
Despite the success of conventional collaborative filtering (CF) approaches for recommendation systems, they exhibit limitations in leveraging semantic knowledge within the textual attributes of users and items. Recent focus on the application of large language models for recommendation (LLM4Rec) ha...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9751 https://ink.library.smu.edu.sg/context/sis_research/article/10751/viewcontent/CIKM24_CCFLLM.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-10751 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-107512024-12-16T03:21:00Z Collaborative cross-modal fusion with Large Language Model for recommendation LIU, Zhongzhou ZHANG, Hao DONG, Kuicai FANG, Yuan Despite the success of conventional collaborative filtering (CF) approaches for recommendation systems, they exhibit limitations in leveraging semantic knowledge within the textual attributes of users and items. Recent focus on the application of large language models for recommendation (LLM4Rec) has highlighted their capability for effective semantic knowledge capture. However, these methods often overlook the collaborative signals in user behaviors. Some simply instruct-tune a language model, while others directly inject the embeddings of a CF-based model, lacking a synergistic fusion of different modalities. To address these issues, we propose a framework of Collaborative Cross-modal Fusion with Large Language Models, termed CCF-LLM, for recommendation. In this framework, we translate the user-item interactions into a hybrid prompt to encode both semantic knowledge and collaborative signals, and then employ an attentive cross-modal fusion strategy to effectively fuse latent embeddings of both modalities. Extensive experiments demonstrate that CCF-LLM outperforms existing methods by effectively utilizing semantic and collaborative signals in the LLM4Rec context. 2024-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9751 info:doi/10.1145/3627673.3679596 https://ink.library.smu.edu.sg/context/sis_research/article/10751/viewcontent/CIKM24_CCFLLM.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Large Language Models Recommendation systems Cross-modal Collaborative filtering Artificial Intelligence and Robotics Computer Sciences |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Large Language Models Recommendation systems Cross-modal Collaborative filtering Artificial Intelligence and Robotics Computer Sciences |
spellingShingle |
Large Language Models Recommendation systems Cross-modal Collaborative filtering Artificial Intelligence and Robotics Computer Sciences LIU, Zhongzhou ZHANG, Hao DONG, Kuicai FANG, Yuan Collaborative cross-modal fusion with Large Language Model for recommendation |
description |
Despite the success of conventional collaborative filtering (CF) approaches for recommendation systems, they exhibit limitations in leveraging semantic knowledge within the textual attributes of users and items. Recent focus on the application of large language models for recommendation (LLM4Rec) has highlighted their capability for effective semantic knowledge capture. However, these methods often overlook the collaborative signals in user behaviors. Some simply instruct-tune a language model, while others directly inject the embeddings of a CF-based model, lacking a synergistic fusion of different modalities. To address these issues, we propose a framework of Collaborative Cross-modal Fusion with Large Language Models, termed CCF-LLM, for recommendation. In this framework, we translate the user-item interactions into a hybrid prompt to encode both semantic knowledge and collaborative signals, and then employ an attentive cross-modal fusion strategy to effectively fuse latent embeddings of both modalities. Extensive experiments demonstrate that CCF-LLM outperforms existing methods by effectively utilizing semantic and collaborative signals in the LLM4Rec context. |
format |
text |
author |
LIU, Zhongzhou ZHANG, Hao DONG, Kuicai FANG, Yuan |
author_facet |
LIU, Zhongzhou ZHANG, Hao DONG, Kuicai FANG, Yuan |
author_sort |
LIU, Zhongzhou |
title |
Collaborative cross-modal fusion with Large Language Model for recommendation |
title_short |
Collaborative cross-modal fusion with Large Language Model for recommendation |
title_full |
Collaborative cross-modal fusion with Large Language Model for recommendation |
title_fullStr |
Collaborative cross-modal fusion with Large Language Model for recommendation |
title_full_unstemmed |
Collaborative cross-modal fusion with Large Language Model for recommendation |
title_sort |
collaborative cross-modal fusion with large language model for recommendation |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2024 |
url |
https://ink.library.smu.edu.sg/sis_research/9751 https://ink.library.smu.edu.sg/context/sis_research/article/10751/viewcontent/CIKM24_CCFLLM.pdf |
_version_ |
1819113127848968192 |