A hybrid approach for detecting prerequisite relations in multi-modal food recipes

Modeling the structure of culinary recipes is the core of recipe representation learning. Current approaches mostly focus on extracting the workflow graph from recipes based on text descriptions. Process images, which constitute an important part of cooking recipes, has rarely been investigated in r...

Full description

Saved in:
Bibliographic Details
Main Authors: PAN, Liangming, CHEN, Jingjing, LIU, Shaoteng, NGO, Chong-wah, KAN, Min-Yen, CHUA, Tat-Seng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2020
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7927
https://ink.library.smu.edu.sg/context/sis_research/article/8930/viewcontent/TMM20_Paper__1_.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8930
record_format dspace
spelling sg-smu-ink.sis_research-89302023-07-14T07:04:39Z A hybrid approach for detecting prerequisite relations in multi-modal food recipes PAN, Liangming CHEN, Jingjing LIU, Shaoteng NGO, Chong-wah KAN, Min-Yen CHUA, Tat-Seng Modeling the structure of culinary recipes is the core of recipe representation learning. Current approaches mostly focus on extracting the workflow graph from recipes based on text descriptions. Process images, which constitute an important part of cooking recipes, has rarely been investigated in recipe structure modeling. We study this recipe structure problem from a multi-modal learning perspective, by proposing a prerequisite tree to represent recipes with cooking images at a step-level granularity. We propose a simple-yet-effective two-stage framework to automatically construct the prerequisite tree for a recipe by (1) utilizing a trained classifier to detect pairwise prerequisite relations that fuses multi-modal features as input; then (2) applying different strategies (greedy method, maximum weight, and beam search) to build the tree structure. Experiments on the MM-ReS dataset demonstrates the advantages of introducing process images for recipe structure modeling. Also, compared with neural methods which require large numbers of training data, we show that our two-stage pipeline can achieve promising results using only 400 labeled prerequisite trees as training data. 2020-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7927 info:doi/10.1109/TMM.2020.3042706 https://ink.library.smu.edu.sg/context/sis_research/article/8930/viewcontent/TMM20_Paper__1_.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Feature extraction Training Task analysis Semantics Pipelines Deep learning Predictive models Food recipes cooking workflow prerequisite trees multi-modal fusion cause-and-effect reasoning deep learning Databases and Information Systems
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Feature extraction
Training
Task analysis
Semantics
Pipelines
Deep learning
Predictive models
Food recipes
cooking workflow
prerequisite trees
multi-modal fusion
cause-and-effect reasoning
deep learning
Databases and Information Systems
spellingShingle Feature extraction
Training
Task analysis
Semantics
Pipelines
Deep learning
Predictive models
Food recipes
cooking workflow
prerequisite trees
multi-modal fusion
cause-and-effect reasoning
deep learning
Databases and Information Systems
PAN, Liangming
CHEN, Jingjing
LIU, Shaoteng
NGO, Chong-wah
KAN, Min-Yen
CHUA, Tat-Seng
A hybrid approach for detecting prerequisite relations in multi-modal food recipes
description Modeling the structure of culinary recipes is the core of recipe representation learning. Current approaches mostly focus on extracting the workflow graph from recipes based on text descriptions. Process images, which constitute an important part of cooking recipes, has rarely been investigated in recipe structure modeling. We study this recipe structure problem from a multi-modal learning perspective, by proposing a prerequisite tree to represent recipes with cooking images at a step-level granularity. We propose a simple-yet-effective two-stage framework to automatically construct the prerequisite tree for a recipe by (1) utilizing a trained classifier to detect pairwise prerequisite relations that fuses multi-modal features as input; then (2) applying different strategies (greedy method, maximum weight, and beam search) to build the tree structure. Experiments on the MM-ReS dataset demonstrates the advantages of introducing process images for recipe structure modeling. Also, compared with neural methods which require large numbers of training data, we show that our two-stage pipeline can achieve promising results using only 400 labeled prerequisite trees as training data.
format text
author PAN, Liangming
CHEN, Jingjing
LIU, Shaoteng
NGO, Chong-wah
KAN, Min-Yen
CHUA, Tat-Seng
author_facet PAN, Liangming
CHEN, Jingjing
LIU, Shaoteng
NGO, Chong-wah
KAN, Min-Yen
CHUA, Tat-Seng
author_sort PAN, Liangming
title A hybrid approach for detecting prerequisite relations in multi-modal food recipes
title_short A hybrid approach for detecting prerequisite relations in multi-modal food recipes
title_full A hybrid approach for detecting prerequisite relations in multi-modal food recipes
title_fullStr A hybrid approach for detecting prerequisite relations in multi-modal food recipes
title_full_unstemmed A hybrid approach for detecting prerequisite relations in multi-modal food recipes
title_sort hybrid approach for detecting prerequisite relations in multi-modal food recipes
publisher Institutional Knowledge at Singapore Management University
publishDate 2020
url https://ink.library.smu.edu.sg/sis_research/7927
https://ink.library.smu.edu.sg/context/sis_research/article/8930/viewcontent/TMM20_Paper__1_.pdf
_version_ 1772829240962580480