T-SciQ: Teaching multimodal Chain-of-Thought reasoning via large language model signals for science question answering
Large Language Models (LLMs) have recently demonstrated exceptional performance in various Natural Language Processing (NLP) tasks. They have also shown the ability to perform chain-of-thought (CoT) reasoning to solve complex problems. Recent studies have explored CoT reasoning in complex multimodal...
Saved in:
Main Authors: | WANG, Lei, HU, Yi, HE, Jiabang, XU, Xing, LIU, Ning, LIU, Hui, SHEN, Heng Tao |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8756 https://ink.library.smu.edu.sg/context/sis_research/article/9759/viewcontent/29884_Article_Text_33938_1_2_20240324_pvoa.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
Unifying text, tables, and images for multimodal question answering
by: LUO, Haohao, et al.
Published: (2023) -
LOVA3 : Learning to visual question answering, asking and assessment
by: ZHAO, Henry Hengyuan, et al.
Published: (2024) -
Large multimodal models for visual reasoning
by: Duong, Ngoc Yen
Published: (2024) -
Q-bench: a benchmark for general-purpose foundation models on low-level vision
by: Wu, Haoning, et al.
Published: (2024) -
Q-instruct: improving low-level visual abilities for multi-modality foundation models
by: Wu, Haoning, et al.
Published: (2024)