Publications

その他 (国際) A Step from VQA to CQA: Adapting Visual QA Models for Community QA Tasks

Avikalp Srivastava (CMU), HsinWen Liu (Waseda Univ.), Sumio Fujita

arXiv.org

2018.8.29

In this paper, we study and develop methods to derive highlevel representations for image-text pairs in image-based community question answering (CQA) for performing tasks of practical significance on such a platform - automated question classification and finding experts for answering a question. Motivated by our aim to also utilize this work as a step towards basic question-answering on image-based CQA, and to utilize the advances in visual question answering models, we analyze the differences between visual QA & community QA datasets, understand the limitations of applying VQA models directly to CQA data and tasks, and make novel augmentations to VQA-inspired models to best exploit the multimodal data from Yahoo! Chiebukuro’s CQA dataset.

Paper : A Step from VQA to CQA: Adapting Visual QA Models for Community QA Tasks新しいタブまたはウィンドウで開く (外部サイト)