{ "id": "1705.01253", "version": "v1", "published": "2017-05-03T04:46:33.000Z", "updated": "2017-05-03T04:46:33.000Z", "title": "The Forgettable-Watcher Model for Video Question Answering", "authors": [ "Hongyang Xue", "Zhou Zhao", "Deng Cai" ], "categories": [ "cs.CV", "cs.CL" ], "abstract": "A number of visual question answering approaches have been proposed recently, aiming at understanding the visual scenes by answering the natural language questions. While the image question answering has drawn significant attention, video question answering is largely unexplored. Video-QA is different from Image-QA since the information and the events are scattered among multiple frames. In order to better utilize the temporal structure of the videos and the phrasal structures of the answers, we propose two mechanisms: the re-watching and the re-reading mechanisms and combine them into the forgettable-watcher model. Then we propose a TGIF-QA dataset for video question answering with the help of automatic question generation. Finally, we evaluate the models on our dataset. The experimental results show the effectiveness of our proposed models.", "revisions": [ { "version": "v1", "updated": "2017-05-03T04:46:33.000Z" } ], "analyses": { "keywords": [ "video question answering", "forgettable-watcher model", "natural language questions", "drawn significant attention", "visual question answering approaches" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }