Segui questo link per vedere altri tipi di pubblicazioni sul tema: Video question answering.

Articoli di riviste sul tema "Video question answering"

Cita una fonte nei formati APA, MLA, Chicago, Harvard e in molti altri stili

Scegli il tipo di fonte:

Vedi i top-50 articoli di riviste per l'attività di ricerca sul tema "Video question answering".

Accanto a ogni fonte nell'elenco di riferimenti c'è un pulsante "Aggiungi alla bibliografia". Premilo e genereremo automaticamente la citazione bibliografica dell'opera scelta nello stile citazionale di cui hai bisogno: APA, MLA, Harvard, Chicago, Vancouver ecc.

Puoi anche scaricare il testo completo della pubblicazione scientifica nel formato .pdf e leggere online l'abstract (il sommario) dell'opera se è presente nei metadati.

Vedi gli articoli di riviste di molte aree scientifiche e compila una bibliografia corretta.

1

Lei, Chenyi, Lei Wu, Dong Liu, Zhao Li, Guoxin Wang, Haihong Tang e Houqiang Li. "Multi-Question Learning for Visual Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 34, n. 07 (3 aprile 2020): 11328–35. http://dx.doi.org/10.1609/aaai.v34i07.6794.

Testo completo
Abstract (sommario):
Visual Question Answering (VQA) raises a great challenge for computer vision and natural language processing communities. Most of the existing approaches consider video-question pairs individually during training. However, we observe that there are usually multiple (either sequentially generated or not) questions for the target video in a VQA task, and the questions themselves have abundant semantic relations. To explore these relations, we propose a new paradigm for VQA termed Multi-Question Learning (MQL). Inspired by the multi-task learning, MQL learns from multiple questions jointly together with their corresponding answers for a target video sequence. The learned representations of video-question pairs are then more general to be transferred for new questions. We further propose an effective VQA framework and design a training procedure for MQL, where the specifically designed attention network models the relation between input video and corresponding questions, enabling multiple video-question pairs to be co-trained. Experimental results on public datasets show the favorable performance of the proposed MQL-VQA framework compared to state-of-the-arts.
Gli stili APA, Harvard, Vancouver, ISO e altri
2

Ruwa, Nelson, Qirong Mao, Liangjun Wang e Jianping Gou. "Affective question answering on video". Neurocomputing 363 (ottobre 2019): 125–39. http://dx.doi.org/10.1016/j.neucom.2019.06.046.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
3

Wang, Yueqian, Yuxuan Wang, Kai Chen e Dongyan Zhao. "STAIR: Spatial-Temporal Reasoning with Auditable Intermediate Results for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 38, n. 17 (24 marzo 2024): 19215–23. http://dx.doi.org/10.1609/aaai.v38i17.29890.

Testo completo
Abstract (sommario):
Recently we have witnessed the rapid development of video question answering models. However, most models can only handle simple videos in terms of temporal reasoning, and their performance tends to drop when answering temporal-reasoning questions on long and informative videos. To tackle this problem we propose STAIR, a Spatial-Temporal Reasoning model with Auditable Intermediate Results for video question answering. STAIR is a neural module network, which contains a program generator to decompose a given question into a hierarchical combination of several sub-tasks, and a set of lightweight neural modules to complete each of these sub-tasks. Though neural module networks are already widely studied on image-text tasks, applying them to videos is a non-trivial task, as reasoning on videos requires different abilities. In this paper, we define a set of basic video-text sub-tasks for video question answering and design a set of lightweight modules to complete them. Different from most prior works, modules of STAIR return intermediate outputs specific to their intentions instead of always returning attention maps, which makes it easier to interpret and collaborate with pre-trained models. We also introduce intermediate supervision to make these intermediate outputs more accurate. We conduct extensive experiments on several video question answering datasets under various settings to show STAIR's performance, explainability, compatibility with pre-trained models, and applicability when program annotations are not available. Code: https://github.com/yellow-binary-tree/STAIR
Gli stili APA, Harvard, Vancouver, ISO e altri
4

Zong, Linlin, Jiahui Wan, Xianchao Zhang, Xinyue Liu, Wenxin Liang e Bo Xu. "Video-Context Aligned Transformer for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 38, n. 17 (24 marzo 2024): 19795–803. http://dx.doi.org/10.1609/aaai.v38i17.29954.

Testo completo
Abstract (sommario):
Video question answering involves understanding video content to generate accurate answers to questions. Recent studies have successfully modeled video features and achieved diverse multimodal interaction, yielding impressive outcomes. However, they have overlooked the fact that the video contains richer instances and events beyond the scope of the stated question. Extremely imbalanced alignment of information from both sides leads to significant instability in reasoning. To address this concern, we propose the Video-Context Aligned Transformer (V-CAT), which leverages the context to achieve semantic and content alignment between video and question. Specifically, the video and text are encoded into a shared semantic space initially. We apply contrastive learning to global video token and context token to enhance the semantic alignment. Then, the pooled context feature is utilized to obtain corresponding visual content. Finally, the answer is decoded by integrating the refined video and question features. We evaluate the effectiveness of V-CAT on MSVD-QA and MSRVTT-QA dataset, both achieving state-of-the-art performance. Extended experiments further analyze and demonstrate the effectiveness of each proposed module.
Gli stili APA, Harvard, Vancouver, ISO e altri
5

Huang, Deng, Peihao Chen, Runhao Zeng, Qing Du, Mingkui Tan e Chuang Gan. "Location-Aware Graph Convolutional Networks for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 34, n. 07 (3 aprile 2020): 11021–28. http://dx.doi.org/10.1609/aaai.v34i07.6737.

Testo completo
Abstract (sommario):
We addressed the challenging task of video question answering, which requires machines to answer questions about videos in a natural language form. Previous state-of-the-art methods attempt to apply spatio-temporal attention mechanism on video frame features without explicitly modeling the location and relations among object interaction occurred in videos. However, the relations between object interaction and their location information are very critical for both action recognition and question reasoning. In this work, we propose to represent the contents in the video as a location-aware graph by incorporating the location information of an object into the graph construction. Here, each node is associated with an object represented by its appearance and location features. Based on the constructed graph, we propose to use graph convolution to infer both the category and temporal locations of an action. As the graph is built on objects, our method is able to focus on the foreground action contents for better video question answering. Lastly, we leverage an attention mechanism to combine the output of graph convolution and encoded question features for final answer reasoning. Extensive experiments demonstrate the effectiveness of the proposed methods. Specifically, our method significantly outperforms state-of-the-art methods on TGIF-QA, Youtube2Text-QA and MSVD-QA datasets.
Gli stili APA, Harvard, Vancouver, ISO e altri
6

Gao, Lianli, Pengpeng Zeng, Jingkuan Song, Yuan-Fang Li, Wu Liu, Tao Mei e Heng Tao Shen. "Structured Two-Stream Attention Network for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 33 (17 luglio 2019): 6391–98. http://dx.doi.org/10.1609/aaai.v33i01.33016391.

Testo completo
Abstract (sommario):
To date, visual question answering (VQA) (i.e., image QA and video QA) is still a holy grail in vision and language understanding, especially for video QA. Compared with image QA that focuses primarily on understanding the associations between image region-level details and corresponding questions, video QA requires a model to jointly reason across both spatial and long-range temporal structures of a video as well as text to provide an accurate answer. In this paper, we specifically tackle the problem of video QA by proposing a Structured Two-stream Attention network, namely STA, to answer a free-form or open-ended natural language question about the content of a given video. First, we infer rich longrange temporal structures in videos using our structured segment component and encode text features. Then, our structured two-stream attention component simultaneously localizes important visual instance, reduces the influence of background video and focuses on the relevant text. Finally, the structured two-stream fusion component incorporates different segments of query and video aware context representation and infers the answers. Experiments on the large-scale video QA dataset TGIF-QA show that our proposed method significantly surpasses the best counterpart (i.e., with one representation for the video input) by 13.0%, 13.5%, 11.0% and 0.3 for Action, Trans., TrameQA and Count tasks. It also outperforms the best competitor (i.e., with two representations) on the Action, Trans., TrameQA tasks by 4.1%, 4.7%, and 5.1%.
Gli stili APA, Harvard, Vancouver, ISO e altri
7

Kumar, Krishnamoorthi Magesh, e P. Valarmathie. "Domain and Intelligence Based Multimedia Question Answering System". International Journal of Evaluation and Research in Education (IJERE) 5, n. 3 (1 settembre 2016): 227. http://dx.doi.org/10.11591/ijere.v5i3.4544.

Testo completo
Abstract (sommario):
Multimedia question answering systems have become very popular over the past few years. It allows users to share their thoughts by answering given question or obtain information from a set of answered questions. However, existing QA systems support only textual answer which is not so instructive for many users. The user’s discussion can be enhanced by adding suitable multimedia data. Multimedia answers offer intuitive information with more suitable image, voice and video. This system includes a set of information as well as classification of question and answer, query generation, multimedia data selection and presentation. This system will take all kinds of media such as text, images, videos, and videos which will be combined with a textual answer. In a way, it automatically collects information from the user to improvising the answer. This method consists of ranking for answers to select the best answer. By dealing out a huge set of QA pairs and adding them to a database, multimedia question answering approach for users which finds multimedia answers by matching their questions with those in the database. The effectiveness of Multimedia system is determined by ranking of text, image, audio and video in users answer. The answer which is given by the user it’s processed by Semantic match algorithm and the best answers can be viewed by Naive Bayesian ranking system.
Gli stili APA, Harvard, Vancouver, ISO e altri
8

Xue, Hongyang, Zhou Zhao e Deng Cai. "Unifying the Video and Question Attentions for Open-Ended Video Question Answering". IEEE Transactions on Image Processing 26, n. 12 (dicembre 2017): 5656–66. http://dx.doi.org/10.1109/tip.2017.2746267.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
9

Jang, Yunseok, Yale Song, Chris Dongjoo Kim, Youngjae Yu, Youngjin Kim e Gunhee Kim. "Video Question Answering with Spatio-Temporal Reasoning". International Journal of Computer Vision 127, n. 10 (18 giugno 2019): 1385–412. http://dx.doi.org/10.1007/s11263-019-01189-x.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
10

Zhuang, Yueting, Dejing Xu, Xin Yan, Wenzhuo Cheng, Zhou Zhao, Shiliang Pu e Jun Xiao. "Multichannel Attention Refinement for Video Question Answering". ACM Transactions on Multimedia Computing, Communications, and Applications 16, n. 1s (28 aprile 2020): 1–23. http://dx.doi.org/10.1145/3366710.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
11

Garcia, Noa, Mayu Otani, Chenhui Chu e Yuta Nakashima. "KnowIT VQA: Answering Knowledge-Based Questions about Videos". Proceedings of the AAAI Conference on Artificial Intelligence 34, n. 07 (3 aprile 2020): 10826–34. http://dx.doi.org/10.1609/aaai.v34i07.6713.

Testo completo
Abstract (sommario):
We propose a novel video understanding task by fusing knowledge-based and video question answering. First, we introduce KnowIT VQA, a video dataset with 24,282 human-generated question-answer pairs about a popular sitcom. The dataset combines visual, textual and temporal coherence reasoning together with knowledge-based questions, which need of the experience obtained from the viewing of the series to be answered. Second, we propose a video understanding model by combining the visual and textual video content with specific knowledge about the show. Our main findings are: (i) the incorporation of knowledge produces outstanding improvements for VQA in video, and (ii) the performance on KnowIT VQA still lags well behind human accuracy, indicating its usefulness for studying current video modelling limitations.
Gli stili APA, Harvard, Vancouver, ISO e altri
12

Mao, Jianguo, Wenbin Jiang, Hong Liu, Xiangdong Wang e Yajuan Lyu. "Inferential Knowledge-Enhanced Integrated Reasoning for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 37, n. 11 (26 giugno 2023): 13380–88. http://dx.doi.org/10.1609/aaai.v37i11.26570.

Testo completo
Abstract (sommario):
Recently, video question answering has attracted growing attention. It involves answering a question based on a fine-grained understanding of video multi-modal information. Most existing methods have successfully explored the deep understanding of visual modality. We argue that a deep understanding of linguistic modality is also essential for answer reasoning, especially for videos that contain character dialogues. To this end, we propose an Inferential Knowledge-Enhanced Integrated Reasoning method. Our method consists of two main components: 1) an Inferential Knowledge Reasoner to generate inferential knowledge for linguistic modality inputs that reveals deeper semantics, including the implicit causes, effects, mental states, etc. 2) an Integrated Reasoning Mechanism to enhance video content understanding and answer reasoning by leveraging the generated inferential knowledge. Experimental results show that our method achieves significant improvement on two mainstream datasets. The ablation study further demonstrates the effectiveness of each component of our approach.
Gli stili APA, Harvard, Vancouver, ISO e altri
13

Jiang, Jianwen, Ziqiang Chen, Haojie Lin, Xibin Zhao e Yue Gao. "Divide and Conquer: Question-Guided Spatio-Temporal Contextual Attention for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 34, n. 07 (3 aprile 2020): 11101–8. http://dx.doi.org/10.1609/aaai.v34i07.6766.

Testo completo
Abstract (sommario):
Understanding questions and finding clues for answers are the key for video question answering. Compared with image question answering, video question answering (Video QA) requires to find the clues accurately on both spatial and temporal dimension simultaneously, and thus is more challenging. However, the relationship between spatio-temporal information and question still has not been well utilized in most existing methods for Video QA. To tackle this problem, we propose a Question-Guided Spatio-Temporal Contextual Attention Network (QueST) method. In QueST, we divide the semantic features generated from question into two separate parts: the spatial part and the temporal part, respectively guiding the process of constructing the contextual attention on spatial and temporal dimension. Under the guidance of the corresponding contextual attention, visual features can be better exploited on both spatial and temporal dimensions. To evaluate the effectiveness of the proposed method, experiments are conducted on TGIF-QA dataset, MSRVTT-QA dataset and MSVD-QA dataset. Experimental results and comparisons with the state-of-the-art methods have shown that our method can achieve superior performance.
Gli stili APA, Harvard, Vancouver, ISO e altri
14

Yang, Saelyne, Sunghyun Park, Yunseok Jang e Moontae Lee. "YTCommentQA: Video Question Answerability in Instructional Videos". Proceedings of the AAAI Conference on Artificial Intelligence 38, n. 17 (24 marzo 2024): 19359–67. http://dx.doi.org/10.1609/aaai.v38i17.29906.

Testo completo
Abstract (sommario):
Instructional videos provide detailed how-to guides for various tasks, with viewers often posing questions regarding the content. Addressing these questions is vital for comprehending the content, yet receiving immediate answers is difficult. While numerous computational models have been developed for Video Question Answering (Video QA) tasks, they are primarily trained on questions generated based on video content, aiming to produce answers from within the content. However, in real-world situations, users may pose questions that go beyond the video's informational boundaries, highlighting the necessity to determine if a video can provide the answer. Discerning whether a question can be answered by video content is challenging due to the multi-modal nature of videos, where visual and verbal information are intertwined. To bridge this gap, we present the YTCommentQA dataset, which contains naturally-generated questions from YouTube, categorized by their answerability and required modality to answer -- visual, script, or both. Experiments with answerability classification tasks demonstrate the complexity of YTCommentQA and emphasize the need to comprehend the combined role of visual and script information in video reasoning. The dataset is available at https://github.com/lgresearch/YTCommentQA.
Gli stili APA, Harvard, Vancouver, ISO e altri
15

Yu, Zhou, Dejing Xu, Jun Yu, Ting Yu, Zhou Zhao, Yueting Zhuang e Dacheng Tao. "ActivityNet-QA: A Dataset for Understanding Complex Web Videos via Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 33 (17 luglio 2019): 9127–34. http://dx.doi.org/10.1609/aaai.v33i01.33019127.

Testo completo
Abstract (sommario):
Recent developments in modeling language and vision have been successfully applied to image question answering. It is both crucial and natural to extend this research direction to the video domain for video question answering (VideoQA). Compared to the image domain where large scale and fully annotated benchmark datasets exists, VideoQA datasets are limited to small scale and are automatically generated, etc. These limitations restrict their applicability in practice. Here we introduce ActivityNet-QA, a fully annotated and large scale VideoQA dataset. The dataset consists of 58,000 QA pairs on 5,800 complex web videos derived from the popular ActivityNet dataset. We present a statistical analysis of our ActivityNet-QA dataset and conduct extensive experiments on it by comparing existing VideoQA baselines. Moreover, we explore various video representation strategies to improve VideoQA performance, especially for long videos.
Gli stili APA, Harvard, Vancouver, ISO e altri
16

Cherian, Anoop, Chiori Hori, Tim K. Marks e Jonathan Le Roux. "(2.5+1)D Spatio-Temporal Scene Graphs for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 36, n. 1 (28 giugno 2022): 444–53. http://dx.doi.org/10.1609/aaai.v36i1.19922.

Testo completo
Abstract (sommario):
Spatio-temporal scene-graph approaches to video-based reasoning tasks, such as video question-answering (QA), typically construct such graphs for every video frame. These approaches often ignore the fact that videos are essentially sequences of 2D ``views'' of events happening in a 3D space, and that the semantics of the 3D scene can thus be carried over from frame to frame. Leveraging this insight, we propose a (2.5+1)D scene graph representation to better capture the spatio-temporal information flows inside the videos. Specifically, we first create a 2.5D (pseudo-3D) scene graph by transforming every 2D frame to have an inferred 3D structure using an off-the-shelf 2D-to-3D transformation module, following which we register the video frames into a shared (2.5+1)D spatio-temporal space and ground each 2D scene graph within it. Such a (2.5+1)D graph is then segregated into a static sub-graph and a dynamic sub-graph, corresponding to whether the objects within them usually move in the world. The nodes in the dynamic graph are enriched with motion features capturing their interactions with other graph nodes. Next, for the video QA task, we present a novel transformer-based reasoning pipeline that embeds the (2.5+1)D graph into a spatio-temporal hierarchical latent space, where the sub-graphs and their interactions are captured at varied granularity. To demonstrate the effectiveness of our approach, we present experiments on the NExT-QA and AVSD-QA datasets. Our results show that our proposed (2.5+1)D representation leads to faster training and inference, while our hierarchical model showcases superior performance on the video QA task versus the state of the art.
Gli stili APA, Harvard, Vancouver, ISO e altri
17

Chu, Wenqing, Hongyang Xue, Zhou Zhao, Deng Cai e Chengwei Yao. "The forgettable-watcher model for video question answering". Neurocomputing 314 (novembre 2018): 386–93. http://dx.doi.org/10.1016/j.neucom.2018.06.069.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
18

Zhu, Linchao, Zhongwen Xu, Yi Yang e Alexander G. Hauptmann. "Uncovering the Temporal Context for Video Question Answering". International Journal of Computer Vision 124, n. 3 (13 luglio 2017): 409–21. http://dx.doi.org/10.1007/s11263-017-1033-7.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
19

Lee, Yue-Shi, Yu-Chieh Wu e Jie-Chi Yang. "BVideoQA: Online English/Chinese bilingual video question answering". Journal of the American Society for Information Science and Technology 60, n. 3 (marzo 2009): 509–25. http://dx.doi.org/10.1002/asi.21002.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
20

Xiao, Junbin, Angela Yao, Zhiyuan Liu, Yicong Li, Wei Ji e Tat-Seng Chua. "Video as Conditional Graph Hierarchy for Multi-Granular Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 36, n. 3 (28 giugno 2022): 2804–12. http://dx.doi.org/10.1609/aaai.v36i3.20184.

Testo completo
Abstract (sommario):
Video question answering requires the models to understand and reason about both the complex video and language data to correctly derive the answers. Existing efforts have been focused on designing sophisticated cross-modal interactions to fuse the information from two modalities, while encoding the video and question holistically as frame and word sequences. Despite their success, these methods are essentially revolving around the sequential nature of video- and question-contents, providing little insight to the problem of question-answering and lacking interpretability as well. In this work, we argue that while video is presented in frame sequence, the visual elements (e.g., objects, actions, activities and events) are not sequential but rather hierarchical in semantic space. To align with the multi-granular essence of linguistic concepts in language queries, we propose to model video as a conditional graph hierarchy which weaves together visual facts of different granularity in a level-wise manner, with the guidance of corresponding textual cues. Despite the simplicity, our extensive experiments demonstrate the superiority of such conditional hierarchical graph architecture, with clear performance improvements over prior methods and also better generalization across different type of questions. Further analyses also demonstrate the model's reliability as it shows meaningful visual-textual evidences for the predicted answers.
Gli stili APA, Harvard, Vancouver, ISO e altri
21

Lee, Kyungjae, Nan Duan, Lei Ji, Jason Li e Seung-won Hwang. "Segment-Then-Rank: Non-Factoid Question Answering on Instructional Videos". Proceedings of the AAAI Conference on Artificial Intelligence 34, n. 05 (3 aprile 2020): 8147–54. http://dx.doi.org/10.1609/aaai.v34i05.6327.

Testo completo
Abstract (sommario):
We study the problem of non-factoid QA on instructional videos. Existing work focuses either on visual or textual modality of video content, to find matching answers to the question. However, neither is flexible enough for our problem setting of non-factoid answers with varying lengths. Motivated by this, we propose a two-stage model: (a) multimodal segmentation of video into span candidates and (b) length-adaptive ranking of the candidates to the question. First, for segmentation, we propose Segmenter for generating span candidates of diverse length, considering both textual and visual modality. Second, for ranking, we propose Ranker to score the candidates, dynamically combining the two models with complementary strength for both short and long spans respectively. Experimental result demonstrates that our model achieves state-of-the-art performance.
Gli stili APA, Harvard, Vancouver, ISO e altri
22

Gao, Feng, Yuanyuan Ge e Yongge Liu. "Remember and forget: video and text fusion for video question answering". Multimedia Tools and Applications 77, n. 22 (27 marzo 2018): 29269–82. http://dx.doi.org/10.1007/s11042-018-5868-x.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
23

Li, Zhangbin, Dan Guo, Jinxing Zhou, Jing Zhang e Meng Wang. "Object-Aware Adaptive-Positivity Learning for Audio-Visual Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 38, n. 4 (24 marzo 2024): 3306–14. http://dx.doi.org/10.1609/aaai.v38i4.28116.

Testo completo
Abstract (sommario):
This paper focuses on the Audio-Visual Question Answering (AVQA) task that aims to answer questions derived from untrimmed audible videos. To generate accurate answers, an AVQA model is expected to find the most informative audio-visual clues relevant to the given questions. In this paper, we propose to explicitly consider fine-grained visual objects in video frames (object-level clues) and explore the multi-modal relations (\textit{i.e.}, the object, audio, and question) in terms of feature interaction and model optimization. For the former, we present an end-to-end object-oriented network that adopts a question-conditioned clue discovery module to concentrate audio/visual modalities on respective keywords of the question and designs a modality-conditioned clue collection module to highlight closely associated audio segments or visual objects. For model optimization, we propose an object-aware adaptive-positivity learning strategy that selects the highly semantic-matched multi-modal pair as \textit{positivity}. Specifically, we design two object-aware contrastive loss functions to identify the highly relevant question-object pairs and audio-object pairs, respectively. These selected pairs are constrained to have larger similarity values than the mismatched pairs. The positivity-selecting process is adaptive as the positivity pairs selected in each video frame may be different. These two object-aware objectives help the model understand \textit{which objects are exactly relevant to the question} and \textit{which are making sounds}. Extensive experiments on the MUSIC-AVQA dataset demonstrate the proposed method is effective in finding favorable audio-visual clues and also achieves new state-of-the-art question-answering performance. The code is available at https://github.com/zhangbin-ai/APL.
Gli stili APA, Harvard, Vancouver, ISO e altri
24

Shao, Zhuang, Jiahui Wan e Linlin Zong. "A Video Question Answering Model Based on Knowledge Distillation". Information 14, n. 6 (12 giugno 2023): 328. http://dx.doi.org/10.3390/info14060328.

Testo completo
Abstract (sommario):
Video question answering (QA) is a cross-modal task that requires understanding the video content to answer questions. Current techniques address this challenge by employing stacked modules, such as attention mechanisms and graph convolutional networks. These methods reason about the semantics of video features and their interaction with text-based questions, yielding excellent results. However, these approaches often learn and fuse features representing different aspects of the video separately, neglecting the intra-interaction and overlooking the latent complex correlations between the extracted features. Additionally, the stacking of modules introduces a large number of parameters, making model training more challenging. To address these issues, we propose a novel multimodal knowledge distillation method that leverages the strengths of knowledge distillation for model compression and feature enhancement. Specifically, the fused features in the larger teacher model are distilled into knowledge, which guides the learning of appearance and motion features in the smaller student model. By incorporating cross-modal information in the early stages, the appearance and motion features can discover their related and complementary potential relationships, thus improving the overall model performance. Despite its simplicity, our extensive experiments on the widely used video QA datasets, MSVD-QA and MSRVTT-QA, demonstrate clear performance improvements over prior methods. These results validate the effectiveness of the proposed knowledge distillation approach.
Gli stili APA, Harvard, Vancouver, ISO e altri
25

Li, Xiangpeng, Jingkuan Song, Lianli Gao, Xianglong Liu, Wenbing Huang, Xiangnan He e Chuang Gan. "Beyond RNNs: Positional Self-Attention with Co-Attention for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 33 (17 luglio 2019): 8658–65. http://dx.doi.org/10.1609/aaai.v33i01.33018658.

Testo completo
Abstract (sommario):
Most of the recent progresses on visual question answering are based on recurrent neural networks (RNNs) with attention. Despite the success, these models are often timeconsuming and having difficulties in modeling long range dependencies due to the sequential nature of RNNs. We propose a new architecture, Positional Self-Attention with Coattention (PSAC), which does not require RNNs for video question answering. Specifically, inspired by the success of self-attention in machine translation task, we propose a Positional Self-Attention to calculate the response at each position by attending to all positions within the same sequence, and then add representations of absolute positions. Therefore, PSAC can exploit the global dependencies of question and temporal information in the video, and make the process of question and video encoding executed in parallel. Furthermore, in addition to attending to the video features relevant to the given questions (i.e., video attention), we utilize the co-attention mechanism by simultaneously modeling “what words to listen to” (question attention). To the best of our knowledge, this is the first work of replacing RNNs with selfattention for the task of visual question answering. Experimental results of four tasks on the benchmark dataset show that our model significantly outperforms the state-of-the-art on three tasks and attains comparable result on the Count task. Our model requires less computation time and achieves better performance compared with the RNNs-based methods. Additional ablation study demonstrates the effect of each component of our proposed model.
Gli stili APA, Harvard, Vancouver, ISO e altri
26

Jiang, Pin, e Yahong Han. "Reasoning with Heterogeneous Graph Alignment for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 34, n. 07 (3 aprile 2020): 11109–16. http://dx.doi.org/10.1609/aaai.v34i07.6767.

Testo completo
Abstract (sommario):
The dominant video question answering methods are based on fine-grained representation or model-specific attention mechanism. They usually process video and question separately, then feed the representations of different modalities into following late fusion networks. Although these methods use information of one modality to boost the other, they neglect to integrate correlations of both inter- and intra-modality in an uniform module. We propose a deep heterogeneous graph alignment network over the video shots and question words. Furthermore, we explore the network architecture from four steps: representation, fusion, alignment, and reasoning. Within our network, the inter- and intra-modality information can be aligned and interacted simultaneously over the heterogeneous graph and used for cross-modal reasoning. We evaluate our method on three benchmark datasets and conduct extensive ablation study to the effectiveness of the network architecture. Experiments show the network to be superior in quality.
Gli stili APA, Harvard, Vancouver, ISO e altri
27

Gu, Mao, Zhou Zhao, Weike Jin, Richang Hong e Fei Wu. "Graph-Based Multi-Interaction Network for Video Question Answering". IEEE Transactions on Image Processing 30 (2021): 2758–70. http://dx.doi.org/10.1109/tip.2021.3051756.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
28

Yu-Chieh Wu e Jie-Chi Yang. "A Robust Passage Retrieval Algorithm for Video Question Answering". IEEE Transactions on Circuits and Systems for Video Technology 18, n. 10 (ottobre 2008): 1411–21. http://dx.doi.org/10.1109/tcsvt.2008.2002831.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
29

Wang, Weining, Yan Huang e Liang Wang. "Long video question answering: A Matching-guided Attention Model". Pattern Recognition 102 (giugno 2020): 107248. http://dx.doi.org/10.1016/j.patcog.2020.107248.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
30

Ye, Yunan, Shifeng Zhang, Yimeng Li, Xufeng Qian, Siliang Tang, Shiliang Pu e Jun Xiao. "Video question answering via grounded cross-attention network learning". Information Processing & Management 57, n. 4 (luglio 2020): 102265. http://dx.doi.org/10.1016/j.ipm.2020.102265.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
31

Zhang, Wenqiao, Siliang Tang, Yanpeng Cao, Shiliang Pu, Fei Wu e Yueting Zhuang. "Frame Augmented Alternating Attention Network for Video Question Answering". IEEE Transactions on Multimedia 22, n. 4 (aprile 2020): 1032–41. http://dx.doi.org/10.1109/tmm.2019.2935678.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
32

Zha, Zheng-Jun, Jiawei Liu, Tianhao Yang e Yongdong Zhang. "Spatiotemporal-Textual Co-Attention Network for Video Question Answering". ACM Transactions on Multimedia Computing, Communications, and Applications 15, n. 2s (12 agosto 2019): 1–18. http://dx.doi.org/10.1145/3320061.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
33

Jiang, Yimin, Tingfei Yan, Mingze Yao, Huibing Wang e Wenzhe Liu. "Cascade transformers with dynamic attention for video question answering". Computer Vision and Image Understanding 242 (maggio 2024): 103983. http://dx.doi.org/10.1016/j.cviu.2024.103983.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
34

Jiao, Guie. "Realization of Video Question Answering System Based on Flash under RIA". Applied Mechanics and Materials 411-414 (settembre 2013): 970–73. http://dx.doi.org/10.4028/www.scientific.net/amm.411-414.970.

Testo completo
Abstract (sommario):
Video question answering system is designed and developed based on Flash platform technology,belongs of B / S (Browser / Server) structure,used to meet the needs of specific groupss teaching and studying.This paper describes the core code of the real-time Q&A video module about video creating and video playback,also describes the relevant realization technology and other functional modules of the system.
Gli stili APA, Harvard, Vancouver, ISO e altri
35

Liu, Mingyang, Ruomei Wang, Fan Zhou e Ge Lin. "Temporally Multi-Modal Semantic Reasoning with Spatial Language Constraints for Video Question Answering". Symmetry 14, n. 6 (31 maggio 2022): 1133. http://dx.doi.org/10.3390/sym14061133.

Testo completo
Abstract (sommario):
Video question answering (QA) aims to understand the video scene and underlying plot by answering video questions. An algorithm that can competently cope with this task needs to be able to: (1) collect multi-modal information scattered in the video frame sequence while extracting, interpreting, and utilizing the potential semantic clues provided by each piece of modal information in the video, (2) integrate the multi-modal context of the above semantic clues and understand the cause and effect of the story as it evolves, and (3) identify and integrate those temporally adjacent or non-adjacent effective semantic clues implied in the above context information to provide reasonable and sufficient visual semantic information for the final question reasoning. In response to the above requirements, a novel temporally multi-modal semantic reasoning with spatial language constraints video QA solution is reported in this paper, which includes a significant feature extraction module used to extract multi-modal features according to a significant sampling strategy, a spatial language constraints module used to recognize and reason spatial dimensions in video frames under the guidance of questions, and a temporal language interaction module used to locate the temporal dimension semantic clues of the appearance features and motion features sequence. Specifically, for a question, the result processed by the spatial language constraints module is to obtain visual clues related to the question from a single image and filter out unwanted spatial information. Further, the temporal language interaction module symmetrically integrates visual clues of the appearance information and motion information scattered throughout the temporal dimensions, obtains the temporally adjacent or non-adjacent effective semantic clue, and filters out irrelevant or detrimental context information. The proposed video QA solution is validated on several video QA benchmarks. Comprehensive ablation experiments have confirmed that modeling the significant video information can improve QA ability. The spatial language constraints module and temporal language interaction module can better collect and summarize visual semantic clues.
Gli stili APA, Harvard, Vancouver, ISO e altri
36

Jin, Yao, Guocheng Niu, Xinyan Xiao, Jian Zhang, Xi Peng e Jun Yu. "Knowledge-Constrained Answer Generation for Open-Ended Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 37, n. 7 (26 giugno 2023): 8141–49. http://dx.doi.org/10.1609/aaai.v37i7.25983.

Testo completo
Abstract (sommario):
Open-ended Video question answering (open-ended VideoQA) aims to understand video content and question semantics to generate the correct answers. Most of the best performing models define the problem as a discriminative task of multi-label classification. In real-world scenarios, however, it is difficult to define a candidate set that includes all possible answers. In this paper, we propose a Knowledge-constrained Generative VideoQA Algorithm (KcGA) with an encoder-decoder pipeline, which enables out-of-domain answer generation through an adaptive external knowledge module and a multi-stream information control mechanism. We use ClipBERT to extract the video-question features, extract framewise object-level external knowledge from a commonsense knowledge base and compute the contextual-aware episode memory units via an attention based GRU to form the external knowledge features, and exploit multi-stream information control mechanism to fuse video-question and external knowledge features such that the semantic complementation and alignment are well achieved. We evaluate our model on two open-ended benchmark datasets to demonstrate that we can effectively and robustly generate high-quality answers without restrictions of training data.
Gli stili APA, Harvard, Vancouver, ISO e altri
37

Patel, Hardik Bhikhabhai, e Sailesh Suryanarayan Iyer. "Comparative Study of Multimedia Question Answering System Models". ECS Transactions 107, n. 1 (24 aprile 2022): 2033–42. http://dx.doi.org/10.1149/10701.2033ecst.

Testo completo
Abstract (sommario):
Question Answering Systems is a mechanized way to deal with finding right solution to the inquiries posed by human in their own language. The enormous measure of data returned by web crawlers like Google, Yahoo, and so on, clients are over-burden to track down the right data from the web search tools. This QA approach takes care of these issues. Rather than returning a rundown of reports and connections from the momentum web crawler, the QA framework gives data from a bunch of replied, alongside reasonable media information. Question answering procedures plans to exploit top to bottom etymological and media content investigation just as area information to return accurate responses to regular language questions. QA framework has rundown of boundary like: question size, m-dimension, technique count, answer cardinality, and answer range. This paper investigates the advancement of mixed media question replying research, their plan boundary, and subtleties its future perspectives. Multiple Line Questions and Video Question Answering methods have been discussed at length and their performance measured on QA framework parameters.
Gli stili APA, Harvard, Vancouver, ISO e altri
38

Gao, Lianli, Yu Lei, Pengpeng Zeng, Jingkuan Song, Meng Wang e Heng Tao Shen. "Hierarchical Representation Network With Auxiliary Tasks for Video Captioning and Video Question Answering". IEEE Transactions on Image Processing 31 (2022): 202–15. http://dx.doi.org/10.1109/tip.2021.3120867.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
39

Yang, Zekun, Noa Garcia, Chenhui Chu, Mayu Otani, Yuta Nakashima e Haruo Takemura. "A comparative study of language transformers for video question answering". Neurocomputing 445 (luglio 2021): 121–33. http://dx.doi.org/10.1016/j.neucom.2021.02.092.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
40

Zhao, Zhou, Zhu Zhang, Shuwen Xiao, Zhenxin Xiao, Xiaohui Yan, Jun Yu, Deng Cai e Fei Wu. "Long-Form Video Question Answering via Dynamic Hierarchical Reinforced Networks". IEEE Transactions on Image Processing 28, n. 12 (dicembre 2019): 5939–52. http://dx.doi.org/10.1109/tip.2019.2922062.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
41

Yin, Chengxiang, Jian Tang, Zhiyuan Xu e Yanzhi Wang. "Memory Augmented Deep Recurrent Neural Network for Video Question Answering". IEEE Transactions on Neural Networks and Learning Systems 31, n. 9 (settembre 2020): 3159–67. http://dx.doi.org/10.1109/tnnls.2019.2938015.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
42

Wang, Zheng, Fangtao Li, Kaoru Ota, Mianxiong Dong e Bin Wu. "ReGR: Relation-aware graph reasoning framework for video question answering". Information Processing & Management 60, n. 4 (luglio 2023): 103375. http://dx.doi.org/10.1016/j.ipm.2023.103375.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
43

Al Mehmadi, Shima M., Yakoub Bazi, Mohamad M. Al Rahhal e Mansour Zuair. "Learning to enhance areal video captioning with visual question answering". International Journal of Remote Sensing 45, n. 18 (30 agosto 2024): 6395–407. http://dx.doi.org/10.1080/01431161.2024.2388875.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
44

Zhuang, Xuqiang, Fang’ai Liu, Jian Hou, Jianhua Hao e Xiaohong Cai. "Modality attention fusion model with hybrid multi-head self-attention for video understanding". PLOS ONE 17, n. 10 (6 ottobre 2022): e0275156. http://dx.doi.org/10.1371/journal.pone.0275156.

Testo completo
Abstract (sommario):
Video question answering (Video-QA) is a subject undergoing intense study in Artificial Intelligence, which is one of the tasks which can evaluate such AI abilities. In this paper, we propose a Modality Attention Fusion framework with Hybrid Multi-head Self-attention (MAF-HMS). MAF-HMS focuses on the task of answering multiple-choice questions regarding a video-subtitle-QA representation by fusion of attention and self-attention between each modality. We use BERT to extract text features, and use Faster R-CNN to ex-tract visual features to provide a useful input representation for our model to answer questions. In addition, we have constructed a Modality Attention Fusion (MAF) framework for the attention fusion matrix from different modalities (video, subtitles, QA), and use a Hybrid Multi-headed Self-attention (HMS) to further determine the correct answer. Experiments on three separate scene datasets show our overall model outperforms the baseline methods by a large margin. Finally, we conducted extensive ablation studies to verify the various components of the network and demonstrate the effectiveness and advantages of our method over existing methods through question type and required modality experimental results.
Gli stili APA, Harvard, Vancouver, ISO e altri
45

Peng, Min, Chongyang Wang, Yu Shi e Xiang-Dong Zhou. "Efficient End-to-End Video Question Answering with Pyramidal Multimodal Transformer". Proceedings of the AAAI Conference on Artificial Intelligence 37, n. 2 (26 giugno 2023): 2038–46. http://dx.doi.org/10.1609/aaai.v37i2.25296.

Testo completo
Abstract (sommario):
This paper presents a new method for end-to-end Video Question Answering (VideoQA), aside from the current popularity of using large-scale pre-training with huge feature extractors. We achieve this with a pyramidal multimodal transformer (PMT) model, which simply incorporates a learnable word embedding layer, a few convolutional and transformer layers. We use the anisotropic pyramid to fulfill video-language interactions across different spatio-temporal scales. In addition to the canonical pyramid, which includes both bottom-up and top-down pathways with lateral connections, novel strategies are proposed to decompose the visual feature stream into spatial and temporal sub-streams at different scales and implement their interactions with the linguistic semantics while preserving the integrity of local and global semantics. We demonstrate better or on-par performances with high computational efficiency against state-of-the-art methods on five VideoQA benchmarks. Our ablation study shows the scalability of our model that achieves competitive results for text-to-video retrieval by leveraging feature extractors with reusable pre-trained weights, and also the effectiveness of the pyramid. Code available at: https://github.com/Trunpm/PMT-AAAI23.
Gli stili APA, Harvard, Vancouver, ISO e altri
46

Kim, Seonhoon, Seohyeong Jeong, Eunbyul Kim, Inho Kang e Nojun Kwak. "Self-supervised Pre-training and Contrastive Representation Learning for Multiple-choice Video QA". Proceedings of the AAAI Conference on Artificial Intelligence 35, n. 14 (18 maggio 2021): 13171–79. http://dx.doi.org/10.1609/aaai.v35i14.17556.

Testo completo
Abstract (sommario):
Video Question Answering (VideoQA) requires fine-grained understanding of both video and language modalities to answer the given questions. In this paper, we propose novel training schemes for multiple-choice video question answering with a self-supervised pre-training stage and a supervised contrastive learning in the main stage as an auxiliary learning. In the self-supervised pre-training stage, we transform the original problem format of predicting the correct answer into the one that predicts the relevant question to provide a model with broader contextual inputs without any further dataset or annotation. For contrastive learning in the main stage, we add a masking noise to the input corresponding to the ground-truth answer, and consider the original input of the ground-truth answer as a positive sample, while treating the rest as negative samples. By mapping the positive sample closer to the masked input, we show that the model performance is improved. We further employ locally aligned attention to focus more effectively on the video frames that are particularly relevant to the given corresponding subtitle sentences. We evaluate our proposed model on highly competitive benchmark datasets related to multiple-choice video QA: TVQA, TVQA+, and DramaQA. Experimental results show that our model achieves state-of-the-art performance on all datasets. We also validate our approaches through further analyses.
Gli stili APA, Harvard, Vancouver, ISO e altri
47

Park, Gyu-Min, A.-Yeong Kim e Seong-Bae Park. "Confident Multiple Choice Learning-based Ensemble Model for Video Question-Answering". Journal of KIISE 49, n. 4 (30 aprile 2022): 284–90. http://dx.doi.org/10.5626/jok.2022.49.4.284.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
48

Liu, Yun, Xiaoming Zhang, Feiran Huang, Bo Zhang e Zhoujun Li. "Cross-Attentional Spatio-Temporal Semantic Graph Networks for Video Question Answering". IEEE Transactions on Image Processing 31 (2022): 1684–96. http://dx.doi.org/10.1109/tip.2022.3142526.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
49

Zhao, Zhou, Zhu Zhang, Xinghua Jiang e Deng Cai. "Multi-Turn Video Question Answering via Hierarchical Attention Context Reinforced Networks". IEEE Transactions on Image Processing 28, n. 8 (agosto 2019): 3860–72. http://dx.doi.org/10.1109/tip.2019.2902106.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
50

Yu, Ting, Jun Yu, Zhou Yu e Dacheng Tao. "Compositional Attention Networks With Two-Stream Fusion for Video Question Answering". IEEE Transactions on Image Processing 29 (2020): 1204–18. http://dx.doi.org/10.1109/tip.2019.2940677.

Testo completo
Gli stili APA, Harvard, Vancouver, ISO e altri
Offriamo sconti su tutti i piani premium per gli autori le cui opere sono incluse in raccolte letterarie tematiche. Contattaci per ottenere un codice promozionale unico!

Vai alla bibliografia