Zeitschriftenartikel zum Thema „Video question answering“
Geben Sie eine Quelle nach APA, MLA, Chicago, Harvard und anderen Zitierweisen an
Machen Sie sich mit Top-50 Zeitschriftenartikel für die Forschung zum Thema "Video question answering" bekannt.
Neben jedem Werk im Literaturverzeichnis ist die Option "Zur Bibliographie hinzufügen" verfügbar. Nutzen Sie sie, wird Ihre bibliographische Angabe des gewählten Werkes nach der nötigen Zitierweise (APA, MLA, Harvard, Chicago, Vancouver usw.) automatisch gestaltet.
Sie können auch den vollen Text der wissenschaftlichen Publikation im PDF-Format herunterladen und eine Online-Annotation der Arbeit lesen, wenn die relevanten Parameter in den Metadaten verfügbar sind.
Sehen Sie die Zeitschriftenartikel für verschiedene Spezialgebieten durch und erstellen Sie Ihre Bibliographie auf korrekte Weise.
Lei, Chenyi, Lei Wu, Dong Liu, Zhao Li, Guoxin Wang, Haihong Tang und Houqiang Li. „Multi-Question Learning for Visual Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 34, Nr. 07 (03.04.2020): 11328–35. http://dx.doi.org/10.1609/aaai.v34i07.6794.
Der volle Inhalt der QuelleRuwa, Nelson, Qirong Mao, Liangjun Wang und Jianping Gou. „Affective question answering on video“. Neurocomputing 363 (Oktober 2019): 125–39. http://dx.doi.org/10.1016/j.neucom.2019.06.046.
Der volle Inhalt der QuelleWang, Yueqian, Yuxuan Wang, Kai Chen und Dongyan Zhao. „STAIR: Spatial-Temporal Reasoning with Auditable Intermediate Results for Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 38, Nr. 17 (24.03.2024): 19215–23. http://dx.doi.org/10.1609/aaai.v38i17.29890.
Der volle Inhalt der QuelleZong, Linlin, Jiahui Wan, Xianchao Zhang, Xinyue Liu, Wenxin Liang und Bo Xu. „Video-Context Aligned Transformer for Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 38, Nr. 17 (24.03.2024): 19795–803. http://dx.doi.org/10.1609/aaai.v38i17.29954.
Der volle Inhalt der QuelleHuang, Deng, Peihao Chen, Runhao Zeng, Qing Du, Mingkui Tan und Chuang Gan. „Location-Aware Graph Convolutional Networks for Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 34, Nr. 07 (03.04.2020): 11021–28. http://dx.doi.org/10.1609/aaai.v34i07.6737.
Der volle Inhalt der QuelleGao, Lianli, Pengpeng Zeng, Jingkuan Song, Yuan-Fang Li, Wu Liu, Tao Mei und Heng Tao Shen. „Structured Two-Stream Attention Network for Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 33 (17.07.2019): 6391–98. http://dx.doi.org/10.1609/aaai.v33i01.33016391.
Der volle Inhalt der QuelleKumar, Krishnamoorthi Magesh, und P. Valarmathie. „Domain and Intelligence Based Multimedia Question Answering System“. International Journal of Evaluation and Research in Education (IJERE) 5, Nr. 3 (01.09.2016): 227. http://dx.doi.org/10.11591/ijere.v5i3.4544.
Der volle Inhalt der QuelleXue, Hongyang, Zhou Zhao und Deng Cai. „Unifying the Video and Question Attentions for Open-Ended Video Question Answering“. IEEE Transactions on Image Processing 26, Nr. 12 (Dezember 2017): 5656–66. http://dx.doi.org/10.1109/tip.2017.2746267.
Der volle Inhalt der QuelleJang, Yunseok, Yale Song, Chris Dongjoo Kim, Youngjae Yu, Youngjin Kim und Gunhee Kim. „Video Question Answering with Spatio-Temporal Reasoning“. International Journal of Computer Vision 127, Nr. 10 (18.06.2019): 1385–412. http://dx.doi.org/10.1007/s11263-019-01189-x.
Der volle Inhalt der QuelleZhuang, Yueting, Dejing Xu, Xin Yan, Wenzhuo Cheng, Zhou Zhao, Shiliang Pu und Jun Xiao. „Multichannel Attention Refinement for Video Question Answering“. ACM Transactions on Multimedia Computing, Communications, and Applications 16, Nr. 1s (28.04.2020): 1–23. http://dx.doi.org/10.1145/3366710.
Der volle Inhalt der QuelleGarcia, Noa, Mayu Otani, Chenhui Chu und Yuta Nakashima. „KnowIT VQA: Answering Knowledge-Based Questions about Videos“. Proceedings of the AAAI Conference on Artificial Intelligence 34, Nr. 07 (03.04.2020): 10826–34. http://dx.doi.org/10.1609/aaai.v34i07.6713.
Der volle Inhalt der QuelleMao, Jianguo, Wenbin Jiang, Hong Liu, Xiangdong Wang und Yajuan Lyu. „Inferential Knowledge-Enhanced Integrated Reasoning for Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 37, Nr. 11 (26.06.2023): 13380–88. http://dx.doi.org/10.1609/aaai.v37i11.26570.
Der volle Inhalt der QuelleJiang, Jianwen, Ziqiang Chen, Haojie Lin, Xibin Zhao und Yue Gao. „Divide and Conquer: Question-Guided Spatio-Temporal Contextual Attention for Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 34, Nr. 07 (03.04.2020): 11101–8. http://dx.doi.org/10.1609/aaai.v34i07.6766.
Der volle Inhalt der QuelleYang, Saelyne, Sunghyun Park, Yunseok Jang und Moontae Lee. „YTCommentQA: Video Question Answerability in Instructional Videos“. Proceedings of the AAAI Conference on Artificial Intelligence 38, Nr. 17 (24.03.2024): 19359–67. http://dx.doi.org/10.1609/aaai.v38i17.29906.
Der volle Inhalt der QuelleYu, Zhou, Dejing Xu, Jun Yu, Ting Yu, Zhou Zhao, Yueting Zhuang und Dacheng Tao. „ActivityNet-QA: A Dataset for Understanding Complex Web Videos via Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 33 (17.07.2019): 9127–34. http://dx.doi.org/10.1609/aaai.v33i01.33019127.
Der volle Inhalt der QuelleCherian, Anoop, Chiori Hori, Tim K. Marks und Jonathan Le Roux. „(2.5+1)D Spatio-Temporal Scene Graphs for Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 36, Nr. 1 (28.06.2022): 444–53. http://dx.doi.org/10.1609/aaai.v36i1.19922.
Der volle Inhalt der QuelleChu, Wenqing, Hongyang Xue, Zhou Zhao, Deng Cai und Chengwei Yao. „The forgettable-watcher model for video question answering“. Neurocomputing 314 (November 2018): 386–93. http://dx.doi.org/10.1016/j.neucom.2018.06.069.
Der volle Inhalt der QuelleZhu, Linchao, Zhongwen Xu, Yi Yang und Alexander G. Hauptmann. „Uncovering the Temporal Context for Video Question Answering“. International Journal of Computer Vision 124, Nr. 3 (13.07.2017): 409–21. http://dx.doi.org/10.1007/s11263-017-1033-7.
Der volle Inhalt der QuelleLee, Yue-Shi, Yu-Chieh Wu und Jie-Chi Yang. „BVideoQA: Online English/Chinese bilingual video question answering“. Journal of the American Society for Information Science and Technology 60, Nr. 3 (März 2009): 509–25. http://dx.doi.org/10.1002/asi.21002.
Der volle Inhalt der QuelleXiao, Junbin, Angela Yao, Zhiyuan Liu, Yicong Li, Wei Ji und Tat-Seng Chua. „Video as Conditional Graph Hierarchy for Multi-Granular Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 36, Nr. 3 (28.06.2022): 2804–12. http://dx.doi.org/10.1609/aaai.v36i3.20184.
Der volle Inhalt der QuelleLee, Kyungjae, Nan Duan, Lei Ji, Jason Li und Seung-won Hwang. „Segment-Then-Rank: Non-Factoid Question Answering on Instructional Videos“. Proceedings of the AAAI Conference on Artificial Intelligence 34, Nr. 05 (03.04.2020): 8147–54. http://dx.doi.org/10.1609/aaai.v34i05.6327.
Der volle Inhalt der QuelleGao, Feng, Yuanyuan Ge und Yongge Liu. „Remember and forget: video and text fusion for video question answering“. Multimedia Tools and Applications 77, Nr. 22 (27.03.2018): 29269–82. http://dx.doi.org/10.1007/s11042-018-5868-x.
Der volle Inhalt der QuelleLi, Zhangbin, Dan Guo, Jinxing Zhou, Jing Zhang und Meng Wang. „Object-Aware Adaptive-Positivity Learning for Audio-Visual Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 38, Nr. 4 (24.03.2024): 3306–14. http://dx.doi.org/10.1609/aaai.v38i4.28116.
Der volle Inhalt der QuelleShao, Zhuang, Jiahui Wan und Linlin Zong. „A Video Question Answering Model Based on Knowledge Distillation“. Information 14, Nr. 6 (12.06.2023): 328. http://dx.doi.org/10.3390/info14060328.
Der volle Inhalt der QuelleLi, Xiangpeng, Jingkuan Song, Lianli Gao, Xianglong Liu, Wenbing Huang, Xiangnan He und Chuang Gan. „Beyond RNNs: Positional Self-Attention with Co-Attention for Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 33 (17.07.2019): 8658–65. http://dx.doi.org/10.1609/aaai.v33i01.33018658.
Der volle Inhalt der QuelleJiang, Pin, und Yahong Han. „Reasoning with Heterogeneous Graph Alignment for Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 34, Nr. 07 (03.04.2020): 11109–16. http://dx.doi.org/10.1609/aaai.v34i07.6767.
Der volle Inhalt der QuelleGu, Mao, Zhou Zhao, Weike Jin, Richang Hong und Fei Wu. „Graph-Based Multi-Interaction Network for Video Question Answering“. IEEE Transactions on Image Processing 30 (2021): 2758–70. http://dx.doi.org/10.1109/tip.2021.3051756.
Der volle Inhalt der QuelleYu-Chieh Wu und Jie-Chi Yang. „A Robust Passage Retrieval Algorithm for Video Question Answering“. IEEE Transactions on Circuits and Systems for Video Technology 18, Nr. 10 (Oktober 2008): 1411–21. http://dx.doi.org/10.1109/tcsvt.2008.2002831.
Der volle Inhalt der QuelleWang, Weining, Yan Huang und Liang Wang. „Long video question answering: A Matching-guided Attention Model“. Pattern Recognition 102 (Juni 2020): 107248. http://dx.doi.org/10.1016/j.patcog.2020.107248.
Der volle Inhalt der QuelleYe, Yunan, Shifeng Zhang, Yimeng Li, Xufeng Qian, Siliang Tang, Shiliang Pu und Jun Xiao. „Video question answering via grounded cross-attention network learning“. Information Processing & Management 57, Nr. 4 (Juli 2020): 102265. http://dx.doi.org/10.1016/j.ipm.2020.102265.
Der volle Inhalt der QuelleZhang, Wenqiao, Siliang Tang, Yanpeng Cao, Shiliang Pu, Fei Wu und Yueting Zhuang. „Frame Augmented Alternating Attention Network for Video Question Answering“. IEEE Transactions on Multimedia 22, Nr. 4 (April 2020): 1032–41. http://dx.doi.org/10.1109/tmm.2019.2935678.
Der volle Inhalt der QuelleZha, Zheng-Jun, Jiawei Liu, Tianhao Yang und Yongdong Zhang. „Spatiotemporal-Textual Co-Attention Network for Video Question Answering“. ACM Transactions on Multimedia Computing, Communications, and Applications 15, Nr. 2s (12.08.2019): 1–18. http://dx.doi.org/10.1145/3320061.
Der volle Inhalt der QuelleJiang, Yimin, Tingfei Yan, Mingze Yao, Huibing Wang und Wenzhe Liu. „Cascade transformers with dynamic attention for video question answering“. Computer Vision and Image Understanding 242 (Mai 2024): 103983. http://dx.doi.org/10.1016/j.cviu.2024.103983.
Der volle Inhalt der QuelleJiao, Guie. „Realization of Video Question Answering System Based on Flash under RIA“. Applied Mechanics and Materials 411-414 (September 2013): 970–73. http://dx.doi.org/10.4028/www.scientific.net/amm.411-414.970.
Der volle Inhalt der QuelleLiu, Mingyang, Ruomei Wang, Fan Zhou und Ge Lin. „Temporally Multi-Modal Semantic Reasoning with Spatial Language Constraints for Video Question Answering“. Symmetry 14, Nr. 6 (31.05.2022): 1133. http://dx.doi.org/10.3390/sym14061133.
Der volle Inhalt der QuelleJin, Yao, Guocheng Niu, Xinyan Xiao, Jian Zhang, Xi Peng und Jun Yu. „Knowledge-Constrained Answer Generation for Open-Ended Video Question Answering“. Proceedings of the AAAI Conference on Artificial Intelligence 37, Nr. 7 (26.06.2023): 8141–49. http://dx.doi.org/10.1609/aaai.v37i7.25983.
Der volle Inhalt der QuellePatel, Hardik Bhikhabhai, und Sailesh Suryanarayan Iyer. „Comparative Study of Multimedia Question Answering System Models“. ECS Transactions 107, Nr. 1 (24.04.2022): 2033–42. http://dx.doi.org/10.1149/10701.2033ecst.
Der volle Inhalt der QuelleGao, Lianli, Yu Lei, Pengpeng Zeng, Jingkuan Song, Meng Wang und Heng Tao Shen. „Hierarchical Representation Network With Auxiliary Tasks for Video Captioning and Video Question Answering“. IEEE Transactions on Image Processing 31 (2022): 202–15. http://dx.doi.org/10.1109/tip.2021.3120867.
Der volle Inhalt der QuelleYang, Zekun, Noa Garcia, Chenhui Chu, Mayu Otani, Yuta Nakashima und Haruo Takemura. „A comparative study of language transformers for video question answering“. Neurocomputing 445 (Juli 2021): 121–33. http://dx.doi.org/10.1016/j.neucom.2021.02.092.
Der volle Inhalt der QuelleZhao, Zhou, Zhu Zhang, Shuwen Xiao, Zhenxin Xiao, Xiaohui Yan, Jun Yu, Deng Cai und Fei Wu. „Long-Form Video Question Answering via Dynamic Hierarchical Reinforced Networks“. IEEE Transactions on Image Processing 28, Nr. 12 (Dezember 2019): 5939–52. http://dx.doi.org/10.1109/tip.2019.2922062.
Der volle Inhalt der QuelleYin, Chengxiang, Jian Tang, Zhiyuan Xu und Yanzhi Wang. „Memory Augmented Deep Recurrent Neural Network for Video Question Answering“. IEEE Transactions on Neural Networks and Learning Systems 31, Nr. 9 (September 2020): 3159–67. http://dx.doi.org/10.1109/tnnls.2019.2938015.
Der volle Inhalt der QuelleWang, Zheng, Fangtao Li, Kaoru Ota, Mianxiong Dong und Bin Wu. „ReGR: Relation-aware graph reasoning framework for video question answering“. Information Processing & Management 60, Nr. 4 (Juli 2023): 103375. http://dx.doi.org/10.1016/j.ipm.2023.103375.
Der volle Inhalt der QuelleAl Mehmadi, Shima M., Yakoub Bazi, Mohamad M. Al Rahhal und Mansour Zuair. „Learning to enhance areal video captioning with visual question answering“. International Journal of Remote Sensing 45, Nr. 18 (30.08.2024): 6395–407. http://dx.doi.org/10.1080/01431161.2024.2388875.
Der volle Inhalt der QuelleZhuang, Xuqiang, Fang’ai Liu, Jian Hou, Jianhua Hao und Xiaohong Cai. „Modality attention fusion model with hybrid multi-head self-attention for video understanding“. PLOS ONE 17, Nr. 10 (06.10.2022): e0275156. http://dx.doi.org/10.1371/journal.pone.0275156.
Der volle Inhalt der QuellePeng, Min, Chongyang Wang, Yu Shi und Xiang-Dong Zhou. „Efficient End-to-End Video Question Answering with Pyramidal Multimodal Transformer“. Proceedings of the AAAI Conference on Artificial Intelligence 37, Nr. 2 (26.06.2023): 2038–46. http://dx.doi.org/10.1609/aaai.v37i2.25296.
Der volle Inhalt der QuelleKim, Seonhoon, Seohyeong Jeong, Eunbyul Kim, Inho Kang und Nojun Kwak. „Self-supervised Pre-training and Contrastive Representation Learning for Multiple-choice Video QA“. Proceedings of the AAAI Conference on Artificial Intelligence 35, Nr. 14 (18.05.2021): 13171–79. http://dx.doi.org/10.1609/aaai.v35i14.17556.
Der volle Inhalt der QuellePark, Gyu-Min, A.-Yeong Kim und Seong-Bae Park. „Confident Multiple Choice Learning-based Ensemble Model for Video Question-Answering“. Journal of KIISE 49, Nr. 4 (30.04.2022): 284–90. http://dx.doi.org/10.5626/jok.2022.49.4.284.
Der volle Inhalt der QuelleLiu, Yun, Xiaoming Zhang, Feiran Huang, Bo Zhang und Zhoujun Li. „Cross-Attentional Spatio-Temporal Semantic Graph Networks for Video Question Answering“. IEEE Transactions on Image Processing 31 (2022): 1684–96. http://dx.doi.org/10.1109/tip.2022.3142526.
Der volle Inhalt der QuelleZhao, Zhou, Zhu Zhang, Xinghua Jiang und Deng Cai. „Multi-Turn Video Question Answering via Hierarchical Attention Context Reinforced Networks“. IEEE Transactions on Image Processing 28, Nr. 8 (August 2019): 3860–72. http://dx.doi.org/10.1109/tip.2019.2902106.
Der volle Inhalt der QuelleYu, Ting, Jun Yu, Zhou Yu und Dacheng Tao. „Compositional Attention Networks With Two-Stream Fusion for Video Question Answering“. IEEE Transactions on Image Processing 29 (2020): 1204–18. http://dx.doi.org/10.1109/tip.2019.2940677.
Der volle Inhalt der Quelle