Artykuły w czasopismach na temat „Video question answering”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Sprawdź 50 najlepszych artykułów w czasopismach naukowych na temat „Video question answering”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Przeglądaj artykuły w czasopismach z różnych dziedzin i twórz odpowiednie bibliografie.
Lei, Chenyi, Lei Wu, Dong Liu, Zhao Li, Guoxin Wang, Haihong Tang i Houqiang Li. "Multi-Question Learning for Visual Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 34, nr 07 (3.04.2020): 11328–35. http://dx.doi.org/10.1609/aaai.v34i07.6794.
Pełny tekst źródłaRuwa, Nelson, Qirong Mao, Liangjun Wang i Jianping Gou. "Affective question answering on video". Neurocomputing 363 (październik 2019): 125–39. http://dx.doi.org/10.1016/j.neucom.2019.06.046.
Pełny tekst źródłaWang, Yueqian, Yuxuan Wang, Kai Chen i Dongyan Zhao. "STAIR: Spatial-Temporal Reasoning with Auditable Intermediate Results for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 38, nr 17 (24.03.2024): 19215–23. http://dx.doi.org/10.1609/aaai.v38i17.29890.
Pełny tekst źródłaZong, Linlin, Jiahui Wan, Xianchao Zhang, Xinyue Liu, Wenxin Liang i Bo Xu. "Video-Context Aligned Transformer for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 38, nr 17 (24.03.2024): 19795–803. http://dx.doi.org/10.1609/aaai.v38i17.29954.
Pełny tekst źródłaHuang, Deng, Peihao Chen, Runhao Zeng, Qing Du, Mingkui Tan i Chuang Gan. "Location-Aware Graph Convolutional Networks for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 34, nr 07 (3.04.2020): 11021–28. http://dx.doi.org/10.1609/aaai.v34i07.6737.
Pełny tekst źródłaGao, Lianli, Pengpeng Zeng, Jingkuan Song, Yuan-Fang Li, Wu Liu, Tao Mei i Heng Tao Shen. "Structured Two-Stream Attention Network for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 33 (17.07.2019): 6391–98. http://dx.doi.org/10.1609/aaai.v33i01.33016391.
Pełny tekst źródłaKumar, Krishnamoorthi Magesh, i P. Valarmathie. "Domain and Intelligence Based Multimedia Question Answering System". International Journal of Evaluation and Research in Education (IJERE) 5, nr 3 (1.09.2016): 227. http://dx.doi.org/10.11591/ijere.v5i3.4544.
Pełny tekst źródłaXue, Hongyang, Zhou Zhao i Deng Cai. "Unifying the Video and Question Attentions for Open-Ended Video Question Answering". IEEE Transactions on Image Processing 26, nr 12 (grudzień 2017): 5656–66. http://dx.doi.org/10.1109/tip.2017.2746267.
Pełny tekst źródłaJang, Yunseok, Yale Song, Chris Dongjoo Kim, Youngjae Yu, Youngjin Kim i Gunhee Kim. "Video Question Answering with Spatio-Temporal Reasoning". International Journal of Computer Vision 127, nr 10 (18.06.2019): 1385–412. http://dx.doi.org/10.1007/s11263-019-01189-x.
Pełny tekst źródłaZhuang, Yueting, Dejing Xu, Xin Yan, Wenzhuo Cheng, Zhou Zhao, Shiliang Pu i Jun Xiao. "Multichannel Attention Refinement for Video Question Answering". ACM Transactions on Multimedia Computing, Communications, and Applications 16, nr 1s (28.04.2020): 1–23. http://dx.doi.org/10.1145/3366710.
Pełny tekst źródłaGarcia, Noa, Mayu Otani, Chenhui Chu i Yuta Nakashima. "KnowIT VQA: Answering Knowledge-Based Questions about Videos". Proceedings of the AAAI Conference on Artificial Intelligence 34, nr 07 (3.04.2020): 10826–34. http://dx.doi.org/10.1609/aaai.v34i07.6713.
Pełny tekst źródłaMao, Jianguo, Wenbin Jiang, Hong Liu, Xiangdong Wang i Yajuan Lyu. "Inferential Knowledge-Enhanced Integrated Reasoning for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 11 (26.06.2023): 13380–88. http://dx.doi.org/10.1609/aaai.v37i11.26570.
Pełny tekst źródłaJiang, Jianwen, Ziqiang Chen, Haojie Lin, Xibin Zhao i Yue Gao. "Divide and Conquer: Question-Guided Spatio-Temporal Contextual Attention for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 34, nr 07 (3.04.2020): 11101–8. http://dx.doi.org/10.1609/aaai.v34i07.6766.
Pełny tekst źródłaYang, Saelyne, Sunghyun Park, Yunseok Jang i Moontae Lee. "YTCommentQA: Video Question Answerability in Instructional Videos". Proceedings of the AAAI Conference on Artificial Intelligence 38, nr 17 (24.03.2024): 19359–67. http://dx.doi.org/10.1609/aaai.v38i17.29906.
Pełny tekst źródłaYu, Zhou, Dejing Xu, Jun Yu, Ting Yu, Zhou Zhao, Yueting Zhuang i Dacheng Tao. "ActivityNet-QA: A Dataset for Understanding Complex Web Videos via Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 33 (17.07.2019): 9127–34. http://dx.doi.org/10.1609/aaai.v33i01.33019127.
Pełny tekst źródłaCherian, Anoop, Chiori Hori, Tim K. Marks i Jonathan Le Roux. "(2.5+1)D Spatio-Temporal Scene Graphs for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 1 (28.06.2022): 444–53. http://dx.doi.org/10.1609/aaai.v36i1.19922.
Pełny tekst źródłaChu, Wenqing, Hongyang Xue, Zhou Zhao, Deng Cai i Chengwei Yao. "The forgettable-watcher model for video question answering". Neurocomputing 314 (listopad 2018): 386–93. http://dx.doi.org/10.1016/j.neucom.2018.06.069.
Pełny tekst źródłaZhu, Linchao, Zhongwen Xu, Yi Yang i Alexander G. Hauptmann. "Uncovering the Temporal Context for Video Question Answering". International Journal of Computer Vision 124, nr 3 (13.07.2017): 409–21. http://dx.doi.org/10.1007/s11263-017-1033-7.
Pełny tekst źródłaLee, Yue-Shi, Yu-Chieh Wu i Jie-Chi Yang. "BVideoQA: Online English/Chinese bilingual video question answering". Journal of the American Society for Information Science and Technology 60, nr 3 (marzec 2009): 509–25. http://dx.doi.org/10.1002/asi.21002.
Pełny tekst źródłaXiao, Junbin, Angela Yao, Zhiyuan Liu, Yicong Li, Wei Ji i Tat-Seng Chua. "Video as Conditional Graph Hierarchy for Multi-Granular Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 3 (28.06.2022): 2804–12. http://dx.doi.org/10.1609/aaai.v36i3.20184.
Pełny tekst źródłaLee, Kyungjae, Nan Duan, Lei Ji, Jason Li i Seung-won Hwang. "Segment-Then-Rank: Non-Factoid Question Answering on Instructional Videos". Proceedings of the AAAI Conference on Artificial Intelligence 34, nr 05 (3.04.2020): 8147–54. http://dx.doi.org/10.1609/aaai.v34i05.6327.
Pełny tekst źródłaGao, Feng, Yuanyuan Ge i Yongge Liu. "Remember and forget: video and text fusion for video question answering". Multimedia Tools and Applications 77, nr 22 (27.03.2018): 29269–82. http://dx.doi.org/10.1007/s11042-018-5868-x.
Pełny tekst źródłaLi, Zhangbin, Dan Guo, Jinxing Zhou, Jing Zhang i Meng Wang. "Object-Aware Adaptive-Positivity Learning for Audio-Visual Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 38, nr 4 (24.03.2024): 3306–14. http://dx.doi.org/10.1609/aaai.v38i4.28116.
Pełny tekst źródłaShao, Zhuang, Jiahui Wan i Linlin Zong. "A Video Question Answering Model Based on Knowledge Distillation". Information 14, nr 6 (12.06.2023): 328. http://dx.doi.org/10.3390/info14060328.
Pełny tekst źródłaLi, Xiangpeng, Jingkuan Song, Lianli Gao, Xianglong Liu, Wenbing Huang, Xiangnan He i Chuang Gan. "Beyond RNNs: Positional Self-Attention with Co-Attention for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 33 (17.07.2019): 8658–65. http://dx.doi.org/10.1609/aaai.v33i01.33018658.
Pełny tekst źródłaJiang, Pin, i Yahong Han. "Reasoning with Heterogeneous Graph Alignment for Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 34, nr 07 (3.04.2020): 11109–16. http://dx.doi.org/10.1609/aaai.v34i07.6767.
Pełny tekst źródłaGu, Mao, Zhou Zhao, Weike Jin, Richang Hong i Fei Wu. "Graph-Based Multi-Interaction Network for Video Question Answering". IEEE Transactions on Image Processing 30 (2021): 2758–70. http://dx.doi.org/10.1109/tip.2021.3051756.
Pełny tekst źródłaYu-Chieh Wu i Jie-Chi Yang. "A Robust Passage Retrieval Algorithm for Video Question Answering". IEEE Transactions on Circuits and Systems for Video Technology 18, nr 10 (październik 2008): 1411–21. http://dx.doi.org/10.1109/tcsvt.2008.2002831.
Pełny tekst źródłaWang, Weining, Yan Huang i Liang Wang. "Long video question answering: A Matching-guided Attention Model". Pattern Recognition 102 (czerwiec 2020): 107248. http://dx.doi.org/10.1016/j.patcog.2020.107248.
Pełny tekst źródłaYe, Yunan, Shifeng Zhang, Yimeng Li, Xufeng Qian, Siliang Tang, Shiliang Pu i Jun Xiao. "Video question answering via grounded cross-attention network learning". Information Processing & Management 57, nr 4 (lipiec 2020): 102265. http://dx.doi.org/10.1016/j.ipm.2020.102265.
Pełny tekst źródłaZhang, Wenqiao, Siliang Tang, Yanpeng Cao, Shiliang Pu, Fei Wu i Yueting Zhuang. "Frame Augmented Alternating Attention Network for Video Question Answering". IEEE Transactions on Multimedia 22, nr 4 (kwiecień 2020): 1032–41. http://dx.doi.org/10.1109/tmm.2019.2935678.
Pełny tekst źródłaZha, Zheng-Jun, Jiawei Liu, Tianhao Yang i Yongdong Zhang. "Spatiotemporal-Textual Co-Attention Network for Video Question Answering". ACM Transactions on Multimedia Computing, Communications, and Applications 15, nr 2s (12.08.2019): 1–18. http://dx.doi.org/10.1145/3320061.
Pełny tekst źródłaJiang, Yimin, Tingfei Yan, Mingze Yao, Huibing Wang i Wenzhe Liu. "Cascade transformers with dynamic attention for video question answering". Computer Vision and Image Understanding 242 (maj 2024): 103983. http://dx.doi.org/10.1016/j.cviu.2024.103983.
Pełny tekst źródłaJiao, Guie. "Realization of Video Question Answering System Based on Flash under RIA". Applied Mechanics and Materials 411-414 (wrzesień 2013): 970–73. http://dx.doi.org/10.4028/www.scientific.net/amm.411-414.970.
Pełny tekst źródłaLiu, Mingyang, Ruomei Wang, Fan Zhou i Ge Lin. "Temporally Multi-Modal Semantic Reasoning with Spatial Language Constraints for Video Question Answering". Symmetry 14, nr 6 (31.05.2022): 1133. http://dx.doi.org/10.3390/sym14061133.
Pełny tekst źródłaJin, Yao, Guocheng Niu, Xinyan Xiao, Jian Zhang, Xi Peng i Jun Yu. "Knowledge-Constrained Answer Generation for Open-Ended Video Question Answering". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 7 (26.06.2023): 8141–49. http://dx.doi.org/10.1609/aaai.v37i7.25983.
Pełny tekst źródłaPatel, Hardik Bhikhabhai, i Sailesh Suryanarayan Iyer. "Comparative Study of Multimedia Question Answering System Models". ECS Transactions 107, nr 1 (24.04.2022): 2033–42. http://dx.doi.org/10.1149/10701.2033ecst.
Pełny tekst źródłaGao, Lianli, Yu Lei, Pengpeng Zeng, Jingkuan Song, Meng Wang i Heng Tao Shen. "Hierarchical Representation Network With Auxiliary Tasks for Video Captioning and Video Question Answering". IEEE Transactions on Image Processing 31 (2022): 202–15. http://dx.doi.org/10.1109/tip.2021.3120867.
Pełny tekst źródłaYang, Zekun, Noa Garcia, Chenhui Chu, Mayu Otani, Yuta Nakashima i Haruo Takemura. "A comparative study of language transformers for video question answering". Neurocomputing 445 (lipiec 2021): 121–33. http://dx.doi.org/10.1016/j.neucom.2021.02.092.
Pełny tekst źródłaZhao, Zhou, Zhu Zhang, Shuwen Xiao, Zhenxin Xiao, Xiaohui Yan, Jun Yu, Deng Cai i Fei Wu. "Long-Form Video Question Answering via Dynamic Hierarchical Reinforced Networks". IEEE Transactions on Image Processing 28, nr 12 (grudzień 2019): 5939–52. http://dx.doi.org/10.1109/tip.2019.2922062.
Pełny tekst źródłaYin, Chengxiang, Jian Tang, Zhiyuan Xu i Yanzhi Wang. "Memory Augmented Deep Recurrent Neural Network for Video Question Answering". IEEE Transactions on Neural Networks and Learning Systems 31, nr 9 (wrzesień 2020): 3159–67. http://dx.doi.org/10.1109/tnnls.2019.2938015.
Pełny tekst źródłaWang, Zheng, Fangtao Li, Kaoru Ota, Mianxiong Dong i Bin Wu. "ReGR: Relation-aware graph reasoning framework for video question answering". Information Processing & Management 60, nr 4 (lipiec 2023): 103375. http://dx.doi.org/10.1016/j.ipm.2023.103375.
Pełny tekst źródłaAl Mehmadi, Shima M., Yakoub Bazi, Mohamad M. Al Rahhal i Mansour Zuair. "Learning to enhance areal video captioning with visual question answering". International Journal of Remote Sensing 45, nr 18 (30.08.2024): 6395–407. http://dx.doi.org/10.1080/01431161.2024.2388875.
Pełny tekst źródłaZhuang, Xuqiang, Fang’ai Liu, Jian Hou, Jianhua Hao i Xiaohong Cai. "Modality attention fusion model with hybrid multi-head self-attention for video understanding". PLOS ONE 17, nr 10 (6.10.2022): e0275156. http://dx.doi.org/10.1371/journal.pone.0275156.
Pełny tekst źródłaPeng, Min, Chongyang Wang, Yu Shi i Xiang-Dong Zhou. "Efficient End-to-End Video Question Answering with Pyramidal Multimodal Transformer". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 2 (26.06.2023): 2038–46. http://dx.doi.org/10.1609/aaai.v37i2.25296.
Pełny tekst źródłaKim, Seonhoon, Seohyeong Jeong, Eunbyul Kim, Inho Kang i Nojun Kwak. "Self-supervised Pre-training and Contrastive Representation Learning for Multiple-choice Video QA". Proceedings of the AAAI Conference on Artificial Intelligence 35, nr 14 (18.05.2021): 13171–79. http://dx.doi.org/10.1609/aaai.v35i14.17556.
Pełny tekst źródłaPark, Gyu-Min, A.-Yeong Kim i Seong-Bae Park. "Confident Multiple Choice Learning-based Ensemble Model for Video Question-Answering". Journal of KIISE 49, nr 4 (30.04.2022): 284–90. http://dx.doi.org/10.5626/jok.2022.49.4.284.
Pełny tekst źródłaLiu, Yun, Xiaoming Zhang, Feiran Huang, Bo Zhang i Zhoujun Li. "Cross-Attentional Spatio-Temporal Semantic Graph Networks for Video Question Answering". IEEE Transactions on Image Processing 31 (2022): 1684–96. http://dx.doi.org/10.1109/tip.2022.3142526.
Pełny tekst źródłaZhao, Zhou, Zhu Zhang, Xinghua Jiang i Deng Cai. "Multi-Turn Video Question Answering via Hierarchical Attention Context Reinforced Networks". IEEE Transactions on Image Processing 28, nr 8 (sierpień 2019): 3860–72. http://dx.doi.org/10.1109/tip.2019.2902106.
Pełny tekst źródłaYu, Ting, Jun Yu, Zhou Yu i Dacheng Tao. "Compositional Attention Networks With Two-Stream Fusion for Video Question Answering". IEEE Transactions on Image Processing 29 (2020): 1204–18. http://dx.doi.org/10.1109/tip.2019.2940677.
Pełny tekst źródła