Статті в журналах з теми "Transformers Multimodaux"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся з топ-50 статей у журналах для дослідження на тему "Transformers Multimodaux".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Переглядайте статті в журналах для різних дисциплін та оформлюйте правильно вашу бібліографію.
Jaiswal, Sushma, Harikumar Pallthadka, Rajesh P. Chinchewadi, and Tarun Jaiswal. "Optimized Image Captioning: Hybrid Transformers Vision Transformers and Convolutional Neural Networks: Enhanced with Beam Search." International Journal of Intelligent Systems and Applications 16, no. 2 (April 8, 2024): 53–61. http://dx.doi.org/10.5815/ijisa.2024.02.05.
Повний текст джерелаBayat, Nasrin, Jong-Hwan Kim, Renoa Choudhury, Ibrahim F. Kadhim, Zubaidah Al-Mashhadani, Mark Aldritz Dela Virgen, Reuben Latorre, Ricardo De La Paz, and Joon-Hyuk Park. "Vision Transformer Customized for Environment Detection and Collision Prediction to Assist the Visually Impaired." Journal of Imaging 9, no. 8 (August 15, 2023): 161. http://dx.doi.org/10.3390/jimaging9080161.
Повний текст джерелаShao, Zilei. "A literature review on multimodal deep learning models for detecting mental disorders in conversational data: Pre-transformer and transformer-based approaches." Applied and Computational Engineering 18, no. 1 (October 23, 2023): 215–24. http://dx.doi.org/10.54254/2755-2721/18/20230993.
Повний текст джерелаHendricks, Lisa Anne, John Mellor, Rosalia Schneider, Jean-Baptiste Alayrac, and Aida Nematzadeh. "Decoupling the Role of Data, Attention, and Losses in Multimodal Transformers." Transactions of the Association for Computational Linguistics 9 (2021): 570–85. http://dx.doi.org/10.1162/tacl_a_00385.
Повний текст джерелаChen, Yu, Ming Yin, Yu Li, and Qian Cai. "CSU-Net: A CNN-Transformer Parallel Network for Multimodal Brain Tumour Segmentation." Electronics 11, no. 14 (July 16, 2022): 2226. http://dx.doi.org/10.3390/electronics11142226.
Повний текст джерелаSun, Qixuan, Nianhua Fang, Zhuo Liu, Liang Zhao, Youpeng Wen, and Hongxiang Lin. "HybridCTrm: Bridging CNN and Transformer for Multimodal Brain Image Segmentation." Journal of Healthcare Engineering 2021 (October 1, 2021): 1–10. http://dx.doi.org/10.1155/2021/7467261.
Повний текст джерелаYu Tian, Qiyang Zhao, Zine el abidine Kherroubi, Fouzi Boukhalfa, Kebin Wu, and Faouzi Bader. "Multimodal transformers for wireless communications: A case study in beam prediction." ITU Journal on Future and Evolving Technologies 4, no. 3 (September 5, 2023): 461–71. http://dx.doi.org/10.52953/jwra8095.
Повний текст джерелаXu, Yifan, Huapeng Wei, Minxuan Lin, Yingying Deng, Kekai Sheng, Mengdan Zhang, Fan Tang, Weiming Dong, Feiyue Huang, and Changsheng Xu. "Transformers in computational visual media: A survey." Computational Visual Media 8, no. 1 (October 27, 2021): 33–62. http://dx.doi.org/10.1007/s41095-021-0247-3.
Повний текст джерелаZhong, Enmin, Carlos R. del-Blanco, Daniel Berjón, Fernando Jaureguizar, and Narciso García. "Real-Time Monocular Skeleton-Based Hand Gesture Recognition Using 3D-Jointsformer." Sensors 23, no. 16 (August 10, 2023): 7066. http://dx.doi.org/10.3390/s23167066.
Повний текст джерелаNia, Zahra Movahedi, Ali Ahmadi, Bruce Mellado, Jianhong Wu, James Orbinski, Ali Asgary, and Jude D. Kong. "Twitter-based gender recognition using transformers." Mathematical Biosciences and Engineering 20, no. 9 (2023): 15957–77. http://dx.doi.org/10.3934/mbe.2023711.
Повний текст джерелаLiang, Yi, Turdi Tohti, and Askar Hamdulla. "False Information Detection via Multimodal Feature Fusion and Multi-Classifier Hybrid Prediction." Algorithms 15, no. 4 (March 29, 2022): 119. http://dx.doi.org/10.3390/a15040119.
Повний текст джерелаDesai, Poorav, Tanmoy Chakraborty, and Md Shad Akhtar. "Nice Perfume. How Long Did You Marinate in It? Multimodal Sarcasm Explanation." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 10 (June 28, 2022): 10563–71. http://dx.doi.org/10.1609/aaai.v36i10.21300.
Повний текст джерелаShan, Qishang, Xiangsen Wei, and Ziyun Cai. "Modality-Invariant and -Specific Representations with Crossmodal Transformer for Multimodal Sentiment Analysis." Journal of Physics: Conference Series 2224, no. 1 (April 1, 2022): 012024. http://dx.doi.org/10.1088/1742-6596/2224/1/012024.
Повний текст джерелаGupta, Arpit, Himanshu Goyal, and Ishita Kohli. "Synthesis of Vision and Language: Multifaceted Image Captioning Application." INTERANTIONAL JOURNAL OF SCIENTIFIC RESEARCH IN ENGINEERING AND MANAGEMENT 07, no. 12 (December 23, 2023): 1–10. http://dx.doi.org/10.55041/ijsrem27770.
Повний текст джерелаLiu, Bo, Lejian He, Yafei Liu, Tianyao Yu, Yuejia Xiang, Li Zhu, and Weijian Ruan. "Transformer-Based Multimodal Infusion Dialogue Systems." Electronics 11, no. 20 (October 20, 2022): 3409. http://dx.doi.org/10.3390/electronics11203409.
Повний текст джерелаWang, LeiChen, Simon Giebenhain, Carsten Anklam, and Bastian Goldluecke. "Radar Ghost Target Detection via Multimodal Transformers." IEEE Robotics and Automation Letters 6, no. 4 (October 2021): 7758–65. http://dx.doi.org/10.1109/lra.2021.3100176.
Повний текст джерелаSalin, Emmanuelle, Badreddine Farah, Stéphane Ayache, and Benoit Favre. "Are Vision-Language Transformers Learning Multimodal Representations? A Probing Perspective." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 10 (June 28, 2022): 11248–57. http://dx.doi.org/10.1609/aaai.v36i10.21375.
Повний текст джерелаZhao, Bin, Maoguo Gong, and Xuelong Li. "Hierarchical multimodal transformer to summarize videos." Neurocomputing 468 (January 2022): 360–69. http://dx.doi.org/10.1016/j.neucom.2021.10.039.
Повний текст джерелаDing, Lan. "Online teaching emotion analysis based on GRU and nonlinear transformer algorithm." PeerJ Computer Science 9 (November 21, 2023): e1696. http://dx.doi.org/10.7717/peerj-cs.1696.
Повний текст джерелаWang, Zhaokai, Renda Bao, Qi Wu, and Si Liu. "Confidence-aware Non-repetitive Multimodal Transformers for TextCaps." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 4 (May 18, 2021): 2835–43. http://dx.doi.org/10.1609/aaai.v35i4.16389.
Повний текст джерелаXiang, Yunfan, Xiangyu Tian, Yue Xu, Xiaokun Guan, and Zhengchao Chen. "EGMT-CD: Edge-Guided Multimodal Transformers Change Detection from Satellite and Aerial Images." Remote Sensing 16, no. 1 (December 25, 2023): 86. http://dx.doi.org/10.3390/rs16010086.
Повний текст джерелаLi, Ning, Jie Chen, Nanxin Fu, Wenzhuo Xiao, Tianrun Ye, Chunming Gao, and Ping Zhang. "Leveraging Dual Variational Autoencoders and Generative Adversarial Networks for Enhanced Multimodal Interaction in Zero-Shot Learning." Electronics 13, no. 3 (January 29, 2024): 539. http://dx.doi.org/10.3390/electronics13030539.
Повний текст джерелаAbdine, Hadi, Michail Chatzianastasis, Costas Bouyioukos, and Michalis Vazirgiannis. "Prot2Text: Multimodal Protein’s Function Generation with GNNs and Transformers." Proceedings of the AAAI Conference on Artificial Intelligence 38, no. 10 (March 24, 2024): 10757–65. http://dx.doi.org/10.1609/aaai.v38i10.28948.
Повний текст джерелаLi, Zuhe, Qingbing Guo, Chengyao Feng, Lujuan Deng, Qiuwen Zhang, Jianwei Zhang, Fengqin Wang, and Qian Sun. "Multimodal Sentiment Analysis Based on Interactive Transformer and Soft Mapping." Wireless Communications and Mobile Computing 2022 (February 3, 2022): 1–12. http://dx.doi.org/10.1155/2022/6243347.
Повний текст джерелаZhang, Yinshuo, Lei Chen, and Yuan Yuan. "Multimodal Fine-Grained Transformer Model for Pest Recognition." Electronics 12, no. 12 (June 10, 2023): 2620. http://dx.doi.org/10.3390/electronics12122620.
Повний текст джерелаZhang, Tianze. "Investigation on task effect analysis and optimization strategy of multimodal large model based on Transformers architecture for various languages." Applied and Computational Engineering 47, no. 1 (March 15, 2024): 213–24. http://dx.doi.org/10.54254/2755-2721/47/20241374.
Повний текст джерелаWang, Zhecan, Haoxuan You, Liunian Harold Li, Alireza Zareian, Suji Park, Yiqing Liang, Kai-Wei Chang, and Shih-Fu Chang. "SGEITL: Scene Graph Enhanced Image-Text Learning for Visual Commonsense Reasoning." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 5 (June 28, 2022): 5914–22. http://dx.doi.org/10.1609/aaai.v36i5.20536.
Повний текст джерелаWei, Jiaqi, Bin Jiang, and Yanxia Zhang. "Identification of Blue Horizontal Branch Stars with Multimodal Fusion." Publications of the Astronomical Society of the Pacific 135, no. 1050 (August 1, 2023): 084501. http://dx.doi.org/10.1088/1538-3873/acea43.
Повний текст джерелаSams, Andrew Steven, and Amalia Zahra. "Multimodal music emotion recognition in Indonesian songs based on CNN-LSTM, XLNet transformers." Bulletin of Electrical Engineering and Informatics 12, no. 1 (February 1, 2023): 355–64. http://dx.doi.org/10.11591/eei.v12i1.4231.
Повний текст джерелаNayak, Roshan, B. S. Ullas Kannantha, Kruthi S, and C. Gururaj. "Multimodal Offensive Meme Classification u sing Transformers and BiLSTM." International Journal of Engineering and Advanced Technology 11, no. 3 (February 28, 2022): 96–102. http://dx.doi.org/10.35940/ijeat.c3392.0211322.
Повний текст джерелаNadal, Clement, and Francois Pigache. "Multimodal electromechanical model of piezoelectric transformers by Hamilton's principle." IEEE Transactions on Ultrasonics, Ferroelectrics, and Frequency Control 56, no. 11 (November 2009): 2530–43. http://dx.doi.org/10.1109/tuffc.2009.1340.
Повний текст джерелаChen, Yunfan, Jinxing Ye, and Xiangkui Wan. "TF-YOLO: A Transformer–Fusion-Based YOLO Detector for Multimodal Pedestrian Detection in Autonomous Driving Scenes." World Electric Vehicle Journal 14, no. 12 (December 18, 2023): 352. http://dx.doi.org/10.3390/wevj14120352.
Повний текст джерелаPezzelle, Sandro, Ece Takmaz, and Raquel Fernández. "Word Representation Learning in Multimodal Pre-Trained Transformers: An Intrinsic Evaluation." Transactions of the Association for Computational Linguistics 9 (2021): 1563–79. http://dx.doi.org/10.1162/tacl_a_00443.
Повний текст джерелаZhang, Yingjie. "The current status and prospects of transformer in multimodality." Applied and Computational Engineering 11, no. 1 (September 25, 2023): 224–30. http://dx.doi.org/10.54254/2755-2721/11/20230240.
Повний текст джерелаHasan, Md Kamrul, Sangwu Lee, Wasifur Rahman, Amir Zadeh, Rada Mihalcea, Louis-Philippe Morency, and Ehsan Hoque. "Humor Knowledge Enriched Transformer for Understanding Multimodal Humor." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 14 (May 18, 2021): 12972–80. http://dx.doi.org/10.1609/aaai.v35i14.17534.
Повний текст джерелаZhang, Xiaojuan, Yongxiu Zhou, Peihao Peng, and Guoyan Wang. "A Novel Multimodal Species Distribution Model Fusing Remote Sensing Images and Environmental Features." Sustainability 14, no. 21 (October 28, 2022): 14034. http://dx.doi.org/10.3390/su142114034.
Повний текст джерелаZhang, Guihao, and Jiangzhong Cao. "Feature Fusion Based on Transformer for Cross-modal Retrieval." Journal of Physics: Conference Series 2558, no. 1 (August 1, 2023): 012012. http://dx.doi.org/10.1088/1742-6596/2558/1/012012.
Повний текст джерелаPark, Junhee, and Nammee Moon. "Design and Implementation of Attention Depression Detection Model Based on Multimodal Analysis." Sustainability 14, no. 6 (March 18, 2022): 3569. http://dx.doi.org/10.3390/su14063569.
Повний текст джерелаQi, Qingfu, Liyuan Lin, Rui Zhang, and Chengrong Xue. "MEDT: Using Multimodal Encoding-Decoding Network as in Transformer for Multimodal Sentiment Analysis." IEEE Access 10 (2022): 28750–59. http://dx.doi.org/10.1109/access.2022.3157712.
Повний текст джерелаLi, Lei, Xiang Chen, Shuofei Qiao, Feiyu Xiong, Huajun Chen, and Ningyu Zhang. "On Analyzing the Role of Image for Visual-Enhanced Relation Extraction (Student Abstract)." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 13 (June 26, 2023): 16254–55. http://dx.doi.org/10.1609/aaai.v37i13.26987.
Повний текст джерелаZhang, Junyan. "Research on transformer and attention in applied algorithms." Applied and Computational Engineering 13, no. 1 (October 23, 2023): 221–28. http://dx.doi.org/10.54254/2755-2721/13/20230737.
Повний текст джерелаGao, Jialin, Jianyu Chen, Jiaqi Wei, Bin Jiang, and A.-Li Luo. "Deep Multimodal Networks for M-type Star Classification with Paired Spectrum and Photometric Image." Publications of the Astronomical Society of the Pacific 135, no. 1046 (April 1, 2023): 044503. http://dx.doi.org/10.1088/1538-3873/acc7ca.
Повний текст джерелаZong, Daoming, and Shiliang Sun. "McOmet: Multimodal Fusion Transformer for Physical Audiovisual Commonsense Reasoning." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 5 (June 26, 2023): 6621–29. http://dx.doi.org/10.1609/aaai.v37i5.25813.
Повний текст джерелаJayaLakshmi, Gundabathina, Abburi Madhuri, Deepak Vasudevan, Balamuralikrishna Thati, Uddagiri Sirisha, and Surapaneni Phani Praveen. "Effective Disaster Management Through Transformer-Based Multimodal Tweet Classification." Revue d'Intelligence Artificielle 37, no. 5 (October 31, 2023): 1263–72. http://dx.doi.org/10.18280/ria.370519.
Повний текст джерелаLiu, Biyuan, Huaixin Chen, Kun Li, and Michael Ying Yang. "Transformer-based multimodal change detection with multitask consistency constraints." Information Fusion 108 (August 2024): 102358. http://dx.doi.org/10.1016/j.inffus.2024.102358.
Повний текст джерелаAbiyev, Rahib H., Mohamad Ziad Altabel, Manal Darwish, and Abdulkader Helwan. "A Multimodal Transformer Model for Recognition of Images from Complex Laparoscopic Surgical Videos." Diagnostics 14, no. 7 (March 23, 2024): 681. http://dx.doi.org/10.3390/diagnostics14070681.
Повний текст джерелаChaudhari, Aayushi, Chintan Bhatt, Achyut Krishna, and Carlos M. Travieso-González. "Facial Emotion Recognition with Inter-Modality-Attention-Transformer-Based Self-Supervised Learning." Electronics 12, no. 2 (January 5, 2023): 288. http://dx.doi.org/10.3390/electronics12020288.
Повний текст джерелаXu, Zhen, David R. So, and Andrew M. Dai. "MUFASA: Multimodal Fusion Architecture Search for Electronic Health Records." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 12 (May 18, 2021): 10532–40. http://dx.doi.org/10.1609/aaai.v35i12.17260.
Повний текст джерелаIlmi, Yuslimu, Pratiwi Retnaningdyah, and Ahmad Munir. "Exploring Digital Multimodal Text in EFL Classroom: Transformed Practice in Multiliteracies Pedagogy." Linguistic, English Education and Art (LEEA) Journal 4, no. 1 (December 28, 2020): 99–108. http://dx.doi.org/10.31539/leea.v4i1.1416.
Повний текст джерелаAmmour, Nassim, Yakoub Bazi, and Naif Alajlan. "Multimodal Approach for Enhancing Biometric Authentication." Journal of Imaging 9, no. 9 (August 22, 2023): 168. http://dx.doi.org/10.3390/jimaging9090168.
Повний текст джерела