Artykuły w czasopismach na temat „Transformers Multimodaux”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Sprawdź 50 najlepszych artykułów w czasopismach naukowych na temat „Transformers Multimodaux”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Przeglądaj artykuły w czasopismach z różnych dziedzin i twórz odpowiednie bibliografie.
Jaiswal, Sushma, Harikumar Pallthadka, Rajesh P. Chinchewadi i Tarun Jaiswal. "Optimized Image Captioning: Hybrid Transformers Vision Transformers and Convolutional Neural Networks: Enhanced with Beam Search". International Journal of Intelligent Systems and Applications 16, nr 2 (8.04.2024): 53–61. http://dx.doi.org/10.5815/ijisa.2024.02.05.
Pełny tekst źródłaBayat, Nasrin, Jong-Hwan Kim, Renoa Choudhury, Ibrahim F. Kadhim, Zubaidah Al-Mashhadani, Mark Aldritz Dela Virgen, Reuben Latorre, Ricardo De La Paz i Joon-Hyuk Park. "Vision Transformer Customized for Environment Detection and Collision Prediction to Assist the Visually Impaired". Journal of Imaging 9, nr 8 (15.08.2023): 161. http://dx.doi.org/10.3390/jimaging9080161.
Pełny tekst źródłaShao, Zilei. "A literature review on multimodal deep learning models for detecting mental disorders in conversational data: Pre-transformer and transformer-based approaches". Applied and Computational Engineering 18, nr 1 (23.10.2023): 215–24. http://dx.doi.org/10.54254/2755-2721/18/20230993.
Pełny tekst źródłaHendricks, Lisa Anne, John Mellor, Rosalia Schneider, Jean-Baptiste Alayrac i Aida Nematzadeh. "Decoupling the Role of Data, Attention, and Losses in Multimodal Transformers". Transactions of the Association for Computational Linguistics 9 (2021): 570–85. http://dx.doi.org/10.1162/tacl_a_00385.
Pełny tekst źródłaChen, Yu, Ming Yin, Yu Li i Qian Cai. "CSU-Net: A CNN-Transformer Parallel Network for Multimodal Brain Tumour Segmentation". Electronics 11, nr 14 (16.07.2022): 2226. http://dx.doi.org/10.3390/electronics11142226.
Pełny tekst źródłaSun, Qixuan, Nianhua Fang, Zhuo Liu, Liang Zhao, Youpeng Wen i Hongxiang Lin. "HybridCTrm: Bridging CNN and Transformer for Multimodal Brain Image Segmentation". Journal of Healthcare Engineering 2021 (1.10.2021): 1–10. http://dx.doi.org/10.1155/2021/7467261.
Pełny tekst źródłaYu Tian, Qiyang Zhao, Zine el abidine Kherroubi, Fouzi Boukhalfa, Kebin Wu i Faouzi Bader. "Multimodal transformers for wireless communications: A case study in beam prediction". ITU Journal on Future and Evolving Technologies 4, nr 3 (5.09.2023): 461–71. http://dx.doi.org/10.52953/jwra8095.
Pełny tekst źródłaXu, Yifan, Huapeng Wei, Minxuan Lin, Yingying Deng, Kekai Sheng, Mengdan Zhang, Fan Tang, Weiming Dong, Feiyue Huang i Changsheng Xu. "Transformers in computational visual media: A survey". Computational Visual Media 8, nr 1 (27.10.2021): 33–62. http://dx.doi.org/10.1007/s41095-021-0247-3.
Pełny tekst źródłaZhong, Enmin, Carlos R. del-Blanco, Daniel Berjón, Fernando Jaureguizar i Narciso García. "Real-Time Monocular Skeleton-Based Hand Gesture Recognition Using 3D-Jointsformer". Sensors 23, nr 16 (10.08.2023): 7066. http://dx.doi.org/10.3390/s23167066.
Pełny tekst źródłaNia, Zahra Movahedi, Ali Ahmadi, Bruce Mellado, Jianhong Wu, James Orbinski, Ali Asgary i Jude D. Kong. "Twitter-based gender recognition using transformers". Mathematical Biosciences and Engineering 20, nr 9 (2023): 15957–77. http://dx.doi.org/10.3934/mbe.2023711.
Pełny tekst źródłaLiang, Yi, Turdi Tohti i Askar Hamdulla. "False Information Detection via Multimodal Feature Fusion and Multi-Classifier Hybrid Prediction". Algorithms 15, nr 4 (29.03.2022): 119. http://dx.doi.org/10.3390/a15040119.
Pełny tekst źródłaDesai, Poorav, Tanmoy Chakraborty i Md Shad Akhtar. "Nice Perfume. How Long Did You Marinate in It? Multimodal Sarcasm Explanation". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 10 (28.06.2022): 10563–71. http://dx.doi.org/10.1609/aaai.v36i10.21300.
Pełny tekst źródłaShan, Qishang, Xiangsen Wei i Ziyun Cai. "Modality-Invariant and -Specific Representations with Crossmodal Transformer for Multimodal Sentiment Analysis". Journal of Physics: Conference Series 2224, nr 1 (1.04.2022): 012024. http://dx.doi.org/10.1088/1742-6596/2224/1/012024.
Pełny tekst źródłaGupta, Arpit, Himanshu Goyal i Ishita Kohli. "Synthesis of Vision and Language: Multifaceted Image Captioning Application". INTERANTIONAL JOURNAL OF SCIENTIFIC RESEARCH IN ENGINEERING AND MANAGEMENT 07, nr 12 (23.12.2023): 1–10. http://dx.doi.org/10.55041/ijsrem27770.
Pełny tekst źródłaLiu, Bo, Lejian He, Yafei Liu, Tianyao Yu, Yuejia Xiang, Li Zhu i Weijian Ruan. "Transformer-Based Multimodal Infusion Dialogue Systems". Electronics 11, nr 20 (20.10.2022): 3409. http://dx.doi.org/10.3390/electronics11203409.
Pełny tekst źródłaWang, LeiChen, Simon Giebenhain, Carsten Anklam i Bastian Goldluecke. "Radar Ghost Target Detection via Multimodal Transformers". IEEE Robotics and Automation Letters 6, nr 4 (październik 2021): 7758–65. http://dx.doi.org/10.1109/lra.2021.3100176.
Pełny tekst źródłaSalin, Emmanuelle, Badreddine Farah, Stéphane Ayache i Benoit Favre. "Are Vision-Language Transformers Learning Multimodal Representations? A Probing Perspective". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 10 (28.06.2022): 11248–57. http://dx.doi.org/10.1609/aaai.v36i10.21375.
Pełny tekst źródłaZhao, Bin, Maoguo Gong i Xuelong Li. "Hierarchical multimodal transformer to summarize videos". Neurocomputing 468 (styczeń 2022): 360–69. http://dx.doi.org/10.1016/j.neucom.2021.10.039.
Pełny tekst źródłaDing, Lan. "Online teaching emotion analysis based on GRU and nonlinear transformer algorithm". PeerJ Computer Science 9 (21.11.2023): e1696. http://dx.doi.org/10.7717/peerj-cs.1696.
Pełny tekst źródłaWang, Zhaokai, Renda Bao, Qi Wu i Si Liu. "Confidence-aware Non-repetitive Multimodal Transformers for TextCaps". Proceedings of the AAAI Conference on Artificial Intelligence 35, nr 4 (18.05.2021): 2835–43. http://dx.doi.org/10.1609/aaai.v35i4.16389.
Pełny tekst źródłaXiang, Yunfan, Xiangyu Tian, Yue Xu, Xiaokun Guan i Zhengchao Chen. "EGMT-CD: Edge-Guided Multimodal Transformers Change Detection from Satellite and Aerial Images". Remote Sensing 16, nr 1 (25.12.2023): 86. http://dx.doi.org/10.3390/rs16010086.
Pełny tekst źródłaLi, Ning, Jie Chen, Nanxin Fu, Wenzhuo Xiao, Tianrun Ye, Chunming Gao i Ping Zhang. "Leveraging Dual Variational Autoencoders and Generative Adversarial Networks for Enhanced Multimodal Interaction in Zero-Shot Learning". Electronics 13, nr 3 (29.01.2024): 539. http://dx.doi.org/10.3390/electronics13030539.
Pełny tekst źródłaAbdine, Hadi, Michail Chatzianastasis, Costas Bouyioukos i Michalis Vazirgiannis. "Prot2Text: Multimodal Protein’s Function Generation with GNNs and Transformers". Proceedings of the AAAI Conference on Artificial Intelligence 38, nr 10 (24.03.2024): 10757–65. http://dx.doi.org/10.1609/aaai.v38i10.28948.
Pełny tekst źródłaLi, Zuhe, Qingbing Guo, Chengyao Feng, Lujuan Deng, Qiuwen Zhang, Jianwei Zhang, Fengqin Wang i Qian Sun. "Multimodal Sentiment Analysis Based on Interactive Transformer and Soft Mapping". Wireless Communications and Mobile Computing 2022 (3.02.2022): 1–12. http://dx.doi.org/10.1155/2022/6243347.
Pełny tekst źródłaZhang, Yinshuo, Lei Chen i Yuan Yuan. "Multimodal Fine-Grained Transformer Model for Pest Recognition". Electronics 12, nr 12 (10.06.2023): 2620. http://dx.doi.org/10.3390/electronics12122620.
Pełny tekst źródłaZhang, Tianze. "Investigation on task effect analysis and optimization strategy of multimodal large model based on Transformers architecture for various languages". Applied and Computational Engineering 47, nr 1 (15.03.2024): 213–24. http://dx.doi.org/10.54254/2755-2721/47/20241374.
Pełny tekst źródłaWang, Zhecan, Haoxuan You, Liunian Harold Li, Alireza Zareian, Suji Park, Yiqing Liang, Kai-Wei Chang i Shih-Fu Chang. "SGEITL: Scene Graph Enhanced Image-Text Learning for Visual Commonsense Reasoning". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 5 (28.06.2022): 5914–22. http://dx.doi.org/10.1609/aaai.v36i5.20536.
Pełny tekst źródłaWei, Jiaqi, Bin Jiang i Yanxia Zhang. "Identification of Blue Horizontal Branch Stars with Multimodal Fusion". Publications of the Astronomical Society of the Pacific 135, nr 1050 (1.08.2023): 084501. http://dx.doi.org/10.1088/1538-3873/acea43.
Pełny tekst źródłaSams, Andrew Steven, i Amalia Zahra. "Multimodal music emotion recognition in Indonesian songs based on CNN-LSTM, XLNet transformers". Bulletin of Electrical Engineering and Informatics 12, nr 1 (1.02.2023): 355–64. http://dx.doi.org/10.11591/eei.v12i1.4231.
Pełny tekst źródłaNayak, Roshan, B. S. Ullas Kannantha, Kruthi S i C. Gururaj. "Multimodal Offensive Meme Classification u sing Transformers and BiLSTM". International Journal of Engineering and Advanced Technology 11, nr 3 (28.02.2022): 96–102. http://dx.doi.org/10.35940/ijeat.c3392.0211322.
Pełny tekst źródłaNadal, Clement, i Francois Pigache. "Multimodal electromechanical model of piezoelectric transformers by Hamilton's principle". IEEE Transactions on Ultrasonics, Ferroelectrics, and Frequency Control 56, nr 11 (listopad 2009): 2530–43. http://dx.doi.org/10.1109/tuffc.2009.1340.
Pełny tekst źródłaChen, Yunfan, Jinxing Ye i Xiangkui Wan. "TF-YOLO: A Transformer–Fusion-Based YOLO Detector for Multimodal Pedestrian Detection in Autonomous Driving Scenes". World Electric Vehicle Journal 14, nr 12 (18.12.2023): 352. http://dx.doi.org/10.3390/wevj14120352.
Pełny tekst źródłaPezzelle, Sandro, Ece Takmaz i Raquel Fernández. "Word Representation Learning in Multimodal Pre-Trained Transformers: An Intrinsic Evaluation". Transactions of the Association for Computational Linguistics 9 (2021): 1563–79. http://dx.doi.org/10.1162/tacl_a_00443.
Pełny tekst źródłaZhang, Yingjie. "The current status and prospects of transformer in multimodality". Applied and Computational Engineering 11, nr 1 (25.09.2023): 224–30. http://dx.doi.org/10.54254/2755-2721/11/20230240.
Pełny tekst źródłaHasan, Md Kamrul, Sangwu Lee, Wasifur Rahman, Amir Zadeh, Rada Mihalcea, Louis-Philippe Morency i Ehsan Hoque. "Humor Knowledge Enriched Transformer for Understanding Multimodal Humor". Proceedings of the AAAI Conference on Artificial Intelligence 35, nr 14 (18.05.2021): 12972–80. http://dx.doi.org/10.1609/aaai.v35i14.17534.
Pełny tekst źródłaZhang, Xiaojuan, Yongxiu Zhou, Peihao Peng i Guoyan Wang. "A Novel Multimodal Species Distribution Model Fusing Remote Sensing Images and Environmental Features". Sustainability 14, nr 21 (28.10.2022): 14034. http://dx.doi.org/10.3390/su142114034.
Pełny tekst źródłaZhang, Guihao, i Jiangzhong Cao. "Feature Fusion Based on Transformer for Cross-modal Retrieval". Journal of Physics: Conference Series 2558, nr 1 (1.08.2023): 012012. http://dx.doi.org/10.1088/1742-6596/2558/1/012012.
Pełny tekst źródłaPark, Junhee, i Nammee Moon. "Design and Implementation of Attention Depression Detection Model Based on Multimodal Analysis". Sustainability 14, nr 6 (18.03.2022): 3569. http://dx.doi.org/10.3390/su14063569.
Pełny tekst źródłaQi, Qingfu, Liyuan Lin, Rui Zhang i Chengrong Xue. "MEDT: Using Multimodal Encoding-Decoding Network as in Transformer for Multimodal Sentiment Analysis". IEEE Access 10 (2022): 28750–59. http://dx.doi.org/10.1109/access.2022.3157712.
Pełny tekst źródłaLi, Lei, Xiang Chen, Shuofei Qiao, Feiyu Xiong, Huajun Chen i Ningyu Zhang. "On Analyzing the Role of Image for Visual-Enhanced Relation Extraction (Student Abstract)". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 13 (26.06.2023): 16254–55. http://dx.doi.org/10.1609/aaai.v37i13.26987.
Pełny tekst źródłaZhang, Junyan. "Research on transformer and attention in applied algorithms". Applied and Computational Engineering 13, nr 1 (23.10.2023): 221–28. http://dx.doi.org/10.54254/2755-2721/13/20230737.
Pełny tekst źródłaGao, Jialin, Jianyu Chen, Jiaqi Wei, Bin Jiang i A.-Li Luo. "Deep Multimodal Networks for M-type Star Classification with Paired Spectrum and Photometric Image". Publications of the Astronomical Society of the Pacific 135, nr 1046 (1.04.2023): 044503. http://dx.doi.org/10.1088/1538-3873/acc7ca.
Pełny tekst źródłaZong, Daoming, i Shiliang Sun. "McOmet: Multimodal Fusion Transformer for Physical Audiovisual Commonsense Reasoning". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 5 (26.06.2023): 6621–29. http://dx.doi.org/10.1609/aaai.v37i5.25813.
Pełny tekst źródłaJayaLakshmi, Gundabathina, Abburi Madhuri, Deepak Vasudevan, Balamuralikrishna Thati, Uddagiri Sirisha i Surapaneni Phani Praveen. "Effective Disaster Management Through Transformer-Based Multimodal Tweet Classification". Revue d'Intelligence Artificielle 37, nr 5 (31.10.2023): 1263–72. http://dx.doi.org/10.18280/ria.370519.
Pełny tekst źródłaLiu, Biyuan, Huaixin Chen, Kun Li i Michael Ying Yang. "Transformer-based multimodal change detection with multitask consistency constraints". Information Fusion 108 (sierpień 2024): 102358. http://dx.doi.org/10.1016/j.inffus.2024.102358.
Pełny tekst źródłaAbiyev, Rahib H., Mohamad Ziad Altabel, Manal Darwish i Abdulkader Helwan. "A Multimodal Transformer Model for Recognition of Images from Complex Laparoscopic Surgical Videos". Diagnostics 14, nr 7 (23.03.2024): 681. http://dx.doi.org/10.3390/diagnostics14070681.
Pełny tekst źródłaChaudhari, Aayushi, Chintan Bhatt, Achyut Krishna i Carlos M. Travieso-González. "Facial Emotion Recognition with Inter-Modality-Attention-Transformer-Based Self-Supervised Learning". Electronics 12, nr 2 (5.01.2023): 288. http://dx.doi.org/10.3390/electronics12020288.
Pełny tekst źródłaXu, Zhen, David R. So i Andrew M. Dai. "MUFASA: Multimodal Fusion Architecture Search for Electronic Health Records". Proceedings of the AAAI Conference on Artificial Intelligence 35, nr 12 (18.05.2021): 10532–40. http://dx.doi.org/10.1609/aaai.v35i12.17260.
Pełny tekst źródłaIlmi, Yuslimu, Pratiwi Retnaningdyah i Ahmad Munir. "Exploring Digital Multimodal Text in EFL Classroom: Transformed Practice in Multiliteracies Pedagogy". Linguistic, English Education and Art (LEEA) Journal 4, nr 1 (28.12.2020): 99–108. http://dx.doi.org/10.31539/leea.v4i1.1416.
Pełny tekst źródłaAmmour, Nassim, Yakoub Bazi i Naif Alajlan. "Multimodal Approach for Enhancing Biometric Authentication". Journal of Imaging 9, nr 9 (22.08.2023): 168. http://dx.doi.org/10.3390/jimaging9090168.
Pełny tekst źródła