Gotowa bibliografia na temat „Multimodal processing”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Spis treści
Zobacz listy aktualnych artykułów, książek, rozpraw, streszczeń i innych źródeł naukowych na temat „Multimodal processing”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Artykuły w czasopismach na temat "Multimodal processing"
Ng, Vincent, i Shengjie Li. "Multimodal Propaganda Processing". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 13 (26.06.2023): 15368–75. http://dx.doi.org/10.1609/aaai.v37i13.26792.
Pełny tekst źródłaSinke, Christopher, Janina Neufeld, Daniel Wiswede, Hinderk M. Emrich, Stefan Bleich i Gregor R. Szycik. "Multisensory processing in synesthesia — differences in the EEG signal during uni- and multimodal processing". Seeing and Perceiving 25 (2012): 53. http://dx.doi.org/10.1163/187847612x646749.
Pełny tekst źródłaD'Ulizia, Arianna, Fernando Ferri i Patrizia Grifoni. "Generating Multimodal Grammars for Multimodal Dialogue Processing". IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans 40, nr 6 (listopad 2010): 1130–45. http://dx.doi.org/10.1109/tsmca.2010.2041227.
Pełny tekst źródłaBarricelli, Barbara Rita, Piero Mussio, Marco Padula i Paolo Luigi Scala. "TMS for multimodal information processing". Multimedia Tools and Applications 54, nr 1 (27.04.2010): 97–120. http://dx.doi.org/10.1007/s11042-010-0527-x.
Pełny tekst źródłaParsons, Aaron D., Stephen W. T. Price, Nicola Wadeson, Mark Basham, Andrew M. Beale, Alun W. Ashton, J. Frederick W. Mosselmans i Paul D. Quinn. "Automatic processing of multimodal tomography datasets". Journal of Synchrotron Radiation 24, nr 1 (1.01.2017): 248–56. http://dx.doi.org/10.1107/s1600577516017756.
Pełny tekst źródłaHoller, Judith, i Stephen C. Levinson. "Multimodal Language Processing in Human Communication". Trends in Cognitive Sciences 23, nr 8 (sierpień 2019): 639–52. http://dx.doi.org/10.1016/j.tics.2019.05.006.
Pełny tekst źródłaFarzin, Faraz, Eric P. Charles i Susan M. Rivera. "Development of Multimodal Processing in Infancy". Infancy 14, nr 5 (1.09.2009): 563–78. http://dx.doi.org/10.1080/15250000903144207.
Pełny tekst źródłaZhang, Ge, Tianxiang Luo, Witold Pedrycz, Mohammed A. El-Meligy, Mohamed Abdel Fattah Sharaf i Zhiwu Li. "Outlier Processing in Multimodal Emotion Recognition". IEEE Access 8 (2020): 55688–701. http://dx.doi.org/10.1109/access.2020.2981760.
Pełny tekst źródłaMetaxakis, Athanasios, Dionysia Petratou i Nektarios Tavernarakis. "Multimodal sensory processing in Caenorhabditis elegans". Open Biology 8, nr 6 (czerwiec 2018): 180049. http://dx.doi.org/10.1098/rsob.180049.
Pełny tekst źródłaNock, Harriet J., Giridharan Iyengar i Chalapathy Neti. "Multimodal processing by finding common cause". Communications of the ACM 47, nr 1 (1.01.2004): 51. http://dx.doi.org/10.1145/962081.962105.
Pełny tekst źródłaRozprawy doktorskie na temat "Multimodal processing"
Cadène, Rémi. "Deep Multimodal Learning for Vision and Language Processing". Electronic Thesis or Diss., Sorbonne université, 2020. http://www.theses.fr/2020SORUS277.
Pełny tekst źródłaDigital technologies have become instrumental in transforming our society. Recent statistical methods have been successfully deployed to automate the processing of the growing amount of images, videos, and texts we produce daily. In particular, deep neural networks have been adopted by the computer vision and natural language processing communities for their ability to perform accurate image recognition and text understanding once trained on big sets of data. Advances in both communities built the groundwork for new research problems at the intersection of vision and language. Integrating language into visual recognition could have an important impact on human life through the creation of real-world applications such as next-generation search engines or AI assistants.In the first part of this thesis, we focus on systems for cross-modal text-image retrieval. We propose a learning strategy to efficiently align both modalities while structuring the retrieval space with semantic information. In the second part, we focus on systems able to answer questions about an image. We propose a multimodal architecture that iteratively fuses the visual and textual modalities using a factorized bilinear model while modeling pairwise relationships between each region of the image. In the last part, we address the issues related to biases in the modeling. We propose a learning strategy to reduce the language biases which are commonly present in visual question answering systems
Hu, Yongtao, i 胡永涛. "Multimodal speaker localization and identification for video processing". Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2014. http://hdl.handle.net/10722/212633.
Pełny tekst źródłaChen, Xun. "Multimodal biomedical signal processing for corticomuscular coupling analysis". Thesis, University of British Columbia, 2014. http://hdl.handle.net/2429/45811.
Pełny tekst źródłaSadr, Lahijany Nadi. "Multimodal Signal Processing for Diagnosis of Cardiorespiratory Disorders". Thesis, The University of Sydney, 2017. http://hdl.handle.net/2123/17636.
Pełny tekst źródłaElshaw, Mark. "Multimodal neural grounding of language processing for robot actions". Thesis, University of Sunderland, 2005. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.420517.
Pełny tekst źródłaFriedel, Paul. "Sensory information processing : detection, feature extraction, & multimodal integration". kostenfrei, 2008. http://mediatum2.ub.tum.de/doc/651333/651333.pdf.
Pełny tekst źródłaSadeghi, Ghandehari Soroush. "Multimodal signal processing in the peripheral and central vestibular pathways". Thesis, McGill University, 2009. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=95559.
Pełny tekst źródłaLes organes sensoriels vestibulaires de l’oreille interne détectent les mouvements de la tète dans r espace. Ces informations sont envoyées aux neurones vestibulaires centraux localises au niveau du tronc cérébral. A ce niveau convergent également d'autres signaux en provenance du cortex, du cervelet. de la moelle ainsi que de divers noyaux du tronc cérébral. Les études présentées ici ont pour but de comprendre le mode de codage et la nature des signaux générés par les neurones vestibulaires périphériques, ainsi que les capacités de traitement des neurones vestibulaires centraux. véritable centres d'intégration sensori-motrice. Ces travaux ont été conduits en condition physiologique et physiopathologique sur le modèle de la compensation vestibulaire. A r aide de mesures issues de la théorie de l'information, nous nous sommes tout d'abord intéresse aux codages effectues par Ies afférences vestibulaires régulières et irrégulières. Ces deux types neuronaux différent notamment par la variabilité de leur fréquence de décharge spontanée (bruit) et leurs sensibilités (signal). Nous avons montre que Ies fibres afférentes régulières utilisent un codage temporel alors que les fibres irrégulières fonctionnent essentiellement sur un codage en modulation de la fréquence, et ce d' autant mieux que les fréquences sont élevées, constituant ainsi de véritables détecteurs d'évènements. Nous avons ensuite étudie les réponses des afférences suite a une stimulation vestibulaire directe ou a une activation du « système efférent ». En conditions physiologiques, nous avons tout d'abord pu démontrer que le système efférent est bien fonctionneI chez le singe éveille. fr
Fateri, Sina. "Advanced signal processing techniques for multimodal ultrasonic guided wave response". Thesis, Brunel University, 2015. http://bura.brunel.ac.uk/handle/2438/11657.
Pełny tekst źródłaCaglayan, Ozan. "Multimodal Machine Translation". Thesis, Le Mans, 2019. http://www.theses.fr/2019LEMA1016/document.
Pełny tekst źródłaMachine translation aims at automatically translating documents from one language to another without human intervention. With the advent of deep neural networks (DNN), neural approaches to machine translation started to dominate the field, reaching state-ofthe-art performance in many languages. Neural machine translation (NMT) also revived the interest in interlingual machine translation due to how it naturally fits the task into an encoder-decoder framework which produces a translation by decoding a latent source representation. Combined with the architectural flexibility of DNNs, this framework paved the way for further research in multimodality with the objective of augmenting the latent representations with other modalities such as vision or speech, for example. This thesis focuses on a multimodal machine translation (MMT) framework that integrates a secondary visual modality to achieve better and visually grounded language understanding. I specifically worked with a dataset containing images and their translated descriptions, where visual context can be useful forword sense disambiguation, missing word imputation, or gender marking when translating from a language with gender-neutral nouns to one with grammatical gender system as is the case with English to French. I propose two main approaches to integrate the visual modality: (i) a multimodal attention mechanism that learns to take into account both sentence and convolutional visual representations, (ii) a method that uses global visual feature vectors to prime the sentence encoders and the decoders. Through automatic and human evaluation conducted on multiple language pairs, the proposed approaches were demonstrated to be beneficial. Finally, I further show that by systematically removing certain linguistic information from the input sentences, the true strength of both methods emerges as they successfully impute missing nouns, colors and can even translate when parts of the source sentences are completely removed
Fridman, Linnea, i Victoria Nordberg. "Two Multimodal Image Registration Approaches for Positioning Purposes". Thesis, Linköpings universitet, Datorseende, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-157424.
Pełny tekst źródłaKsiążki na temat "Multimodal processing"
Renals, Steve, Herve Bourlard, Jean Carletta i Andrei Popescu-Belis, red. Multimodal Signal Processing. Cambridge: Cambridge University Press, 2009. http://dx.doi.org/10.1017/cbo9781139136310.
Pełny tekst źródłaMaragos, Petros, Alexandros Potamianos i Patrick Gros, red. Multimodal Processing and Interaction. Boston, MA: Springer US, 2008. http://dx.doi.org/10.1007/978-0-387-76316-3.
Pełny tekst źródłaRenals, Steve. Multimodal signal processing: Human interactions in meetings. Cambridge: Cambridge University Press, 2012.
Znajdź pełny tekst źródłaFerran, Marques, Knovel (Firm) i ScienceDirect (Online service), red. Multimodal signal processing: Theory and applications for human-computer interaction. Amsterdam: Academic, 2010.
Znajdź pełny tekst źródłaGavrilova, Marina L. Multimodal biometrics and intelligent image processing for security systems. Hershey, PA: Information Science Reference, 2013.
Znajdź pełny tekst źródła1959-, Grifoni Patrizia, red. Multimodal human computer interaction and pervasive services. Hershey PA: Information Science Reference, 2009.
Znajdź pełny tekst źródłaSappa, Angel D. Multimodal Interaction in Image and Video Applications. Berlin, Heidelberg: Springer Berlin Heidelberg, 2013.
Znajdź pełny tekst źródła1953-, Zimmer H. D., red. Human memory: A multimodal approach. Seattle: Hogrefe & Huber Publishers, 1994.
Znajdź pełny tekst źródłaAdams, Teresa M. Guidelines for the implementation of multimodal transportation location referencing systems. Washington, D.C: National Academy Press, 2001.
Znajdź pełny tekst źródłaKühnel, Christine. Quantifying Quality Aspects of Multimodal Interactive Systems. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012.
Znajdź pełny tekst źródłaCzęści książek na temat "Multimodal processing"
Huang, Lihe. "Collecting and processing multimodal data". W Toward Multimodal Pragmatics, 99–108. London: Routledge, 2021. http://dx.doi.org/10.4324/9781003251774-5.
Pełny tekst źródłaGullberg, Marianne. "Studying Multimodal Language Processing". W The Routledge Handbook of Second Language Acquisition and Psycholinguistics, 137–49. New York: Routledge, 2022. http://dx.doi.org/10.4324/9781003018872-14.
Pełny tekst źródłaWaibel, Alex, Minh Tue Vo, Paul Duchnowski i Stefan Manke. "Multimodal Interfaces". W Integration of Natural Language and Vision Processing, 299–319. Dordrecht: Springer Netherlands, 1996. http://dx.doi.org/10.1007/978-94-009-1716-3_9.
Pełny tekst źródłaRivet, Bertrand, i Jonathon Chambers. "Multimodal Speech Separation". W Advances in Nonlinear Speech Processing, 1–11. Berlin, Heidelberg: Springer Berlin Heidelberg, 2010. http://dx.doi.org/10.1007/978-3-642-11509-7_1.
Pełny tekst źródłaNéel, Françoise D., i Wolfgang M. Minker. "Multimodal Speech Systems". W Computational Models of Speech Pattern Processing, 404–30. Berlin, Heidelberg: Springer Berlin Heidelberg, 1999. http://dx.doi.org/10.1007/978-3-642-60087-6_34.
Pełny tekst źródłaMaragos, Petros, Patrick Gros, Athanassios Katsamanis i George Papandreou. "Cross-Modal Integration for Performance Improving in Multimedia: A Review". W Multimodal Processing and Interaction, 1–46. Boston, MA: Springer US, 2008. http://dx.doi.org/10.1007/978-0-387-76316-3_1.
Pełny tekst źródłaFerecatu, Marin, Nozha Boujemaa i Michel Crucianu. "Interactive Image Retrieval Using a Hybrid Visual and Conceptual Content Representation". W Multimodal Processing and Interaction, 1–20. Boston, MA: Springer US, 2008. http://dx.doi.org/10.1007/978-0-387-76316-3_10.
Pełny tekst źródłaNeumayer, Robert, i Andreas Rauber. "Multimodal Analysis of Text and Audio Features for Music Information Retrieval". W Multimodal Processing and Interaction, 1–17. Boston, MA: Springer US, 2008. http://dx.doi.org/10.1007/978-0-387-76316-3_11.
Pełny tekst źródłaPetrakis, Euripides G. M. "Intelligent Search for Image Information on the Web through Text and Link Structure Analysis". W Multimodal Processing and Interaction, 1–17. Boston, MA: Springer US, 2008. http://dx.doi.org/10.1007/978-0-387-76316-3_12.
Pełny tekst źródłaPotamianos, Alexandros, i Manolis Perakakis. "IDesign Principles for Multimodal Spoken Dialogue Systems". W Multimodal Processing and Interaction, 1–18. Boston, MA: Springer US, 2008. http://dx.doi.org/10.1007/978-0-387-76316-3_13.
Pełny tekst źródłaStreszczenia konferencji na temat "Multimodal processing"
Potamianos, Alexandros. "Cognitive Multimodal Processing". W the 2014 Workshop. New York, New York, USA: ACM Press, 2014. http://dx.doi.org/10.1145/2666253.2666264.
Pełny tekst źródłaJohnston, Michael. "Multimodal language processing". W 5th International Conference on Spoken Language Processing (ICSLP 1998). ISCA: ISCA, 1998. http://dx.doi.org/10.21437/icslp.1998-278.
Pełny tekst źródłaYang, Lixin, Genshe Chen, Ronghua Xu, Sherry Chen i Yu Chen. "Decentralized autonomous imaging data processing using blockchain". W Multimodal Biomedical Imaging XIV, redaktorzy Fred S. Azar, Xavier Intes i Qianqian Fang. SPIE, 2019. http://dx.doi.org/10.1117/12.2513243.
Pełny tekst źródłaBarricelli, Barbara Rita, Marco Padula i Paolo Luigi Scala. "TMS for Multimodal Information Processing". W 2009 20th International Workshop on Database and Expert Systems Application. IEEE, 2009. http://dx.doi.org/10.1109/dexa.2009.34.
Pełny tekst źródłaZhu, Junnan, Haoran Li, Tianshang Liu, Yu Zhou, Jiajun Zhang i Chengqing Zong. "MSMO: Multimodal Summarization with Multimodal Output". W Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2018. http://dx.doi.org/10.18653/v1/d18-1448.
Pełny tekst źródłaDamian, Ionut, Michael Dietz, Frank Gaibler i Elisabeth André. "Social signal processing for dummies". W ICMI '16: INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION. New York, NY, USA: ACM, 2016. http://dx.doi.org/10.1145/2993148.2998527.
Pełny tekst źródłaKokkinidis, K., A. Stergiaki i A. Tsagaris. "Machine learning via multimodal signal processing". W 2017 6th International Conference on Modern Circuits and Systems Technologies (MOCAST). IEEE, 2017. http://dx.doi.org/10.1109/mocast.2017.7937653.
Pełny tekst źródłaOviatt, Sharon. "Multimodal system processing in mobile environments". W the 13th annual ACM symposium. New York, New York, USA: ACM Press, 2000. http://dx.doi.org/10.1145/354401.354408.
Pełny tekst źródłaKharinov, Mikhail V., i Aleksandr N. Bykov. "Data Structure for Multimodal Signal Processing". W 2019 International Russian Automation Conference. IEEE, 2019. http://dx.doi.org/10.1109/rusautocon.2019.8867769.
Pełny tekst źródłaBangalore, Srinivas, i Michael Johnston. "Robust gesture processing for multimodal interaction". W the 10th international conference. New York, New York, USA: ACM Press, 2008. http://dx.doi.org/10.1145/1452392.1452439.
Pełny tekst źródłaRaporty organizacyjne na temat "Multimodal processing"
Gazzaniga, Michael S. Multimodal Interactions in Sensory-Motor Processing. Fort Belvoir, VA: Defense Technical Information Center, czerwiec 1992. http://dx.doi.org/10.21236/ada255780.
Pełny tekst źródłaHughes, H. C., P. A. Reuter-Lorenz, R. Fendrich, G. Nozawa i M. S. Gazzaniga. Multimodal Interactions in Sensory-Motor Processing. Fort Belvoir, VA: Defense Technical Information Center, wrzesień 1990. http://dx.doi.org/10.21236/ada229111.
Pełny tekst źródłaLewandowski, Lawrence J., Susan B. Hursh i David A. Kobus. Multimodal versus Unimodal Information Processing of Words. Fort Belvoir, VA: Defense Technical Information Center, lipiec 1985. http://dx.doi.org/10.21236/ada160517.
Pełny tekst źródłaVarshney, Pramod K. Multimodal Signal Processing for Personnel Detection and Activity Classification for Indoor Surveillance. Fort Belvoir, VA: Defense Technical Information Center, listopad 2013. http://dx.doi.org/10.21236/ada606602.
Pełny tekst źródłaHamlin, Alexandra, Erik Kobylarz, James Lever, Susan Taylor i Laura Ray. Assessing the feasibility of detecting epileptic seizures using non-cerebral sensor. Engineer Research and Development Center (U.S.), grudzień 2021. http://dx.doi.org/10.21079/11681/42562.
Pełny tekst źródłaLee, W. S., Victor Alchanatis i Asher Levi. Innovative yield mapping system using hyperspectral and thermal imaging for precision tree crop management. United States Department of Agriculture, styczeń 2014. http://dx.doi.org/10.32747/2014.7598158.bard.
Pełny tekst źródłaFederal Information Processing Standards Publication: detail specification for 62.5?m core diameter125-?m cladding diameter class IA multimode, graded-index optical waveguide fibers. Gaithersburg, MD: National Institute of Standards and Technology, 1989. http://dx.doi.org/10.6028/nist.fips.159.
Pełny tekst źródłaFederal Information Processing Standards Publication: detail specification for 62.5?m core diameter125-?m cladding diameter class IA multimode, graded-index optical waveguide fibers. Gaithersburg, MD: National Institute of Standards and Technology, 1990. http://dx.doi.org/10.6028/nist.fips.159-1990.
Pełny tekst źródła