Littérature scientifique sur le sujet « Multimodal Concepts »
Créez une référence correcte selon les styles APA, MLA, Chicago, Harvard et plusieurs autres
Consultez les listes thématiques d’articles de revues, de livres, de thèses, de rapports de conférences et d’autres sources académiques sur le sujet « Multimodal Concepts ».
À côté de chaque source dans la liste de références il y a un bouton « Ajouter à la bibliographie ». Cliquez sur ce bouton, et nous générerons automatiquement la référence bibliographique pour la source choisie selon votre style de citation préféré : APA, MLA, Harvard, Vancouver, Chicago, etc.
Vous pouvez aussi télécharger le texte intégral de la publication scolaire au format pdf et consulter son résumé en ligne lorsque ces informations sont inclues dans les métadonnées.
Articles de revues sur le sujet "Multimodal Concepts"
Van der Linden, P. « Multimodal Blood Sparing Concepts ». ains · Anästhesiologie · Intensivmedizin · Notfallmedizin · Schmerztherapie 36, Suppl 2 (novembre 2001) : 101. http://dx.doi.org/10.1055/s-2001-18197.
Texte intégralMangin, Olivier. « Multimodal concepts for social robots ». AI Matters 3, no 1 (25 mai 2017) : 19–20. http://dx.doi.org/10.1145/3054837.3054844.
Texte intégralMangin, Olivier. « Multimodal concepts for social robots ». AI Matters 3, no 1 (20 mars 2017) : 19–20. http://dx.doi.org/10.1145/3067682.3067688.
Texte intégralGulen, Elvan, Turgay Yilmaz et Adnan Yazici. « Multimodal Information Fusion for Semantic Video Analysis ». International Journal of Multimedia Data Engineering and Management 3, no 4 (octobre 2012) : 52–74. http://dx.doi.org/10.4018/jmdem.2012100103.
Texte intégralWust, P., B. Rau, M. Gremmler, P. Schlag, A. Jordan, J. Löffel, H. Riess et R. Felix. « Radio-Thermotherapy in Multimodal Surgical Treatment Concepts ». Oncology Research and Treatment 18, no 2 (1995) : 110–21. http://dx.doi.org/10.1159/000218570.
Texte intégralLedezma, Carlos J., et Max Wintermark. « Multimodal CT in Stroke Imaging : New Concepts ». Radiologic Clinics of North America 47, no 1 (janvier 2009) : 109–16. http://dx.doi.org/10.1016/j.rcl.2008.10.008.
Texte intégralVidosavljević, Milena. « Multimodal and digital literacy as new concepts in education ». Zbornik radova Filozofskog fakulteta u Pristini 52, no 4 (2022) : 129–44. http://dx.doi.org/10.5937/zrffp52-37477.
Texte intégralLou, Adrian. « Multimodal simile ». English Text Construction 10, no 1 (15 juin 2017) : 106–31. http://dx.doi.org/10.1075/etc.10.1.06lou.
Texte intégralShcherbin, Vyacheslav K. « Methods of modelling of socio-economic phenomena – method of constructing multiple spirals and multimodal analysis ». Journal of the Belarusian State University. Sociology, no 4 (16 décembre 2021) : 15–25. http://dx.doi.org/10.33581/2521-6821-2021-4-15-25.
Texte intégralDaulay, Nahdyah Sari, SitiIsma Sari Lubis et Widya Wulandari. « MULTIMODAL METAPHOR IN ADVERTISEMENT ». AICLL : ANNUAL INTERNATIONAL CONFERENCE ON LANGUAGE AND LITERATURE 1, no 1 (17 avril 2018) : 170–75. http://dx.doi.org/10.30743/aicll.v1i1.24.
Texte intégralThèses sur le sujet "Multimodal Concepts"
Schmüdderich, Jens M. [Verfasser]. « Multimodal Learning of Grounded Concepts in Embodied Systems / Jens M Schmüdderich ». Aachen : Shaker, 2010. http://d-nb.info/1120864771/34.
Texte intégralNguyen, Nhu Van. « Représentations visuelles de concepts textuels pour la recherche et l'annotation interactives d'images ». Phd thesis, Université de La Rochelle, 2011. http://tel.archives-ouvertes.fr/tel-00730707.
Texte intégralFeuerstein, Marco. « Augmented reality in laparoscopic surgery new concepts and methods for intraoperative multimodal imaging and hybrid tracking in computer aided surgery ». Saarbrücken VDM Verlag Dr. Müller, 2007. http://d-nb.info/991301250/04.
Texte intégralMangin, Olivier. « Emergence de concepts multimodaux : de la perception de mouvements primitifs à l'ancrage de mots acoustiques ». Thesis, Bordeaux, 2014. http://www.theses.fr/2014BORD0002/document.
Texte intégralThis thesis focuses on learning recurring patterns in multimodal perception. For that purpose it develops cognitive systems that model the mechanisms providing such capabilities to infants; a methodology that fits into thefield of developmental robotics.More precisely, this thesis revolves around two main topics that are, on the one hand the ability of infants or robots to imitate and understand human behaviors, and on the other the acquisition of language. At the crossing of these topics, we study the question of the how a developmental cognitive agent can discover a dictionary of primitive patterns from its multimodal perceptual flow. We specify this problem and formulate its links with Quine's indetermination of translation and blind source separation, as studied in acoustics.We sequentially study four sub-problems and provide an experimental formulation of each of them. We then describe and test computational models of agents solving these problems. They are particularly based on bag-of-words techniques, matrix factorization algorithms, and inverse reinforcement learning approaches. We first go in depth into the three separate problems of learning primitive sounds, such as phonemes or words, learning primitive dance motions, and learning primitive objective that compose complex tasks. Finally we study the problem of learning multimodal primitive patterns, which corresponds to solve simultaneously several of the aforementioned problems. We also details how the last problems models acoustic words grounding
Myers, Isaac [Verfasser]. « Improved survival of patients with HCC through new therapeutic options and the use of multimodal therapy concepts : data from a large German university hospital / Isaac Myers ». Berlin : Medizinische Fakultät Charité - Universitätsmedizin Berlin, 2014. http://d-nb.info/1061023567/34.
Texte intégralDíaz, Silva Luis Eduardo, et Antezana Diego Federico Rioseco. « El transporte multimodal : concepto, problemática y proyección ». Tesis, Universidad de Chile, 2001. http://www.repositorio.uchile.cl/handle/2250/114557.
Texte intégralEl éxito de las transacciones comerciales nacionales e internacionales depende, entre muchos otros factores, de la eficiencia de las cadenas de transporte de las mercaderías transadas. En la actualidad el transporte ha tenido que adaptarse a las exigencias de la “nueva economía”, cuya demanda por la integración económica, permiten la creación de la figura jurídica del transporte multimodal de mercancías. El transporte multimodal de mercancías es aquel por el cual se transportan mercancías, por dos o más modos diferentes de transporte, de acuerdo a un sólo contrato. La presente obra tiene como objetivo presentar e introducir el novedoso concepto del transporte multimodal, diferenciándolo del tradicional transporte unimodal y segmentado; plantear sus actuales dificultades y divergencias, como también proponer los cambios necesarios para la implementación del mismo como nuevo sistema de transporte que aporte al crecimiento económico de los países en vías de desarrollo.
Bellik, Yacine. « Interfaces multimodales : concepts, modeles et architectures ». Paris 11, 1995. http://www.theses.fr/1995PA112178.
Texte intégralHalonen, Maria. « Design för lärande och multimodala texter i svenskämnet : En produktorienterad studie av två läromedel i svenska ». Thesis, Högskolan i Gävle, Avdelningen för humaniora, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:hig:diva-18844.
Texte intégralLiu, Yuanting. « Multimodal interaction : developing an interaction concept for a touchscreen incorporating tactile feedback ». Diss., lmu, 2012. http://nbn-resolving.de/urn:nbn:de:bvb:19-138991.
Texte intégralLiu, Ningning. « Contributions to generic and affective visual concept recognition ». Thesis, Ecully, Ecole centrale de Lyon, 2013. http://www.theses.fr/2013ECDL0038.
Texte intégralThis Ph.D thesis is dedicated to visual concept recognition (VCR). Due to many realistic difficulties, it is still considered to be one of the most challenging problems in computer vision and pattern recognition. In this context, we have proposed some innovative contributions for the task of VCR, particularly in building multimodal approaches that efficiently combine visual and textual information. Firstly, we have proposed semantic features for VCR and have investigated the efficiency of different types of low-level visual features for VCR including color, texture and shape. Specifically, we believe that different concepts require different features to efficiently characterize them for the recognition. Therefore, we have investigated in the context of VCR various visual representations, not only global features including color, shape and texture, but also the state-of-the-art local visual descriptors such as SIFT, Color SIFT, HOG, DAISY, LBP, Color LBP. To help bridging the semantic gap between low-level visual features and high level semantic concepts, and particularly those related to emotions and feelings, we have proposed mid-level visual features based on the visual harmony and dynamism semantics using Itten’s color theory and psychological interpretations. Moreover, we have employed a spatial pyramid strategy to capture the spatial information when building our mid-level features harmony and dynamism. We have also proposed a new representation of color HSV histograms by employing a visual attention model to identify the regions of interest in images. Secondly, we have proposed a novel textual feature designed for VCR. Indeed, most of online-shared photos provide textual descriptions in the form of tags or legends. In fact, these textual descriptions are a rich source of semantic information on visual data that is interesting to consider for the purpose of VCR or multimedia information retrieval. We propose the Histograms of Textual Concepts (HTC) to capture the semantic relatedness of concepts. The general idea behind HTC is to represent a text document as a histogram of textual concepts towards a vocabulary or dictionary, whereas its value is the accumulation of the contribution of each word within the text document toward the underlying concept according to a predefined semantic similarity measure. Several variants of HTC have been proposed that revealed to be very efficient for VCR. Inspired by the Cepstral speech analysis process, we have also developed Cepstral HTC to capture both term frequency-based information (like TF-IDF) and the relatedness of semantic concepts in the sparse image tags, which overcomes the HTC’s shortcoming of ignoring term frequency-based information. Thirdly, we have proposed a fusion scheme to combine different sources of Later Fusion, (SWLF) is designed to select the best features and to weight their scores for each concept to be recognized. SWLF proves particularly efficient for fusing visual and textual modalities in comparison with some other standard fusion schemes. While a late fusion at score level is reputed as a simple and effective way to fuse features of different nature for machine-learning problems, the proposed SWLF builds on two simple insights. First, the score delivered by a feature type should be weighted by its intrinsic quality for the classification problem at hand. Second, in a multi-label scenario where several visual concepts may be assigned to an image, different visual concepts may require different features which best recognize them. In addition to SWLF, we also propose a novel combination approach based on Dempster-Shafer’s evidence theory, whose interesting properties allow fusing different ambiguous sources of information for visual affective recognition. [...]
Livres sur le sujet "Multimodal Concepts"
Brown, J. Martin, Minesh P. Mehta et Carsten Nieder, dir. Multimodal Concepts for Integration of Cytotoxic Drugs. Berlin, Heidelberg : Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/3-540-35662-2.
Texte intégralPattern Mining and Concept Discovery for Multimodal Content Analysis. [New York, N.Y.?] : [publisher not identified], 2016.
Trouver le texte intégralNes, Ir R. Design of multimodal transport systems : Setting the scene, review of literature and basic concept. Delft : TRAIL Research School, 2000.
Trouver le texte intégralMehta, M. P., L. W. Brady, J. M. Brown, C. Nieder et H. P. Heilmann. Multimodal Concepts for Integration of Cytotoxic Drugs. Springer Berlin / Heidelberg, 2010.
Trouver le texte intégralGórska, Elżbieta. Understanding Abstract Concepts across Modes in Multimodal Discourse. Routledge, 2019. http://dx.doi.org/10.4324/9780429282737.
Texte intégralGórska, Elzbieta. Understanding Abstract Concepts Across Modes in Multimodal Discourse. Taylor & Francis Group, 2021.
Trouver le texte intégralUnderstanding Abstract Concepts Across Modes in Multimodal Discourse. Taylor & Francis Group, 2019.
Trouver le texte intégralMultimodal Concepts for Integration of Cytotoxic Drugs (Medical Radiology). Springer, 2006.
Trouver le texte intégral(Foreword), L. W. Brady, H. P. Heilmann (Foreword), M. Molls (Foreword), J. M. Brown (Editor), M. P. Mehta (Editor) et C. Nieder (Editor), dir. Multimodal Concepts for Integration of Cytotoxic Drugs (Medical Radiology / Radiation Oncology). Springer, 2006.
Trouver le texte intégralGórska, Elżbieta. Understanding Abstract Concepts Across Modes in Multimodal Discourse : A Cognitive Linguistic Approach. Taylor & Francis Group, 2019.
Trouver le texte intégralChapitres de livres sur le sujet "Multimodal Concepts"
Sheppard, Jennifer. « Issues in Digital and Multimodal Writing ». Dans Concepts in Composition, 379–445. Third edition. | New York : Routledge, 2019. : Routledge, 2019. http://dx.doi.org/10.4324/9780203728659-10.
Texte intégralCarroll, Michael A., et Ema C. Yamamoto. « Level of Service Concepts in Multimodal Environments ». Dans Traffic Engineering Handbook, 149–76. Hoboken, NJ, USA : John Wiley & Sons, Inc, 2016. http://dx.doi.org/10.1002/9781119174738.ch5.
Texte intégralHürst, Wolfgang, et Casper van Wezel. « Multimodal Interaction Concepts for Mobile Augmented Reality Applications ». Dans Lecture Notes in Computer Science, 157–67. Berlin, Heidelberg : Springer Berlin Heidelberg, 2011. http://dx.doi.org/10.1007/978-3-642-17829-0_15.
Texte intégralGhasemAghaei, Reza, Ali Arya et Robert Biddle. « MADE Ratio : Affective Multimodal Software for Mathematical Concepts ». Dans Lecture Notes in Computer Science, 487–98. Cham : Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-39483-1_44.
Texte intégralStopka, Ulrike. « Multimodal Mobility Packages – Concepts and Methodological Design Approaches ». Dans HCI in Mobility, Transport, and Automotive Systems. Driving Behavior, Urban and Smart Mobility, 318–39. Cham : Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-50537-0_24.
Texte intégralFyles, Anthony W., Michael Milosevic et Amit Oza. « Applications to Gynecological Cancers ». Dans Multimodal Concepts for Integration of Cytotoxic Drugs, 303–15. Berlin, Heidelberg : Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/3-540-35662-2_20.
Texte intégralSignorello, Giovanni, Giovanni Maria Farinella, Giovanni Gallo, Luciano Santo, Antonino Lopes et Emanuele Scuderi. « Exploring Protected Nature Through Multimodal Navigation of Multimedia Contents ». Dans Advanced Concepts for Intelligent Vision Systems, 841–52. Cham : Springer International Publishing, 2015. http://dx.doi.org/10.1007/978-3-319-25903-1_72.
Texte intégralChetty, Girija, Julian Goodwin et Monica Singh. « Digital Image Tamper Detection Based on Multimodal Fusion of Residue Features ». Dans Advanced Concepts for Intelligent Vision Systems, 79–87. Berlin, Heidelberg : Springer Berlin Heidelberg, 2010. http://dx.doi.org/10.1007/978-3-642-17691-3_8.
Texte intégralLinebarger, Deborah L., et Lori Norton-Meier. « Scientific Concepts, Multiple Modalities, and Young Children ». Dans Using Multimodal Representations to Support Learning in the Science Classroom, 97–116. Cham : Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-16450-2_6.
Texte intégralÁlvarez-Carmona, Miguel Á., Esaú Villatoro-Tello, Luis Villaseñor-Pineda et Manuel Montes-y-Gómez. « Classifying the Social Media Author Profile Through a Multimodal Representation ». Dans Intelligent Technologies : Concepts, Applications, and Future Directions, 57–81. Singapore : Springer Nature Singapore, 2022. http://dx.doi.org/10.1007/978-981-19-1021-0_3.
Texte intégralActes de conférences sur le sujet "Multimodal Concepts"
Wang, Zhichun, et Minqiang Li. « A Coevolution Approach for Learning Multimodal Concepts ». Dans Third International Conference on Natural Computation (ICNC 2007). IEEE, 2007. http://dx.doi.org/10.1109/icnc.2007.12.
Texte intégralRajaby Faghihi, Hossein, Roshanak Mirzaee, Sudarshan Paliwal et Parisa Kordjamshidi. « Latent Alignment of Procedural Concepts in Multimodal Recipes ». Dans Proceedings of the First Workshop on Advances in Language and Vision Research. Stroudsburg, PA, USA : Association for Computational Linguistics, 2020. http://dx.doi.org/10.18653/v1/2020.alvr-1.5.
Texte intégralLi, PengYuan, et YongLi Wang. « A Multimodal Entity Linking Approach Incorporating Topic Concepts ». Dans 2021 International Conference on Computer Information Science and Artificial Intelligence (CISAI). IEEE, 2021. http://dx.doi.org/10.1109/cisai54367.2021.00100.
Texte intégralMaiera, Kathrin, Jürgen Hellbrücka et Heike Sacherb. « A Visuohaptic Collision Warning Approach for High-Priority Braking Scenarios ». Dans Applied Human Factors and Ergonomics Conference. AHFE International, 2022. http://dx.doi.org/10.54941/ahfe100751.
Texte intégralNakamura, Tomoaki, Takayuki Nagai et Naoto Iwahashi. « Grounding of word meanings in multimodal concepts using LDA ». Dans 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2009). IEEE, 2009. http://dx.doi.org/10.1109/iros.2009.5354736.
Texte intégralZhu, Qiang, Mei-Chen Yeh et Kwang-Ting Cheng. « Multimodal fusion using learned text concepts for image categorization ». Dans the 14th annual ACM international conference. New York, New York, USA : ACM Press, 2006. http://dx.doi.org/10.1145/1180639.1180698.
Texte intégralCalabrese, Agostina, Michele Bevilacqua et Roberto Navigli. « EViLBERT : Learning Task-Agnostic Multimodal Sense Embeddings ». Dans Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20}. California : International Joint Conferences on Artificial Intelligence Organization, 2020. http://dx.doi.org/10.24963/ijcai.2020/67.
Texte intégralChaplot, Devendra Singh, Lisa Lee, Ruslan Salakhutdinov, Devi Parikh et Dhruv Batra. « Embodied Multimodal Multitask Learning ». Dans Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20}. California : International Joint Conferences on Artificial Intelligence Organization, 2020. http://dx.doi.org/10.24963/ijcai.2020/338.
Texte intégralWang, Liming, et Mark A. Hasegawa-Johnson. « Multimodal Word Discovery and Retrieval with Phone Sequence and Image Concepts ». Dans Interspeech 2019. ISCA : ISCA, 2019. http://dx.doi.org/10.21437/interspeech.2019-1487.
Texte intégralBendre, Nihar, Kevin Desai et Peyman Najafirad. « Generalized Zero-Shot Learning Using Multimodal Variational Auto-Encoder With Semantic Concepts ». Dans 2021 IEEE International Conference on Image Processing (ICIP). IEEE, 2021. http://dx.doi.org/10.1109/icip42928.2021.9506108.
Texte intégral