Thèses sur le sujet « Dati multimodali »
Créez une référence correcte selon les styles APA, MLA, Chicago, Harvard et plusieurs autres
Consultez les 50 meilleures thèses pour votre recherche sur le sujet « Dati multimodali ».
À côté de chaque source dans la liste de références il y a un bouton « Ajouter à la bibliographie ». Cliquez sur ce bouton, et nous générerons automatiquement la référence bibliographique pour la source choisie selon votre style de citation préféré : APA, MLA, Harvard, Vancouver, Chicago, etc.
Vous pouvez aussi télécharger le texte intégral de la publication scolaire au format pdf et consulter son résumé en ligne lorsque ces informations sont inclues dans les métadonnées.
Parcourez les thèses sur diverses disciplines et organisez correctement votre bibliographie.
GIANSANTI, VALENTINA. « Integration of heterogeneous single cell data with Wasserstein Generative Adversarial Networks ». Doctoral thesis, Università degli Studi di Milano-Bicocca, 2023. https://hdl.handle.net/10281/404516.
Texte intégralTissues, organs and organisms are complex biological systems. They are objects of many studies aiming at characterizing their biological processes. Understanding how they work and how they interact in healthy and unhealthy samples gives the possibility to interfere, correcting and preventing dysfunctions, possibly leading to diseases. Recent advances in single-cell technologies are expanding our capabilities to profile at single-cell resolution various molecular layers, by targeting the transcriptome, the genome, the epigenome and the proteome. The number of single-cell datasets, their size and the diverse modalities they describe is continuously increasing, prompting the need to develop robust methods to integrate multiomic datasets, whether paired from the same cells or, most challenging, from unpaired separate experiments. The integration of different source of information results in a more comprehensive description of the whole system. Most published methods allow the integration of limited number of omics (generally two) and make assumptions about their inter-relationships. They often impose the conversion of a data modality into the other one (e.g., ATAC peaks converted in a gene activity matrix). This step introduces an important level of approximation, which could affect the analysis later performed. Here we propose MOWGAN (Multi Omic Wasserstein Generative Adversarial Network), a deep-learning based framework to simulate paired multimodal data supporting high number of modalities (more than two) and agnostic about their relationships (no assumption is imposed). Each modality is embedded into feature spaces with same dimensionality across all modalities. This step prevents any conversion between data modalities. The embeddings are sorted based on the first Laplacian Eigenmap. Mini-batches are selected by a Bayesian ridge regressor to train a Wasserstein Generative Adversarial Network with gradient penalty. The output of the generative network is used to bridge real unpaired data. MOWGAN was prototyped on public data for which paired and unpaired RNA and ATAC experiments exists. Evaluation was conducted on the ability to produce data integrable with the original ones, on the amount of shared information between synthetic layers and on the ability to impose association between molecular layers that are truly connected. The organization of the embeddings in mini-batches allows MOWGAN to have a network architecture independent of the number of modalities evaluated. Indeed, the framework was also successfully applied to integrate three (e.g., RNA, ATAC and protein or histone modification data) and four modalities (e.g., RNA, ATAC, protein, histone modifications). MOWGAN’s performance was evaluated in terms of both computational scalability and biological meaning, being the latter the most important to avoid erroneous conclusion. A comparison was conducted with published methods, concluding that MOWGAN performs better when looking at the ability to retrieve the correct biological identity (e.g., cell types) and associations. In conclusion, MOWGAN is a powerful tool for multi-omics data integration in single-cell, which answer most of the critical issues observed in the field.
Medjahed, Hamid. « Distress situation identification by multimodal data fusion for home healthcare telemonitoring ». Thesis, Evry, Institut national des télécommunications, 2010. http://www.theses.fr/2010TELE0002/document.
Texte intégralThe population age increases in all societies throughout the world. In Europe, for example, the life expectancy for men is about 71 years and for women about 79 years. For North America the life expectancy, currently is about 75 for men and 81 for women. Moreover, the elderly prefer to preserve their independence, autonomy and way of life living at home the longest time possible. The current healthcare infrastructures in these countries are widely considered to be inadequate to meet the needs of an increasingly older population. Home healthcare monitoring is a solution to deal with this problem and to ensure that elderly people can live safely and independently in their own homes for as long as possible. Automatic in-home healthcare monitoring is a technological approach which helps people age in place by continuously telemonitoring. In this thesis, we explore automatic in-home healthcare monitoring by conducting a study of professionals who currently perform in-home healthcare monitoring, by combining and synchronizing various telemonitoring modalities,under a data synchronization and multimodal data fusion platform, FL-EMUTEM (Fuzzy Logic Multimodal Environment for Medical Remote Monitoring). This platform incorporates algorithms that process each modality and providing a technique of multimodal data fusion which can ensures a pervasive in-home health monitoring for elderly people based on fuzzy logic.The originality of this thesis which is the combination of various modalities in the home, about its inhabitant and their surroundings, will constitute an interesting benefit and impact for the elderly person suffering from loneliness. This work complements the stationary smart home environment in bringing to bear its capability for integrative continuous observation and detection of critical situations
Vielzeuf, Valentin. « Apprentissage neuronal profond pour l'analyse de contenus multimodaux et temporels ». Thesis, Normandie, 2019. http://www.theses.fr/2019NORMC229/document.
Texte intégralOur perception is by nature multimodal, i.e. it appeals to many of our senses. To solve certain tasks, it is therefore relevant to use different modalities, such as sound or image.This thesis focuses on this notion in the context of deep learning. For this, it seeks to answer a particular problem: how to merge the different modalities within a deep neural network?We first propose to study a problem of concrete application: the automatic recognition of emotion in audio-visual contents.This leads us to different considerations concerning the modeling of emotions and more particularly of facial expressions. We thus propose an analysis of representations of facial expression learned by a deep neural network.In addition, we observe that each multimodal problem appears to require the use of a different merge strategy.This is why we propose and validate two methods to automatically obtain an efficient fusion neural architecture for a given multimodal problem, the first one being based on a central fusion network and aimed at preserving an easy interpretation of the adopted fusion strategy. While the second adapts a method of neural architecture search in the case of multimodal fusion, exploring a greater number of strategies and therefore achieving better performance.Finally, we are interested in a multimodal view of knowledge transfer. Indeed, we detail a non-traditional method to transfer knowledge from several sources, i.e. from several pre-trained models. For that, a more general neural representation is obtained from a single model, which brings together the knowledge contained in the pre-trained models and leads to state-of-the-art performances on a variety of facial analysis tasks
Lazarescu, Mihai M. « Incremental learning for querying multimodal symbolic data ». Thesis, Curtin University, 2000. http://hdl.handle.net/20.500.11937/1660.
Texte intégralLazarescu, Mihai M. « Incremental learning for querying multimodal symbolic data ». Curtin University of Technology, School of Computing, 2000. http://espace.library.curtin.edu.au:80/R/?func=dbin-jump-full&object_id=10010.
Texte intégralDA, CRUZ GARCIA NUNO RICARDO. « Learning with Privileged Information using Multimodal Data ». Doctoral thesis, Università degli studi di Genova, 2020. http://hdl.handle.net/11567/997636.
Texte intégralXin, Bowen. « Multimodal Data Fusion and Quantitative Analysis for Medical Applications ». Thesis, The University of Sydney, 2021. https://hdl.handle.net/2123/26678.
Texte intégralPOLSINELLI, MATTEO. « Modelli di Intelligenza Artificiale per l'analisi di dati da neuroimaging multimodale ». Doctoral thesis, Università degli Studi dell'Aquila, 2022. http://hdl.handle.net/11697/192072.
Texte intégralKhan, Mohd Tauheed. « Multimodal Data Fusion Using Voice and Electromyography Data for Robotic Control ». University of Toledo / OhioLINK, 2019. http://rave.ohiolink.edu/etdc/view?acc_num=toledo156440368925597.
Texte intégralOztarak, Hakan. « Structural And Event Based Multimodal Video Data Modeling ». Master's thesis, METU, 2005. http://etd.lib.metu.edu.tr/upload/12606919/index.pdf.
Texte intégralMcLaughlin, N. R. « Robust multimodal person identification given limited training data ». Thesis, Queen's University Belfast, 2013. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.579747.
Texte intégralKäshammer, Philipp Florian. « A Semantic Interpreter for Multimodal and Multirobot Data ». Thesis, Luleå tekniska universitet, Institutionen för system- och rymdteknik, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:ltu:diva-36896.
Texte intégralRao, Dushyant. « Multimodal learning from visual and remotely sensed data ». Thesis, The University of Sydney, 2015. http://hdl.handle.net/2123/15535.
Texte intégralNourbakhsh, Nargess. « Multimodal Physiological Cognitive Load Measurement ». Thesis, The University of Sydney, 2015. http://hdl.handle.net/2123/14294.
Texte intégralHe, Linbo. « Improving 3D Point Cloud Segmentation Using Multimodal Fusion of Projected 2D Imagery Data : Improving 3D Point Cloud Segmentation Using Multimodal Fusion of Projected 2D Imagery Data ». Thesis, Linköpings universitet, Datorseende, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-157705.
Texte intégralSperandeo, Marco. « Smart mobility : percorsi multimodali in ambiente urbano ». Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2018. http://amslaurea.unibo.it/16178/.
Texte intégralHan, Bote. « The Multimodal Interaction through the Design of Data Glove ». Thesis, Université d'Ottawa / University of Ottawa, 2015. http://hdl.handle.net/10393/32529.
Texte intégralSun, Feng-Tso. « Nonparametric Discovery of Human Behavior Patterns from Multimodal Data ». Research Showcase @ CMU, 2014. http://repository.cmu.edu/dissertations/359.
Texte intégralDamoni, Arben. « Multimodal segmentation for data mining applications in multimedia engineering ». Thesis, London South Bank University, 2012. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.631732.
Texte intégralBalakrishnan, Arjun. « Integrity Analysis of Data Sources in Multimodal Localization System ». Thesis, université Paris-Saclay, 2020. http://www.theses.fr/2020UPASG060.
Texte intégralIntelligent vehicles are a key component in humanity’s vision for safer, efficient, and accessible transportation systems across the world. Due to the multitude of data sources and processes associated with Intelligent vehicles, the reliability of the total system is greatly dependent on the possibility of errors or poor performances observed in its components. In our work, we focus on the critical task of localization of intelligent vehicles and address the challenges in monitoring the integrity of data sources used in localization. The primary contribution of our research is the proposition of a novel protocol for integrity by combining integrity concepts from information systems with the existing integrity concepts in the field of Intelligent Transport Systems (ITS). An integrity monitoring framework based on the theorized integrity protocol that can handle multimodal localization problems is formalized. As the first step, a proof of concept for this framework is developed based on cross-consistency estimation of data sources using polynomial models. Based on the observations from the first step, a 'Feature Grid' data representation is proposed in the second step and a generalized prototype for the framework is implemented. The framework is tested in highways as well as complex urban scenarios to demonstrate that the proposed framework is capable of providing continuous integrity estimates of multimodal data sources used in intelligent vehicle localization
Spechler, Philip. « Predictive Modeling of Adolescent Cannabis Use From Multimodal Data ». ScholarWorks @ UVM, 2017. http://scholarworks.uvm.edu/graddis/690.
Texte intégralNobis, Claudia. « Multimodale Vielfalt ». Doctoral thesis, Humboldt-Universität zu Berlin, Mathematisch-Naturwissenschaftliche Fakultät II, 2015. http://dx.doi.org/10.18452/17194.
Texte intégralMultimodality, the use of several modes of transportation during a specified time period, is a general term for a wide variety of everyday mobility behaviors. It is perceived as an alternative to one-sided use of private cars, and one which has attracted great hopes for the future development of transportation. Based on the research which has been done in the past, people almost always limit themselves to a particular form of multimodal behavior, most often to use of cars and public transportation. The starting point of the present paper is to present and examine the various facets of multimodal behavior in their entirety. To this end, a method of classification will be developed which is derived from the selection of modes of transportation. The analysis of mobility behavior will be based on the data of the German Mobility Panel from 1999 to 2008 and the Mobility in Germany study from the years 2002 and 2008. Subjects will be assigned to modal groups depending on which of the modes of transportation, motorized individual traffic, public transportation and bicycle, are used in the course of a week. The analysis reveals the enormously diverse nature of multimodal behavior. In general, multimodal behavior is an urban phenomenon which is increasingly characterizing the everyday urban routine, especially for younger persons. In aggregate, multimodal persons drive fewer kilometers by car than monomodal car drivers. Their carbon footprint is 20-34 percent less than that of exclusive car drivers, depending on the data set. Nevertheless, many multimodal persons do use cars for a considerable portion of their travel needs. How the relative share of the various modes of transportation will change in the future, especially with respect to long-distance travel, and the impact of the currently observable changes in supply and demand will be decisive factors in the future.
Fernández, Carbonell Marcos. « Automated Multimodal Emotion Recognition ». Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-282534.
Texte intégralAtt kunna läsa och tolka affektiva tillstånd spelar en viktig roll i det mänskliga samhället. Detta är emellertid svårt i vissa situationer, särskilt när information är begränsad till antingen vokala eller visuella signaler. Många forskare har undersökt de så kallade grundläggande känslorna på ett övervakat sätt. Det här examensarbetet innehåller resultaten från en multimodal övervakad och oövervakad studie av ett mer realistiskt antal känslor. För detta ändamål extraheras ljud- och videoegenskaper från GEMEP-data med openSMILE respektive OpenFace. Det övervakade tillvägagångssättet inkluderar jämförelse av flera lösningar och visar att multimodala pipelines kan överträffa unimodala sådana, även med ett större antal affektiva tillstånd. Den oövervakade metoden omfattar en konservativ och en utforskande metod för att hitta meningsfulla mönster i det multimodala datat. Den innehåller också ett innovativt förfarande för att bättre förstå resultatet av klustringstekniker.
Woodcock, Anna, et Elin Salemyr. « Kampen om kommunikationen : En kvalitativ studie av Försvarsmaktens kommunikation och uppdrag ». Thesis, Uppsala universitet, Medier och kommunikation, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-434366.
Texte intégralKulikova, Sofya. « Integration of multimodal imaging data for investigation of brain development ». Thesis, Sorbonne Paris Cité, 2015. http://www.theses.fr/2015PA05T021/document.
Texte intégralMagnetic Resonance Imaging (MRI) is a fundamental tool for in vivo investigation of brain development in newborns, infants and children. It provides several quantitative parameters that reflect changes in tissue properties during development depending on different undergoing maturational processes. However, reliable evaluation of the white matter maturation is still an open question: on one side, none of these parameters can describe the whole complexity of the undergoing changes; on the other side, neither of them is specific to any particular developmental process or tissue property. Developing multiparametric approaches combining complementary information from different MRI parameters is expected to improve our understanding of brain development. In this PhD work, I present two examples of such approaches and demonstrate their relevancy for investigation of maturation across different white matter bundles. The first approach provides a global measure of maturation based on the Mahalanobis distance calculated from different MRI parameters (relaxation times T1 and T2, longitudinal and transverse diffusivities from Diffusion Tensor Imaging, DTI) in infants (3-21 weeks) and adults. This approach provides a better description of the asynchronous maturation across the bundles than univariate approaches. Furthermore, it allows estimating the relative maturational delays between the bundles. The second approach aims at quantifying myelination of brain tissues by calculating Myelin Water Fraction (MWF) in each image voxel. This approach is based on a 3-component tissue model, with each model component having specific relaxation characteristics that were pre-calibrated in three healthy adult subjects. This approach allows fast computing of the MWF maps from infant data and could reveal progression of the brain myelination. The robustness of this approach was further investigated using computer simulations. Another important issue for studying white matter development in children is bundles identification. In the last part of this work I also describe creation of a preliminary atlas of white matter structural connectivity in children aged 17-81 months. This atlas allows automatic extraction of the bundles from tractography datasets. This approach demonstrated its relevance for evaluation of regional maturation of normal white matter in children. Finally, in the last part of the manuscript I describe potential future applications of the previously developed methods to investigation of the white matter in cases of two specific pathologies: focal epilepsy and metachromatic leukodystrophy
Gunapati, Venkat Yashwanth. « CLOUD BASED DISTRIBUTED COMPUTING PLATFORM FOR MULTIMODAL ENERGY DATA STREAMS ». Case Western Reserve University School of Graduate Studies / OhioLINK, 2014. http://rave.ohiolink.edu/etdc/view?acc_num=case1399373847.
Texte intégralJayapandian, Catherine Praveena. « Cloudwave : A Cloud Computing Framework for Multimodal Electrophysiological Big Data ». Case Western Reserve University School of Graduate Studies / OhioLINK, 2014. http://rave.ohiolink.edu/etdc/view?acc_num=case1405516626.
Texte intégralDiehn, Sabrina Maria. « Analysis of data from multimodal chemical characterizations of plant tissues ». Doctoral thesis, Humboldt-Universität zu Berlin, 2021. http://dx.doi.org/10.18452/23065.
Texte intégralThe pre-processing and analysis of spectrometric and spectroscopic data of plant tissue are important in a wide variety of research areas, such as plant biology, agricultural science, and climate research. The focus of the thesis is the optimized utilization of data from plant tissues, which includes data from Matrix-Assisted-Laser Desorption/Ionization time of flight mass spectrometry, Raman spectroscopy, and Fourier transform infrared spectroscopy. The ability to attain a classification using these methods is compared, in particular after combination of the data with each other and with additional chemical and biological information. The discussed examples are concerned with the investigation and classification within a particular plant species, such as the distinction of samples from different populations, growth conditions, or tissue substructures. The data were analyzed by exploratory tools such as principal component analysis and hierarchical cluster analysis, as well as by predictive tools that included partial least square-discriminant analysis and machine learning approaches. Specifically, the results show that combination of the methods with additional plant-related information in a consensus principal component analysis leads to a comprehensive characterization of the samples. Different data pre-treatment strategies are discussed to reduce non-relevant spectral information, e.g., from maps of plant tissues or embedded pollen grains. The results in this work indicate the relevance of the targeted utilization of spectrometric and spectroscopic data and could be applied not only to plant-related topics but also to other analytical classification problems.
Ming, Joy Carol. « #Autism Versus 299.0 : Topic Model Exploration of Multimodal Autism Data ». Thesis, Harvard University, 2015. http://nrs.harvard.edu/urn-3:HUL.InstRepos:14398542.
Texte intégralRastgoo, Mohammad Naim. « Driver stress level detection based on multimodal measurements ». Thesis, Queensland University of Technology, 2019. https://eprints.qut.edu.au/134144/1/Mohammad%20Naim%20Rastgoo%20Thesis_Redacted.pdf.
Texte intégralBöckmann, Christine, Jens Biele, Roland Neuber et Jenny Niebsch. « Retrieval of multimodal aerosol size distribution by inversion of multiwavelength data ». Universität Potsdam, 1997. http://opus.kobv.de/ubp/volltexte/2007/1436/.
Texte intégralSalami, Alireza. « Decoding the complex brain : multivariate and multimodal analyses of neuroimaging data ». Doctoral thesis, Umeå universitet, Institutionen för integrativ medicinsk biologi (IMB), 2012. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-51842.
Texte intégralVukotic, Verdran. « Deep Neural Architectures for Automatic Representation Learning from Multimedia Multimodal Data ». Thesis, Rennes, INSA, 2017. http://www.theses.fr/2017ISAR0015/document.
Texte intégralIn this dissertation, the thesis that deep neural networks are suited for analysis of visual, textual and fused visual and textual content is discussed. This work evaluates the ability of deep neural networks to learn automatic multimodal representations in either unsupervised or supervised manners and brings the following main contributions:1) Recurrent neural networks for spoken language understanding (slot filling): different architectures are compared for this task with the aim of modeling both the input context and output label dependencies.2) Action prediction from single images: we propose an architecture that allow us to predict human actions from a single image. The architecture is evaluated on videos, by utilizing solely one frame as input.3) Bidirectional multimodal encoders: the main contribution of this thesis consists of neural architecture that translates from one modality to the other and conversely and offers and improved multimodal representation space where the initially disjoint representations can translated and fused. This enables for improved multimodal fusion of multiple modalities. The architecture was extensively studied an evaluated in international benchmarks within the task of video hyperlinking where it defined the state of the art today.4) Generative adversarial networks for multimodal fusion: continuing on the topic of multimodal fusion, we evaluate the possibility of using conditional generative adversarial networks to lean multimodal representations in addition to providing multimodal representations, generative adversarial networks permit to visualize the learned model directly in the image domain
Zhu, Meng. « Cross-modal semantic-associative labelling, indexing and retrieval of multimodal data ». Thesis, University of Reading, 2010. http://centaur.reading.ac.uk/24828/.
Texte intégralSamper, González Jorge Alberto. « Learning from multimodal data for classification and prediction of Alzheimer's disease ». Electronic Thesis or Diss., Sorbonne université, 2019. http://www.theses.fr/2019SORUS361.
Texte intégralAlzheimer's disease (AD) is the first cause of dementia worldwide, affecting over 20 million people. Its diagnosis at an early stage is essential to ensure a proper care of patients, and to develop and test novel treatments. AD is a complex disease that has to be characterized by the use of different measurements: cognitive and clinical tests, neuroimaging including magnetic resonance imaging (MRI) and positron emission tomography (PET), genotyping, etc. There is an interest in exploring the discriminative and predictive capabilities of these diverse markers, which reflect different aspects of the disease and potentially carry complementary information, from an early stage of the disease. The objective of this PhD thesis was thus to assess the potential and to integrate multiple modalities using machine learning methods, in order to automatically classify patients with AD and predict the development of the disease from the earliest stages. More specifically, we aimed to make progress toward the translation of such approaches toward clinical practice. The thesis comprises three main studies. The first one tackles the differential diagnosis between different forms of dementia from MRI data. This study was performed using clinical routine data, thereby providing a more realistic evaluation scenario. The second one proposes a new framework for reproducible evaluation of AD classification algorithms from MRI and PET data. Indeed, while numerous approaches have been proposed for AD classification in the literature, they are difficult to compare and to reproduce. The third part is devoted to the prediction of progression to AD in patients with mild cognitive impairment through the integration of multimodal data, including MRI, PET, clinical/cognitive evaluations and genotyping. In particular, we systematically assessed the added value of neuroimaging over clinical/cognitive data only. Since neuroimaging is more expensive and less widely available, this is important to justify its use as input of classification algorithms
Bao, Guoqing. « End-to-End Machine Learning Models for Multimodal Medical Data Analysis ». Thesis, The University of Sydney, 2022. https://hdl.handle.net/2123/28153.
Texte intégralFaghihi, Reza. « Mise en correspondance SPECT-CT par conditions de consistance ». Université Joseph Fourier (Grenoble), 2002. http://www.theses.fr/2002GRE19011.
Texte intégralMolins, Jiménez Antonio. « Multimodal integration of EEG and MEG data using minimum ℓ₂-norm estimates ». Thesis, Massachusetts Institute of Technology, 2007. http://hdl.handle.net/1721.1/40528.
Texte intégralIncludes bibliographical references (leaves 69-74).
The aim of this thesis was to study the effects of multimodal integration of electroencephalography (EEG) and magnetoencephalography (MEG) data on the minimum ℓ₂-norm estimates of cortical current densities. We investigated analytically the effect of including EEG recordings in MEG studies versus the addition of new MEG channels. To further confirm these results, clinical datasets comprising concurrent MEG/EEG acquisitions were analyzed. Minimum ℓ₂-norm estimates were computed using MEG alone, EEG alone, and the combination of the two modalities. Localization accuracy of responses to median-nerve stimulation was evaluated to study the utility of combining MEG and EEG.
by Antonio Molins Jiménez.
S.M.
Bießmann, Felix Verfasser], et Klaus-Robert [Akademischer Betreuer] [Müller. « Data-driven analysis for multimodal neuroimaging / Felix Bießmann. Betreuer : Klaus-Robert Müller ». Berlin : Universitätsbibliothek der Technischen Universität Berlin, 2012. http://d-nb.info/1018985220/34.
Texte intégralYoung, J. M. « Probabilistic prediction of Alzheimer's disease from multimodal image data with Gaussian processes ». Thesis, University College London (University of London), 2015. http://discovery.ucl.ac.uk/1461115/.
Texte intégralGimenes, Gabriel Perri. « Advanced techniques for graph analysis : a multimodal approach over planetary-scale data ». Universidade de São Paulo, 2015. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-26062015-105026/.
Texte intégralAplicações como comércio eletrônico, redes de computadores, redes sociais e biologia (interação proteica), entre outras, levaram a produção de dados que podem ser representados como grafos à escala planetária { podendo possuir milhões de nós e bilhões de arestas. Tais aplicações apresentam problemas desafiadores quando a tarefa consiste em usar as informações contidas nos grafos para auxiliar processos de tomada de decisão através da descoberta de padrões não triviais e potencialmente utéis. Para processar esses grafos em busca de padrões, tanto pesquisadores como a indústria tem usado recursos de processamento distribuído organizado em clusters computacionais. Entretanto, a construção e manutenção desses clusters pode ser complexa, trazendo tanto problemas técnicos como financeiros que podem ser proibitivos em diversos casos. Por isso, torna-se desejável a capacidade de se processar grafos em larga escala usando somente um nó computacional. Para isso, foram desenvolvidos processos e algoritmos seguindo três abordagens diferentes, visando a definição de um arcabouço de análise capaz de revelar padrões, compreensão e auxiliar na tomada de decisão sobre grafos em escala planetária.
Quack, Till. « Large scale mining and retrieval of visual data in a multimodal context ». Konstanz Hartung-Gorre, 2009. http://d-nb.info/993614620/04.
Texte intégralLabourey, Quentin. « Fusions multimodales pour la recherche d'humains par un robot mobile ». Thesis, Université Grenoble Alpes (ComUE), 2017. http://www.theses.fr/2017GREAM020/document.
Texte intégralIn this work, we consider the case of mobile robot that aims at detecting and positioning itself with respect to humans in its environment. In order to fulfill this mission, the robot is equipped with various sensors (RGB-Depth, microphones, laser telemeter). This thesis contains contributions of various natures:Sound classification in indoor environments: A small taxonomy is proposed in a classification method destined to enable a robot to detect human presence. Uncertainty of classification is taken into account through the use of belief functions, allowing us to label a sound as "unknown".Speaker tracking thanks to audiovisual data fusion: The robot is witness to a social interaction and tracks the successive speakers with probabilistic audiovisual data fusion. The proposed method was tested on videos extracted from the robot's sensors.Navigation dedicated to human detection thanks to a multimodal fusion:} The robot autonomously navigates in a known environment to detect humans thanks to heterogeneous sensors. The data is fused to create a multimodal perception grid. This grid enables the robot to chose its destinations, depending on the priority of perceived information. This system was implemented and tested on a Q.bo robot.Credibilist modelization of the environment for navigation: The creation of the multimodal perception grid is improved by the use of credibilist fusion. This enables the robot to maintain an evidential grid in time, containing the perceived information and its uncertainty. This system was implemented in simulation first, and then on a Q.bo robot
Masri, Ali. « Multi-Network integration for an Intelligent Mobility ». Thesis, Université Paris-Saclay (ComUE), 2017. http://www.theses.fr/2017SACLV091/document.
Texte intégralMultimodality requires the integration of heterogeneous transportation data and services to construct a broad view of the transportation network. Many new transportation services (e.g. ridesharing, car-sharing, bike-sharing) are emerging and gaining a lot of popularity since in some cases they provide better trip solutions.However, these services are still isolated from the existing multimodal solutions and are proposed as alternative plans without being really integrated in the suggested plans. The concept of open data is raising and being adopted by many companies where they publish their data sources to the web in order to gain visibility. The goal of this thesis is to use these data to enable multimodality by constructing an extended transportation network that links these new services to existing ones.The challenges we face mainly arise from the integration problem in both transportation services and transportation data
Heyder, Jakob Wendelin. « Knowledge Base Augmentation from Spreadsheet Data : Combining layout inference with multimodal candidate classification ». Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-278824.
Texte intégralKalkylblad består av ett värdefullt och särskilt stort datasätt av dokument inom många företagsorganisationer och på webben. Även om kalkylblad är intuitivt att använda och är utrustad med kraftfulla funktioner, utvinning och transformation av data är fortfarande en besvärlig och manuell uppgift. Den stora flexibiliteten som de ger användaren resulterar i data som är godtyckligt strukturerade och svåra att bearbeta för andra applikationer. I det här förslaget föreslår vi en ny arkitektur som kombinerar övervakad layoutinferens och multimodal kandidatklassificering för att tillåta kunskapsbasförstärkning från godtyckliga kalkylblad. I vår design överväger vi behovet av att reparera felklassificeringar och möjliggöra verifiering och rangordning av tvetydiga kandidater. Vi utvärderar systemets utförande på två datasätt, en med singeltabellkalkylblad, en annan med kalkylblad av godtyckligt format. Utvärderingsresultatet visar att det föreslagna systemet uppnår liknande prestanda på singel-tabellkalkylblad jämfört med state-of-the-art regelbaserade lösningar. Dessutom tillåter systemets flexibilitet oss att bearbeta godtyckliga kalkylark format, inklusive horisontella och vertikala inriktade tabeller, flera kalkylblad och sammanhangsförande metadata. Detta var inte möjligt med existerande rent textbaserade eller tabellbaserade lösningar. Experimenten visar att det kan uppnå hög effektivitet med en F1-poäng på 95.71 på godtyckliga kalkylblad som kräver tolkning av omgivande metadata. Systemets precision kan ökas ytterligare genom att applicera schema-matchning av kandidater baserat på semantisk likhet mellan kolumnrubriker.
Gómez, Bruballa Raúl Álamo. « Exploiting the Interplay between Visual and Textual Data for Scene Interpretation ». Doctoral thesis, Universitat Autònoma de Barcelona, 2020. http://hdl.handle.net/10803/670533.
Texte intégralLa experimentación en aprendizaje automático en escenarios controlados y con bases de datos estándares es necesaria para comparar el desempeño entre algoritmos evaluándolos en las mismas condiciones. Sin embargo, también en necesaria experimentación en cómo se comportan estos algoritmos cuando son entrenados con datos menos controlados y aplicados a problemas reales para indagar en cómo los avances en investigación pueden contribuir a nuestra sociedad. En esta tesis experimentamos con los algoritmos más recientes de visión por ordenador y procesado del lenguaje natural aplicándolos a la interpretación de escenas multimodales. En particular, investigamos en cómo la interpretación automática de imagen y texto se puede explotar conjuntamente para resolver problemas reales, enfocándonos en aprender de datos de redes sociales. Encaramos diversas tareas que implican información visual y textual, discutimos sus características y retos y exponemos nuestras conclusiones experimentales. Primeramente trabajamos en la detección de texto en imágenes. A continuación, trabajamos con publicaciones de redes sociales, usando las leyendas textuales de imágenes como supervisión para aprender características visuales, que aplicamos a la búsqueda de imágenes semántica con consultas multimodales. Después, trabajamos con imágenes de redes sociales geolocalizadas con etiquetas textuales asociadas, experimentando en cómo usar las etiquetas como supervisión, en búsqueda de imágenes sensible a localización, y en explotar la localización para el etiquetado de imágenes. Finalmente, encaramos un problema de clasificación específico de publicaciones de redes sociales formadas por una imagen y un texto: Clasificación de discurso del odio multimodal.
Machine learning experimentation under controlled scenarios and standard datasets is necessary to compare algorithms performance by evaluating all of them in the same setup. However, experimentation on how those algorithms perform on unconstrained data and applied tasks to solve real world problems is also a must to ascertain how that research can contribute to our society. In this dissertation we experiment with the latest computer vision and natural language processing algorithms applying them to multimodal scene interpretation. Particularly, we research on how image and text understanding can be jointly exploited to address real world problems, focusing on learning from Social Media data. We address several tasks that involve image and textual information, discuss their characteristics and offer our experimentation conclusions. First, we work on detection of scene text in images. Then, we work with Social Media posts, exploiting the captions associated to images as supervision to learn visual features, which we apply to multimodal semantic image retrieval. Subsequently, we work with geolocated Social Media images with associated tags, experimenting on how to use the tags as supervision, on location sensitive image retrieval and on exploiting location information for image tagging. Finally, we work on a specific classification problem of Social Media publications consisting on an image and a text: Multimodal hate speech classification.
Kirchler, Dominik. « Routage efficace sur réseaux de transport multimodaux ». Phd thesis, Ecole Polytechnique X, 2013. http://pastel.archives-ouvertes.fr/pastel-00877450.
Texte intégralSaragiotis, Panagiotis. « Cross-modal classification and retrieval of multimodal data using combinations of neural networks ». Thesis, University of Surrey, 2006. http://epubs.surrey.ac.uk/843338/.
Texte intégralLu, Pascal. « Statistical Learning from Multimodal Genetic and Neuroimaging data for prediction of Alzheimer's Disease ». Electronic Thesis or Diss., Sorbonne université, 2019. http://www.theses.fr/2019SORUS636.
Texte intégralAlzheimer's Disease (AD) is nowadays the main cause of dementia in the world. It provokes memory and behavioural troubles in elderly people. The early diagnosis of Alzheimer's Disease is an active topic of research. Three different types of data play a major role when it comes to its diagnosis: clinical tests, neuroimaging and genetics. The two first data bring informations concerning the patient's current state. On the contrary, genetic data help to identify whether a patient could develop AD in the future. Furthermore, during the past decade, researchers have created longitudinal dataset on A and important advances for processing and analyse of complex and high-dimensional data have been made. The first contribution of this thesis will be to study how to combine different modalities in order to increase their predictive power in the context of classification. We will focus on hierarchical models that capture potential interactions between modalities. Moreover, we will adequately modelled the structure of each modality (genomic structure, spatial structure for brain images), through the use of adapted penalties such as the ridge penalty for images and the group lasso penalty for genetic data. The second contribution of this thesis will be to explore models for predict the conversion date to Alzheimer's Disease for mild cognitive impairment subjects. Such problematic has been enhanced by the TADPOLE challenge. We will use the framework provided by survival analysis. Starting from basic models such as the Cox proportional hasard model, the additive Aalen model, and the log-logistic model, we will develop other survival models for combining different modalities, such as a multilevel log-logistic model or a multilevel Cox model
Diehn, Sabrina Maria [Verfasser]. « Analysis of data from multimodal chemical characterizations of plant tissues / Sabrina Maria Diehn ». Berlin : Humboldt-Universität zu Berlin, 2021. http://d-nb.info/1238074006/34.
Texte intégral