Dissertations / Theses on the topic 'Raffinement des graphes de connaissances'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Raffinement des graphes de connaissances.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Khajeh, Nassiri Armita. "Expressive Rule Discovery for Knowledge Graph Refinement." Electronic Thesis or Diss., université Paris-Saclay, 2023. http://www.theses.fr/2023UPASG045.
Full textKnowledge graphs (KGs) are heterogeneous graph structures representing facts in a machine-readable format. They find applications in tasks such as question answering, disambiguation, and entity linking. However, KGs are inherently incomplete, and refining them is crucial to improve their effectiveness in downstream tasks. It's possible to complete the KGs by predicting missing links within a knowledge graph or integrating external sources and KGs. By extracting rules from the KG, we can leverage them to complete the graph while providing explainability. Various approaches have been proposed to mine rules efficiently. Yet, the literature lacks effective methods for effectively incorporating numerical predicates in rules. To address this gap, we propose REGNUM, which mines numerical rules with interval constraints. REGNUM builds upon the rules generated by an existing rule mining system and enriches them by incorporating numerical predicates guided by quality measures. Additionally, the interconnected nature of web data offers significant potential for completing and refining KGs, for instance, by data linking, which is the task of finding sameAs links between entities of different KGs. We introduce RE-miner, an approach that mines referring expressions (REs) for a class in a knowledge graph and uses them for data linking. REs are rules that are only applied to one entity. They support knowledge discovery and serve as an explainable way to link data. We employ pruning strategies to explore the search space efficiently, and we define characteristics to generate REs that are more relevant for data linking. Furthermore, we aim to explore the advantages and opportunities of fine-tuning language models to bridge the gap between KGs and textual data. We propose GilBERT, which leverages fine-tuning techniques on language models like BERT using a triplet loss. GilBERT demonstrates promising results for refinement tasks of relation prediction and triple classification tasks. By considering these challenges and proposing novel approaches, this thesis contributes to KG refinement, particularly emphasizing explainability and knowledge discovery. The outcomes of this research open doors to more research questions and pave the way for advancing towards more accurate and comprehensive KGs
Nedjari, Tayeb. "Réseaux de neurones artificiels et connaissances symboliques : insertion, raffinement et extraction." Paris 13, 1998. http://www.theses.fr/1998PA132024.
Full textRaad, Joe. "Gestion d'identité dans des graphes de connaissances." Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLA028/document.
Full textIn the absence of a central naming authority on the Web of data, it is common for different knowledge graphs to refer to the same thing by different names (IRIs). Whenever multiple names are used to denote the same thing, owl:sameAs statements are needed in order to link the data and foster reuse. Such identity statements have strict logical semantics, indicating that every property asserted to one name, will also be inferred to the other, and vice versa. While such inferences can be extremely useful in enabling and enhancing knowledge-based systems such as search engines and recommendation systems, incorrect use of identity can have wide-ranging effects in a global knowledge space like the Web of data. With several studies showing that owl:sameAs is indeed misused for different reasons, a proper approach towards the handling of identity links is required in order to make the Web of data succeed as an integrated knowledge space. This thesis investigates the identity problem at hand, and provides different, yet complementary solutions. Firstly, it presents the largest dataset of identity statements that has been gathered from the LOD Cloud to date, and a web service from which the data and its equivalence closure can be queried. Such resource has both practical impacts (it helps data users and providers to find different names for the same entity), as well as analytical value (it reveals important aspects of the connectivity of the LOD Cloud). In addition, by relying on this collection of 558 million identity statements, we show how network metrics such as the community structure of the owl:sameAs graph can be used in order to detect possibly erroneous identity assertions. For this, we assign an error degree for each owl:sameAs based on the density of the community(ies) in which they occur, and their symmetrical characteristics. One benefit of this approach is that it does not rely on any additional knowledge. Finally, as a way to limit the excessive and incorrect use of owl:sameAs, we define a new relation for asserting the identity of two ontology instances in a specific context (a sub-ontology). This identity relation is accompanied with an approach for automatically detecting these links, with the ability of using certain expert constraints for filtering irrelevant contexts. As a first experiment, the detection and exploitation of the detected contextual identity links are conducted on two knowledge graphs for life sciences, constructed in a mutual effort with domain experts from the French National Institute of Agricultural Research (INRA)
Saidouni, Djamel-Eddine. "Sémantique de maximalité : application au raffinement d'actions dans LOTOS." Toulouse 3, 1996. http://www.theses.fr/1996TOU30040.
Full textKerdiles, Gwenael Nang. "Logiques de graphes conceptuels." Montpellier 2, 2001. http://www.theses.fr/2001MON20185.
Full textGandon, Fabien. "Graphes RDF et leur Manipulation pour la Gestion de Connaissances." Habilitation à diriger des recherches, Université de Nice Sophia-Antipolis, 2008. http://tel.archives-ouvertes.fr/tel-00351772.
Full textDans le deuxième chapitre, nous rappelons comment les formalismes à base de graphes peuvent être utilisés pour représenter des connaissances avec un degré variable de formalisation en fonction des besoins identifiés dans les scénarios d'application et des traitements à effectuer notamment pour la mise en place de webs sémantiques. Nous identifierons brièvement les caractéristiques de certains de ces formalismes qui sont utilisés dans nos travaux et les opportunités d'extensions qu'ils offrent. Nous synthétiserons aussi une initiative en cours pour factoriser la définition des structures mathématiques partagées par ces formalismes et réutiliser l'algorithmique des traitements communs à ces structures.
Dans le troisième chapitre nous expliquons que l'ontologie offre un support à d'autres types de raisonnement que la dérivation logique. Par exemple, la hiérarchie de notions contenue dans une ontologie peut être vue comme un espace métrique permettant de définir des distances pour comparer la proximité sémantique de deux notions. Nous avons mis en œuvre cette idée dans plusieurs scénarios comme l'allocation distribuée d'annotations, la recherche approchée ou le clustering. Nous résumons dans ce troisième chapitre diverses utilisations que nous avons faites des distances sémantiques et discutons notre position sur ce domaine. Nous donnons les scénarios d'utilisation et les distances utilisées dans un échantillon représentatif de projets que nous avons menés. Pour nous, cette première série d'expériences a permis de démontrer l'intérêt et le potentiel des distances, et aussi de souligner l'importance du travail restant à faire pour identifier et caractériser les familles de distances existantes et leur adéquation respective aux tâches pour lesquelles nos utilisateurs souhaitent être assistés.
Dans le quatrième chapitre, nous rappelons qu'un web sémantique, tel que nous en utilisons dans nos scénarios, qu'il soit public ou sur l'intranet d'une entreprise, repose généralement sur plusieurs serveurs web qui proposent chacun différentes ontologies et différentes bases d'annotations utilisant ces ontologies pour décrire des ressources. Les scénarios d'usage amènent souvent un utilisateur à formuler des requêtes dont les réponses combinent des éléments d'annotation distribués entre plusieurs de ces serveurs.
Ceci demande alors d'être capable :
(1) d'identifier les serveurs susceptibles d'avoir des éléments de réponse ;
(2) d'interroger des serveurs distants sur les éléments qu'ils connaissent sans surcharger le réseau;
(3) de décomposer la requête et router les sous-requêtes vers les serveurs idoines ;
(4) de recomposer les résultats à partir des réponses partielles.
Nous avons, avec le web sémantique, les briques de base d'une architecture distribuée. Le quatrième chapitre résume un certain nombre d'approches que nous avons proposées pour tenir compte de la distribution et gérer des ressources distribuées dans les webs sémantiques que nous concevons.
Les ontologies et les représentations de connaissances sont souvent dans le cœur technique de nos architectures, notamment lorsqu'elles utilisent des représentations formelles. Pour interagir avec le web sémantique et ses applications, le cinquième chapitre rappelle que nous avons besoin d'interfaces qui les rendent intelligibles pour les utilisateurs finaux. Dans nos systèmes d'inférences des éléments de connaissances sont manipulés et combinés, et même si les éléments de départ étaient intelligibles, l'intelligibilité des résultats, elle, n'est pas préservée par ces transformations.
Actuellement, et dans le meilleur des cas, les concepteurs d'interfaces mettent en œuvre des transformations ad hoc des structures de données internes en représentations d'interface en oubliant souvent les capacités de raisonnement que pourraient fournir ces représentations pour construire de telles interfaces. Dans le pire des cas, et encore trop souvent, les structures de représentation normalement internes sont directement mises à nu dans des widgets sans que cela soit justifié et, au lieu d'assister l'interaction, ces représentations alourdissent les interfaces.
Puisqu'elles reçoivent les contributions d'un monde ouvert, les interfaces du web sémantique devront être, au moins en partie, générées dynamiquement et rendues pour chaque structure devant rentrer en contact avec les utilisateurs. Le cinquième et dernier chapitre souligne cette opportunité croissante d'utiliser des systèmes à base d'ontologies dans l'assistance aux interactions avec nos utilisateurs.
Trouillon, Théo. "Modèles d'embeddings à valeurs complexes pour les graphes de connaissances." Thesis, Université Grenoble Alpes (ComUE), 2017. http://www.theses.fr/2017GREAM048/document.
Full textThe explosion of widely available relational datain the form of knowledge graphsenabled many applications, including automated personalagents, recommender systems and enhanced web search results.The very large size and notorious incompleteness of these data basescalls for automatic knowledge graph completion methods to make these applicationsviable. Knowledge graph completion, also known as link-prediction,deals with automatically understandingthe structure of large knowledge graphs---labeled directed graphs---topredict missing entries---labeled edges. An increasinglypopular approach consists in representing knowledge graphs as third-order tensors,and using tensor factorization methods to predict their missing entries.State-of-the-art factorization models propose different trade-offs between modelingexpressiveness, and time and space complexity. We introduce a newmodel, ComplEx---for Complex Embeddings---to reconcile both expressivenessand complexity through the use of complex-valued factorization, and exploreits link with unitary diagonalization.We corroborate our approach theoretically and show that all possibleknowledge graphs can be exactly decomposed by the proposed model.Our approach based on complex embeddings is arguably simple,as it only involves a complex-valued trilinear product,whereas other methods resort to more and more complicated compositionfunctions to increase their expressiveness. The proposed ComplEx model isscalable to large data sets as it remains linear in both space and time, whileconsistently outperforming alternative approaches on standardlink-prediction benchmarks. We also demonstrateits ability to learn useful vectorial representations for other tasks,by enhancing word embeddings that improve performanceson the natural language problem of entailment recognitionbetween pair of sentences.In the last part of this thesis, we explore factorization models abilityto learn relational patterns from observed data.By their vectorial nature, it is not only hard to interpretwhy this class of models works so well,but also to understand where they fail andhow they might be improved. We conduct an experimentalsurvey of state-of-the-art models, not towardsa purely comparative end, but as a means to get insightabout their inductive abilities.To assess the strengths and weaknesses of each model, we create simple tasksthat exhibit first, atomic properties of knowledge graph relations,and then, common inter-relational inference through synthetic genealogies.Based on these experimental results, we propose new researchdirections to improve on existing models, including ComplEx
Archer, Vincent. "Graphes linguistiques multiniveau pour l'extraction de connaissances : l'exemple des collocations." Phd thesis, Université Joseph Fourier (Grenoble ; 1971-2015), 2009. http://www.theses.fr/2009GRE10127.
Full textIn order to model at best linguistic phenomena, natural language processing systems need to have quality ressources at their disposal, yet existing ressources are most often incomplete and do not allow to treat data in an adequate manner in process like translation, analysis, etc. This thesis is about acquisition of linguistic knowledge, and more precisely about the extraction of that knowledge from corpora where it appears. We study especially the problem of the collocations, these couple of terms where one term is chosen in function of the other one to express a particular meaning (as « driving rain », where « driving » is used to express the intensification). To allow large-scale data acquisition, it is necessary to make it easy to realize in an automatic manner, and simple to configure by linguists with limited knowledge in computer programming. For that reason, we have to rely on a precise and suitable model for data and process. We describe MuLLinG, the multilevel linguistic graph we realized, where each level represents information in a different manner, and operations for the manipulation of these graphs. That model, based on a simple structure (the graph one), allows to represent, treat, and manage diverse kinds of ressources. Indeed, associated operations were written in order to be as most generic as possible, which means that they are independent of what nodes and edges represents, and of the task to realize. That enables our model, which has been implemented and used for several experiments, some concerning collocation extraction, to view a process (sometimes complex) of linguistic knowledge extraction, as a succession of small graph manipulation operations
Naudin, Antoine. "Impact des connaissances initiales sur la calculabilité distribuée." Thesis, Aix-Marseille, 2017. http://www.theses.fr/2017AIXM0259/document.
Full textIn this study, we show how knowledge impacts the computability in distributed systems. First, we characterize what we need to know to elect in the unknown participant model. This model is a natural extension for the message passing model that formalises dynamicity that occurs in some networks. We give a necessary and sufficient condition on the knowledge needed to solve the following fundamental problems : map construction, leader election and k-leader election. For each of them, we provide an algorithm solving the problem using any knowledge satisfying our condition. Then, we extend the model to anonymous networks. We characterize, with the same methodology, the knowledge needed to solve an election in this model and we provide an algorithm using such a knowledge and a bound on the size of the network. In the second part, we study the impact of local knowledge on the computability of the anonymous graph exploration problem. We introduce a new model of mobile agents where an agent is endowed with binoculars, a local sensor permitting to perceive the graph induced by the vertices adjacent to its location. In this model, we characterize the graphs that can be explored by a single mobile agent without any global information and we provide an algorithm exploring all of them. Unfortunately, universal algorithm has a cost : The number of moves required by such an algorithm cannot be bounded by a computable function. Finally, we prove that large classes of graphs like chordal graphs, Johnson graphs, . . . can be explored in a linear number of moves using binoculars by providing an exploration algorithm for the family of Weetman graphs
Egyed-Zsigmond, Elöd. "Gestion des connaissances dans une base de documents multimédias." Lyon, INSA, 2003. http://theses.insa-lyon.fr/publication/2003ISAL0055/these.pdf.
Full textOur research has been carried out as a part of the RNRT project RECIS in collaboration with FT R&D and the INRIA. The goal of this project is to study different techniques to enrich multimedia content access services. We created an audiovisual annotation model : E-SIA. This model represents documents fragments, annotations as well as annotation structuring elements as graph. The model enables to document and guide annotation providing in meantime a large degree of freedom to users. We have developed a client-server prototype application enabling manual and automatic video annotation as well as the navigation an search in these annotations. We have extended our annotation model in order to create an experience tracing, capitalization and reuse framework: Trèfle (Club). This model enables to trace in a contextual manner the use of a computer program following a use and an observation model in a graph. We consider that in a computer program users manipulate objects using procedures
Egyed-Zsigmond, Elod Pinon Jean-Marie Mille Alain. "Gestion des connaissances dans une base de documents multimédias." Villeurbanne : Doc'INSA, 2005. http://docinsa.insa-lyon.fr/these/pont.php?id=egyed_zsigmond.
Full textGazzotti, Raphaël. "Prédiction d’hospitalisation par la génération de caractéristiques extraites de graphes de connaissances." Thesis, Université Côte d'Azur, 2020. http://www.theses.fr/2020COAZ4018.
Full textThe use of electronic medical records (EMRs) and electronic prescribing are priorities in the various European action plans on connected health. The development of the EMR is a tremendous source of data; it captures all symptomatic episodes in a patient’s life and should lead to improved medical and care practices, as long as automatic treatment procedures are set up.As such, we are working on hospitalization prediction based on EMRs and after having represented them in vector form, we enrich these models in order to benefit from the knowledge resulting from referentials, whether generalist or specific in the medical field, in order to improve the predictive power of automatic classification algorithms. Determining the knowledge to be extracted with the objective of integrating it into vector representations is both a subjective task and intended for experts, we will see a semi-supervised procedure to partially automate this process.As a result of our research, we designed a product for general practitioners to prevent their patients from being hospitalized or at least improve their health. Thus, through a simulation, it will be possible for the doctor to evaluate the factors involved on the risk of hospitalization of his patient and to define the preventive actions to be planned to avoid the occurrence of this event.This decision support algorithm is intended to be directly integrated into the physician consultation software. For this purpose, we have developed in collaboration with many professional bodies, including the first to be concerned, general practitioners
Faci, Adam. "Représentation, simulation et exploitation de connaissances dans le formalisme des graphes conceptuels." Electronic Thesis or Diss., Sorbonne université, 2022. http://www.theses.fr/2022SORUS419.
Full textThis thesis addresses the field of knowledge representation in conceptual graphs, a structured knowledge representation formalism that provides efficient manipulation tools. There are many challenges in symbolic artificial intelligence concerning the representation of knowledge in general, and in particular the concise representation of a large amount of information. Conceptual graphs, by their ability to visually represent different kinds of knowledge, mainly ontological and factual, and by offering efficient manipulation tools, provide a perfect framework for the answer to these problems. We conduct a comparative study of fuzzy extensions of conceptual graphs, then we propose an algorithm for conceptual graphs simulation as well as an efficient algorithm for extracting frequent patterns that are not redundant with ontological knowledge
Leclère, Michel. "Les connaissances du niveau terminologique du modèle des graphes conceptuels : construction et exploitation." Montpellier 2, 1995. http://www.theses.fr/1995MON20257.
Full textRaimbault, Thomas. "Transition de modèles de connaissances : un système de connaissance fondé sur OWL, graphes conceptuels et UML." Phd thesis, Nantes, 2008. https://archive.bu.univ-nantes.fr/pollux/show/show?id=4ef8d797-9884-4506-8973-e5bc095e2459.
Full textThe purpose of this thesis is using multiple knowledge models for representating knowledge and reasoning on this represented knowledge. This thesis proposes transitions between several knowledge models\string: OWL, Conceptual Graphs and UML. The originality of this thesis lies both in the centralized modeling knowledge within a knowledge system and in action to pass this knowledge from one model to another of system as needs required by modeling and reasoning. The main goal of these transitions knowledge models is twofold. On the one hand, it is to benefit from relatively easy using expressiveness of models to obtain as result a strong expressive power. On the other hand, it helps in the design and operation of a modeling, using best knowed models or best suited models. The tools of each model can then be used on represented knowledge, providing complementary use of these models
Guinaldo, Olivier. "Etude d'un gestionnaire d'ensembles de graphes conceptuels." Montpellier 2, 1996. http://www.theses.fr/1996MON20189.
Full textCarbonneill, Boris. "Vers un système de représentation de connaissances et de raisonnement fondé sur les graphes conceptuels." Montpellier 2, 1996. http://www.theses.fr/1996MON20015.
Full textDe, Oliveira Joffrey. "Gestion de graphes de connaissances dans l'informatique en périphérie : gestion de flux, autonomie et adaptabilité." Electronic Thesis or Diss., Université Gustave Eiffel, 2023. http://www.theses.fr/2023UEFL2069.
Full textThe research work carried out as part of this PhD thesis lies at the interface between the Semantic Web, databases and edge computing. Indeed, our objective is to design, develop and evaluate a database management system (DBMS) based on the W3C Resource Description Framework (RDF) data model, which must be adapted to the terminals found in Edge computing.The possible applications of such a system are numerous and cover a wide range of sectors such as industry, finance and medicine, to name but a few. As proof of this, the subject of this thesis was defined with the team from the Computer Science and Artificial Intelligence Laboratory (CSAI) at ENGIE Lab CRIGEN. The latter is ENGIE's research and development centre dedicated to green gases (hydrogen, biogas and liquefied gases), new uses of energy in cities and buildings, industry and emerging technologies (digital and artificial intelligence, drones and robots, nanotechnologies and sensors). CSAI financed this thesis as part of a CIFRE-type collaboration.The functionalities of a system satisfying these characteristics must enable anomalies and exceptional situations to be detected in a relevant and effective way from measurements taken by sensors and/or actuators. In an industrial context, this could mean detecting excessively high measurements, for example of pressure or flow rate in a gas distribution network, which could potentially compromise infrastructure or even the safety of individuals. This detection must be carried out using a user-friendly approach to enable as many users as possible, including non-programmers, to describe risk situations. The approach must therefore be declarative, not procedural, and must be based on a query language, such as SPARQL.We believe that Semantic Web technologies can make a major contribution in this context. Indeed, the ability to infer implicit consequences from explicit data and knowledge is a means of creating new services that are distinguished by their ability to adjust to the circumstances encountered and to make autonomous decisions. This can be achieved by generating new queries in certain alarming situations, or by defining a minimal sub-graph of knowledge that an instance of our DBMS needs in order to respond to all of its queries.The design of such a DBMS must also take into account the inherent constraints of Edge computing, i.e. the limits in terms of computing capacity, storage, bandwidth and sometimes energy (when the terminal is powered by a solar panel or a battery). Architectural and technological choices must therefore be made to meet these limitations. With regard to the representation of data and knowledge, our design choice fell on succinct data structures (SDS), which offer, among other advantages, the fact that they are very compact and do not require decompression during querying. Similarly, it was necessary to integrate data flow management within our DBMS, for example with support for windowing in continuous SPARQL queries, and for the various services supported by our system. Finally, as anomaly detection is an area where knowledge can evolve, we have integrated support for modifications to the knowledge graphs stored on the client instances of our DBMS. This support translates into an extension of certain SDS structures used in our prototype
Truong, Quoc Dinh. "Approches par les graphes pour l'aide à la sélection, l'analyse et la visualisation des connaissances." Toulouse 2, 2008. http://www.theses.fr/2008TOU20053.
Full textKnowledges discovery in databases is a decision support process for which useful data is retrieved from available data before being transformed to information from which useful knowledge will be extracted. However, the intelligibility and the usability of obtained knowledges depend crucially on the way we represent these knowledges. The research works we present deal with this important issue and use graph theory as a mean to achieve good knowledge presentation along with problem modeling and solving. This work deals with information retrieval for raw data selection and filtering, information analysis and knowledge visualization. Knowledge visualization is intended not only to help to represent the obtained knowledge but also to retrieve this knowledge
Raimbault, Thomas. "Transition de modèles de connaissances - Un système de connaissance fondé sur OWL, Graphes conceptuels et UML." Phd thesis, Université de Nantes, 2008. http://tel.archives-ouvertes.fr/tel-00482664.
Full textSimonet, Geneviève. "Héritage non monotone à base de chemins et de graphes partiels." Montpellier 2, 1994. http://www.theses.fr/1994MON20151.
Full textPinaud, Bruno. "Contribution à la visualisation des connaissances par des graphes dans une mémoire d'entreprise : application sur le serveur Atanor." Phd thesis, Université de Nantes, 2006. http://tel.archives-ouvertes.fr/tel-00335934.
Full textLe passage au modèle de graphes pose le problème de sa représentation visuelle. Les tracés doivent rester lisibles et compréhensibles par les utilisateurs. Ceci se traduit notamment par le respect de critères esthétiques qui permettent de modéliser un problème d'optimisation combinatoire consistant à trouver un ordre optimal des sommets dans chaque niveau. Pour résoudre ce problème, nous avons développé un algorithme génétique qui possède deux particularités : deux opérateurs de croisements spécifiques et une hybridation par une recherche locale. Les expérimentations montrent que pour des graphes de taille standard, l'algorithme génétique donne de meilleurs résultats que les autres méthodes que nous connaissons. La comparaison des modèles de représentation des connaissances sur un exemple industriel montre qu'en plus de faciliter la lecture, Graph'Atanor permet de facilement suivre la trace des utilisateurs et de mettre en avant les sommets critiques.
Auillans, Pascal. "Modélisation de réseaux sémantiques par des hypergraphes et applications." Bordeaux 1, 2005. http://www.theses.fr/2005BOR12966.
Full textThe goal of the Web evolutions planned by the W3C is to enable the improvement of web services quality. To this end, W3C has added to the Web architecture a knowledge management system named Semantic Web, which implements a theoretical model relying on descriptive logic. This thesis, of which researches are more specifically applied to another knowledge representation system named Topic Maps, aims to rovide an alternative to the use of descriptive logic. We will show how graph theory can be used to structure the knowledge, hence benefit to the field of knowledge representation. This thesis initialy stands within the european project KePT, which aimed to implement a visualization interface for knowledge, structured according to the norm ISO 13250 Topic Maps, in Mondeca's ITM application. Research on graph clustering made for this project raised the need of both a better understanding of the topic maps structure, and tools that enable implementation of efficient processing. Therefore, we propose a formal model relying on graph theory that enables to express structural properties, beyond the expressive power of first order logic. Our model is not only suited for theoretical studies, but also for the adaptation of fast graph theory algorithms to knowledge processing. These processing were previously hardly implementable in industrial applications
Jacquelinet, Christian. "Modélisation du langage naturel et représentation des connaissances par graphes conceptuels : applications au domaine de l'information médicale." Rennes 1, 2002. http://www.theses.fr/2002REN1B060.
Full textAyats, H. Ambre. "Construction de graphes de connaissances à partir de textes avec une intelligence artificielle explicable et centrée-utilisateur·ice." Electronic Thesis or Diss., Université de Rennes (2023-....), 2023. http://www.theses.fr/2023URENS095.
Full textWith recent advances in artificial intelligence, the question of human control has become central. Today, this involves both research into explainability and designs centered around interaction with the user. What's more, with the expansion of the semantic web and automatic natural language processing methods, the task of constructing knowledge graphs from texts has become an important issue. This thesis presents a user-centered system for the construction of knowledge graphs from texts. This thesis presents several contributions. First, we introduce a user-centered workflow for the aforementioned task, having the property of progressively automating the user's actions while leaving them a fine-grained control over the outcome. Next, we present our contributions in the field of formal concept analysis, used to design an explainable instance-based learning module for relation classification. Finally, we present our contributions in the field of relation extraction, and how these fit into the presented workflow
Karouach, Saïd. "Visualisations interactives pour la découverte de connaissances, concepts, méthodes et outils." Toulouse 3, 2003. http://www.theses.fr/2003TOU30082.
Full textDoumbouya, Mamadou Bilo. "Système d'argumentation pour la collaboration en télémédecine." Thesis, Toulouse 1, 2016. http://www.theses.fr/2016TOU10030/document.
Full textTelemedicine involves the practice of medical procedures remotely through the use of new information and communications technology. Among these medical procedures, we looked at the tele-expertise which is a kind of collaborative activity consisting of collecting the opinions of medical experts facing a particular health problem. In our work, we have chosen to model these collaborative activities by Dung argumentation system based on mathematical foundations and illustrates the interactions between the different stakeholders and at the same time provides mathematical tools decisions. We opted for a semantic modeling with conceptual graphs as one of our objectives is to ensure semantic interoperability. This modeling can often include inconsistencies (poor relations of attacks in argumentation system) which will be verified by the use of constraints in conceptual graphs. To solve these inconsistency problems, two major solutions have been proposed : (i) the weight of the arguments of different health professionals, (ii) modeling some aspects of medical law as constraints. This work demonstrates a computer application of logical reasoning in a judicial medical setting where it sheds light on the verification of information, argumentation and interaction. It aims to ensure good cooperation in order to guard against possible financial and legal consequences
Martin, Philippe. "Exploitation de graphes conceptuels et de documents structurés et hypertextes pour l'acquisition de connaissances et la recherche d'informations." Nice, 1996. http://www.theses.fr/1996NICE4992.
Full textTognetti, Yannick. "Contribution à la modélisation des systèmes d'information chimique par la théorie et l'algorithmique de graphes." Montpellier 2, 2002. http://www.theses.fr/2002MON20182.
Full textBuron, Maxime. "Raisonnement efficace sur des grands graphes hétérogènes." Thesis, Institut polytechnique de Paris, 2020. http://www.theses.fr/2020IPPAX061.
Full textThe Semantic Web offers knowledge representations, which allow to integrate heterogeneous data from several sources into a unified knowledge base. In this thesis, we investigate techniques for querying such knowledge bases.The first part is devoted to query answering techniques on a knowledge base, represented by an RDF graph subject to ontological constraints. Implicit information entailed by the reasoning, enabled by the set of RDFS entailment rules, has to be taken into account to correctly answer such queries. First, we present a sound and complete query reformulation algorithm for Basic Graph Pattern queries, which exploits a partition of RDFS entailment rules into assertion and constraint rules. Second, we introduce a novel RDF storage layout, which combines two well-known layouts. For both contributions, our experiments assess our theoretical and algorithmic results.The second part considers the issue of querying heterogeneous data sources integrated into an RDF graph, using BGP queries. Following the Ontology-Based Data Access paradigm, we introduce a framework of data integration under an RDFS ontology, using the Global-Local-As-View mappings, rarely considered in the literature.We present several query answering strategies, which may materialize the integrated RDF graph or leave it virtual, and differ on how and when RDFS reasoning is handled. We implement these strategies in a platform, in order to conduct experiments, which demonstrate the particular interest of one of the strategies based on mapping saturation. Finally, we show that mapping saturation can be extended to reasoning defined by a subset of existential rules
Pennerath, Frédéric. "Méthodes d'extraction de connaissances à partir de données modélisables par des graphes : Application à des problèmes de synthèse organique." Phd thesis, Université Henri Poincaré - Nancy I, 2009. http://tel.archives-ouvertes.fr/tel-00436568.
Full textPennerath, Frédéric Napoli Amedeo. "Méthodes d'extraction de connaissances à partir de données modélisables par des graphes Application à des problèmes de synthèse organique /." S. l. : S. n, 2009. http://www.scd.uhp-nancy.fr/docnum/SCD_T_2009_0061_PENNERATH.pdf.
Full textLoudcher, Rabaseda Sabine. "Contributions à l'extraction automatique de connaissances : application à l'analyse clinique de la marche." Lyon 1, 1996. http://www.theses.fr/1996LYO10309.
Full textQuiniou, Solen. "Intégration de connaissances linguistiques pour la reconnaissance de textes manuscrits en-ligne." Phd thesis, Rennes, INSA, 2007. ftp://ftp.irisa.fr/techreports/theses/2007/quiniou.pdf.
Full textThe aim of this thesis is to build a sentence regognition system based on an existing word regognition system. Two research axes are considered: the sentence segmentation int words as well as the integration of linguistic knowledge to take into account the context of the sentences. We studied several types of statistic language models by comparing their respective impact on the recognition system performances. We also tried to find the best strategy to introduce them efficiently into the whole recognition system. One of the originality of this study is the integration of a representation of the different sentence hypotheses in the form of a confusion network; which is then used to detect and correct the remaining regognition errors. Using the aforementioned techniques allowed us to considerably reduce the number of recognition errors among the words of the sentences
Carloni, Olivier. "Introduction de raisonnement dans un outil industriel de gestion des connaissances." Phd thesis, Université Montpellier II - Sciences et Techniques du Languedoc, 2008. http://tel.archives-ouvertes.fr/tel-00387017.
Full textLi, Jinpeng. "Extraction de connaissances symboliques et relationnelles appliquée aux tracés manuscrits structurés en-ligne." Phd thesis, Nantes, 2012. http://tel.archives-ouvertes.fr/tel-00785984.
Full textTalon, Bénédicte. "Un système d'aide à l'acquisition de concepts nouveaux pour un outil d'analyse du langage naturel." Compiègne, 1991. http://www.theses.fr/1991COMPD378.
Full textCaillaut, Gaëtan. "Apprentissage d'espaces prétopologiques pour l'extraction de connaissances structurées." Electronic Thesis or Diss., Orléans, 2019. http://www.theses.fr/2019ORLE3208.
Full textPretopology is a mathematical theory whose goal is to relax the set of axioms governing the well known topology theory. Weakening the set of axioms mainly consists in redefining the pseudo-closure operator which is idempotent in topology. The non-idempotence of the pretopological pseudo-closure operator offers an appropriate framework for the modeling of various phenomena, such as iterative processes evolving throughout time. Pretopology is the outcome of the generalisation of several concepts, amongst topology but also graph theory. This thesis is divided in four main parts. The first one is an introduction to the theoretical framework of the pretopology, as well as an overview of several applications in domains where the pretopology theory shines, such as machine learning, image processing or complex systems analysis.The second part will settle the logical modeling of pretopological spaces which allows to define pretopological spaces by a logical and multi-criteria combination. This modeling enables learning algorithms to define pretopological spaces by learning a logical formula. This part will also present an unrestricted pretopological spaces learning algorithm. Unrestricted pretopological spaces can be quite hard to manipulate, especially when the studied population has some structural properties that can be described in a more restricted space. This is why the third part is dedicated to the automatic learning of pretopological spaces of type V. These spaces are defined by a set of prefilters which impose a particular structure. The LPSMI algorithm, which is the main contribution of this work, is presented in this part. This algorithm relies on the multi-instance learning principles to accurately capture the structural properties of pretopological spaces of type V. Finally, the last part consists of multiple applications of the theoretical framework presented in this thesis. Applications to lexical taxonomies extraction, community detection and extraction of temporal relations, as part of a NLP process, will be presented in order to show the usefulness, the relevance and the flexibility of pretopological spaces learning
Quiniou, Solen. "Intégration de connaissances linguistiques pour la reconnaissance de textes manuscrits en-ligne." Phd thesis, INSA de Rennes, 2007. http://tel.archives-ouvertes.fr/tel-00580623.
Full textKherroubi, Souad. "Un cadre formel pour l'intégration de connaissances du domaine dans la conception des systèmes : application au formalisme Event-B." Thesis, Université de Lorraine, 2018. http://www.theses.fr/2018LORR0230/document.
Full textThis thesis aims at defining techniques to better exploit the knowledge provided from the domain in order to account for the reality of systems described as complex and critical. Modeling is an essential step in performing verifications and expressing properties that a system must satisfy according to the needs and requirements established in the specifications. Modeling is a representation that simplifies the reality of a system. However, a complex system can not be reduced to a model. A model that represents a system must always fit into its observational theory to account for any anomalies that it may contain. Our study clearly shows that the context is the first issue to deal with as the main source of conflict in the design process of a system. The approach adopted in this thesis is that of integrating knowledge of the domain by associating the system to design with declarative formalisms qualified of descriptive ones that we call ontologies. We pay a particular attention to the Event-B formalism, whose correct-by-construction approach called refinement is the main mechanism at the heart of this formalism, which makes it possible to make proofs on abstract representations of systems for expressing and verifying properties of safety and invariance. The first problem treated is the representation and modeling of contextual knowledge in V&V of models. Following to the study looked at the different sources of conflict, we established new definitions and rules for a refinement context knowledge extraction for Event-B V&V. A study of logical formalisms that represent and interpret the context allowed us to define a new mechanism for better structuring Event-B models. A second study concerns the contribution that domain knowledge can make to the V&V of models. We define a logic for the Event-B formalism with domain constraints based on the description logic, and we define rules to integrate domain knowledge for model V&V. The evaluation of the proposals made deal with very complex case studies such as voting systems whose design patterns are also developed in this thesis. We raise fundamental issues about the complementarity that the integration of domain knowledge can bring to Event-B models by refinement using ontological reasoning, and we propose to define a new structures for a partially automated extraction on both levels, namely the V&V
Oshurko, Ievgeniia. "Knowledge representation and curation in hierarchies of graphs." Thesis, Lyon, 2020. http://www.theses.fr/2020LYSEN024.
Full textThe task of automatically extracting insights or building computational models fromknowledge on complex systems greatly relies on the choice of appropriate representation.This work makes an effort towards building a framework suitable for representation offragmented knowledge on complex systems and its semi-automated curation---continuouscollation, integration, annotation and revision.We propose a knowledge representation system based on hierarchies of graphs relatedwith graph homomorphisms. Individual graphs situated in such hierarchies representdistinct fragments of knowledge and the homomorphisms allow relating these fragments.Their graphical structure can be used efficiently to express entities and their relations. Wefocus on the design of mathematical mechanisms, based on algebraic approaches to graphrewriting, for transformation of individual graphs in hierarchies that maintain consistentrelations between them. Such mechanisms provide a transparent audit trail, as well as aninfrastructure for maintaining multiple versions of knowledge.We describe how the developed theory can be used for building schema-aware graphdatabases that provide schema-data co-evolution capabilities. The proposed knowledgerepresentation framework is used to build the KAMI (Knowledge Aggregation and ModelInstantiation) framework for curation of cellular signalling knowledge. The frameworkallows for semi-automated aggregation of individual facts on protein-protein interactionsinto knowledge corpora, reuse of this knowledge for instantiation of signalling models indifferent cellular contexts and generation of executable rule-based models
Chikhi, Nacim Fateh. "Calcul de centralité et identification de structures de communautés dans les graphes de documents." Phd thesis, Université Paul Sabatier - Toulouse III, 2010. http://tel.archives-ouvertes.fr/tel-00619177.
Full textBelhadj, Ihssen. "Ingénierie des connaissances pour l’épidémiologie et l’aide à la décision en santé publique : Analyse des besoins potentiels et expérimentations dans le contexte du registre français des maladies rénales." Thesis, Paris 13, 2014. http://www.theses.fr/2014PA132068.
Full textExpressing terms referring to pathological conceptualization is an important issue toward the development of clinical research and public health decision support systems. From the context of the French Registry of End Stage Renal Disease, requirements for disease terms representation are anlysed highlighting the acute and hidden problem of statistical continuity in disease data and knowledge representation. The underpinned assumption relies on the idea of ensuring terminological continuity through agenerative method of building Ontology Based Terminological systems. Rather than looking at all the terms that are necessary to describe a domain, we focused solely on the modeling of basic and definitional knowledge about disease. A set ontological rules for diseases hierachies were defined. Eperiments have been designed and implemented taking advantage of GC formalism and a logic programming toll called prolog-GC. The results confimed that such method allow performing two major activities that are carried out in the conventional building process of medical terminologies : refinement of disease terms granularity and consistency improvement. Terminological continuity needs to be considered as major criteria in disease terminological building. Generative approaches helps to improve the terminological continuity as imposes to create news terms of the bases of existing ones formal definitions
Thierry, Eric. "Sur quelques interactions entre structures de données et algorithmes efficaces." Montpellier 2, 2001. http://www.theses.fr/2001MON20079.
Full textBugeaud, Florie. "ISamsara : pour une ingénierie des systèmes de services à base de méréologie et d'hypergraphes." Troyes, 2011. http://www.theses.fr/2011TROY0005.
Full textThe « service orientation » and its challenges are now well known within companies. How-ever, the analysis of the activities of services designers at a telecommunications operator, as well as the state of the art about the service theories and models show the still ill-specified nature of that term. Our work is anchored in the discipline of computer science, but it pursues the discussion of the SSME multidisciplinary approach (Service Science Management and Engineering). It first redefines the concept of service according to a systemic perspective. It also adopts a theory of the service relationship as a « traffic relationship », and various philosophical inspirations that help us to ontologically conceive reality as nested and interconnected processes. Therefore, we formalize the service as a dynamic phenomenon emerging from the coordination of heterogeneous entities, for which the « agencements theory » seems useful. Then, we propose to represent and simulate these « servicial agencements » with a hypergraph and a mereo-relational ontology revealing the structure but also to the calculated experience’s scenarios. The proposed method and tool’s proof of concept for the modeling and simulation of servicial agencements (iSamsara) are part of the problem of constructing a real Services Systems Engineering. Moreover, this work takes place in a con-text of knowledge engineering related to the process of opportunities study of a telecom operator
Carloni, Olivier. "Introduction de raisonnement dans un outil industriel de gestion des connaissances." Phd thesis, Montpellier 2, 2008. http://www.theses.fr/2008MON20101.
Full textBaalbaki, Hussein. "Designing Big Data Frameworks for Quality-of-Data Controlling in Large-Scale Knowledge Graphs." Electronic Thesis or Diss., Sorbonne université, 2023. http://www.theses.fr/2023SORUS697.
Full textKnowledge Graphs (KGs) are the most used representation of structured information about a particular domain consisting of billions of facts in the form of entities (nodes) and relations (edges) between them. Additionally, the semantic type information of the entities is also contained in the KGs. The number of KGs has steadily increased over the past 20 years in a variety of fields, including government, academic research, the biomedical fields, etc. Applications based on machine learning that use KGs include entity linking, question-answering systems, recommender systems, etc. Open KGs are typically produced heuristically, automatically from a variety of sources, including text, photos, and other resources, or are hand-curated. However, these KGs are often incomplete, i.e., there are missing links between the entities and missing links between the entities and their corresponding entity types. In this thesis, we are addressing one of the most challenging issues facing Knowledge Graph Completion (KGC) which is link prediction. General Link Prediction in KGs that include head and tail prediction, triple classification. In recent years, KGE have been trained to represent the entities and relations in the KG in a low-dimensional vector space preserving the graph structure. In most published works such as the translational models, neural network models and others, the triple information is used to generate the latent representation of the entities and relations. In this dissertation, several methods have been proposed for KGC and their effectiveness is shown empirically in this thesis. Firstly, a novel KG embedding model TransModE is proposed for Link Prediction. TransModE projects the contextual information of the entities to modular space, while considering the relation as transition vector that guide the head to the tail entity. Secondly, we worked on building a simple low complexity KGE model, meanwhile preserving its efficiency. KEMA is a novel KGE model among the lowest KGE models in terms of complexity, meanwhile it obtains promising results. Finally, KEMA++ is proposed as an upgrade of KEMA to predict the missing triples in KGs using product arithmetic operation in modular space. The extensive experiments and ablation studies show efficiency of the proposed model, which compete the current state of the art models and set new baselines for KGC. The proposed models establish new way in solving KGC problem other than transitional, neural network, or tensor factorization based approaches. The promising results and observations open up interesting scopes for future research involving exploiting the proposed models in domain-specific KGs such as scholarly data, biomedical data, etc. Furthermore, the link prediction model can be exploited as a base model for the entity alignment task as it considers the neighborhood information of the entities
Chikhi, Nacim Fateh. "Calcul de centralité et identification de structures de communautés dans les graphes de documents." Phd thesis, Toulouse 3, 2010. http://thesesups.ups-tlse.fr/1364/.
Full textIn this thesis, we are interested in characterizing large collections of documents (using the links between them) in order to facilitate their use and exploitation by humans or by software tools. Initially, we addressed the problem of centrality computation in document graphs. We described existing centrality algorithms by focusing on the TKC (Tightly Knit Community) problem which affects most existing centrality measures. Then, we proposed three new centrality algorithms (MHITS, NHITS and DocRank) which tackle the TKC effect. The proposed algorithms were evaluated and compared to existing approaches using several graphs and evaluation measures. In a second step, we investigated the problem of document clustering. Specifically, we considered this clustering as a task of community structure identification (CSI) in document graphs. We described the existing CSI approaches by distinguishing those based on a generative model from the algorithmic or traditional ones. Then, we proposed a generative model (SPCE) based on smoothing and on an appropriate initialization for CSI in sparse graphs. The SPCE model was evaluated and validated by comparing it to other CSI approaches. Finally, we showed that the SPCE model can be extended to take into account simultaneously the links and content of documents
Simonne, Lucas. "Mining differential causal rules in knowledge graphs." Electronic Thesis or Diss., université Paris-Saclay, 2023. http://www.theses.fr/2023UPASG008.
Full textThe mining of association rules within knowledge graphs is an important area of research.Indeed, this type of rule makes it possible to represent knowledge, and their application makes it possible to complete a knowledge graph by adding missing triples or to remove erroneous triples.However, these rules express associations and do not allow the expression of causal relations, whose semantics differ from an association or a correlation.In a system, a causal link between variable A and variable B is a relationship oriented from A to B. It indicates that a change in A causes a change in B, with the other variables in the system maintaining the same values.Several frameworks exist for determining causal relationships, including the potential outcome framework, which involves matching similar instances with different values on a variable named treatment to study the effect of that treatment on another variable named the outcome.In this thesis, we propose several approaches to define rules representing a causal effect of a treatment on an outcome.This effect can be local, i.e., valid for a subset of instances of a knowledge graph defined by a graph pattern, or average, i.e., valid on average for the whole set of graph instances.The discovery of these rules is based on the framework of studying potential outcomes by matching similar instances and comparing their RDF descriptions or their learned vectorial representations through graph embedding models
Giorgino, Mathieu. "Inductive representation, proofs and refinement of pointer structures." Toulouse 3, 2013. http://thesesups.ups-tlse.fr/2076/.
Full textThis thesis stands in the general domain of formal methods that gives semantics to programs to formally prove properties about them. It originally draws its motivation from the need for certification of systems in an industrial context where Model Driven Engineering (MDE) and object-oriented (OO) languages are common. In order to obtain efficient transformations on models (graphs), we can represent them as pointer structures, allowing space and time savings through the sharing of nodes. However verification of properties on programs manipulating pointer structures is still hard. To ease this task, we propose to start the development with a high-level implementation embodied by functional programs manipulating inductive data-structures, that are easily verified in proof assistants such as Isabelle/HOL. Pointer structures are represented by a spanning tree adorned with additional references. These functional programs are then refined - if necessary - to imperative programs thanks to the library Imperative_HOL. These programs are finally extracted to Scala code (OO). This thesis describes this kind of representation and refinement and provides tools to manipulate and prove OO programs in Isabelle/HOL. This approach is put in practice with several examples, and especially with the Schorr-Waite algorithm and the construction of Binary Decision Diagrams (BDDs)