Tesis sobre el tema "Base de données SEER"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte los 50 mejores tesis para su investigación sobre el tema "Base de données SEER".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Explore tesis sobre una amplia variedad de disciplinas y organice su bibliografía correctamente.
Li, Honghao. "Interpretable biological network reconstruction from observational data". Electronic Thesis or Diss., Université Paris Cité, 2021. http://www.theses.fr/2021UNIP5207.
Texto completoThis thesis is focused on constraint-based methods, one of the basic types of causal structure learning algorithm. We use PC algorithm as a representative, for which we propose a simple and general modification that is applicable to any PC-derived methods. The modification ensures that all separating sets used during the skeleton reconstruction step to remove edges between conditionally independent variables remain consistent with respect to the final graph. It consists in iterating the structure learning algorithm while restricting the search of separating sets to those that are consistent with respect to the graph obtained at the end of the previous iteration. The restriction can be achieved with limited computational complexity with the help of block-cut tree decomposition of the graph skeleton. The enforcement of separating set consistency is found to increase the recall of constraint-based methods at the cost of precision, while keeping similar or better overall performance. It also improves the interpretability and explainability of the obtained graphical model. We then introduce the recently developed constraint-based method MIIC, which adopts ideas from the maximum likelihood framework to improve the robustness and overall performance of the obtained graph. We discuss the characteristics and the limitations of MIIC, and propose several modifications that emphasize the interpretability of the obtained graph and the scalability of the algorithm. In particular, we implement the iterative approach to enforce separating set consistency, and opt for a conservative rule of orientation, and exploit the orientation probability feature of MIIC to extend the edge notation in the final graph to illustrate different causal implications. The MIIC algorithm is applied to a dataset of about 400 000 breast cancer records from the SEER database, as a large-scale real-life benchmark
Dehainsala, Hondjack. "Explicitation de la sémantique dans lesbases de données : Base de données à base ontologique et le modèle OntoDB". Phd thesis, Université de Poitiers, 2007. http://tel.archives-ouvertes.fr/tel-00157595.
Texto completoen termes de classes et de propriétés, ainsi que des relations qui les lient. Avec le développement de
modèles d'ontologies stables dans différents domaines, OWL dans le domaine duWeb sémantique,
PLIB dans le domaine technique, de plus en plus de données (ou de métadonnées) sont décrites par référence à ces ontologies. La taille croissante de telles données rend nécessaire de les gérer au sein de bases de données originales, que nous appelons bases de données à base ontologique (BDBO), et qui possèdent la particularité de représenter, outre les données, les ontologies qui en définissent le sens. Plusieurs architectures de BDBO ont ainsi été proposées au cours des dernières années. Les chémas qu'elles utilisent pour la représentation des données sont soit constitués d'une unique table de triplets de type (sujet, prédicat, objet), soit éclatés en des tables unaires et binaires respectivement pour chaque classe et pour chaque propriété. Si de telles représentations permettent une grande flexibilité dans la structure des données représentées, elles ne sont ni susceptibles de passer à grande échelle lorsque chaque instance est décrite par un nombre significatif de propriétés, ni adaptée à la structure des bases de données usuelles, fondée sur les relations n-aires. C'est ce double inconvénient que vise à résoudre le modèle OntoDB. En introduisant des hypothèses de typages qui semblent acceptables dans beaucoup de domaine d'application, nous proposons une architecture de BDBO constituée de quatre parties : les deux premières parties correspondent à la structure usuelle des bases de données : données reposant sur un schéma logique de données, et méta-base décrivant l'ensemble de la structure de tables.
Les deux autres parties, originales, représentent respectivement les ontologies, et le méta-modèle
d'ontologie au sein d'un méta-schéma réflexif. Des mécanismes d'abstraction et de nomination permettent respectivement d'associer à chaque donnée le concept ontologique qui en définit le sens, et d'accéder aux données à partir des concepts, sans se préoccuper de la représentation des données. Cette architecture permet à la fois de gérer de façon efficace des données de grande taille définies par référence à des ontologies (données à base ontologique), mais aussi d'indexer des bases de données usuelles au niveau connaissance en leur adjoignant les deux parties : ontologie et méta-schéma. Le modèle d'architecture que nous proposons a été validé par le développement d'un prototype opérationnel implanté sur le système PostgreSQL avec le modèle d'ontologie PLIB. Nous présentons également une évaluation comparative de nos propositions aux modèles présentés antérieurement.
Bounar, Boualem. "Génération automatique de programmes sur une base de données en réseau : couplage PROLOG-Base de données en réseau". Lyon 1, 1986. http://www.theses.fr/1986LYO11703.
Texto completoEl, Khalil Firas. "Sécurité de la base de données cadastrales". Thesis, Polynésie française, 2015. http://www.theses.fr/2015POLF0001/document.
Texto completoQuantity Based Aggregation (QBA) controls closely related to inference control database and has been rarely addressed by the scientific community. Let us consider a set S of N elements. The aggregation of k elements, at most, out of N is not considered sensitive, while the aggregation of mor than k out of N elements is considered sensitive and should be prevented. The role of QBA control is to make sure the number of disclosed elements of S is less than or equal to k, where k
Dehainsala, Hondjack. "Explicitation de la sémantique dans les bases de données : base de données à base ontologique et le modèle OntoDB". Poitiers, 2007. http://www.theses.fr/2007POIT2270.
Texto completoAn Ontology–Based DataBase (OBDB) is a database which allows to store both data and ontologies that define data meaning. In this thesis, we propose a new architecture model for OBDB, called OntoDB. This model has two main original features. First, like usual databases, each stored entity is associated with a logical schema which define the structure of all its instances. Thus, our approach provides for adding ontology to existing database for semantic indexation of its content. Second, meta-model of the ontology model is also represented in the same database. This allows to support change and evolution of ontology models. The OntoDB model has been validated by a prototype. Performance evaluation of this prototype has been done and has shown that our approach allows to manage very large data and supports scalability much better than the previously proposed approaches
Lemaire, Pierre. "Base de données informatique : application aux leucémies aigue͏̈s". Paris 5, 1997. http://www.theses.fr/1997PA05P039.
Texto completoKindombi, Lola Ndontoni. "Communications interactives avec une machine base de données". Paris 11, 1985. http://www.theses.fr/1985PA112379.
Texto completoHsu, Lung-Cheng. "Pbase : une base de données déductive en Prolog". Compiègne, 1988. http://www.theses.fr/1988COMPD126.
Texto completoThis thesis describes a relational database system coupling PROLOG II and VAX RMS (Record Management Services). The SQL-like DDL (Data Definition Language) and DML (Data Manipulation Language) are implemented in PROLOG and the management of storage and research of fact record is delegated to RMS. The indexed file organization is adopted to provide a satisfactory response time. An interface written in PASCAL is called to enable the communication between PROLOG and RMS. Once the interface is established, access to the database is transparent. No precompilation is requiert. PBASE can be used as a general DBMS or it can cooperate with an expert system (Our SQL translation module can be considered as such) to manage the voluminous facts stored in the secondary memory. It can also cooperate with VAX RDB (Relational DataBase) to constitute a powerful deductive database. Although PBASE works for normalized relations as well as non-normalized ones, a normalization module is included to avoid the problems caused by the redundancy of data
Fankam, Nguemkam Chimène. "OntoDB2 : un système flexible et efficient de base de données à base ontologique pour le web sémantique et les données techniques". Chasseneuil-du-Poitou, Ecole nationale supérieure de mécanique et d'aéronautique, 2009. https://tel.archives-ouvertes.fr/tel-00452533.
Texto completoThe need to represent the semantics of data in various scientific fields (medicine, geography, engineering, etc…) has resulted in the definition of data referring to ontologies, also called ontology-based data. With the proliferation of domain ontologies, and the increasing volume of data to handle, has emerge the need to define systems capable of managing large size of ontology-based data. Such systems are called Ontology Based DataBase (OBDB) Management Systems. The main limitations of existing OBDB systems are (1) their rigidity, (2) lack of support for non standard data (spatial, temporal, etc…) and (3) their lack of effectiveness to manage large size data. In this thesis, we propose a new OBDB called OntoDB2, allowing (1) the support of ontologies based on different ontology models, (2) the extension of its model to meet specific applications requirements, and (3) an original management of ontology-based data facilitating scalability. Onto DB2 is based on the existence of a kernel ontology, and model-based techniques to enable a flexible extension of this kernel. We propose to represent only canonical data by transforming, under certain conditions, any given non-canonical data to its canonical representation. We propose to use the ontology query language to (1) to access non-canonical data thereby transform and, (2) index and pre-calculate the reasoning operations by using the mechanisms of the underlying DBMS
Jouzier, Cécile. "Constitution d'une base de données d'histoire médico-pharmaceutique bordelaise". Bordeaux 2, 2000. http://www.theses.fr/2000BOR2P107.
Texto completoDevulder, Grégory. "Base de données de séquences, phylogénie et identification bactérienne". Lyon 1, 2004. http://www.theses.fr/2004LYO10164.
Texto completoOuld, Yahia Sabiha. "Interrogation multi-critères d'une base de données spatio-temporelles". Troyes, 2005. http://www.theses.fr/2005TROY0006.
Texto completoThe study of the human behavior in driving situations is of primary importance for the improvement of drivers security. This study is complex because of the numerous situations in which the driver may be involved. The objective of the CASSICE project (Symbolic Characterization of Driving Situations) is to elaborate a tool in order to simplify the analysis task of the driver's behavior. In this paper, we will mainly take an interest in the indexation and querying of a multimedia database including the numerical data and the video sequences relating to a type of driving situations. We will put the emphasis on the queries to this database. They are often complex because they are formulated according to criteria depending on time, space and they use terms of the natural language
Vachey, Françoise. "Les suffixes toponymiques français : atlas et base de données". Nancy 2, 1999. http://www.theses.fr/1999NAN21036.
Texto completoPloquin, Catherine. "LAB langage d'analyse associé à une base de données". Bordeaux 1, 1985. http://www.theses.fr/1985BOR10534.
Texto completoBec, Xavier. "Une base de données pour les effets spéciaux numériques". Paris 8, 2000. http://www.theses.fr/2000PA081818.
Texto completoAbdelhédi, Fatma. "Conception assistée d’entrepôts de données et de documents XML pour l’analyse OLAP". Thesis, Toulouse 1, 2014. http://www.theses.fr/2014TOU10005/document.
Texto completoToday, data warehouses are a major issue for business intelligence applications within companies. Sources of a warehouse, i.e. the origin of data that feed, are diverse and heterogeneous sequential files, spreadsheets, relational databases, Web documents. The complexity is such that the software on the market only partially meets the needs of decision makers when they want to analyze the data. Therefore, our work is within the decision support systems context that integrate all data types (mainly extracted from relational databases and XML documents databases) for decision makers. They aim to provide models, methods and software tools to elaborate and manipulate data warehouses. Our work has specifically focused on two complementary issues: aided data warehouse and modeling and OLAP analysis of XML documents
Grignard, Arnaud. "Modèles de visualisation à base d'agents". Electronic Thesis or Diss., Paris 6, 2015. http://www.theses.fr/2015PA066268.
Texto completoInformation visualization is the study of interactive visual representations of abstract data to reinforce human cognition. It is very closely associated with data mining issues which allow to explore, understand and analyze phenomena, systems or data masses whose complexity continues to grow today. However, most existing visualization techniques are not suited to the exploration and understanding of datasets that consist of a large number of individual data from heterogeneous sources that share many properties with what are commonly called "complex systems". The reason is often the use of monolithic and centralized approaches. This situation is reminiscent of the modeling of complex systems (social sciences, chemistry, ecology, and many other fields) before progress represented by the generalization of agent-based approaches twenty years ago. In this thesis, I defend the idea that the same approach can be applied with the same success to the field of information visualization. By starting from the now commonly accepted idea that the agent-based models offer appropriate representations the complexity of a real system, I propose to use an approach based on the definition of agent-based visualization models to facilitate visual representation of complex data and to provide innovative support which allows to explore, programmatically and visually, their underlying dynamics. Just like their software counterparts, agent-based visualization models are composed of autonomous graphical entities that can interact and organize themselves, learn from the data they process and as a result adapt their behavior and visual representations. By providing a user the ability to describe visualization tasks in this form, my goal is to allow them to benefit from the flexibility, modularity and adaptability inherent in agent-based approaches. These concepts have been implemented and experimented on the GAMA modeling and simulation platform in which I developed a 3D immersive environment offering the user different point of views and way to interact with agents. Their implementation is validated on models chosen for their properties, supports a linear progression in terms of complexity, allowing us to highlight the concepts of flexibility, modularity and adaptability. Finally, I demonstrate through the particular case of data visualization, how my approach allows, in real time, to represent, to clarify, or even discover their dynamics and how that progress in terms of visualization can contributing,in turn, to improve the modeling of complex systems
Pineau, Nicolas. "La performance en analyse sensorielle : une approche base de données". Phd thesis, Université de Bourgogne, 2006. http://tel.archives-ouvertes.fr/tel-00125171.
Texto completoGagnon, Bertrand. "Gestion d'information sur les procédés thermiques par base de données". Thesis, McGill University, 1986. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=65447.
Texto completoFolio, Patrice. "Etablissement d'une base de données protéomique de Listeria monocytogenes EGDe". Clermont-Ferrand 2, 2003. http://www.theses.fr/2003CLF21478.
Texto completoPersyn, Emmanuel. "Base de données informatiques sur la première guerre du Golfe". Lille 3, 2003. http://www.theses.fr/2003LIL30018.
Texto completoTahir, Hassane. "Aide à la contextualisation de l’administration de base de données". Paris 6, 2013. http://www.theses.fr/2013PA066789.
Texto completoThe complexity of database administration tasks requires the development of tools for supporting database experts. When problems occur, the database administrator (DBA) is frequently the first person blamed. Most DBAs work in a fire-fighting mode and have little opportunity to be proactive. They must be constantly ready to analyze and correct failures based on a large set of procedures. In addition, they are continually readjusting these procedures and developing practices to manage a multitude of specific situations that differ from the generic situation by some few contextual elements. These practices have to deal with these contextual elements in order to solve the problem at hand. This thesis aims to use Contextual Graphs formalism in order to improve existing procedures used in database administration. The thesis shows also the benefits of using Contextual Graphs to capture user practices in order to be reused in the working contexts. Up to now, this improvement is achieved by a DBA through practices that adapt procedures to the context in which tasks should be performed and the incidents appear. This work will be the basis for designing and implementing a Context-Based Intelligent Assistant System (CBIAS) for supporting DBAs
Treger, Michèle. "Spécification et implantation d'une base de données des contacts intermoléculaires". Université Louis Pasteur (Strasbourg) (1971-2008), 1991. http://www.theses.fr/1991STR13089.
Texto completoÉtat-Le, Blanc Marie-Sylvie d'. "Une base de données sédimentologiques : structure, mise en place, applications". Bordeaux 1, 1986. http://www.theses.fr/1986BOR10565.
Texto completoPeerbocus, Mohamed Ally. "Gestion de l'évolution spatiotemporelle dans une base de données géographiques". Paris 9, 2001. https://portail.bu.dauphine.fr/fileviewer/index.php?doc=2001PA090055.
Texto completoCuré, Olivier. "Relations entre bases de données et ontologies dans le cadre du web des données". Habilitation à diriger des recherches, Université Paris-Est, 2010. http://tel.archives-ouvertes.fr/tel-00843284.
Texto completoDe, Vlieger P. "Création d'un environnement de gestion de base de données " en grille ". Application à l'échange de données médicales". Phd thesis, Université d'Auvergne - Clermont-Ferrand I, 2011. http://tel.archives-ouvertes.fr/tel-00654660.
Texto completoDe, Vlieger Paul. "Création d'un environnement de gestion de base de données "en grille" : application à l'échange de données médicales". Phd thesis, Université d'Auvergne - Clermont-Ferrand I, 2011. http://tel.archives-ouvertes.fr/tel-00719688.
Texto completoGrignard, Arnaud. "Modèles de visualisation à base d'agents". Thesis, Paris 6, 2015. http://www.theses.fr/2015PA066268/document.
Texto completoInformation visualization is the study of interactive visual representations of abstract data to reinforce human cognition. It is very closely associated with data mining issues which allow to explore, understand and analyze phenomena, systems or data masses whose complexity continues to grow today. However, most existing visualization techniques are not suited to the exploration and understanding of datasets that consist of a large number of individual data from heterogeneous sources that share many properties with what are commonly called "complex systems". The reason is often the use of monolithic and centralized approaches. This situation is reminiscent of the modeling of complex systems (social sciences, chemistry, ecology, and many other fields) before progress represented by the generalization of agent-based approaches twenty years ago. In this thesis, I defend the idea that the same approach can be applied with the same success to the field of information visualization. By starting from the now commonly accepted idea that the agent-based models offer appropriate representations the complexity of a real system, I propose to use an approach based on the definition of agent-based visualization models to facilitate visual representation of complex data and to provide innovative support which allows to explore, programmatically and visually, their underlying dynamics. Just like their software counterparts, agent-based visualization models are composed of autonomous graphical entities that can interact and organize themselves, learn from the data they process and as a result adapt their behavior and visual representations. By providing a user the ability to describe visualization tasks in this form, my goal is to allow them to benefit from the flexibility, modularity and adaptability inherent in agent-based approaches. These concepts have been implemented and experimented on the GAMA modeling and simulation platform in which I developed a 3D immersive environment offering the user different point of views and way to interact with agents. Their implementation is validated on models chosen for their properties, supports a linear progression in terms of complexity, allowing us to highlight the concepts of flexibility, modularity and adaptability. Finally, I demonstrate through the particular case of data visualization, how my approach allows, in real time, to represent, to clarify, or even discover their dynamics and how that progress in terms of visualization can contributing,in turn, to improve the modeling of complex systems
Ponchateau, Cyrille. "Conception et exploitation d'une base de modèles : application aux data sciences". Thesis, Chasseneuil-du-Poitou, Ecole nationale supérieure de mécanique et d'aérotechnique, 2018. http://www.theses.fr/2018ESMA0005/document.
Texto completoIt is common practice in experimental science to use time series to represent experimental results, that usually come as a list of values in chronological order (indexed by time) and generally obtained via sensors connected to the studied physical system. Those series are analyzed to obtain a mathematical model that allow to describe the data and thus to understand and explain the behavio rof the studied system. Nowadays, storage and analyses technologies for time series are numerous and mature, but the storage and management technologies for mathematical models and their linking to experimental numerical data are both scarce and recent. Still, mathematical models have an essential role to play in the interpretation and validation of experimental results. Consequently, an adapted storage system would ease the management and re-usability of mathematical models. This work aims at developing a models database to manage mathematical models and provide a “query by data” system, to help retrieve/identify a model from an experimental time series. In this work, I will describe the conception (from the modeling of the system, to its software architecture) of the models database and its extensions to allow the “query by data”. Then, I will describe the prototype of models database,that I implemented and the results obtained by tests performed on the latter
Devogele, Thomas. "Processus d'intégration et d'appariement de bases de données géographiques : application à une base de données routières multi-échelles". Versailles-St Quentin en Yvelines, 1997. https://tel.archives-ouvertes.fr/tel-00085113.
Texto completoJean, Stéphane. "OntoQL, un langage d'exploitation des bases de données à base ontologique". Phd thesis, Université de Poitiers, 2007. http://tel.archives-ouvertes.fr/tel-00201777.
Texto completoKratky, Andreas. "Les auras numériques : pour une poétique de la base de données". Thesis, Paris 1, 2013. http://www.theses.fr/2013PA010561/document.
Texto completoDatabase are ubiquitous in our lives and play an important rôle in many aspects of our daily activities. Conceived as a technical support to facilitate the efficient management of information and as the preferred means of storage, the database has gained a level of importance with aesthetic and political implications that go far beyond purely technical questions.Both theorical and practical in its approach, our research investigates the database as a means of expressive and poetic creation and reveals its specific character, in particular the discretization of data and the establishment of flexible relationships between them. In order to develop a poetics of the database we will reconsider the term « aura », which was utilized by walter Benjamin to analyse the transformations of the nature of aesthetic experience brought about by industrial rationalisation and technology at the end of the nineteenth century. The practical part of our research consists of two interactive projects based on the poetic principles elaborated in context of this dissertation
Alfonso, Espinosa-Oviedo Javier. "Coordination fiable de services de données à base de politiques active". Phd thesis, Université de Grenoble, 2013. http://tel.archives-ouvertes.fr/tel-01011464.
Texto completoDubois, Jean-Christophe. "Vers une interrogation en langage naturel d'une base de données image". Nancy 1, 1998. http://www.theses.fr/1998NAN10044.
Texto completoDa, Costa David. "Visualisation et fouille interactive de données à base de points d'intérêts". Tours, 2007. http://www.theses.fr/2007TOUR4021.
Texto completoIn this thesis, we present the problem of the visual data mining. We generally notice that it is specific to the types of data and that it is necessary to spend a long time to analyze the results in order to obtain an answer on the aspect of data. In this thesis, we have developed an interactive visualization environment for data exploration using points of interest. This tool visualizes all types of data and is generic because it uses only one similarity measure. These methods must be able to deal with large data sets. We also sought to improve the performances of our visualization algorithms, thus we managed to represent one million data. We also extended our tool to the data clustering. Most existing data clustering methods work in an automatic way, the user is not implied iin the process. We try to involve more significantly the user role in the data clustering process in order to improve his comprehensibility of the data results
Rolbert, Monique. "Résolution des formes pronominales dans l'interface d'interrogation d'une base de données". Aix-Marseille 2, 1989. http://www.theses.fr/1989AIX22018.
Texto completoPinasseau, Marianne. "Elaboration d'une base de données bibliographique sur les eaux minérales embouteillées". Bordeaux 2, 1999. http://www.theses.fr/1999BOR2P012.
Texto completoEspinosa, Oviedo Javier Alfonso. "Coordination fiable de services de données à base de politiques actives". Thesis, Grenoble, 2013. http://www.theses.fr/2013GRENM022/document.
Texto completoWe propose an approach for adding non-functional properties (exception handling, atomicity, security, persistence) to services' coordinations. The approach is based on an Active Policy Model (AP Model) for representing services' coordinations with non-functional properties as a collection of types. In our model, a services' coordination is represented as a workflow composed of an ordered set of activities, each activity in charge of implementing a call to a service' operation. We use the type Activity for representing a workflow and its components (i.e., the workflow' activities and the order among them). A non-functional property is represented as one or several Active Policy types, each policy composed of a set of event-condition-action rules in charge of implementing an aspect of the property. Instances of active policy and activity types are considered in the model as entities that can be executed. We use the Execution Unit type for representing them as entities that go through a series of states at runtime. When an active policy is associated to one or several execution units, its rules verify whether each unit respects the implemented non-functional property by evaluating their conditions over their execution unit state, and when the property is not verified, the rules execute their actions for enforcing the property at runtime. We also proposed a proof of concept Active Policy Execution Engine for executing an active policy oriented workflow modelled using our AP Model. The engine implements an execution model that determines how AP, Rule and Activity instances interact among each other for adding non-functional properties (NFPs) to a workflow at execution time. We validated the AP Model and the Active Policy Execution Engine by defining active policy types for addressing exception handling, atomicity, state management, persistency and authentication properties. These active policy types were used for implementing reliable service oriented applications, and mashups for integrating data from services
Jouis, Christophe. "Contributions à la conceptualisation et à la Modélisation des connaissances à partir d'une analyse linguistique de textes : réalisation d'un prototype : le système SEEK". Paris, EHESS, 1993. http://www.theses.fr/1993EHES0051.
Texto completoWe present a linguistic and computer model the aim of which is the understanding of linguistic items inserted in their context. This model is constituted by knowledge based systems of contextual exploration which consists in seeking linguistic clues in texts. It is shown that nothing else than a basic morpho-syntactical analysis and the use of the context of an examined linguistic item is required to build semantic representations. These contextual data express a knowledge of the language without using any other knowledge of the world. We present in details a program based on this model : seek. It is a help tool for knowledge extraction forme texts in natural language. This latter has been integrated in a workshop of cognitive engineering associated to a methodology of knowledge acquisition and modelling called metodac
Toumani, Farouk. "Le raisonnement taxinomique dans les modèles conceptuels de données : application à la retro-conception des bases de données relationnelles". Lyon, INSA, 1997. http://www.theses.fr/1997ISAL0051.
Texto completoTerrninological logics, as modem knowledge representation formalisms, are acknowledged to be one of the most promising artificial intelligence techniques in database applications. They allow the development of new data models equipped with taxonomic reasoning ability. However, these languages turned out to be inadequate in conceptual modelling area where emphasis must be put on the accurate and natural desc1iption of the universe of discourse. In this work, we first examine the features of terminological logics with respect to the requirements of conceptual modelling. We show that terminological logics do not support direct mode/ling requirement and constructs in these formalisms are semantically overloaded. Then we propose a model, called as a formalization of a semantic model, namely an Entity Relationship (E/R) model, using terminological logics. We demonstrate that E/R schemas and schemas are equivalent with respect to their formation capacity measure. This result ensure that the reasoning on an E/R schema can be reduced to a reasoning on its equivalent M schemas. As an application of this work, we propose to use in a relational database reverse engineering process in order to support automatic construction and enrichment of conceptual schemas and to maintain their correctness (consistency and minimality)
Verlaine, Lionel. "Optimisation des requêtes dans une machine bases de données". Paris 6, 1986. http://www.theses.fr/1986PA066532.
Texto completoBrahimi, Lahcene. "Données de tests non fonctionnels de l'ombre à la lumière : une approche multidimensionnelle pour déployer une base de données". Thesis, Chasseneuil-du-Poitou, Ecole nationale supérieure de mécanique et d'aérotechnique, 2017. http://www.theses.fr/2017ESMA0009/document.
Texto completoChoosing appropriate database management systems (DBMS) and/or execution platforms for given database (DB) is complex and tends to be time- and effort-intensive since this choice has an important impact on the satisfaction of non-functional requirements (e.g., temporal performance or energy consumption). lndeed, a large number of tests have been performed for assessing the quality of developed DB. This assessment often involves metrics associated with non-functional requirement. That leads to a mine of tests covering all life-cycle phases of the DB's design. Tests and their environments are usually published in scientific articles or specific websites such as Transaction Processing Council (TPC). Therefore, this thesis bas taken a special interest to the capitalization and the reutilization of performed tests to reduce and mastery the complexity of the DBMS/platforms selection process. By analyzing the test accurately, we identify that tests concem: the data set, the execution platform, the addressed non-functional requirements, the used queries, etc. Thus, we propose an approach of conceptualization and persistence of all dimensions as well as the results of tests. Conseguently, this thesis leads to the following contributions. (1) The design model based on descriptive, prescriptive and ontological concepts to raise the different dimensions. (2) The development of a multidimensional repository to store the test environments and their results. (3) The development of a decision making methodology based on a recommender system for DBMS and platforms selection
Dârlea, Georgiana-Lavinia. "Un système de classification supervisée à base de règles implicatives". Chambéry, 2010. http://www.theses.fr/2010CHAMS001.
Texto completoThis PhD thesis presents a series of research works done in the field of supervised data classification more precisely in the domain of semi-automatic learning of fuzzy rules-based classifiers. The prepared manuscript presents first an overview of the classification problem, and also of the main classification methods that have already been implemented and certified in order to place the proposed method in the general context of the domain. Once the context established, the actual research work is presented: the definition of a formal background for representing an elementary fuzzy rule-based classifier in a bi-dimensional space, the description of a learning algorithm for these elementary classifiers for a given data set and the conception of a multi-dimensional classification system which is able to handle multi-classes problems by combining the elementary classifiers. The implementation and testing of all these functionalities and finally the application of the resulted classifier on two real-world digital image problems are finally presented: the analysis of the quality of industrial products using 3D tomographic images and the identification of regions of interest in radar satellite images
Abecassis, Eric. "Yooda : un système de gestion de base de données orienté objet adapté à la distribution". Paris 6, 1995. http://www.theses.fr/1995PA066819.
Texto completoVallès-Parlangeau, Nathalie. "Recherche d'événements articulatoires pertinents pour l'annotation automatique de la base de données multisensorielles ACCOR". Toulouse 3, 1998. http://www.theses.fr/1998TOU30295.
Texto completoLe, Thanh Nhan. "Contribution à l'étude de la généralisation et de l'association dans une base de données relationnelle : Les ISO-dépendances et le modèle B-relationnel". Nice, 1986. http://www.theses.fr/1986NICE4014.
Texto completoJacques, Julien. "Contribution à l'apprentissage statistique à base de modèles génératifs pour données complexes". Habilitation à diriger des recherches, Université des Sciences et Technologie de Lille - Lille I, 2012. http://tel.archives-ouvertes.fr/tel-00761184.
Texto completoPeltier, Caroline. "L'analyse statistique du profil sensoriel revisitée par une approche base de données". Thesis, Dijon, 2015. http://www.theses.fr/2015DIJOS084/document.
Texto completoDuring a sensory evaluation, trained subjects taste and score products on differentdescriptors in order to obtain a descriptive profile of these products. The data are thenanalyzed by several statistical methods (PCA, ANOVA…) in order to monitor the subjectperformances and map the product space.This work aims to revisit these statistical methods thanks to the analysis of a large number ofdatasets of sensory profiling data from the SensoBase (a database containing more than athousand of datasets). Different statistical methods of mapping and analysis ofperformances were compared, then improved. We mainly focused on the so-called scalingeffect (tendency of the subject to spread his scores more or less than the panel) and onmaps related to specific statistical models.This works led to some findings about sensory analysis and recommendations for panelleaders. Furthermore, different statistical tools (R packages such as CVAPack, MAMCAP andMultiMAM) were offered. These tools could also be useful in other scientific fields usinghuman beings as a measurement instrument (psychology, economics, marketing,… )
Feuillard, Vincent. "Analyse d'une base de données pour la calibration d'un code de calcul". Phd thesis, Université Pierre et Marie Curie - Paris VI, 2007. http://tel.archives-ouvertes.fr/tel-00809048.
Texto completo