Dissertationen zum Thema „Data management“
Geben Sie eine Quelle nach APA, MLA, Chicago, Harvard und anderen Zitierweisen an
Machen Sie sich mit Top-50 Dissertationen für die Forschung zum Thema "Data management" bekannt.
Neben jedem Werk im Literaturverzeichnis ist die Option "Zur Bibliographie hinzufügen" verfügbar. Nutzen Sie sie, wird Ihre bibliographische Angabe des gewählten Werkes nach der nötigen Zitierweise (APA, MLA, Harvard, Chicago, Vancouver usw.) automatisch gestaltet.
Sie können auch den vollen Text der wissenschaftlichen Publikation im PDF-Format herunterladen und eine Online-Annotation der Arbeit lesen, wenn die relevanten Parameter in den Metadaten verfügbar sind.
Sehen Sie die Dissertationen für verschiedene Spezialgebieten durch und erstellen Sie Ihre Bibliographie auf korrekte Weise.
Morshedzadeh, Iman. „Data Classification in Product Data Management“. Thesis, Högskolan i Skövde, Institutionen för teknik och samhälle, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-14651.
Der volle Inhalt der QuelleYang, Ying. „Interactive Data Management and Data Analysis“. Thesis, State University of New York at Buffalo, 2017. http://pqdtopen.proquest.com/#viewpdf?dispub=10288109.
Der volle Inhalt der QuelleEveryone today has a big data problem. Data is everywhere and in different formats, they can be referred to as data lakes, data streams, or data swamps. To extract knowledge or insights from the data or to support decision-making, we need to go through a process of collecting, cleaning, managing and analyzing the data. In this process, data cleaning and data analysis are two of the most important and time-consuming components.
One common challenge in these two components is a lack of interaction. The data cleaning and data analysis are typically done as a batch process, operating on the whole dataset without any feedback. This leads to long, frustrating delays during which users have no idea if the process is effective. Lacking interaction, human expert effort is needed to make decisions on which algorithms or parameters to use in the systems for these two components.
We should teach computers to talk to humans, not the other way around. This dissertation focuses on building systems --- Mimir and CIA --- that help user conduct data cleaning and analysis through interaction. Mimir is a system that allows users to clean big data in a cost- and time-efficient way through interaction, a process I call on-demand ETL. Convergent inference algorithms (CIA) are a family of inference algorithms in probabilistic graphical models (PGM) that enjoys the benefit of both exact and approximate inference algorithms through interaction.
Mimir provides a general language for user to express different data cleaning needs. It acts as a shim layer that wraps around the database making it possible for the bulk of the ETL process to remain within a classical deterministic system. Mimir also helps users to measure the quality of an analysis result and provides rankings for cleaning tasks to improve the result quality in a cost efficient manner. CIA focuses on providing user interaction through the process of inference in PGMs. The goal of CIA is to free users from the upfront commitment to either approximate or exact inference, and provide user more control over time/accuracy trade-offs to direct decision-making and computation instance allocations. This dissertation describes the Mimir and CIA frameworks to demonstrate that it is feasible to build efficient interactive data management and data analysis systems.
Mathew, Avin D. „Asset management data warehouse data modelling“. Thesis, Queensland University of Technology, 2008. https://eprints.qut.edu.au/19310/1/Avin_Mathew_Thesis.pdf.
Der volle Inhalt der QuelleMathew, Avin D. „Asset management data warehouse data modelling“. Queensland University of Technology, 2008. http://eprints.qut.edu.au/19310/.
Der volle Inhalt der QuelleSehat, Mahdis, und FLORES RENÉ PAVEZ. „Customer Data Management“. Thesis, KTH, Industriell ekonomi och organisation (Avd.), 2012. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-109251.
Der volle Inhalt der QuelleScott, Mark. „Research data management“. Thesis, University of Southampton, 2014. https://eprints.soton.ac.uk/374711/.
Der volle Inhalt der QuelleTran, Viet-Trung. „Scalable data-management systems for Big Data“. Phd thesis, École normale supérieure de Cachan - ENS Cachan, 2013. http://tel.archives-ouvertes.fr/tel-00920432.
Der volle Inhalt der QuelleSchnyder, Martin. „Web 2.0 data management“. Zürich : ETH, Eidgenössische Technische Hochschule Zürich, Department of Computer Science, Institute of Information Systems, Global Information Systems Group, 2008. http://e-collection.ethbib.ethz.ch/show?type=dipl&nr=403.
Der volle Inhalt der QuelleHe, Ying Surveying & Spatial Information Systems Faculty of Engineering UNSW. „Spatial data quality management“. Publisher:University of New South Wales. Surveying & Spatial Information Systems, 2008. http://handle.unsw.edu.au/1959.4/43323.
Der volle Inhalt der QuelleVoigt, Hannes. „Flexibility in Data Management“. Doctoral thesis, Saechsische Landesbibliothek- Staats- und Universitaetsbibliothek Dresden, 2014. http://nbn-resolving.de/urn:nbn:de:bsz:14-qucosa-136681.
Der volle Inhalt der QuelleNguyen, Benjamin. „Privacy-Centric Data Management“. Habilitation à diriger des recherches, Université de Versailles-Saint Quentin en Yvelines, 2013. http://tel.archives-ouvertes.fr/tel-00936130.
Der volle Inhalt der QuelleMonk, Kitty A. „Data management in MARRS“. Thesis, Kansas State University, 1986. http://hdl.handle.net/2097/9939.
Der volle Inhalt der QuelleUichanco, Joline Ann Villaranda. „Data-driven revenue management“. Thesis, Massachusetts Institute of Technology, 2007. http://hdl.handle.net/1721.1/41728.
Der volle Inhalt der QuelleIncludes bibliographical references (p. 125-127).
In this thesis, we consider the classical newsvendor model and various important extensions. We do not assume that the demand distribution is known, rather the only information available is a set of independent samples drawn from the demand distribution. In particular, the variants of the model we consider are: the classical profit-maximization newsvendor model, the risk-averse newsvendor model and the price-setting newsvendor model. If the explicit demand distribution is known, then the exact solutions to these models can be found either analytically or numerically via simulation methods. However, in most real-life settings, the demand distribution is not available, and usually there is only historical demand data from past periods. Thus, data-driven approaches are appealing in solving these problems. In this thesis, we evaluate the theoretical and empirical performance of nonparametric and parametric approaches for solving the variants of the newsvendor model assuming partial information on the distribution. For the classical profit-maximization newsvendor model and the risk-averse newsvendor model we describe general non-parametric approaches that do not make any prior assumption on the true demand distribution. We extend and significantly improve previous theoretical bounds on the number of samples required to guarantee with high probability that the data-driven approach provides a near-optimal solution. By near-optimal we mean that the approximate solution performs arbitrarily close to the optimal solution that is computed with respect to the true demand distributions.
(cont.) For the price-setting newsvendor problem, we analyze a previously proposed simulation-based approach for a linear-additive demand model, and again derive bounds on the number of samples required to ensure that the simulation-based approach provides a near-optimal solution. We also perform computational experiments to analyze the empirical performance of these data-driven approaches.
by Joline Ann Villaranda Uichanco.
S.M.
Garling, James, und David Cahill. „ENTERPRISE DATA MANAGEMENT SYSTEMS“. International Foundation for Telemetering, 2003. http://hdl.handle.net/10150/605813.
Der volle Inhalt der QuelleThis paper discusses ongoing regulatory effects on efforts aimed at developing data infrastructures that assist test engineers in achieving information superiority and for maintaining their information, and on possible architectural frameworks for resolving the engineer’s need versus the regulatory requirements. Since current commercial-off-the-shelf (COTS) Enterprise Content Management (ECM) systems are targeted primarily at business environments such as back office applications, financial sectors, and manufacturing, these COTS systems do not provide sufficient focus for managing the unique aspects of flight test data and associated artifacts (documents, drawings, pretest data, etc.). This paper presents our ongoing efforts for deploying a storage infrastructure independent enterprise data management system for maintaining vital up-to-date information and for managing the archival of such data.
Anumalla, Kalyani. „DATA PREPROCESSING MANAGEMENT SYSTEM“. University of Akron / OhioLINK, 2007. http://rave.ohiolink.edu/etdc/view?acc_num=akron1196650015.
Der volle Inhalt der QuelleDomingues, Sérgio Rafael de Oliveira. „Market Data information management“. Master's thesis, Universidade de Aveiro, 2009. http://hdl.handle.net/10773/1740.
Der volle Inhalt der QuelleO presente relatório de projecto resulta de um estágio curricular realizado na empresa Bosch Termotecnologia, S.A. em Cacia, ao longo de oito meses. O trabalho realizado consistiu numa recolha de informação oriunda de vários mercados, a fim de se organizar e disponibilizar a mesma, inicialmente a nível interno e depois às restantes filiais. A apresentação deste relatório de projecto foi feita em quatro capítulos. No primeiro foi feito um enquadramento teórico dos Sistemas de Informação (SI) nas organizações, no segundo foi efectuada a apresentação e evolução da empresa e no terceiro capítulo foi apresentado o projecto que se pretende implementar. Na sequência dos capítulos anteriormente referidos houve lugar às conclusões e recomendações que são apresentadas no quarto capítulo.
This present work is the outcome of an academic internship that took place at Bosch Termotecnologia, S.A. in Cacia, over eight months. The work has consisted in gathering information from different markets in order to organize and provide the results, initually, internally and, later on, to the other fimr’s subsidiaries. This present work is presented in three different chapters. The first chapter addresses a theoretical framework of Information Systems (IS) in organizations. The second chapter introduces and presents the evolution of the company and in the third chapter presents how the project was implemented. Following the abovementioned chapters, chapter four presents the conclusions and recommendations.
Jäkel, Tobias. „Role-based Data Management“. Doctoral thesis, Saechsische Landesbibliothek- Staats- und Universitaetsbibliothek Dresden, 2017. http://nbn-resolving.de/urn:nbn:de:bsz:14-qucosa-224416.
Der volle Inhalt der QuelleÖhman, Mikael. „a Data-Warehouse Solution for OMS Data Management“. Thesis, Umeå universitet, Institutionen för datavetenskap, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-80688.
Der volle Inhalt der QuelleWang, Yi. „Data Management and Data Processing Support on Array-Based Scientific Data“. The Ohio State University, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=osu1436157356.
Der volle Inhalt der QuelleParviainen, A. (Antti). „Product portfolio management requirements for product data management“. Master's thesis, University of Oulu, 2014. http://urn.fi/URN:NBN:fi:oulu-201409021800.
Der volle Inhalt der QuelleDonohue, Christine M. Hayes Gregory A. Dolk Daniel R. Bui Tung X. „Data management : implementation and lessons learned from Department of the Army data management program /“. Monterey, Calif. : Springfield,Va. : Naval Postgraduate School; Available from the National Technical Information Service, 1992. http://handle.dtic.mil/100.2/ADA257858.
Der volle Inhalt der QuelleDonohue, Christine M. „Data management : implementation and lessons learned from Department of the Army data management program“. Thesis, Monterey, California. Naval Postgraduate School, 1992. http://hdl.handle.net/10945/30610.
Der volle Inhalt der QuelleInformation systems executives within Department of Defense (DoD) activities are being challenged to develop innovative ways in which information technology can contribute to the streamlining of DoD organizations. A key step in developing information systems that will meet the future needs of DoD organizations is to manage the data resource. This thesis examines the concepts, implementation strategies, and issues relating to data management and illustrates, using a case study of the Department of the Army data management methodology, the critical success factors required to implement data management programs throughout the DoD. Data management, Data standardization, Information resource management.
Saravanan, Mahesh. „Expressions as Data in Relational Data Base Management Systems“. ScholarWorks@UNO, 2006. http://scholarworks.uno.edu/td/500.
Der volle Inhalt der QuelleMelander, Lars. „Integrating Visual Data Flow Programming with Data Stream Management“. Doctoral thesis, Uppsala universitet, Datalogi, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-286536.
Der volle Inhalt der QuelleTatarinov, Igor. „Semantic data sharing with a peer data management system /“. Thesis, Connect to this title online; UW restricted, 2004. http://hdl.handle.net/1773/6942.
Der volle Inhalt der QuelleOfori-Duodu, Michael Samuel. „Exploring Data Security Management Strategies for Preventing Data Breaches“. ScholarWorks, 2019. https://scholarworks.waldenu.edu/dissertations/7947.
Der volle Inhalt der QuelleTudoran, Radu-Marius. „High-Performance Big Data Management Across Cloud Data Centers“. Electronic Thesis or Diss., Rennes, École normale supérieure, 2014. http://www.theses.fr/2014ENSR0004.
Der volle Inhalt der QuelleThe easily accessible computing power offered by cloud infrastructures, coupled with the "Big Data" revolution, are increasing the scale and speed at which data analysis is performed. Cloud computing resources for compute and storage are spread across multiple data centers around the world. Enabling fast data transfers becomes especially important in scientific applications where moving the processing close to data is expensive or even impossible. The main objectives of this thesis are to analyze how clouds can become "Big Data - friendly", and what are the best options to provide data management services able to meet the needs of applications. In this thesis, we present our contributions to improve the performance of data management for applications running on several geographically distributed data centers. We start with aspects concerning the scale of data processing on a site, and continue with the development of MapReduce type solutions allowing the distribution of calculations between several centers. Then, we present a transfer service architecture that optimizes the cost-performance ratio of transfers. This service is operated in the context of real-time data streaming between cloud data centers. Finally, we study the viability, for a cloud provider, of the solution consisting in integrating this architecture as a service based on a flexible pricing paradigm, qualified as "Transfer-as-a-Service"
Heerde, Harold Johann Wilhelm van. „Privacy-aware data management by means of data degradation“. Versailles-St Quentin en Yvelines, 2010. http://www.theses.fr/2010VERS0031.
Der volle Inhalt der QuelleService-providers collect more and more privacy-sensitive information, even though it is hard to protect this information against hackers, abuse of weak privacy policies, negligence, and malicious database administrators. In this thesis, we take the position that endless retention of privacy-sensitive information will inevitably lead to unauthorized data disclosure. Limiting the retention of privacy-sensitive information limits the amount of stored data and therefore the impact of such a disclosure. Removing data from a database system is not a straightforward task; data degradation has an impact on the storage structure, indexing, transaction management, and logging mechanisms. To show the feasibility of data degradation, we provide several techniques to implement it; mainly, a combination of keeping data sorted on degradation time and using encryption techniques where possible. The techniques are founded with a prototype implementation and a theoretical analysis
Lee, Yong Woo. „Data aggregation for capacity management“. Thesis, [College Station, Tex. : Texas A&M University, 2003. http://hdl.handle.net/1969.1/90.
Der volle Inhalt der Quelle"Major Subject: Industrial Engineering" Title from author supplied metadata (automated record created on Jul. 18, 2005.) Vita. Abstract. Includes bibliographical references.
Herrmann, Kai. „Multi-Schema-Version Data Management“. Doctoral thesis, Saechsische Landesbibliothek- Staats- und Universitaetsbibliothek Dresden, 2017. http://nbn-resolving.de/urn:nbn:de:bsz:14-qucosa-231946.
Der volle Inhalt der QuelleMatus, Castillejos Abel, und n/a. „Management of Time Series Data“. University of Canberra. Information Sciences & Engineering, 2006. http://erl.canberra.edu.au./public/adt-AUC20070111.095300.
Der volle Inhalt der QuelleDeshmukh, Pritam. „Data uncertainity in bridge management“. Diss., Columbia, Mo. : University of Missouri-Columbia, 2006. http://hdl.handle.net/10355/4510.
Der volle Inhalt der QuelleThe entire dissertation/thesis text is included in the research.pdf file; the official abstract appears in the short.pdf file (which also appears in the research.pdf); a non-technical general description, or public abstract, appears in the public.pdf file. Title from title screen of research.pdf file viewed on (May 20, 2007) Vita. Includes bibliographical references.
Wang, Yanchao. „Protein Structure Data Management System“. Digital Archive @ GSU, 2007. http://digitalarchive.gsu.edu/cs_diss/20.
Der volle Inhalt der QuelleChiu, Chao-Ying. „Visualization of construction management data“. Thesis, University of British Columbia, 2011. http://hdl.handle.net/2429/37903.
Der volle Inhalt der QuelleBansal, Dheeraj Kumar. „Non-identifying Data Management Systems“. Thesis, KTH, Skolan för informations- och kommunikationsteknik (ICT), 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-172351.
Der volle Inhalt der QuelleI denna avhandling genomförde vi en inledande analys av en affärsprocess i administrativdatahantering. Med fokus på att kartlägga behovet av att bindaautentiserade identiteter till åtgärder på de olika stegen i processen. Baserat pådenna analys har vi föreslagit en ny modell för affärsprocessen. Vi utvärderadevår modell med olika utvärderingskriterier, som fastställdes under den inledandefasen. Baserat på diskussioner med de berörda parterna, kom vi fram till slutsatsenatt även om vårt föreslagna systemet löser en hel del av integritet relateradeproblem för de berörda parterna, i fråga om affärsprocesser, är det inte lätt attändra befintliga system. Vi fann också en intressant uppsättning av problemsom kan uppstå med sådana system.
Harley, Samuel, Michael Reil, Thea Blunt-Henderson und George Bartlett. „Data, Information, and Knowledge Management“. International Foundation for Telemetering, 2005. http://hdl.handle.net/10150/604784.
Der volle Inhalt der QuelleThe Aberdeen Test Center Versatile Information System – Integrated, ONline (VISION) project has developed and deployed a telemetry capability based upon modular instrumentation, seamless communications, and the VISION Digital Library. Each of the three key elements of VISION contributes to a holistic solution to the data collection, distribution, and management requirements of Test and Evaluation. This paper provides an overview of VISION instrumentation, communications, and overall data management technologies, with a focus on engineering performance data.
Okkonen, O. (Olli). „RESTful clinical data management system“. Master's thesis, University of Oulu, 2015. http://urn.fi/URN:NBN:fi:oulu-201505291735.
Der volle Inhalt der QuelleIn the era of digitalization, clinical trials have often been left behind in adoption of automation and cost-efficiency offered by computerized systems. Poor implementations, lack of technical experience, and inertia caused by overlapping old and new procedures have failed to prove the business value of data management systems. This has led into settling for inadequate tools for data management, leaving many studies struggling with traditional approaches involving heavy paper usage further complicating the management and drastically slowing preparations for final analysis. This Master’s Thesis presents Genesis, a web-based clinical data management system development for the LIRA-study, which will take place in Finland and Sweden. Genesis has been developed to address the aforementioned obstacles with adopting information technology solutions in an agile manner with the integration of security concerns. Furthermore, Genesis has been designed to offer the long term value through reusability in terms of effortless portability for upcoming studies and interconnectability with web-enabled legacy system and handheld devices via a uniform interface. In addition to representing the design, implementation and evaluation of Genesis, the future prospects of Genesis are discussed, noting the preliminary interest of utilizing Genesis in additional studies, including the world’s largest type-1 diabetes study
Owen, J. „Data management in engineering design“. Thesis, University of Southampton, 2015. https://eprints.soton.ac.uk/385838/.
Der volle Inhalt der QuelleIngnäs, Joakim, Mikael Söderberg, Nicole Tutsch und Conrad Åslund. „Digitized management of flight data“. Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-327984.
Der volle Inhalt der QuelleDigitala flygloggar skapades genom att använda ett redan existerande system kallat Open Glider Network. Open Glider Network tillåter insamlingen av data från luftfartyg som har ett FLARM system installerat. FLARM, vilket är ett akronym för Flight Alarm, är det nuvarande säkerhetssystemet för att förhindra kollisioner mellan mindre luftfartyg. Utöver detta registerar FLARM även uppmätningar, såsom luftfartygets hastighet, höjd och position. Projektet fokuserar på användandet av insamlad data från FLARM systemet för att generera digitaliserade flygloggar för Stockholms Segelflygklubb, vilka görs tillgängliga via en hemsida. Detta kommer göra det möjligt för klubben att jämföra det digitala systemet med det analoga för att vara säkra på att säkerhetsaspekterna är desamma. Genom att göra detta kan de minimera riskerna för de mänskliga fel som rör flygloggarna när systemet används.
Mühlberger, Ralf Maximilian. „Data management for interoperable systems /“. [St. Lucia, Qld.], 2001. http://www.library.uq.edu.au/pdfserve.php?image=thesisabs/absthe16277.pdf.
Der volle Inhalt der QuelleEl, Husseini Wafaa. „Efficient ontology-based data management“. Electronic Thesis or Diss., Université de Rennes (2023-....), 2023. https://ged.univ-rennes1.fr/nuxeo/site/esupversions/afaf2edb-f3f2-4765-b1e1-9c960c6a60b4.
Der volle Inhalt der QuelleOntology-mediated query answering (OMQA) consists in asking database queries on knowledge bases (KBs); a KB is a set of facts called a database, which is described by a domain knowledge called an ontology. A main OMQA technique is FO-rewriting, which reformulates a query asked on a KB \wrt to the KB's ontology; query answers are then computed through the relational evaluation of the query reformulation on the KB's database. Essentially, because FO-rewriting compiles the domain knowledge relevant to queries into their reformulations, query reformulations may be complex and their optimization is the crux of efficiency. We devise a novel optimization framework for a large set of OMQA settings that enjoy FO-rewriting : conjunctive queries, i.e., the core select-project-join queries, asked on KBs expressed in datalog$\pm$ and existential rules, description logic and OWL, or RDF/S. We optimize the query reformulations produced by any state-of-the-art algorithm for FO-rewriting by computing rapidly, using a KB's database summary, simpler queries with same answers that can be evaluated faster by DBMSs. We show on a well-established OMQA benchmark that time performance is significantly improved by our optimization framework in general, up to three orders of magnitude
Antonov, Anton. „Product Information Management“. Master's thesis, Vysoká škola ekonomická v Praze, 2012. http://www.nusl.cz/ntk/nusl-150108.
Der volle Inhalt der QuelleAngeles, Maria del Pilar. „Management of data quality when integrating data with known provenance“. Thesis, Heriot-Watt University, 2007. http://hdl.handle.net/10399/64.
Der volle Inhalt der QuelleKalibjian, Jeff. „"Big Data" Management and Security Application to Telemetry Data Products“. International Foundation for Telemetering, 2013. http://hdl.handle.net/10150/579664.
Der volle Inhalt der Quelle"Big Data" [1] and the security challenge of managing "Big Data" is a hot topic in the IT world. The term "Big Data" is used to describe very large data sets that cannot be processed by traditional database applications in "tractable" periods of time. Securing data in a conventional database is challenge enough; securing data whose size may exceed hundreds of terabytes or even petabytes is even more daunting! As the size of telemetry product and telemetry post-processed product continues to grow, "Big Data" management techniques and the securing of that data may have ever increasing application in the telemetry realm. After reviewing "Big Data", "Big Data" security and management basics, potential application to telemetry post-processed product will be explored.
Tatikonda, Shirish. „Towards Efficient Data Analysis and Management of Semi-structured Data“. The Ohio State University, 2010. http://rave.ohiolink.edu/etdc/view?acc_num=osu1275414859.
Der volle Inhalt der QuelleDiallo, Thierno Mahamoudou. „Discovering data quality rules in a master data management context“. Thesis, Lyon, INSA, 2013. http://www.theses.fr/2013ISAL0067.
Der volle Inhalt der QuelleDirty data continues to be an important issue for companies. The datawarehouse institute [Eckerson, 2002], [Rockwell, 2012] stated poor data costs US businesses $611 billion dollars annually and erroneously priced data in retail databases costs US customers $2.5 billion each year. Data quality becomes more and more critical. The database community pays a particular attention to this subject where a variety of integrity constraints like Conditional Functional Dependencies (CFD) have been studied for data cleaning. Repair techniques based on these constraints are precise to catch inconsistencies but are limited on how to exactly correct data. Master data brings a new alternative for data cleaning with respect to it quality property. Thanks to the growing importance of Master Data Management (MDM), a new class of data quality rule known as Editing Rules (ER) tells how to fix errors, pointing which attributes are wrong and what values they should take. The intuition is to correct dirty data using high quality data from the master. However, finding data quality rules is an expensive process that involves intensive manual efforts. It remains unrealistic to rely on human designers. In this thesis, we develop pattern mining techniques for discovering ER from existing source relations with respect to master relations. In this set- ting, we propose a new semantics of ER taking advantage of both source and master data. Thanks to the semantics proposed in term of satisfaction, the discovery problem of ER turns out to be strongly related to the discovery of both CFD and one-to-one correspondences between sources and target attributes. We first attack the problem of discovering CFD. We concentrate our attention to the particular class of constant CFD known as very expressive to detect inconsistencies. We extend some well know concepts introduced for traditional Functional Dependencies to solve the discovery problem of CFD. Secondly, we propose a method based on INclusion Dependencies to extract one-to-one correspondences from source to master attributes before automatically building ER. Finally we propose some heuristics of applying ER to clean data. We have implemented and evaluated our techniques on both real life and synthetic databases. Experiments show both the feasibility, the scalability and the robustness of our proposal
Schubert, Chris, Georg Seyerl und Katharina Sack. „Dynamic Data Citation Service-Subset Tool for Operational Data Management“. MDPI, 2019. http://dx.doi.org/10.3390/data4030115.
Der volle Inhalt der QuelleFernández, Moctezuma Rafael J. „A Data-Descriptive Feedback Framework for Data Stream Management Systems“. PDXScholar, 2012. https://pdxscholar.library.pdx.edu/open_access_etds/116.
Der volle Inhalt der QuelleZhang, Yanling. „From theory to practice : environmental management in China /“. Berlin : wvb, 2005. http://www.wvberlin.de/data/inhalt/zhang.htm.
Der volle Inhalt der QuelleLehmann, Marek. „Data access in workflow management systems /“. Berlin : Aka, 2006. http://aleph.unisg.ch/hsgscan/hm00172711.pdf.
Der volle Inhalt der Quelle