Thèses sur le sujet « DATA INTEGRATION APPROACH »

Pour voir les autres types de publications sur ce sujet consultez le lien suivant : DATA INTEGRATION APPROACH.

Créez une référence correcte selon les styles APA, MLA, Chicago, Harvard et plusieurs autres

Choisissez une source :

Consultez les 50 meilleures thèses pour votre recherche sur le sujet « DATA INTEGRATION APPROACH ».

À côté de chaque source dans la liste de références il y a un bouton « Ajouter à la bibliographie ». Cliquez sur ce bouton, et nous générerons automatiquement la référence bibliographique pour la source choisie selon votre style de citation préféré : APA, MLA, Harvard, Vancouver, Chicago, etc.

Vous pouvez aussi télécharger le texte intégral de la publication scolaire au format pdf et consulter son résumé en ligne lorsque ces informations sont inclues dans les métadonnées.

Parcourez les thèses sur diverses disciplines et organisez correctement votre bibliographie.

1

Fan, Hao. « Investigating a heterogeneous data integration approach for data warehousing ». Thesis, Birkbeck (University of London), 2005. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.424299.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
2

Ziegler, Patrick. « The SIRUP approach to personal semantic data integration / ». [S.l. : s.n.], 2007. http://bvbr.bib-bvb.de:8991/F?func=service&doc_library=BVB01&doc_number=016357341&line_number=0001&func_code=DB_RECORDS&service_type=MEDIA.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
3

Criollo, Manjarrez Rotman A. « An approach for hydrogeological data management, integration and analysis ». Doctoral thesis, Universitat Politècnica de Catalunya, 2019. http://hdl.handle.net/10803/666507.

Texte intégral
Résumé :
The conceptualisation of a groundwater system involves continuous monitoring and evaluation of a large number of parameters (e.g., hydraulic parameters). Regarding hydraulic properties of the aquifers, their quantification is one of the most common problems in groundwater resources and it is recognised that all methods to obtain them have their limitations and are scale dependants. Therefore, it is necessary to have methods and tools to estimate them within a spatial context and to validate their uncertainty when they are applied in an upper scale. All these datasets collected and generated to perform a groundwater conceptual model are often stored in different scales and formats (e.g., maps, spreadsheets or databases). This continuous growing volume of data entails further improving on how it is compiled, stored and integrated for their analysis. This thesis contributes to: (i) provide dynamic and scalable methodologies for migrating and integrating multiple data infrastructures (data warehouses, spatial data infrastructures, ICT tools); (ii) to gain higher performance of their analysis within their spatial context; (iii) to provide specific tools to analyse hydrogeological processes and to obtain hydraulic parameters that have a key role in groundwater studies; and (iv) to share open-source and user-friendly software that allows standardisation, management, analysis, interpretation and sharing of hydrogeological data with a numerical model within a unique geographical platform (GIS platform). A dynamic and scalable methodology has been designed to harmonise and standardise multiple datasets and third-party databases from different origins, or to connect them with ICT tools. This methodology can be widely applied in any kind of data migration and integration (DMI) process, to develop Data warehouses, Spatial Data Infrastructures or to implement ICT tools on existing data infrastructures for further analyses, improving data governance. A higher performance to obtain hydraulic parameters of the aquifer has been addressed from the development of a GIS-based tool. The interpretation of pumping tests within its spatial context can reduce the uncertainty of its analysis with an accurate knowledge of the aquifer geometry and boundaries. This software designed to collect, manage, visualise and analyse pumping tests in a GIS environment supports the hydraulic parameterization of groundwater flow and transport models. To enhance the hydraulic parameters quantification, a compilation, revision and analysis of the hydraulic conductivity based on grain size methodologies have been performed. Afterwards, the uncertainty of applying these methods on a larger scale has been addressed and discussed by comparison of the upscaling results with pumping tests. Finally, a sharing, open-source and user-friendly GIS-based tool is presented. This new generation of GIS-based tool aims at simplifying the characterisation of groundwater bodies for the purpose of building rigorous and data-based environmental conceptual models. It allows to standardise, manage, analyse and interpret hydrogeological and hydrochemical data. Due to its free and open-source architecture, it can be updated and extended depending on the tailored applications.
La conceptualització d’un sistema hidrogeològic implica una continua monitorització i avaluació d’una gran quantitat de paràmetres (e.g., paràmetres hidràulics). Pel que fa als paràmetres hidràulics de l’aqüífer, la seva quantificació és un dels problemes més comuns als estudis hidrogeològics. És àmpliament reconegut que els mètodes per obtenir aquest tipus de paràmetres tenen les seves limitacions i són dependents de l’escala d’anàlisi. Per aquest motiu, cal disposar de mètodes i eines per estimar-los dins del seu context espacial i validar la seva incertesa quan s’apliquen en una escala superior d’anàlisi. Les dades recopilades i generades per realitzar un model conceptual hidrogeològic sovint s'emmagatzemen en diferents escales i formats (e.g., mapes, fulls de càlcul o bases de dades). Aquest volum de dades en continu creixement requereix d'eines i metodologies que millorin la seva compilació i gestió per al seu posterior anàlisi. Les contribucions realitzades en aquesta tesi son: (i) proporcionar metodologies dinàmiques i escalables per migrar i integrar múltiples infraestructures de dades (infraestructures de dades espacials i no espacials, o la implementació d'eines TIC); (ii) obtenir un major rendiment de l'anàlisi hidrogeològic tenint en compte el seu context espacial; (iii) proporcionar eines específiques per analitzar processos hidrogeològics i obtenir paràmetres hidràulics que tenen un paper clau en els estudis d'aigües subterrànies; i (iv) difondre software de codi lliure i de fàcil accés que permeti l'estandardització, gestió, anàlisi, interpretació i intercanvi de dades hidrogeològiques amb un model numèric dins d'una única plataforma de informació geogràfica (SIG). S'ha dissenyat una metodologia dinàmica i escalable per harmonitzar i estandarditzar múltiples conjunts de dades de diferents orígens, o bé per connectar aquestes infraestructures de dades amb eines TIC. Aquesta metodologia pot ser implementada en qualsevol tipus de procés de migració i integració de dades (DMI), per a desenvolupar infraestructures de dades espacials i no espacials, o bé per implementar eines TIC a les infraestructures de dades existents per a anàlisi addicionals; millorant així la governança de les dades. Un major rendiment per obtenir els paràmetres hidràulics de l'aqüífer s'adreça des del desenvolupament d'una eina SIG. La interpretació dels assaigs de bombament dins del seu context espacial, pot reduir la incertesa del seu anàlisi amb un coneixement precís de la geometria i els límits de l'aqüífer. Aquest software dissenyat per recopilar, administrar, visualitzar i analitzar els assaigs de bombament en un entorn GIS, dóna suport a la parametrització hidràulica dels models de flux i transport d'aigües subterrànies. Per millorar la quantificació dels paràmetres hidràulics, es va realitzar una compilació, revisió i anàlisi de la conductivitat hidràulica basada en metodologies de mida de gra. Posteriorment, s'ha considerat i discutit la incertesa d'aplicar aquests mètodes en una escala major comparant els resultats de la millora d'escala amb les proves de bombament. Finalment, es presenta una eina SIG lliure, de codi obert i de fàcil aplicació. Aquesta nova generació d'eines SIG pretenen simplificar la caracterització de les masses d'aigua subterrània amb el propòsit de construir models conceptuals ambientals rigorosos. A més, aquesta eina permet estandarditzar, gestionar, analitzar i interpretar dades hidrogeològiques i hidroquímiques. Donat que la seva arquitectura és de codi lliure i obert, es pot actualitzar i ampliar segons les aplicacions personalitzades que cada usuari requereixi.
La conceptualización de un sistema hidrogeológico implica el continuo monitoreo y evaluación de una gran cantidad de parámetros (e.g., parámetros hidráulicos). Con respecto a los parámetros hidráulicos, su cuantificación es uno de los problemas más comunes en los estudios hidrogeológicos. Es ampliamente reconocido que los métodos para obtener este tipo de parámetros tienen sus limitaciones y son dependientes de la escala de análisis. En este sentido, es necesario disponer de métodos y herramientas para estimarlos dentro de su contexto espacial y validar su incertidumbre cuando se aplican en una escala superior de análisis. Los datos recopilados y generados para realizar un modelo conceptual hidrogeológico a menudo se almacenan en diferentes escalas y formatos (e.g., mapas, hojas de cálculo o bases de datos). Este volumen de datos en continuo crecimiento requiere de herramientas y metodologías que mejoren su compilación y gestión para su posterior análisis. Las contribuciones realizadas son: (i) proporcionar metodologías dinámicas y escalables para migrar e integrar múltiples infraestructuras de datos (ya sean infraestructuras de datos espaciales y no espaciales, o la implementación de herramientas TIC); (ii) obtener un mayor rendimiento del análisis hidrogeológico teniendo en cuenta su contexto espacial; (iii) proporcionar herramientas específicas para analizar procesos hidrogeológicos y obtener parámetros hidráulicos que desempeñan un papel clave en los estudios de aguas subterráneas; y (iv) difundir software de código abierto y de fácil acceso que permita la estandarización, gestión, análisis, interpretación e intercambio de datos hidrogeológicos con un modelo numérico dentro de una única plataforma de información geográfica (SIG). Se ha diseñado una metodología dinámica y escalable para armonizar y estandarizar múltiples conjuntos de datos de diferentes orígenes, o bien para conectar éstas infraestructuras de datos con herramientas TIC. Esta metodología puede ser implementada en cualquier tipo de proceso de migración e integración de datos (DMI), para desarrollar infraestructuras de datos espaciales y no espaciales, o para implementar herramientas TIC en las infraestructuras de datos existentes para análisis adicionales; mejorando así la gobernanza de los datos. Un mayor rendimiento para obtener los parámetros hidráulicos del acuífero se ha abordado desde el desarrollo de una herramienta SIG. La interpretación de ensayos de bombeo dentro de su contexto espacial, puede reducir la incertidumbre de su análisis con un conocimiento preciso de la geometría y los límites del acuífero. Este software diseñado para recopilar, administrar, visualizar y analizar las pruebas de bombeo en un entorno SIG, apoya la parametrización hidráulica de los modelos de flujo y transporte de aguas subterráneas. Para mejorar la cuantificación de los parámetros hidráulicos, se ha realizado una compilación, revisión y análisis de la conductividad hidráulica basada en metodologías de tamaño de grano. Posteriormente, se ha considerado y discutido la incertidumbre de aplicar estos métodos en una escala mayor comparando los resultados de la mejora de escala con los obtenidos en ensayos de bombeo. Finalmente, se presenta una herramienta SIG libre, de código abierto y de fácil aplicación. Esta nueva generación de herramienta SIG pretende simplificar la caracterización de los cuerpos de agua subterránea con el propósito de construir modelos conceptuales ambientales rigurosos. Además, esta herramienta permite estandarizar, gestionar, analizar e interpretar datos hidrogeológicos e hidroquímicos. Gracias a su arquitectura de código libre y abierto, se puede actualizar y ampliar según las aplicaciones personalizadas que cada usuario requiera
Styles APA, Harvard, Vancouver, ISO, etc.
4

Faulstich, Lukas C. « The HyperView approach to the integration of semistructured data ». [S.l. : s.n.], 2000. http://www.diss.fu-berlin.de/2000/33/index.html.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
5

Kittivoravitkul, Sasivimol. « A bi-directional transformation approach for semistructured data integration ». Thesis, Imperial College London, 2007. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.444093.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
6

GHEZZI, ANNALISA. « A new approach to data integration in Archaeological geophysics ». Doctoral thesis, Università degli Studi di Camerino, 2020. http://hdl.handle.net/11581/447386.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
7

Lewis, Richard. « A semantic approach to railway data integration and decision support ». Thesis, University of Birmingham, 2015. http://etheses.bham.ac.uk//id/eprint/5959/.

Texte intégral
Résumé :
The work presented in this thesis was motivated by the desire of the railway industry to capitalise on new technology developments promising seamless integration of distributed data. This includes systems that generate, consume and transmit data for asset decision support. The primary aim of the research was to investigate the limitations of previous syntactic data integration exercises, creating a foundation for semantic system development. The objective was to create a modelling process enabling domain experts to provide the information concepts and semantic relationships between those concepts. The resulting model caters for the heterogeneity between systems supplying data that previous syntactic approaches failed to achieve and integrate data from multiple systems such that the context of data is not lost when centralised in a repository. The essence of this work is founded on two characteristics of distributed data management; the first is that current Web tools, like XML, are not effective for all aspects of technical interoperability because they do not capture the context of the data; and second, there is little relationship between conventional database management systems and the data structures that are utilised in Web based data exchange which means that a different set of architecture components are required.
Styles APA, Harvard, Vancouver, ISO, etc.
8

Mireku, Kwakye Michael. « A Practical Approach to Merging Multidimensional Data Models ». Thesis, Université d'Ottawa / University of Ottawa, 2011. http://hdl.handle.net/10393/20457.

Texte intégral
Résumé :
Schema merging is the process of incorporating data models into an integrated, consistent schema from which query solutions satisfying all incorporated models can be derived. The efficiency of such a process is reliant on the effective semantic representation of the chosen data models, as well as the mapping relationships between the elements of the source data models. Consider a scenario where, as a result of company mergers or acquisitions, a number of related, but possible disparate data marts need to be integrated into a global data warehouse. The ability to retrieve data across these disparate, but related, data marts poses an important challenge. Intuitively, forming an all-inclusive data warehouse includes the tedious tasks of identifying related fact and dimension table attributes, as well as the design of a schema merge algorithm for the integration. Additionally, the evaluation of the combined set of correct answers to queries, likely to be independently posed to such data marts, becomes difficult to achieve. Model management refers to a high-level, abstract programming language designed to efficiently manipulate schemas and mappings. Particularly, model management operations such as match, compose mappings, apply functions and merge, offer a way to handle the above-mentioned data integration problem within the domain of data warehousing. In this research, we introduce a methodology for the integration of star schema source data marts into a single consolidated data warehouse based on model management. In our methodology, we discuss the development of three (3) main streamlined steps to facilitate the generation of a global data warehouse. That is, we adopt techniques for deriving attribute correspondences, and for schema mapping discovery. Finally, we formulate and design a merge algorithm, based on multidimensional star schemas; which is primarily the core contribution of this research. Our approach focuses on delivering a polynomial time solution needed for the expected volume of data and its associated large-scale query processing. The experimental evaluation shows that an integrated schema, alongside instance data, can be derived based on the type of mappings adopted in the mapping discovery step. The adoption of Global-And-Local-As-View (GLAV) mapping models delivered a maximally-contained or exact representation of all fact and dimensional instance data tuples needed in query processing on the integrated data warehouse. Additionally, different forms of conflicts, such as semantic conflicts for related or unrelated dimension entities, and descriptive conflicts for differing attribute data types, were encountered and resolved in the developed solution. Finally, this research has highlighted some critical and inherent issues regarding functional dependencies in mapping models, integrity constraints at the source data marts, and multi-valued dimension attributes. These issues were encountered during the integration of the source data marts, as it has been the case of evaluating the queries processed on the merged data warehouse as against that on the independent data marts.
Styles APA, Harvard, Vancouver, ISO, etc.
9

Mukviboonchai, Suvimol. « The mediated data integration (MeDInt) : An approach to the integration of database and legacy systems ». Thesis, Edith Cowan University, Research Online, Perth, Western Australia, 2003. https://ro.ecu.edu.au/theses/1308.

Texte intégral
Résumé :
The information required for decision making by executives in organizations is normally scattered across disparate data sources including databases and legacy systems. To gain a competitive advantage, it is extremely important for executives to be able to obtain one unique view of information in an accurate and timely manner. To do this, it is necessary to interoperate multiple data sources, which differ structurally and semantically. Particular problems occur when applying traditional integration approaches, for example, the global schema needs to be recreated when the component schema has been modified. This research investigates the following heterogeneities between heterogeneous data sources: Data Model Heterogeneities, Schematic Heterogeneities and Semantic Heterogeneities. The problems of existing integration approaches are reviewed and solved by introducing and designing a new integration approach to logically interoperate heterogeneous data sources and to resolve three previously classified heterogeneities. The research attempts to reduce the complexity of the integration process by maximising the degree of automation. Mediation and wrapping techniques are employed in this research. The Mediated Data Integration (MeDint) architecture has been introduced to integrate heterogeneous data sources. Three major elements, the MeDint Mediator, wrappers, and the Mediated Data Model (MDM) play important roles in the integration of heterogeneous data sources. The MeDint Mediator acts as an intermediate layer transforming queries to sub-queries, resolving conflicts, and consolidating conflict-resolved results. Wrappers serve as translators between the MeDint Mediator and data sources. Both the mediator and wrappers arc well-supported by MDM, a semantically-rich data model which can describe or represent heterogeneous data schematically and semantically. Some organisational information systems have been tested and evaluated using the MeDint architecture. The results have addressed all the research questions regarding the interoperability of heterogeneous data sources. In addition, the results also confirm that the Me Dint architecture is able to provide integration that is transparent to users and that the schema evolution does not affect the integration.
Styles APA, Harvard, Vancouver, ISO, etc.
10

Engström, Henrik. « Selection of maintenance policies for a data warehousing environment : a cost based approach to meeting quality of service requirements ». Thesis, University of Exeter, 2002. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.269667.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
11

Sannellappanavar, Vijaya Laxmankumar. « DATAWAREHOUSE APPROACH TO DECISION SUPPORT SYSTEM FROM DISTRIBUTED, HETEROGENEOUS SOURCES ». University of Akron / OhioLINK, 2006. http://rave.ohiolink.edu/etdc/view?acc_num=akron1153506475.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
12

Sehlstedt, Jonas. « Replacing qpcr non-detects with microarray expression data : An initialized approach towards microarray and qPCR data integration ». Thesis, Högskolan i Skövde, Institutionen för biovetenskap, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-15790.

Texte intégral
Résumé :
Gene expression analysis can be performed by a number of methods. One of the most common methods is using relative qPCR  to assess the relative expression of a determined set of genes compared to a reference gene. Analysis methods benefits from an as homogeneous sample set as possible, as great variety in original sample disease status, quality, type, or distribution may yield an uneven base expression between replicates. Additionally normalization of qPCR data will not work if there are missing values in the data. There are methods for handling non-detects (i.e. missing values) in the data, where most of them are only recommended to use when there is a single, or very few, value missing. By integrating microarray expression data with qPCR data, the data quality could be improved on, eradicating the need to redo an entire experiment when too much data is missing or sample data too is heterogeneous. In this project, publically available microarray data, with similar sample status of a given qPCR dataset, was downloaded and processed. The qPCR dataset included 51 genes, where a set of four DLG genes has been chosen for in-depth analysis. For handling missing values, mean imputation and inserting Cq value 40 were used, as well as a novel method initialized where microarray data was used to replace missing values. In summary replacing missing values with microarray data did not show any significant difference to the other two methods in three of the four DLG genes. From this project, it is also suggested an initialized approach towards testing the possibility of qPCR and microarray data integration.
Styles APA, Harvard, Vancouver, ISO, etc.
13

Shea, Geoffrey Yu Kai Surveying &amp Spatial Information Systems Faculty of Engineering UNSW. « A Web-Based Approach to the Integration of Diverse Data Sources for GIS ». Awarded by:University of New South Wales. Surveying and Spatial Information Systems, 2001. http://handle.unsw.edu.au/1959.4/17855.

Texte intégral
Résumé :
The rigorous developments of GIS over the past decades have enabled application developers to create powerful systems that are used to facilitate the management of spatial data. Unfortunately, each one of these systems is specific to a local service, with little or no interconnection with services in other locales. This makes it virtually impossible to perform dynamic and interactive GIS operations across multiple locales which have similar or dissimilar system configurations. The Spatial Data Transfer Standard (SDTS) resolved the problems partially by offering excellent conceptual and logical abstraction model for data exchange. Recent advancements of the Internet enlightened the GIS community as to the realization of an ideal concept of information interchange. A suite of new technologies that embraces Extensible Markup Language (XML), Scalable Vector Graphics (SVG), Portable Network Graphics (PNG) and Java creates a powerful and new perspective that can be applied to all phases of online GIS system development. The online GIS is a Web-based approach to integrating diverse spatial data sources for GIS applications. To address the spatial data integration options and implications related to the Web-based approach the investigation was undertaken in 5 phases: (1) Determine the mapping requirements of graphic and non-graphic spatial data for online GIS application; (2) Analyze the requirements of spatial data integration for online environments; (3) Investigate a suitable method for integrating different formats of spatial data; (4) Study the feasibility and applicability of setting up the online GIS; and (5) Develop a prototype for online sharing of teaching resources. Resulting from the critical review on current Internet technology, a conceptual framework for spatial data integration was proposed. This framework was based on the emerging Internet technology on XML, SVG, PNG, and Java. It was comprised of four loosely coupled modules, namely, Application Interface, Presentation, Integrator, and Data module. This loosely coupled framework provides an environment that will be independent of the underlying GIS data structure and makes it easy to change or update the system as a new task or knowledge is acquired. A feasibility study was conducted to test the applicability for the proposed conceptual framework. A detailed user requirements and system specification was thus devised from the feasibility study. These user requirements and system specification provided some guidelines for online GIS application development. They were expressed specifically in terms of six aspects: (1) User; (2) Teaching resources management; (3) Data; (4) Cartography; (5) Functions; and (6) Software development configuration. A prototype system based on some of the devised system specifications was developed. In the prototype software design, the architecture of a Three-Tier Client-Server computing model was adopted. Due to the inadequacy of native support for SVG and PNG in all currently available Web browsers, the prototype was thus implemented in HTML, Java and vendor specific vector format. The prototype demonstrated how teaching resources from a variety of sources and format (including map data and non-map resources) were integrated and shared. The implementation of the prototype revealed that the Web is still an ideal medium for providing wider accessibility of geographical information to a larger number of users through a corporate intranet or the Internet cost-effectively. The investigation concluded that current WWW technology is limited in its capability for spatial data integration and delivering online functionality. However, developing of XML-based GIS data model and graphic standards SVG and PNG for structuring and transferring spatial data on the Internet appear to be providing solutions to the current limitations. It is believed that the ideal world where everyone retrieving spatial information contextually through a Web browser disregarding the information format and location will eventually become true.
Styles APA, Harvard, Vancouver, ISO, etc.
14

Shea, Geoffrey Yu Kai. « A web-based approach to the integration of diverse data sources for GIS / ». Sydney : School of Surveying and Spatial Information Systems, University of New South Wales, 2001. http://www.library.unsw.edu.au/~thesis/adt-NUN/public/adt-NUN20011018.170350/index.html.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
15

Jakobsson, Erik. « A new approach to Pairs Trading : Using fundamental data to find optimal portfolios ». Thesis, Umeå universitet, Institutionen för matematik och matematisk statistik, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-104314.

Texte intégral
Résumé :
Since its’ invention at Morgan Stanley in 1987 pairs trading has grown to be one of the most common and most researched strategies for market neutral returns. The strategy identifies stocks, or other financial securities, that historically has co-moved and forms a trading pair. If the price relation is broken a short position is entered in the overperforming stock and a long in the underperforming. The positions are closed when the spread returns to the long-term relation. A pairs trading portfolio is formed by combining a number of pairs. To detect adequate pairs different types of data analysis has been used. The most common way has been to study historical price data with different statistical models such as the distance method. Gatev et al (2006) used this method and provided the most extensive research on the subject and this study will follow the standards set by that article and add new interesting factors. This is done through an investigation on how the analysis can be improved by using the stocks fundamental data, e.g. P/E, P/B, leverage, industry classification. This data is used to set up restrictions and Lasso models (type of regression) to optimize the trading portfolio and achieve higher returns. All models have been back-tested using S&P 500 stocks between 2001-04-01 and 2015-04-01 with portfolios changed every six months. The most important finding of the study is that restricting stocks to have close P/E-ratios combined with traditional price series analysis increases returns. The most conservative measure gives annual returns of 3.99% to 4.98% depending on the trading rules for this portfolio. The returns are significantly (5%-level) higher than those obtained by the traditional distance method. Considerable variations in return levels is shown to be created when capital commitments are changed and trading rules, transaction costs and restrictions on unique portfolio stocks are implemented. Further research regarding how analysis of P/E-ratios can improve pairs trading is suggested. The thesis has been written independently without an external client and studied an area that the author found interesting.
Styles APA, Harvard, Vancouver, ISO, etc.
16

Proesser, Malte. « A new approach to systems integration in the mechatronic engineering design process of manufacturing systems ». Thesis, De Montfort University, 2014. http://hdl.handle.net/2086/10492.

Texte intégral
Résumé :
Creating flexible and automated production facilities is a complex process that requires high levels of cooperation involving all mechatronics disciplines, where software tools being utilised have to work as closely as their users. Some of these tools are well-integrated but others can hardly exchange any data. This research aims to integrate the software systems applied by the mechatronic engineering disciplines to enable an enhanced design process characterised by a more parallel and iterative work flow. This thesis approaches systems integration from a data modelling point of view because it sees information transfer between heterogeneous data models as a key element of systems integration. A new approach has been developed which is called middle-in data modelling strategy since it is a combination of currently applied top-down and bottom-up approaches. It includes the separation of data into core design data which is modelled top-down and detailed design data modules which are modelled bottom-up. The effectiveness of the integration approach has been demonstrated in a case study undertaken for the mechatronic engineering design process of body shop production lines in the automotive industry. However, the application of the middle-in data modelling strategy is not limited to this use case: it can be used to enhance a variety of system integration tasks. The middle-in data modelling strategy is tested and evaluated in comparison with present top-down and bottom-up data modelling strategies on the basis of three test cases. These test cases simulated how the systems integration solutions based on the different data modelling strategies react to certain disturbances in the data exchange process as they would likely occur during industrial engineering design work. The result is that the top-down data modelling strategy is best in maintaining data integrity and consistency while the bottom-up strategy is most flexibly adaptable to further developments of systems integration solutions. The middle-in strategy combines the advantages of top-down and bottom-up approaches while their weaknesses and disadvantages are kept at a minimum. Hence, it enables the maintenance of data modelling consistency while being responsive to multidisciplinary requirements and adaptive during its step-by-step introduction into an industrial engineering process.
Styles APA, Harvard, Vancouver, ISO, etc.
17

Kabir, Sami. « BRB based Deep Learning Approach with Application in Sensor Data Streams ». Thesis, Luleå tekniska universitet, Datavetenskap, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:ltu:diva-75974.

Texte intégral
Résumé :
Predicting events based on available data is an effective way to protect human lives. Issuing health alert based on prediction of environmental pollution, executing timely evacuation of people from vulnerable areas based on prediction of natural disasters are the application areas of sensor data stream where accurate and timely prediction is crucial to safeguard people and assets. Thus, prediction accuracy plays a significant role to take precautionary measures and minimize the extent of damage. Belief rule-based Expert System (BRBES) is a rule-driven approach to perform accurate prediction based on knowledge base and inference engine. It outperforms other such knowledge-driven approaches, such as, fuzzy logic, Bayesian probability theory in terms of dealing with uncertainties. On the other hand, Deep Learning is a data-driven approach which belongs to Artificial Intelligence (AI) domain. Deep Learning discovers hidden data pattern by performing analytics on huge amount of data. Thus, Deep Learning is also an effective way to predict events based on available data, such as, historical data and sensor data streams. Integration of Deep Learning with BRBES can improve prediction accuracy further as one can address the inefficiency of the other to bring down error gap. We have taken air pollution prediction as the application area of our proposed integrated approach. Our combined approach has shown higher accuracy than relying only on BRBES and only on Deep Learning.

This is a Master Thesis Report as part of degree requirement of Erasmus Mundus Joint Master Degree (EMJMD) in Pervasive Computing and Communications for Sustainable Development (PERCCOM).

Styles APA, Harvard, Vancouver, ISO, etc.
18

Coronado, Zamora Marta. « Mapping natural selection through the drosophila melanogaster development following a multiomics data integration approach ». Doctoral thesis, Universitat Autònoma de Barcelona, 2018. http://hdl.handle.net/10803/666761.

Texte intégral
Résumé :
La teoria de l'evolució de Charles Darwin proposa que les adaptacions dels organismes sorgeixen com a conseqüència del procés de la selecció natural. La selecció natural deixa una empremta característica en els patrons de variació genètica que pot detectar-se mitjançant mètodes estadístics d'anàlisi genòmica. Avui en dia podem inferir l'acció de la selecció natural en el genoma i fins i tot quantificar quina proporció de les noves variants genètiques que incorpora una espècie són adaptatives. L’era genòmica ha conduït a la situació paradoxal en la qual disposem de més informació sobre la selecció en el genoma que sobre el fenotip de l’organisme, l’objectiu principal de la selecció natural. El desenvolupament de les tecnologies de seqüenciació de nova generació (NGS, per les seves sigles en anglès) està proporcionant una gran quantitat de dades -òmiques, incrementant notablement la disponibilitat de sèries transcriptòmiques del desenvolupament. A diferència del genoma d'un organisme, el transcriptoma és un fenotip que varia al llarg de la vida i en diferents parts del cos. L'estudi d'un transcriptoma des d'una perspectiva genòmica-poblacional i espai-temporal és un enfocament prometedor per comprendre les bases genètiques i del desenvolupament del canvi fenotípic. Aquesta tesi és un projecte integrador de genòmica de poblacions i biologia evolutiva seguint un enfocament bioinformàtic. Es compon de tres passos seqüencials: (i) la comparativa d'un conjunt de mètodes de McDonald i Kreitman (MKT), un test per detectar selecció positiva recurrent en seqüències codificants a nivell molecular, utilitzant tant dades empíriques d'una població nord-americana de D. melanogaster i dades simulades, (ii) la inferència de les característiques del genoma que es correlacionen amb la tassa evolutiva dels gens codificadors de proteïnes, i (iii) la integració de patrons de variació genòmica amb anotacions de grans conjunts de dades espai-temporals del desenvolupament (evo-dev-omics). Com a resultat d'aquest enfocament hem dut a terme dos estudis diferents que integren els patrons de diversitat genòmica amb capes multiòmiques al llarg del desenvolupament, tant en el temps com en l'espai. En el primer estudi, donem una perspectiva global sobre com actua la selecció natural durant tot el cicle de vida de D. melanogaster, avaluant com els diferents règims de selecció actuen a través dels diferents estadis del desenvolupament. En el segon estudi, tracem un mapa exhaustiu de com la selecció actua sobre l'anatomia completa de l'embrió de D. melanogaster. En conjunt, els nostres resultats mostren que els gens expressats en el desenvolupament embrionari mitjà i tardà exhibeixen la major conservació a nivell de seqüència i una estructura gènica més complexa: són més llargs, contenen més exons i introns més llargs, codifiquen un gran nombre de isoformes i, de mitjana, tenen més expressió. El constrenyiment selectiu és ubic, especialment afectant els sistemes digestiu i nerviós. D'altra banda, els primers estadis del desenvolupament embrionari són els més divergents, i sembla ser degut a una menor eficàcia de la selecció natural sobre els gens d'efecte matern. A més, els gens expressats en aquestes primeres etapes tenen, de mitjana, els introns més curts, probablement degut a la necessitat d'expressar-se ràpidament i eficientment durant els cicles cel·lulars curts. L'adaptació es produeix en aquelles estructures que també mostren evidències d'adaptació en l'adult, el sistema immunològic i el sistema reproductiu. Finalment, els gens que s’expressen en una o unes poques estructures anatòmiques són evolutivament més joves i exhibeixen unes taxes d'evolució més altes, a diferència dels gens que s’expressen en totes o gairebé totes les estructures. La genòmica de poblacions ja no és una ciència teòrica, s’ha convertit en un camp interdisciplinari on la bioinformàtica, grans conjunts de dades -òmiques, models estadístics i evolutius i tècniques moleculars emergents s’integren per obtenir una visió sistèmica de les causes i les conseqüències de l’evolució. La integració de la genòmica de poblacions amb altres dades fenotípiques multiòmiques és un pas necessari per obtenir una visió global de com l’adaptació ocorre en la natura.
Charles Darwin's theory of evolution proposes that the adaptations of organisms arise because of the process of natural selection. Natural selection leaves a characteristic footprint on the patterns of genetic variation that can be detected by means of statistical methods of genomic analysis. Today, we can infer the action of natural selection in a genome and even quantify what proportion of the incorporated genetic variants in the populations are adaptive. The genomic era has led to the paradoxical situation in which much more evidence of selection is available on the genome than on the phenotype of the organism, the primary target of natural selection. The advent of next-generation sequencing (NGS) technologies is providing a vast amount of -omics data, especially increasing the breadth of available developmental transcriptomic series. In contrast to the genome of an organism, the transcriptome is a phenotype that varies during the lifetime and across different body parts. Studying a developmental transcriptome from a population genomic and spatio-temporal perspective is a promising approach to understand the genetic and developmental basis of the phenotypic change. This thesis is an integrative population genomics and evolutionary biology project following a bioinformatic approach. It is performed in three sequential steps: (i) the comparison of different variations of the McDonald and Kreitman test (MKT), a method to detect recurrent positive selection on coding sequences at the molecular level, using empirical data from a North American population of D. melanogaster and simulated data, (ii) the inference of the genome features correlated with the evolutionary rate of protein-coding genes, and (iii) the integration of patterns of genomic variation with annotations of large sets of spatio-temporal developmental data (evo-dev-omics). As a result of this approach, we have carried out two different studies integrating the patterns of genomic diversity with multiomics layers across developmental time and space. In the first study we give a global perspective on how natural selection acts during the whole life cycle of D. melanogaster, assessing whether different regimes of selection act through the developmental stages. In the second study, we draw an exhaustive map of selection acting on the complete embryo anatomy of D. melanogaster. Taking all together, our results show that genes expressed in mid- and late-embryonic development stages exhibit the highest sequence conservation and the most complex structure: they are larger, consist of more exons and longer introns, encode a large number of isoforms and, on average, are highly expressed. Selective constraint is pervasive, particularly on the digestive and nervous systems. On the other hand, earlier stages of embryonic development are the most divergent, which seems to be due to the diminished efficiency of natural selection on maternal-effect genes. Additionally, genes expressed in these first stages have on average the shortest introns, probably due to the need for a rapid and efficient expression during the short cell cycles. Adaptation is found in the structures that also show evidence of adaptation in the adult, the immune and reproductive systems. Finally, genes that are expressed in one or a few different anatomical structures are younger and have higher rates of evolution, unlike genes that are expressed in all or almost all structures. Population genomics is no longer a theoretical science, it has become an interdisciplinary field where bioinformatics, large functional -omics datasets, statistical and evolutionary models and emerging molecular techniques are all integrated to get a systemic view of the causes and consequences of evolution. The integration of population genomics with other phenotypic multiomics data is the necessary step to gain a global picture of how adaptation occurs in nature.
Styles APA, Harvard, Vancouver, ISO, etc.
19

Samuel, John. « Feeding a data warehouse with data coming from web services. A mediation approach for the DaWeS prototype ». Thesis, Clermont-Ferrand 2, 2014. http://www.theses.fr/2014CLF22493/document.

Texte intégral
Résumé :
Cette thèse traite de l’établissement d’une plateforme logicielle nommée DaWeS permettant le déploiement et la gestion en ligne d’entrepôts de données alimentés par des données provenant de services web et personnalisés à destination des petites et moyennes entreprises. Ce travail s’articule autour du développement et de l’expérimentation de DaWeS. L’idée principale implémentée dans DaWeS est l’utilisation d’une approche virtuelle d’intégration de données (la médiation) en tant queprocessus ETL (extraction, transformation et chargement des données) pour les entrepôts de données gérés par DaWeS. A cette fin, un algorithme classique de réécriture de requêtes (l’algorithme inverse-rules) a été adapté et testé. Une étude théorique sur la sémantique des requêtes conjonctives et datalog exprimées avec des relations munies de limitations d’accès (correspondant aux services web) a été menée. Cette dernière permet l’obtention de bornes supérieures sur les nombres d’appels aux services web requis dans l’évaluation de telles requêtes. Des expérimentations ont été menées sur des services web réels dans trois domaines : le marketing en ligne, la gestion de projets et les services d’aide aux utilisateurs. Une première série de tests aléatoires a été effectuée pour tester le passage à l’échelle
The role of data warehouse for business analytics cannot be undermined for any enterprise, irrespective of its size. But the growing dependence on web services has resulted in a situation where the enterprise data is managed by multiple autonomous and heterogeneous service providers. We present our approach and its associated prototype DaWeS [Samuel, 2014; Samuel and Rey, 2014; Samuel et al., 2014], a DAta warehouse fed with data coming from WEb Services to extract, transform and store enterprise data from web services and to build performance indicators from them (stored enterprise data) hiding from the end users the heterogeneity of the numerous underlying web services. Its ETL process is grounded on a mediation approach usually used in data integration. This enables DaWeS (i) to be fully configurable in a declarative manner only (XML, XSLT, SQL, datalog) and (ii) to make part of the warehouse schema dynamic so it can be easily updated. (i) and (ii) allow DaWeS managers to shift from development to administration when they want to connect to new web services or to update the APIs (Application programming interfaces) of already connected ones. The aim is to make DaWeS scalable and adaptable to smoothly face the ever-changing and growing web services offer. We point out the fact that this also enables DaWeS to be used with the vast majority of actual web service interfaces defined with basic technologies only (HTTP, REST, XML and JSON) and not with more advanced standards (WSDL, WADL, hRESTS or SAWSDL) since these more advanced standards are not widely used yet to describe real web services. In terms of applications, the aim is to allow a DaWeS administrator to provide to small and medium companies a service to store and query their business data coming from their usage of third-party services, without having to manage their own warehouse. In particular, DaWeS enables the easy design (as SQL Queries) of personalized performance indicators. We present in detail this mediation approach for ETL and the architecture of DaWeS. Besides its industrial purpose, working on building DaWeS brought forth further scientific challenges like the need for optimizing the number of web service API operation calls or handling incomplete information. We propose a bound on the number of calls to web services. This bound is a tool to compare future optimization techniques. We also present a heuristics to handle incomplete information
Styles APA, Harvard, Vancouver, ISO, etc.
20

Werlang, Ricardo. « Ontology-based approach for standard formats integration in reservoir modeling ». reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2015. http://hdl.handle.net/10183/115196.

Texte intégral
Résumé :
A integração de dados oriundos de fontes autônomas e heterogêneas ainda é um grande problema para diversas aplicações. Na indústria de petróleo e gás, uma grande quantidade de dados é gerada diariamente a partir de múltiplas fontes, tais como dados sísmicos, dados de poços, dados de perfuração, dados de transporte e dados de marketing. No entanto, estes dados são adquiridos através da aplicação de diferentes técnicas e representados em diferentes formatos e padrões. Assim, estes dados existem de formas estruturadas em banco de dados e de formas semi-estruturadas em planilhas e documentos, tais como relatórios e coleções multimídia. Para lidar com a heterogeneidade dos formatos de dados, a informação precisa ser padronizada e integrada em todos os sistemas, disciplinas e fronteiras organizacionais. Como resultado, este processo de integração permitirá uma melhor tomada de decisão dentro de colaborações, uma vez que dados de alta qualidade poderão ser acessados em tempo hábil. A indústria do petróleo depende do uso eficiente desses dados para a construção de modelos computacionais, a fim de simplificar a realidade geológica e para ajudar a compreende-la. Tal modelo, que contém objetos geológicos analisados por diferentes profissionais—geólogos, geofísicos e engenheiros — não representa a realidade propriamente dita, mas a conceitualização do especialista. Como resultado, os objetos geológicos modelados assumem representações semânticas distintas e complementares no apoio à tomada de decisões. Para manter os significados pretendidos originalmente, ontologias estão sendo usadas para explicitar a semântica dos modelos e para integrar os dados e arquivos gerados nas etapas da cadeia de exploração. A principal reivindicação deste trabalho é que a interoperabilidade entre modelos da terra construídos e manipulados por diferentes profissionais e sistemas pode ser alcançada evidenciando o significado dos objetos geológicos representados nos modelos. Nós mostramos que ontologias de domínio desenvolvidas com o apoio de conceitos teórico de ontologias de fundamentação demonstraram ser uma ferramenta adequada para esclarecer a semântica dos conceitos geológicos. Nós exemplificamos essa capacidade através da análise dos formatos de comunicação padrões mais utilizados na cadeia de modelagem (LAS, WITSML e RESQML), em busca de entidades semanticamente relacionadas com os conceitos geológicos descritos em ontologias de Geociências. Mostramos como as noções de identidade, rigidez, essencialidade e unidade, aplicadas a conceitos ontológicos, conduzem o modelador à definir mais precisamente os objetos geológicos no modelo. Ao tornar explícitas as propriedades de identidade dos objetos modelados, o modelador pode superar as ambiguidades da terminologia geológica. Ao fazer isso, explicitamos os objetos e propriedades relevantes que podem ser mapeados a partir de um modelo para outro, mesmo quando eles estão representados em diferentes nomes e formatos.
The integration of data issued from autonomous and heterogeneous sources is still a significant problem for an important number of applications. In the oil and gas industry, a large amount of data is generated every day from multiple sources such as seismic data, well data, drilling data, transportation data, and marketing data. However, these data are acquired by the application of different techniques and represented in different standards and formats. Thus, these data exist in a structured form in databases, and in semi-structured forms in spreadsheets and documents such as reports and multimedia collections. To deal with this large amount of information, as well as the heterogeneous data formats of the data, the information needs to be standardized and integrated across systems, disciplines and organizational boundaries. As a result, this information integration will enable better decision making within collaborations, once high quality data will be accessible timely. The petroleum industry depends on the efficient use of these data to the construction of computer models in order to simplify the geological reality and to help understanding it. Such a model, which contains geological objects analyzed by different professionals – geologists, geophysicists and engineers – does not represent the reality itself, but the expert’s conceptualization. As a result, the geological objects modeled assume distinct semantic representations and complementary in supporting decision-making. For keeping the original intended meanings, ontologies were used for expliciting the semantic of the models and for integrating the data and files generated in the various stages of the exploration chain. The major claim of this work is that interoperability among earth models built and manipulated by different professionals and systems can be achieved by making apparent the meaning of the geological objects represented in the models. We show that domain ontologies developed with support of theoretical background of foundational ontologies show to be an adequate tool to clarify the semantic of geology concepts. We exemplify this capability by analyzing the communication standard formats most used in the modeling chain (LAS,WITSML, and RESQML), searching for entities semantically related with the geological concepts described in ontologies for Geosciences. We show how the notions of identity, rigidity, essentiality and unity applied to ontological concepts lead the modeler to more precisely define the geological objects in the model. By making explicit the identity properties of the modeled objects, the modeler who applies data standards can overcome the ambiguities of the geological terminology. In doing that, we clarify which are the relevant objects and properties that can be mapped from one model to another, even when they are represented with different names and formats.
Styles APA, Harvard, Vancouver, ISO, etc.
21

Prieto, Barja Pablo 1986. « NGS applications in genome evolution and adaptation : A reproducible approach to NGS data analysis and integration ». Doctoral thesis, Universitat Pompeu Fabra, 2017. http://hdl.handle.net/10803/565601.

Texte intégral
Résumé :
In this PhD I have used NGS technologies in different organisms and scenarios such as in ENCODE, comparing the conservation and evolution of long non-coding RNA sequences between human and mouse, using experimental evidences from genome, transcriptome and chromatin. A similar approach was followed in other organisms such as the mesoamerican common bean and in chicken. Other analysis carried with NGS data involved the well known parasite, Leishmania Donovani, the causative agent of Leishmaniasis. I used NGS data obtained from genome and transcriptome to study the fate of its genome in survival strategies for adaptation and long term evolution. All this work was approached while working in tools and strategies to efficiently design and implement the bioinformatics analysis also known as pipelines or workflows, in order to make them easy to use, easily deployable, accessible and highly performing. This work has provided several strategies in order to avoid lack of reproducibility and inconsistency in scientific research with real biological applications towards sequence analysis and genome evolution.
En aquest doctorat he utilitzat tecnologies NGS en diferents organismes i projectes com l'ENCODE, comparant la conservació i evolució de seqüències de RNA llargs no codificant entre el ratolí i l'humà, utilitzant evidències experimentals del genoma, transcriptoma i cromatina. He seguit una estratègia similar en altres organismes com són la mongeta mesoamericana i el pollastre. En altres anàlisis he hagut d'utilitzar dades NGS en l'estudi del conegut paràsit leishmània Donovani, l'agent causatiu de la malaltia Leishmaniosis. Utilitzant dades NGS obtingudes del genoma i transcriptoma he estudiat les conseqüències del genoma en estratègies d'adaptació i evolució a llarg termini. Aquest treball es va realitzar mentre treballava en eines i estratègies per dissenyar eficientment i implementar els anàlisis bioinformàtics coneguts com a diagrames de treball, per tal de fer-los fàcils d'utilitzar, fàcilment realitzables, accessibles i amb un alt rendiment. Aquest treball present diverses estratègies per tal d'evitar la falta de reproductibilitat i consistència en la investigació científica amb aplicacions reals a la biologia de l'anàlisi de seqüències i evolució de genomes.
Styles APA, Harvard, Vancouver, ISO, etc.
22

Saal, Petronella Elize. « Integrating computers into mathematics education in South African Schools ». Diss., University of Pretoria, 2017. http://hdl.handle.net/2263/62904.

Texte intégral
Résumé :
The purpose of the study was to determine how South African mathematics teachers were integrating computers into their classrooms. The study was a response to the low achievement scores in mathematics as attained by grade nine learners in the 2011 Trends in International Mathematics and Science Study (TIMSS). TIMSS 2011 assessed Grade four and eight learners. However, South Africa as well as Botswana and Honduras opted to administer the Grade eight assessment to their Grade nine learners instead. South Africa’s Grade nine learners achieved an average score of 352 (35.2%) out of a possible 1 000 points. This quantitative secondary data analysis study utilised data collected from mathematics teachers from 298 schools in South Africa. The dataset was analysed using descriptive analysis that included percentages as well as the Pearson two-way Chi-square tabulations. The major finding of the study is that 73. 9% of South African mathematics teachers are still not integrating computers into mathematics education. Results showed that teachers are mostly using computers for preparation (35.5%) and administration (65.3%) purposes. Even though 45.5% of the teachers reported that they feel comfortable using computers, others feel that they are still in need of technical support. Moreover, the findings showed that 64.8% of the teachers do not attend professional development programmes that focus on the integration of Information Technology (IT) into mathematics.
Dissertation (MEd)--University of Pretoria, 2017.
Science, Mathematics and Technology Education
MEd
Unrestricted
Styles APA, Harvard, Vancouver, ISO, etc.
23

Swint, Galen Steen. « Clearwater an extensible, pliable, and customizable approach to code generation / ». Diss., Available online, Georgia Institute of Technology, 2006, 2006. http://etd.gatech.edu/theses/available/etd-07082006-012732/.

Texte intégral
Résumé :
Thesis (Ph. D.)--Computing, Georgia Institute of Technology, 2007.
Calton Pu, Committee Chair ; Ling Liu, Committee Member ; Karsten Schwan, Committee Member ; Olin Shivers, Committee Member ; Donald F. Ferguson, Committee Member.
Styles APA, Harvard, Vancouver, ISO, etc.
24

Watanabe, Toyohide, Yuuji Yoshida et Teruo Fukumura. « Editing model based on the object-oriented approach ». IEEE, 1988. http://hdl.handle.net/2237/6930.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
25

Tessier, Sean Michael. « Ontology-based approach to enable feature interoperability between CAD systems ». Thesis, Georgia Institute of Technology, 2011. http://hdl.handle.net/1853/41118.

Texte intégral
Résumé :
Data interoperability between computer-aided design (CAD) systems remains a major obstacle in the information integration and exchange in a collaborative engineering environment. The standards for CAD data exchange have remained largely restricted to geometric representations, causing the design intent portrayed through construction history, features, parameters, and constraints to be discarded in the exchange process. In this thesis, an ontology-based framework is proposed to allow for the full exchange of semantic feature data. A hybrid ontology approach is proposed, where a shared base ontology is used to convey the concepts that are common amongst different CAD systems, while local ontologies are used to represent the feature libraries of individual CAD systems as combinations of these shared concepts. A three-branch CAD feature model is constructed to reduce ambiguity in the construction of local ontology feature data. Boundary representation (B-Rep) data corresponding to the output of the feature operation is incorporated into the feature data to enhance data exchange. The Ontology Web Language (OWL) is used to construct a shared base ontology and a small feature library, which allows the use of existing ontology reasoning tools to infer new relationships and information between heterogeneous data. A combination of OWL and SWRL (Semantic Web Rule Language) rules are developed to allow a feature from an arbitrary source system expressed via the shared base ontology to be automatically classified and translated into the target system. These rules relate input parameters and reference types to expected B-Rep objects, allowing classification even when feature definitions vary or when little is known about the source system. In cases when the source system is well known, this approach also permits direct translation rules to be implemented. With such a flexible framework, a neutral feature exchange format could be developed.
Styles APA, Harvard, Vancouver, ISO, etc.
26

Phalak, Kashmira. « Cognitively-inspired Architecture for Wireless Sensor Networks : A Model Driven Approach for Data Integration in a Traffic Monitoring System ». Thesis, Virginia Tech, 2006. http://hdl.handle.net/10919/33675.

Texte intégral
Résumé :

We describe CoSMo, a Cognitively Inspired Service and Model Architecture for situational awareness and monitoring of vehicular traffic in urban transportation systems using a network of wireless sensors. The system architecture combines (i) a cognitively inspired internal representation for analyzing and answering queries concerning the observed system and (ii) a service oriented architecture that facilitates interaction among individual modules, of the internal representation, the observed system and the user. The cognitively inspired model architecture allows effective deductive as well as inductive reasoning by combining simulation based dynamic models for planning with traditional relational databases for knowledge and data representation. On the other hand the service oriented design of interaction allows one to build flexible, extensible and scalable systems that can be deployed in practical settings. To illustrate our concepts and the novel features of our architecture, we have recently completed a prototype implementation of CoSMo. The prototype illustrates advantages of our approach over other traditional approaches for designing scalable software for situational awareness in large complex systems. The basic architecture and its prototype implementation are generic and can be applied for monitoring other complex systems. This thesis describes the design of cognitively-inspired model architecture and its corresponding prototype. Two important contributions include the following:

  1. The cognitively-inspired architecture: In contrast to earlier work in model driven architecture, CoSMo contains a number of cognitively inspired features, including perception, memory and learning. Apart from illustrating interesting trade-offs between computational cost (e.g. access time, memory), and correctness available to a user, it also allows users specified deductive and inductive queries.
  2. Distributed Data Integration and Fusion: In keeping with the cognitively-inspired model-driven approach, the system allows for an efficient data fusion from heterogeneous sensors, simulation based dynamic models and databases that are continually updated with real world and simulated data. It is capable of supporting a rich class of queries.

Master of Science
Styles APA, Harvard, Vancouver, ISO, etc.
27

Du, Preez Jacobus Frederick. « The integration of informal minibus-taxi transport services into formal public transport planning and operations - A data driven approach ». Master's thesis, University of Cape Town, 2018. http://hdl.handle.net/11427/29885.

Texte intégral
Résumé :
The MiniBus Taxi (MBT) mode is poorly understood by planning and operational authorities, yet plays a big role in the economies of developing countries transporting the workforce to and from their places of employment and offering employment to thousands in the operations of these services, as well as the numerous rank-side services and amenities offered to patrons. In recent years, research focussed on mapping paratransit services, including MBTs, in cities of the developing world has contributed significantly to the understanding of the mode in terms of its spatial extent in its respective service areas. In South Africa, experience has shown that the wholesale replacement of MBTs with scheduled services is an unattainable goal. Instead, planning authorities and researchers have, more recently, shown interest in investigating feasible methods of integrating the scheduled and unscheduled services as hybrid planned-trunk and paratransit-feeder networks. The objective of this research is to present the case for simple methods of planning and carrying out onboard surveys of paratransit services to classify and to better understand the operations of individual routes, identified route classes, the network as a whole, as well as revealed passenger demand for the services and, ultimately, how this information can be wielded in the planning and implementation of hybrid routes or networks. The data central to this study consist of onboard captured MBT data, which was collected with a public transport data capturing application using GPS enabled smartphones in the City of Cape Town from April to August 2017 as part of a City of Cape Town’s Transport and Urban Development Authority (TDA) data collection project. The purpose of the project was to clarify the actual extent of MBT services within the City and to improve the representation of the MBT mode in the City of Cape Town’s travel demand model. An Android smartphone application, purpose-built for collecting operational information onboard public transport vehicles, was used to collect spatial and temporal data on the operations of a sample of active MBT routes in Cape Town. The application, which saw some functionality updates specifically for the project, was used to collect the following information per MBT trip: · Location of stops; · Time of arrival and departure at stops; · Number of passengers boarding and alighting at each stop; · The relative boarding and alighting stop of each specific passenger; · The amount paid in fare money per passenger at each stop; · The actual path travelled by the vehicle as a GPS route trace; and · The origin and destination route description of each route captured. It is estimated that there are more than 800 active and operational routes in the Cape Town. The objective of the data collection project was to survey each one of these routes for a prespecified number of trips. As the project was still underway when this research was carried out, the information listed above collected for a sample of trips for 278 routes (556 if the reverse direction is considered as a unique route designation) formed the basis of this study. During the course of this study, the analyses of these data have shown that while the operational characteristics of individual routes are relatively consistent and stable, it is possible to distinguish between different service typologies within the larger route network. From the raw data structure listed above, the operational characteristics that were calculated for each trip and aggregated at the route level included: · Trip and route distances; · Average operating speeds; · Travel times; · Number of stops per trip; · Load factors between stops along the route; and · Fare rates and trip revenues. In addition to the identification of the operational characteristics of the MBT network, service classes and routes, the outcomes of the study include providing a framework of methods for the collection, extraction, cleansing, analysis and visualisation of the data. It also includes the identification of metrics which are key in describing the difference in service types. The descriptive operational characteristics that were calculated for each trip record, inbound and outbound per route, were evaluated to establish whether they can be used to determine if different service typologies can be observed in the data. It was found that simple k-means clustering procedures may be used to classify the routes into separate, distinguishable service classes. For the purpose of this study, it was decided, nominally, that the classification should be executed for three classes. Three was subjectively considered a good value to be inclusive of traditional Trunk and Feeder or Distribution, route types as well as the possibility of the existence of a yet to be defined third type. The clustering procedures were carried out for different combinations of the operational variables for which the most consistent results were obtained for the combination distance – stop density1 – passenger turnover. Analysis of the within-class operational characteristics indicates that these three service classes clearly differ in terms of their stop frequencies, distances, speeds and their spatial network coverage. The study furthermore provides evidence that the understanding of the MBT network and sub-networks of service classes within this network, including its interaction with other public transport modes and infrastructure, provides planning and operating authorities with key information for effectively planning and implementing hybrid networks. Finally, the study demonstrates many additional insights can be garnered from these data by implementing improved statistical sampling and survey methods at the route level and by analysing aspects of the data that were not considered central to the research. These aspects include route adherence studies, origin – destination studies and methods of expanding the onboard data samples accurately by marrying it with data collected during static rank departure and arrival counts. Ultimately, the study shows that an unprecedented knowledge of the operations of MBT routes and networks may be obtained through detailed yet simple analysis of onboard data and that this knowledge may be very useful in the planning and operations of integrated public transport networks.
Styles APA, Harvard, Vancouver, ISO, etc.
28

SELICATI, VALERIA. « Innovative thermodynamic hybrid model-based and data-driven techniques for real time manufacturing sustainability assessment ». Doctoral thesis, Università degli studi della Basilicata, 2022. http://hdl.handle.net/11563/157566.

Texte intégral
Résumé :
This doctoral thesis is the result of the supervision and collaboration of the University of Basilicata, the Polytechnic of Bari, and the enterprise Master Italy s.r.l. The main research lines explored and discussed in the thesis are: sustainability in general and, more specifically, manufacturing sustainability, the Industry 4.0 paradigm linked to smart (green) manufacturing, model-based assessment techniques of manufacturing processes, and data-driven analysis methodologies. These seemingly unrelated topics are handled throughout the thesis in such a way that it reveal how strongly interwoven and characterised by transversality they are. The goal of the PhD programme was to design and validate innovative assessment models in order to investigate the nature of manufacturing processes and rationalize the relationships and correlations between the different stages of the process. This composite model may be utilized as a tool in political decision-making about the long-term development of industrial processes and the continuous improvement of manufacturing processes. The overarching goal of this research is to provide strategies for real-time monitoring of manufacturing performance and sustainability based on hybrid thermodynamic models of the first and second order, as well as those based on data and machine learning. The proposed model is tested on a real industrial case study using a systemic approach: the phases of identifying the requirements, data inventory (materials, energetic, geometric, physical, economic, social, qualitative, quantitative), modelling, analysis, ad hoc algorithm adjustment (tuning), implementation, and validation are developed for the aluminium alloy die-casting processes of Master Italy s.r.l., a southern Italian SME which designs and produces the accessories and metal components for windows since 1986. The thesis digs in the topic of the sustainability of smart industrial processes from each and every perspective, including both the quantity and quality of resources used throughout the manufacturing process's life cycle. Traditional sustainability analysis models (such as life cycle analysis, LCA) are combined with approaches based on the second law of thermodynamics (exergetic analysis); they are then complemented by models based on information technology (big-data analysis). A full analysis of the potential of each strategy, whether executed alone or in combination, is provided. Following a summary of the metrics relevant for determining the degree of sustainability of industrial processes, the case study is demonstrated using modelling and extensive analysis of the processes, namely aluminium alloy die casting. After assessing the sustainability of production processes using a model-based approach, we move on to the real-time application of machine learning analyses with the goal of identifying downtime and failures during the production cycle and predicting their occurrence well in advance using real-time process thermodynamic parameter values and automatic learning. Finally, the thesis suggests the use of integrated models on various case studies, such as laser deposition processes and the renovation of existing buildings, to demonstrate the multidisciplinarity and transversality of these issues. The thesis reveals fascinating findings derived from the use of a hybrid method to assessing the sustainability of manufacturing processes, combining exergetic analysis with life cycle assessment. The proposed theme is completely current and relevant to the most recent developments in the field of industrial sustainability, combining traditional model-based approaches with innovative approaches based on the collection of big data and its analysis using the most appropriate machine learning methodologies. Furthermore, the thesis demonstrates a highly promising application of machine learning approaches to real-time data collected in order to identify any fault source in the manufacturing line beginning with sustainability measures generated from exergetic analysis and life cycle analysis. As such, it unquestionably represents an advancement above earlier information depicted in the initial state of the art. In actuality, manufacturing companies that implement business strategies based on smart models and key enabling technologies today have a higher market value in terms of quality, customisation, flexibility, and sustainability.
Styles APA, Harvard, Vancouver, ISO, etc.
29

Koga, Ivo Kenji 1981. « An event-based approach to process environmental data = Um enfoque baseado em eventos para processar dados ambientais ». [s.n.], 2013. http://repositorio.unicamp.br/jspui/handle/REPOSIP/275618.

Texte intégral
Résumé :
Orientador: Claudia Maria Bauzer Medeiros
Tese (doutorado) - Universidade Estadual de Campinas, Instituto de Computação
Made available in DSpace on 2018-08-23T23:06:49Z (GMT). No. of bitstreams: 1 Koga_IvoKenji_D.pdf: 2109870 bytes, checksum: 7ac5400b2e71be3e15b3bdf5504e3adf (MD5) Previous issue date: 2013
Resumo: O resumo poderá ser visualizado no texto completo da tese digital
Abstract: The complete abstract is available with the full electronic document.
Doutorado
Ciência da Computação
Doutor em Ciência da Computação
Styles APA, Harvard, Vancouver, ISO, etc.
30

Singh, Shikhar. « An approach to automate the adaptor software generation for tool integration in Application/ Product Lifecycle Management tool chains ». Thesis, KTH, Skolan för informations- och kommunikationsteknik (ICT), 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-193919.

Texte intégral
Résumé :
An emerging problem in organisations is that there exist a large number of tools storing data that communicate with each other too often, throughout the process of an application or product development. However, no means of communication without the intervention of a central entity (usually a server) or storing the schema at a central repository exist. Accessing data among tools and linking them is tough and resource intensive. As part of the thesis, we develop a software (also referred to as ‘adaptor’ in the thesis), which, when implemented in the lifecycle management systems, integrates data seamlessly. This will eliminate the need of storing database schemas at a central repository and make the process of accessing data within tools less resource intensive. The adaptor acts as a wrapper to the tools and allows them to directly communicate with each other and exchange data. When using the developed adaptor for communicating data between various tools, the data in relational databases is first converted into RDF format and is then sent or received. Hence, RDF forms the crucial underlying concept on which the software will be based. The Resource description framework (RDF) provides the functionality of data integration irrespective of underlying schemas by treating data as resource and representing it as URIs. The model of RDF is a data model that is used for exchange and communication of data on the Internet and can be used in solving other real world problems like tool integration and automation of communication in relational databases. However, developing this adaptor for every tool requires understanding the individual schemas and structure of each of the tools’ database. This again requires a lot of effort for the developer of the adaptor. So, the main aim of the thesis will be to automate the development of such adaptors. With this automation, the need for anyone to manually assess the database and then develop the adaptor specific to the database is eliminated. Such adaptors and concepts can be used to implement similar solutions in other organisations faced with similar problems. In the end, the output of the thesis is an approachwhich automates the process of generating these adaptors.
Resource Description Framework (RDF) ger funktionaliteten av dataintegration, oberoende av underliggande scheman genom att behandla uppgifter som resurs och representerar det som URI. Modellen för Resource Description Framework är en datamodell som används för utbyte och kommunikation av uppgifter om Internet och kan användas för att lösa andra verkliga problem som integrationsverktyg och automatisering av kommunikation i relationsdatabaser. Ett växande problem i organisationer är att det finns ett stort antal verktyg som lagrar data och som kommunicerar med varandra alltför ofta, under hela processen för ett program eller produktutveckling. Men inga kommunikationsmedel utan ingripande av en central enhet (oftast en server) finns. Åtkomst av data mellan verktyg och länkningar mellan dem är resurskrävande. Som en del av avhandlingen utvecklar vi en programvara (även hänvisad till som "adapter" i avhandlingen), som integrerar data utan större problem. Detta kommer att eliminera behovet av att lagra databasscheman på en central lagringsplats och göra processen för att hämta data inom verktyg mindre resurskrävande. Detta kommer att ske efter beslut om en särskild strategi för att uppnå kommunikation mellan olika verktyg som kan vara en sammanslagning av många relevanta begrepp, genom studier av nya och kommande metoder som kan hjälpa i nämnda scenarier. Med den utvecklade programvaran konverteras först datat i relationsdatabaserna till RDF form och skickas och tas sedan emot i RDF format. Således utgör RDF det viktiga underliggande konceptet för programvaran. Det främsta målet med avhandlingen är att automatisera utvecklingen av ett sådant verktyg (adapter). Med denna automatisering elimineras behovet att av någon manuellt behöver utvärdera databasen och sedan utveckla adaptern enligt databasen. Ett sådant verktyg kan användas för att implementera liknande lösningar i andra organisationer som har liknande problem. Således är resultatet av avhandlingen en algoritm eller ett tillvägagångssätt för att automatisera processen av att skapa adaptern.
Styles APA, Harvard, Vancouver, ISO, etc.
31

Zhu, Junxiang. « Integration of Building Information Modelling and Geographic Information System at Data Level Using Semantics and Geometry Conversion Approach Towards Smart Infrastructure Management ». Thesis, Curtin University, 2018. http://hdl.handle.net/20.500.11937/74945.

Texte intégral
Résumé :
This study integrates Building Information Modelling (BIM)and Geographic Information System (GIS) at data level using an open source approach for geometry transformation and an automatic attribute searching algorithm for semantics transfer for the purpose of facilitating data transformation from BIM to GIS. Based on that, an infrastructure management system has been developed using Web GIS technology in conjunction with the models created by BIM and transformed into GIS using the proposed approach.
Styles APA, Harvard, Vancouver, ISO, etc.
32

Reda, Roberto. « A Semantic Web approach to ontology-based system : integrating, sharing and analysing IoT health and fitness data ». Master's thesis, Alma Mater Studiorum - Università di Bologna, 2017. http://amslaurea.unibo.it/14645/.

Texte intégral
Résumé :
With the rapid development of fitness industry, Internet of Things (IoT) technology is becoming one of the most popular trends for the health and fitness areas. IoT technologies have revolutionised the fitness and the sport industry by giving users the ability to monitor their health status and keep track of their training sessions. More and more sophisticated wearable devices, fitness trackers, smart watches and health mobile applications will appear in the near future. These systems do collect data non-stop from sensors and upload them to the Cloud. However, from a data-centric perspective the landscape of IoT fitness devices and wellness appliances is characterised by a plethora of representation and serialisation formats. The high heterogeneity of IoT data representations and the lack of common accepted standards, keep data isolated within each single system, preventing users and health professionals from having an integrated view of the various information collected. Moreover, in order to fully exploit the potential of the large amounts of data, it is also necessary to enable advanced analytics over it, thus achieving actionable knowledge. Therefore, due the above situation, the aim of this thesis project is to design and implement an ontology based system to (1) allow data interoperability among heterogeneous IoT fitness and wellness devices, (2) facilitate the integration and the sharing of information and (3) enable advanced analytics over the collected data (Cognitive Computing). The novelty of the proposed solution lies in exploiting Semantic Web technologies to formally describe the meaning of the data collected by the IoT devices and define a common communication strategy for information representation and exchange.
Styles APA, Harvard, Vancouver, ISO, etc.
33

Ruppert, Jan Christian [Verfasser], Michael [Akademischer Betreuer] Bonkowski, Hartmut [Akademischer Betreuer] Arndt et Stefan [Akademischer Betreuer] Porembski. « Advancing Functional Understanding of Primary Production in Drylands : Insights from a Data-Integration Approach / Jan Christian Ruppert. Gutachter : Michael Bonkowski ; Hartmut Arndt ; Stefan Porembski ». Köln : Universitäts- und Stadtbibliothek Köln, 2014. http://d-nb.info/1054445044/34.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
34

Dabous, Feras Taleb Abdel Rahman School of Information Systems Technology &amp Management UNSW. « A pattern based approach for the architectural design of e-business applications ». Awarded by:University of New South Wales. School of Information Systems Technology and Management, 2005. http://handle.unsw.edu.au/1959.4/22047.

Texte intégral
Résumé :
With the widespread use of the Internet and its associated technologies, enterprises have to evolve in the way they are conducting business. 'e-business applications' refer to a new class of distributed applications that involves the Internet as a communication platform. Each e-business application supports the full automation of business processes that can span across multiple enterprises. For a given application domain that involves e-business application development, a number of design decisions that best fullfil stakeholders requirements have to be made. One important issue is the reuse of functionality which exists within legacy systems that can belong to one or more enterprises within the same domain. Most existing design approaches are inadequate in supporting the exploration of all design combinations. Moreover, there is little work on how to identify the best design decisions systematically for a given application domain. In this thesis we present a pattern-based approach that addresses the architectural design of e-business applications. We identify a number of architectural patterns whose instantiation on a given design problem correspond to different architectural design alternatives. We also identify models that enable the estimation of quality attributes for such alternatives. Then we investigate and utilise methods to select the best pattern for a given design problem. We also describe the process of generating the alternative architectures, estimating their qualities, and then ranking them with respect to any quality attribute or a combination of quality attributes. We validate this approach on a real life case study in the area of capital markets. The case study concerns realistic e-business applications that rely on existing legacy applications. The validation exercise has produced predictions which have been compared with actual design decisions that have been made. The thesis also proposes a framework for the systematic identification of architectural patterns. An additional set of architectural patterns and their impact on the case study are discussed. The main contribution of this thesis is in the identification of patterns and quality attributes models for the architectural design of e-business applications that aid in the systematic selection of the most appropriate architectural pattern for a given problem context.
Styles APA, Harvard, Vancouver, ISO, etc.
35

Wang, Penghao. « An integrative approach for phylogenetic inference ». Thesis, The University of Sydney, 2009. https://hdl.handle.net/2123/28166.

Texte intégral
Résumé :
Molecular evolution is the key to explain the divergence of species and the origin of life on earth. The main task in the study of molecular evolution is the reconstruction of evolutionary trees from sequences data of current species. This is a huge, complicated and highly contentious field and in the literature a large number of phylogenetic reconstruction methods have been proposed to address such a problem. This thesis first researches into the mathematical and statistical foundations of the subject and discusses the limitations of existing methods. All the methods reconstruct the phylogeny optimising an explicit criterion. These criteria attempt to explain the real evolutionary process by simplified assumptions under a certain number of models and parameters, all of which are more or less biased in practice. Therefore, all current available phylogenetic reconstruction methods have limitations. This has become a serious issue for many important real-life applications which demand very accurate results. Therefore, it is desirable to have an integrative approach which can effectively incorporate different methods, useful information and analytical results from various sources during dynamic computation and to allow biologists to interact with the computation to control the quality of the analysis at various processing stages. In this way, we may achieve biologically higher accuracy and reliability for large and complex real-life problems. This thesis presents a novel integrative and interactive distributed high performance computing system for phylogenetic analysis. It is a cross-disciplinary project and involves sophisticated research, design and development of algorithms and systems: ( l) the design of new algorithms which are able to effectively integrate heterogeneous information from various computational and experimental studies, (2) the design of new service-oriented high performance computing technologies which enable efficient dynamic integration and interaction, and (3) the development of a new distributed high performance computing platform which allows biologists to interact effectively with the computation to control the quality of the analysis in solving their large and complex problems. This work also provides a preliminary algorithmic framework for the coestimation of multiple sequence alignment and phylogenetic reconstruction.
Styles APA, Harvard, Vancouver, ISO, etc.
36

Guan, Xiaowei. « Bioinformatics Approaches to Heterogeneous Omic Data Integration ». Case Western Reserve University School of Graduate Studies / OhioLINK, 2012. http://rave.ohiolink.edu/etdc/view?acc_num=case1340302883.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
37

Jana, Vimbai Lisa Michelle. « Adopting a harmonised regional approach to customs regulation for the tripartite free trade agreement ». Thesis, University of the Western Cape, 2013. http://etd.uwc.ac.za/index.php?module=etd&action=viewtitle&id=gen8Srv25Nme4_8861_1380710167.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
38

Xiao, Hui. « Network-based approaches for multi-omic data integration ». Thesis, University of Cambridge, 2019. https://www.repository.cam.ac.uk/handle/1810/289716.

Texte intégral
Résumé :
The advent of advanced high-throughput biological technologies provides opportunities to measure the whole genome at different molecular levels in biological systems, which produces different types of omic data such as genome, epigenome, transcriptome, translatome, proteome, metabolome and interactome. Biological systems are highly dynamic and complex mechanisms which involve not only the within-level functionality but also the between-level regulation. In order to uncover the complexity of biological systems, it is desirable to integrate multi-omic data to transform the multiple level data into biological knowledge about the underlying mechanisms. Due to the heterogeneity and high-dimension of multi-omic data, it is necessary to develop effective and efficient methods for multi-omic data integration. This thesis aims to develop efficient approaches for multi-omic data integration using machine learning methods and network theory. We assume that a biological system can be represented by a network with nodes denoting molecules and edges indicating functional links between molecules, in which multi-omic data can be integrated as attributes of nodes and edges. We propose four network-based approaches for multi-omic data integration using machine learning methods. Firstly, we propose an approach for gene module detection by integrating multi-condition transcriptome data and interactome data using network overlapping module detection method. We apply the approach to study the transcriptome data of human pre-implantation embryos across multiple development stages, and identify several stage-specific dynamic functional modules and genes which provide interesting biological insights. We evaluate the reproducibility of the modules by comparing with some other widely used methods and show that the intra-module genes are significantly overlapped between the different methods. Secondly, we propose an approach for gene module detection by integrating transcriptome, translatome, and interactome data using multilayer network. We apply the approach to study the ribosome profiling data of mTOR perturbed human prostate cancer cells and mine several translation efficiency regulated modules associated with mTOR perturbation. We develop an R package, TERM, for implementation of the proposed approach which offers a useful tool for the research field. Next, we propose an approach for feature selection by integrating transcriptome and interactome data using network-constrained regression. We develop a more efficient network-constrained regression method eGBL. We evaluate its performance in term of variable selection and prediction, and show that eGBL outperforms the other related regression methods. With application on the transcriptome data of human blastocysts, we select several interested genes associated with time-lapse parameters. Finally, we propose an approach for classification by integrating epigenome and transcriptome data using neural networks. We introduce a superlayer neural network (SNN) model which learns DNA methylation and gene expression data parallelly in superlayers but with cross-connections allowing crosstalks between them. We evaluate its performance on human breast cancer classification. The SNN provides superior performances and outperforms several other common machine learning methods. The approaches proposed in this thesis offer effective and efficient solutions for integration of heterogeneous high-dimensional datasets, which can be easily applied to other datasets presenting the similar structures. They are therefore applicable to many fields including but not limited to Bioinformatics and Computer Science.
Styles APA, Harvard, Vancouver, ISO, etc.
39

Zheng, Feng. « A data exchange approach to integrating autonomous manufacturing databases ». Thesis, Loughborough University, 1996. https://dspace.lboro.ac.uk/2134/27164.

Texte intégral
Résumé :
It has become evident in recent years that due to the complexity and diversity of computer applications in manufacturing environments, not only will previously established 'islands of automation' continue to exist, but new islands will emerge as a result of system expansion and technical renovation. Therefore, it is vitally important that systems integration methods are capable of supporting pre-existing manufacturing application systems as well as incremental system growth.
Styles APA, Harvard, Vancouver, ISO, etc.
40

Tous, Ruben. « Data integration with XML and semantic web technologies novel approaches in the design of modern data integration systems ». Saarbrücken VDM Verlag Dr. Müller, 2006. http://d-nb.info/991303105/04.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
41

He, Zhong. « Integration of dynamic data into reservoir description using streamline approaches ». Texas A&M University, 2003. http://hdl.handle.net/1969.1/1188.

Texte intégral
Résumé :
Integration of dynamic data is critical for reliable reservoir description and has been an outstanding challenge for the petroleum industry. This work develops practical dynamic data integration techniques using streamline approaches to condition static geological models to various kinds of dynamic data, including two-phase production history, interference pressure observations and primary production data. The proposed techniques are computationally efficient and robust, and thus well-suited for large-scale field applications. We can account for realistic field conditions, such as gravity, and changing field conditions, arising from infill drilling, pattern conversion, and recompletion, etc., during the integration of two-phase production data. Our approach is fast and exhibits rapid convergence even when the initial model is far from the solution. The power and practical applicability of the proposed techniques are demonstrated with a variety of field examples. To integrate two-phase production data, a travel-time inversion analogous to seismic inversion is adopted. We extend the method via a 'generalized travel-time' inversion to ensure matching of the entire production response rather than just a single time point while retaining most of the quasi-linear property of travel-time inversion. To integrate the interference pressure data, we propose an alternating procedure of travel-time inversion and peak amplitude inversion or pressure inversion to improve the overall matching of the pressure response. A key component of the proposed techniques is the efficient computation of the sensitivities of dynamic responses with respect to reservoir parameters. These sensitivities are calculated analytically using a single forward simulation. Thus, our methods can be orders of magnitude faster than finite-difference based numerical approaches that require multiple forward simulations. Streamline approach has also been extended to identify reservoir compartmentalization and flow barriers using primary production data in conjunction with decline type-curve analysis. The streamline 'diffusive' time of flight provides an effective way to calculate the drainage volume in 3D heterogeneous reservoirs. The flow barriers and reservoir compartmentalization are inferred based on the matching of drainage volumes from streamline-based calculation and decline type-curve analysis. The proposed approach is well-suited for application in the early stages of field development with limited well data and has been illustrated using a field example from the Gulf of Mexico.
Styles APA, Harvard, Vancouver, ISO, etc.
42

PATRIZI, SARA. « Multi-omics approaches to complex diseases in children ». Doctoral thesis, Università degli Studi di Trieste, 2022. http://hdl.handle.net/11368/3015193.

Texte intégral
Résumé :
Le tecnologie “-omiche” studiano l’insieme delle molecole presenti nel campione biologico di interesse, in maniera completamente agnostica. L’integrazione di diversi tipi di dati omici, chiamata “multi-omica” o “omica verticale”, fornisce indicazioni importanti su come le cause di una malattia portano alle sue conseguenze funzionali. Queste indicazioni sono particolarmente utili nel caso delle malattie complesse, che sono causate dall’interazione di vari fattori genetici e regolatori con vari contributi ambientali. In questo lavoro, degli approcci multi-omici appropriati sono stati applicati a due malattie complesse che di solito iniziano a manifestarsi durante l’infanzia, hanno un’incidenza crescente, e hanno vari elementi sconosciuti nella loro patologia molecolare, ovvero le malformazioni polmonari congenite e la celiachia. Gli scopi dei due progetti sono, rispettivamente, di verificare se nel tessuto polmonare malformato ci sono varianti genetiche o alterazioni della metilazione del DNA associate al cancro, e di trovare alterazioni comuni nel metiloma e nel trascrittoma di cellule epiteliali dell’intestino tenue di bambini affetti da celiachia. Per quanto riguarda i metodi, nel progetto sulle malformazioni polmonari sono stati usati microarray di metilazione whole genome e sequenziamento dell’intero genoma, mentre nel progetto sulla celiachia sono stati usati microarray di metilazione whole genome e sequenziamento dell’mRNA totale. In tutte le 20 malformazioni polmonari incluse nello studio sono state trovate regioni differenzialmente metilate in geni probabilmente legati al cancro del polmone. Inoltre, 5 campioni malformati avevano almeno una variante somatica missenso in un gene noto come driver del tumore del polmone, e 5 altri campioni avevano un totale di 2 delezioni di oncosoppressori driver del tumore del polmone e 10 amplificazioni di oncogeni driver del tumore del polmone. Questi dati suggeriscono che le malformazioni polmonari congenite possono avere alterazioni genetiche ed epigenetiche di tipo pre-maligno, la cui presenza è impossibile da prevedere sulla base delle sole informazioni cliniche. Nel secondo progetto, una Principal Component Analysis dei dati di metilazione ha mostrato che i pazienti celiaci si dividono in due cluster, di cui uno si sovrappone ai controlli. 174 geni erano differenzialmente metilati rispetto ai controlli in entrambi i cluster. Una Principal Component Analysis dei dati di espressione genica (mRNA-Seq) ha mostrato una distribuzione simile a quella dei dati di metilazione, e 442 geni erano differenzialmente espressi in entrambi i cluster. Sei geni, principalmente coinvolti nella risposta interferonica e nel processo di processamento e presentazione degli antigeni, erano sia differenzialmente espressi che differenzialmente metilati in entrambi i cluster. Questi risultati indicano che le cellule epiteliali dell’intestino tenue di bambini affetti da celiachia sono altamente variabili da un punto di vista molecolare, ma condividono delle differenze fondamentali che le rendono in grado di rispondere agli interferoni e di processare e presentare antigeni con maggiore efficienza rispetto ai controlli. Nonostante le loro limitazioni, gli studi presentati mostrano che degli approcci multi-omici specifici possono essere usati per rispondere alle domande ancora aperte riguardo a diverse malattie, studiando più funzioni cellulari contemporaneamente e spesso portando anche alla generazione di nuove ipotesi e a scoperte inaspettate.
“-Omic” technologies can detect the entirety of the molecules in the biological sample of interest, in a non-targeted and non-biased fashion. The integration of multiple types of omics data, known as “multi-omics” or “vertical omics”, can provide a better understanding of how the cause of disease leads to its functional consequences, which is particularly valuable in the study of complex diseases, that are caused by the interaction of multiple genetic and regulatory factors with contributions from the environment. In the present work appropriate multi-omics approaches are applied to two complex conditions that usually first manifest in childhood, have rising incidence and gaps in the knowledge of their molecular pathology, specifically Congenital Lung Malformations and Coeliac Disease. The aims are, respectively, to verify if cancer-associated genomic variants or DNA methylation features exist in the malformed lung tissue and to find common alterations in the methylome and the transcriptome of small intestine epithelial cells of children with CD. The methods used in the Congenital Lung Malformations project are Whole Genome Methylation microarrays and Whole Genome Sequencing, and for the Coeliac Disease the whole genome methylation microarrays and mRNA sequencing. Differentially methylated regions in possibly cancer-related genes were found in each one of the 20 lung malformation samples included. Moreover, 5 malformed samples had at least one somatic missense single nucleotide variant in genes known as lung cancer drivers, and 5 malformed samples had a total of 2 deletions of lung cancer driver tumour suppressor and 10 amplifications of lung cancer driver oncogenes. The data showed that congenital lung malformations can have premalignant genetic and epigenetic features, that are impossible to predict with clinical information only. In the second project, Principal Component Analysis of the whole genome methylation data showed that CD patients divide into two clusters, one of which overlaps with controls. 174 genes were differentially methylated compared to the controls in both clusters. Principal Component Analysis of gene expression data (mRNA-Seq) showed a distribution that is similar to the methylation data, and 442 genes were differentially expressed in both clusters. Six genes, mainly related to interferon response and antigen processing and presentation, were differentially expressed and methylated in both clusters. These results show that the intestinal epithelial cells of individuals with CD are highly variable from a molecular point of view, but they share some fundamental differences that make them able to respond to interferons, process, and present antigens more efficiently than controls. Despite the limitations of the present studies, they have shown that targeted multi-omics approaches can be set up to answer the relevant disease-specific questions by investigating many cellular functions at once, often generating new hypotheses and making unexpected discoveries in the process.
Styles APA, Harvard, Vancouver, ISO, etc.
43

Chen, Wen-Chih. « Integrating approaches to efficiency and productivity measurement ». Diss., Georgia Institute of Technology, 2003. http://hdl.handle.net/1853/25422.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
44

Laurén, Carl-Fredrik. « Current approaches on how to acquire and integrate external data into Data Warehouses ». Thesis, University of Skövde, Department of Computer Science, 2003. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-754.

Texte intégral
Résumé :

Integration of internal data is often mentioned in literature as the most demanding task when building or maintaining a DW. There is no literature that outlines the approach for the integration of external data into a DW. The integration of external data has increased during the last years enabling corporations to understand the opportunities in the market and to be able to better plan for future success of the corporation. The aim of this work is to exploratory outline current approaches for acquiring and integrating external data into DW and to give a brief overview of the future trends for external data integration. This aim was researched using an interview study. The results show that how to integrate external data is depending on what the corporations purpose with the external data is. Additional results show that how to integrate external data also depends on how the data is acquired.

Styles APA, Harvard, Vancouver, ISO, etc.
45

Do, Hong-Hai. « Schema matching and mapping based data integration architecture, approaches and evaluation ». Saarbrücken VDM, Müller, 2006. http://deposit.d-nb.de/cgi-bin/dokserv?id=2863983&prov=M&dok_var=1&dok_ext=htm.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
46

Fleischman, David R. « A data oriented approach to integrating manufacturing functions in flexible Manufacturing Systems ». Thesis, Monterey, California. Naval Postgraduate School, 1988. http://hdl.handle.net/10945/23150.

Texte intégral
Résumé :
Computer Integrated Manufacturing (CIM) seeks to integrate computers into the manufacturing environment, with the end result being a more efficient and productive factory. Current approaches to CIM generally fail to truly integrate the various manufacturing functions (design, scheduling, planning, manufacture, business, etc.) and instead result in self-sufficient, computer- served 'islands of automation.' In these systems, data must be translated before moving from one manufacturing function to another. Wu and Madison have approached data modeling in a CIM environment from a new perspective. Their approach seeks to provide one data model that meets the needs of all manufacturing functions within a factory, negating the need for human or machine data translators. Keywords: Computer integrated manufacturing; Flexible manufacturing systems; Data modeling
Styles APA, Harvard, Vancouver, ISO, etc.
47

Kettouch, Mohamed Salah. « A new approach for interlinking and integrating semi-structured and linked data ». Thesis, Anglia Ruskin University, 2017. http://arro.anglia.ac.uk/702722/.

Texte intégral
Résumé :
This work focuses on improving data integration and interlinking systems targeting semi-structured and Linked Data. It aims at facilitating the exploitation of semi-structured and Linked Data by addressing the problems of heterogeneity, complexity, scalability and the degree of automation. Technologies, such as the Resource Description Framework (RDF), enabled new data spaces and concept descriptors to define an increasing complex and heterogeneous web of data. Many data providers, however, continue to publish their data using classic models and formats. In addition, a significant amount of the data released before the existence of the Linked Data movement have not emigrated and still have a high value. Hence, as a long term solution, an interlinking system has been designed to contribute to the publishing of semi-structured data as Linked Data. Simultaneously, to utilise these growing data resource spaces, a data integration middleware has been proposed as an immediate solution. The proposed interlinking system verifies in the first place the existence of the Uniform Resource Identifier (URI) of the resource being published in the cloud in order to establish links with it. It uses the domain information in defining and matching the datasets. Its main aim is facilitating following best practice recommendations in publishing data into the Linked Data cloud. The results of this interlinking approach show that it can target large amounts of data whilst preserving good precision and recall. The new approach for integrating semi-structured and Linked Data is a mediator-based architecture. It enables the integration, on-the-fly, of semi-structured heterogeneous data sources with large-scale Linked Data sources. Complexity is tackled through a usable and expressive interface. The evaluation of the proposed architecture shows high performance, precision and adaptability.
Styles APA, Harvard, Vancouver, ISO, etc.
48

Gadaleta, Emanuela. « A multidisciplinary computational approach to model cancer-omics data : organising, integrating and mining multiple sources of data ». Thesis, Queen Mary, University of London, 2015. http://qmro.qmul.ac.uk/xmlui/handle/123456789/8141.

Texte intégral
Résumé :
It is imperative that the cancer research community has the means with which to effectively locate, access, manage, analyse and interpret the plethora of data values being generated by novel technologies. This thesis addresses this unmet requirement by using pancreatic cancer and breast cancer as prototype malignancies to develop a generic integrative transcriptomic model. The analytical workflow was initially applied to publicly available pancreatic cancer data from multiple experimental types. The transcriptomic landscape of comparative groups was examined both in isolation and relative to each other. The main observations included (i) a clear separation of profiles based on experimental type, (ii) identification of three subgroups within normal tissue samples resected adjacent to pancreatic cancer, each showing disruptions to biofunctions previously associated with pancreatic cancer (iii) and that cell lines and xenograft models are not representative of changes occurring during pancreatic tumourigenesis. Previous studies examined transcriptomic profiles across 306 biological and experimental samples, including breast cancer. The plethora of clinical and survival data readily available for breast cancer, compared to the paucity of publicly available pancreatic cancer data, allowed for expansion of the pipeline’s infrastructure to include functionalities for cross-platform and survival analysis. Application of this enhanced pipeline to multiple cohorts of triple negative and basal-like breast cancers identified differential risk groups within these breast cancer subtypes. All of the main experimental findings of this thesis are being integrated with the Pancreatic Expression Database and the Breast Cancer Campaign Tissue Bank bioinformatics portal, which enhances the sharing capacity of this information and ensures its exposure to a wider audience.
Styles APA, Harvard, Vancouver, ISO, etc.
49

Al, Shekaili Dhahi. « Integrating Linked Data search results using statistical relational learning approaches ». Thesis, University of Manchester, 2017. https://www.research.manchester.ac.uk/portal/en/theses/integrating-linked-data-search-results-using-statistical-relational-learning-approaches(3f77386b-a38a-4110-8ce1-bda6340e6f0b).html.

Texte intégral
Résumé :
Linked Data (LD) follows the web in providing low barriers to publication, and in deploying web-scale keyword search as a central way of identifying relevant data. As in the web, searchesinitially identify results in broadly the form in which they were published, and the published form may be provided to the user as the result of a search. This will be satisfactory in some cases, but the diversity of publishers means that the results of the search may be obtained from many different sources, and described in many different ways. As such, there seems to bean opportunity to add value to search results by providing userswith an integrated representation that brings together features from different sources. This involves an on-the-fly and automated data integration process being applied to search results, which raises the question as to what technologies might bemost suitable for supporting the integration of LD searchresults. In this thesis we take the view that the problem of integrating LD search results is best approached by assimilating different forms ofevidence that support the integration process. In particular, thisdissertation shows how Statistical Relational Learning (SRL) formalisms (viz., Markov Logic Networks (MLN) and Probabilistic Soft Logic (PSL)) can beexploited to assimilate different sources of evidence in a principledway and to beneficial effect for users. Specifically, in this dissertation weconsider syntactic evidence derived from LD search results and from matching algorithms, semantic evidence derived from LD vocabularies, and user evidence,in the form of feedback. This dissertation makes the following key contributions: (i) a characterisation of key features of LD search results that are relevant to their integration, and a description of some initial experiences in the use of MLN for interpreting search results; (ii)a PSL rule-base that models the uniform assimilation of diverse kinds of evidence;(iii) an empirical evaluation of how the contributed MLN and PSL approaches perform in terms of their ability to infer a structure for integrating LD search results;and (iv) concrete examples of how populating such inferred structures for presentation to the end user is beneficial, as well as guiding the collection of feedbackwhose assimilation further improves search results presentation.
Styles APA, Harvard, Vancouver, ISO, etc.
50

Mönchgesang, Susann [Verfasser]. « Metabolomics and biochemical omics data - integrative approaches : [kumulative Dissertation] / Susann Mönchgesang ». Halle, 2017. http://d-nb.info/1131075994/34.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
Nous offrons des réductions sur tous les plans premium pour les auteurs dont les œuvres sont incluses dans des sélections littéraires thématiques. Contactez-nous pour obtenir un code promo unique!

Vers la bibliographie