Зміст
Добірка наукової літератури з теми "Profilage des données"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся зі списками актуальних статей, книг, дисертацій, тез та інших наукових джерел на тему "Profilage des données".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Статті в журналах з теми "Profilage des données"
Nault, Geneviève, Émilie Couture-Glassco, and Katharine Larose-Hébert. "Le mal caché de la rue : le poids de l’étiquette." Reflets 22, no. 1 (July 28, 2016): 56–82. http://dx.doi.org/10.7202/1037163ar.
Повний текст джерелаReys, Victor, and Gilles Labesse. "Profilage in silico des inhibiteurs de protéine kinases." médecine/sciences 36 (October 2020): 38–41. http://dx.doi.org/10.1051/medsci/2020182.
Повний текст джерелаLouche, B., and V. Hallet. "Détermination de la structure tectonique de l'aquifère crayeux du littoral Nord Pas-de-Calais par prospection géophysique couplée à des observations par forage. Conséquence sur la répartition d'eau salée." Revue des sciences de l'eau 14, no. 3 (April 12, 2005): 265–80. http://dx.doi.org/10.7202/705420ar.
Повний текст джерелаDIEYE, Pape Issakha, Seni NDIAYE, Fode DIONE, Abdoulaye DIOP, Assane DIENG, Amadou DIOP, Bara NDIAYE, Yerim Mbagnick DIOP, and Serigne Omar SARR. "Étude corrélée de l’activité antibactérienne et antifongique des extraits de Jatropha chevalieri et de Cordylla pinnata, et de leurs profils chromatographiques." Journal of Applied Biosciences 159 (March 31, 2021): 16396–410. http://dx.doi.org/10.35759/jabs.159.4.
Повний текст джерелаDIEYE, Pape Issakha, Seni NDIAYE, Fode DIONE, Abdoulaye DIOP, Assane DIENG, Amadou DIOP, Bara NDIAYE, Yerim Mbagnick DIOP, and Serigne Omar SARR. "Étude corrélée de l’activité antibactérienne et antifongique des extraits de Jatropha chevalieri et de Cordylla pinnata, et de leurs profils chromatographiques." Journal of Applied Biosciences 159 (March 31, 2021): 16396–410. http://dx.doi.org/10.35759/jabs.159.4.
Повний текст джерелаEstabrooks, Carole A., Jeff W. Poss, Janet E. Squires, Gary F. Teare, Debra G. Morgan, Norma Stewart, Malcolm B. Doupe, Greta G. Cummings, and Peter G. Norton. "A Profile of Residents in Prairie Nursing Homes." Canadian Journal on Aging / La Revue canadienne du vieillissement 32, no. 3 (August 6, 2013): 223–31. http://dx.doi.org/10.1017/s0714980813000287.
Повний текст джерелаZIDAOUI, I., C. JOANNIS, J. WERTEL, S. ISEL, C. WEMMERT, J. VAZQUEZ, and M. DUFRESNE. "Utilisation de l’intelligence artificielle pour la validation des mesures en continu de la pollution des eaux usées." Techniques Sciences Méthodes 11 (November 21, 2022): 39–51. http://dx.doi.org/10.36904/tsm/202211039.
Повний текст джерелаD’Ascoli, Yannick, and Jean-Louis Berger. "Les déterminants du choix de carrière des enseignants de la formation professionnelle et leur relation aux caractéristiques sociodémographiques." Nouveaux cahiers de la recherche en éducation 15, no. 2 (September 20, 2013): 1–33. http://dx.doi.org/10.7202/1018455ar.
Повний текст джерелаRekwot, Z. G., O. Oyinbo, and N. P. Achi. "Poverty reduction among beef cattle value chain actors in North-West zone of Nigeria." Nigerian Journal of Animal Production 48, no. 6 (January 18, 2022): 363–73. http://dx.doi.org/10.51791/njap.v48i6.3324.
Повний текст джерелаEmery, Yves, and Armand Brice Kouadio. "Marque employeur et stratégies RH pour les employeurs publics. Le cas du bassin d’emploi Franco-Valdo-Genevois." Les nouvelles frontières du management public 21, no. 2 (October 16, 2018): 47–59. http://dx.doi.org/10.7202/1052686ar.
Повний текст джерелаДисертації з теми "Profilage des données"
Chevallier, Marc. "L’Apprentissage artificiel au service du profilage des données." Electronic Thesis or Diss., Paris 13, 2022. http://www.theses.fr/2022PA131060.
Повний текст джерелаThe digital transformation that has been rapidly happening within companies over the last few decades has led to a massive production of data. Once the problems related to the storage of those data have been solved, its use within Business Intelligence (BI) or Machine Learning (ML) has become a major objective for companies in order to make their data profitable. But the exploitation of the data is complex because it is not well documented and often contains many errors. It is in this context that the fields of data profiling and data quality (DQ) have become increasingly important. Profiling aims at extracting informative metadata from the data and data quality aims at quantifying the errors in the data.Profiling being a prerequisite to data quality, we have focused our work on this subject through the use of metadata vectors resulting from simple profiling actions. These simple information vectors have allowed us to perform advanced profiling tasks, in particular the prediction of complex semantic types using machine learning. The metadata vectors we used are large and are therefore affected by the curse of dimensionality. This term refers to a set of performance problems that occur in machine learning when the number of dimensions of the problem increases. One method to solve these problems is to use genetic algorithms to select a subset of dimensions with good properties. In this framework we have proposed improvements: on one hand, a non-random initialization of the individuals composing the initial population of the genetic algorithm, on the other hand, a modification to the genetic algorithm with aggressive mutations in order to improve its performance (GAAM)
Ben, Ellefi Mohamed. "La recommandation des jeux de données basée sur le profilage pour le liage des données RDF." Thesis, Montpellier, 2016. http://www.theses.fr/2016MONTT276/document.
Повний текст джерелаWith the emergence of the Web of Data, most notably Linked Open Data (LOD), an abundance of data has become available on the web. However, LOD datasets and their inherent subgraphs vary heavily with respect to their size, topic and domain coverage, the schemas and their data dynamicity (respectively schemas and metadata) over the time. To this extent, identifying suitable datasets, which meet specific criteria, has become an increasingly important, yet challenging task to supportissues such as entity retrieval or semantic search and data linking. Particularlywith respect to the interlinking issue, the current topology of the LOD cloud underlines the need for practical and efficient means to recommend suitable datasets: currently, only well-known reference graphs such as DBpedia (the most obvious target), YAGO or Freebase show a high amount of in-links, while there exists a long tail of potentially suitable yet under-recognized datasets. This problem is due to the semantic web tradition in dealing with "finding candidate datasets to link to", where data publishers are used to identify target datasets for interlinking.While an understanding of the nature of the content of specific datasets is a crucial prerequisite for the mentioned issues, we adopt in this dissertation the notion of "dataset profile" - a set of features that describe a dataset and allow the comparison of different datasets with regard to their represented characteristics. Our first research direction was to implement a collaborative filtering-like dataset recommendation approach, which exploits both existing dataset topic proles, as well as traditional dataset connectivity measures, in order to link LOD datasets into a global dataset-topic-graph. This approach relies on the LOD graph in order to learn the connectivity behaviour between LOD datasets. However, experiments have shown that the current topology of the LOD cloud group is far from being complete to be considered as a ground truth and consequently as learning data.Facing the limits the current topology of LOD (as learning data), our research has led to break away from the topic proles representation of "learn to rank" approach and to adopt a new approach for candidate datasets identication where the recommendation is based on the intensional profiles overlap between differentdatasets. By intensional profile, we understand the formal representation of a set of schema concept labels that best describe a dataset and can be potentially enriched by retrieving the corresponding textual descriptions. This representation provides richer contextual and semantic information and allows to compute efficiently and inexpensively similarities between proles. We identify schema overlap by the help of a semantico-frequential concept similarity measure and a ranking criterion based on the tf*idf cosine similarity. The experiments, conducted over all available linked datasets on the LOD cloud, show that our method achieves an average precision of up to 53% for a recall of 100%. Furthermore, our method returns the mappings between the schema concepts across datasets, a particularly useful input for the data linking step.In order to ensure a high quality representative datasets schema profiles, we introduce Datavore| a tool oriented towards metadata designers that provides rankedlists of vocabulary terms to reuse in data modeling process, together with additional metadata and cross-terms relations. The tool relies on the Linked Open Vocabulary (LOV) ecosystem for acquiring vocabularies and metadata and is made available for the community
Ammous, Karim. "Compression par profilage du code Java compilé pour les systèmes embarqués." Valenciennes, 2007. http://ged.univ-valenciennes.fr/nuxeo/site/esupversions/a56319aa-b36f-46ed-b617-a1464a995056.
Повний текст джерелаThe embedded systems are characterized by reduced hardware resources. Although these resources are constantly increasing, they remain insufficient. The memory space is one of the most critical resources. The compression of the code designed for embedded systems constitutes an interesting solution to reduce the memory footprint. Our study focuses on the compression of Java code represented by Java Class format files. Our contribution consists in designing and implementing a profiler based system in order to guide the compression of Java class files. Our profiler enables us to set up, on the basis of elementary compression techniques, an efficient compression strategy which delivers the best rate of compression. This strategy takes into consideration the features of the code given in input and dependencies between compression techniques. Our approach is based on four points: 1 - the study of the input files in order to extract the necessary information for the guidance of the compression process. 2 - the analysis of compression techniques dependencies in terms of effects produced by each technique to the others. To do so, we developed two methods: one numerical, based on the estimation of performance, the other analytical in order to determine whether there are common points between the different compression methods. 3 - the statistic performance assessment which allows to choose a strategy of compression: we have identified the parameters, related to each method, that enable this assessment. 4 - the definition of heuristics in order to identify the most efficient compression path in a research space characterized by an oriented graph
Ben, salem Aïcha. "Qualité contextuelle des données : détection et nettoyage guidés par la sémantique des données." Thesis, Sorbonne Paris Cité, 2015. http://www.theses.fr/2015USPCD054/document.
Повний текст джерелаNowadays, complex applications such as knowledge extraction, data mining, e-learning or web applications use heterogeneous and distributed data. The quality of any decision depends on the quality of the used data. The absence of rich, accurate and reliable data can potentially lead an organization to make bad decisions.The subject covered in this thesis aims at assisting the user in its quality ap-proach. The goal is to better extract, mix, interpret and reuse data. For this, the data must be related to its semantic meaning, data types, constraints and comments.The first part deals with the semantic schema recognition of a data source. This enables the extraction of data semantics from all the available information, inculding the data and the metadata. Firstly, it consists of categorizing the data by assigning it to a category and possibly a sub-category, and secondly, of establishing relations between columns and possibly discovering the semantics of the manipulated data source. These links detected between columns offer a better understanding of the source and the alternatives for correcting data. This approach allows automatic detection of a large number of syntactic and semantic anomalies.The second part is the data cleansing using the reports on anomalies returned by the first part. It allows corrections to be made within a column itself (data homogeni-zation), between columns (semantic dependencies), and between lines (eliminating duplicates and similar data). Throughout all this process, recommendations and analyses are provided to the user
Bakiri, Ali. "Développements informatiques de déréplication et de classification de données spectroscopiques pour le profilage métabolique d’extraits d'algues." Thesis, Reims, 2018. http://www.theses.fr/2018REIMS013.
Повний текст джерелаThe emergence of dereplication strategies as a new tool for the rapid identification of the natural products from complex natural extracts has unveiled a great need for cheminformatic tools for the treatment and analysis of the spectral data. The present thesis deals with the development of in silico dereplication methods based on Nuclear Magnetic Resonance (NMR). The first method, DerepCrud, is based on 13C NMR spectroscopy. It identifies the major compounds contained in a crude natural extract without any need for fractionation. The principle of the method is to compare the 13C NMR spectrum of the analyzed mixture to a series of 13C NMR chemical shifts of natural compounds stored in a local database. The second method, BCNet, is designed to exploit the richness of 2D NMR data (HMBC and HSQC) for the dereplication of the natural products. BCNet traces back the network formed by the HMBC correlations of the molecules present in a naturel extract, then isolates the groups of correlations belonging to the individual molecules using a community detection algorithm. The molecules are identified by searching these correlations within a locally constructed database that associates natural product structures and 2D NMR peak positions. Finally, the HSQC correlations of the molecules identified during the previous step are compared to the experimental HSQC correlations of the studied extract in order to increase the quality of identification accuracy
Lagraa, Sofiane. "New MP-SoC profiling tools based on data mining techniques." Thesis, Grenoble, 2014. http://www.theses.fr/2014GRENM026/document.
Повний текст джерелаMiniaturization of electronic components has led to the introduction of complex electronic systems which are integrated onto a single chip with multiprocessors, so-called Multi-Processor System-on-Chip (MPSoC). The majority of recent embedded systems are based on massively parallel MPSoC architectures, hence the necessity of developing embedded parallel applications. Embedded parallel application design becomes more challenging: It becomes a parallel programming for non-trivial heterogeneous multiprocessors with diverse communication architectures and design constraints such as hardware cost, power, and timeliness. A challenge faced by many developers is the profiling of embedded parallel applications so that they can scale over more and more cores. This is especially critical for embedded systems powered by MPSoC, where ever demanding applications have to run smoothly on numerous cores, each with modest power budget. Moreover, application performance does not necessarily improve as more cores are added. Application performance can be limited due to multiple bottlenecks including contention for shared resources such as caches and memory. It becomes time consuming for a developer to pinpoint in the source code the bottlenecks decreasing the performance. To overcome these issues, in this thesis, we propose a fully three automatic methods which detect the instructions of the code which lead to a lack of performance due to contention and scalability of processors on a chip. The methods are based on data mining techniques exploiting gigabytes of low level execution traces produced by MPSoC platforms. Our profiling approaches allow to quantify and pinpoint, automatically the bottlenecks in source code in order to aid the developers to optimize its embedded parallel application. We performed several experiments on several parallel application benchmarks. Our experiments show the accuracy of the proposed techniques, by quantifying and pinpointing the hotspot in the source code
Brunie, Hugo. "Optimisation des allocations de données pour des applications du Calcul Haute Performance sur une architecture à mémoires hétérogènes." Thesis, Bordeaux, 2019. http://www.theses.fr/2019BORD0014/document.
Повний текст джерелаHigh Performance Computing, which brings together all the players responsible for improving the computing performance of scientific applications on supercomputers, aims to achieve exaflopic performance. This race for performance is today characterized by the manufacture of heterogeneous machines in which each component is specialized. Among these components, system memories specialize too, and the trend is towards an architecture composed of several memories with complementary characteristics. The question arises then of these new machines use whose practical performance depends on the application data placement on the different memories. Compromising code update against performance is challenging. In this thesis, we have developed a data allocation on Heterogeneous Memory Architecture problem formulation. In this formulation, we have shown the benefit of a temporal analysis of the problem, because many studies were based solely on a spatial approach this result highlight their weakness. From this formulation, we developed an offline profiling tool to approximate the coefficients of the objective function in order to solve the allocation problem and optimize the allocation of data on a composite architecture composed of two main memories with complementary characteristics. In order to reduce the amount of code changes needed to execute an application according to our toolbox recommended allocation strategy, we have developed a tool that can automatically redirect data allocations from a minimum source code instrumentation. The performance gains obtained on mini-applications representative of the scientific applications coded by the community make it possible to assert that intelligent data allocation is necessary to fully benefit from heterogeneous memory resources. On some problem sizes, the gain between a naive data placement strategy, and an educated data allocation one, can reach up to ×3.75 speedup
Haine, Christopher. "Kernel optimization by layout restructuring." Thesis, Bordeaux, 2017. http://www.theses.fr/2017BORD0639/document.
Повний текст джерелаCareful data layout design is crucial for achieving high performance, as nowadays processors waste a considerable amount of time being stalled by memory transactions, and in particular spacial and temporal locality have to be optimized. However, data layout transformations is an area left largely unexplored by state-of-the-art compilers, due to the difficulty to evaluate the possible performance gains of transformations. Moreover, optimizing data layout is time-consuming, error-prone, and layout transformations are too numerous tobe experimented by hand in hope to discover a high performance version. We propose to guide application programmers through data layout restructuring with an extensive feedback, firstly by providing a comprehensive multidimensional description of the initial layout, built via analysis of memory traces collected from the application binary textit {in fine} aiming at pinpointing problematic strides at the instruction level, independently of theinput language. We choose to focus on layout transformations,translatable to C-formalism to aid user understanding, that we apply and assesson case study composed of two representative multithreaded real-lifeapplications, a cardiac wave simulation and lattice QCD simulation, with different inputs and parameters. The performance prediction of different transformations matches (within 5%) with hand-optimized layout code
Jouravel, Glorianne. "Stratégies innovantes pour une valorisation d’extraits de plantes en cosmétique : Mise en oeuvre d’un outil de profilage métabolique et recherche de nouvelles activités biologiques." Thesis, Orléans, 2018. http://www.theses.fr/2018ORLE2017.
Повний текст джерелаThe cosmetic field valorizes plant extracts by integrating them in care products. These extracts constitute the active ingredients of the cosmetic formulation. Plants are diverse, rich and contain numerous compounds of biological interest. Phytochemistry is a way to describe the metabolic content of plant extracts. But molecular characterization of these complex matrices remains a major challenge nowadays. Indeed,steps of data treatment are time-consuming and laborious. In this way, a tool of metabolic profiling, GAINS, has been developed in order to treat in an automatized way data from analyses performed in liquid chromatography coupled with high-resolution mass spectrometry. It constitutes a real support for phytochemists because automatized data treatment allows gaining time compared to manual treatment. This tool, associated with a large database of natural compounds make possible to assign potential candidates to detected peaks. GAINS appeals a module of in silico fragmentation for holding candidates assignments up.This permits to compare modeled spectrum of fragmentation of candidates with experimental spectrum off ragmentation.The whole set of phytochemical studies realized to identify or isolate compounds goes hand in hand with the study of potential biological effects of extracts to the skin, targeted organ by skin-care products. This allows the discovery of beneficial actions that the extract could have. By knowing the phytochemical content, it is possible to explain and rationalize assays about biological activities. The development of an anti-aging ingredient from purple loosestrife, a plant occurring in the region Centre-Val de Loire, is an example of it
Awwad, Tarek. "Context-aware worker selection for efficient quality control in crowdsourcing." Thesis, Lyon, 2018. http://www.theses.fr/2018LYSEI099/document.
Повний текст джерелаCrowdsourcing has proved its ability to address large scale data collection tasks at a low cost and in a short time. However, due to the dependence on unknown workers, the quality of the crowdsourcing process is questionable and must be controlled. Indeed, maintaining the efficiency of crowdsourcing requires the time and cost overhead related to this quality control to stay low. Current quality control techniques suffer from high time and budget overheads and from their dependency on prior knowledge about individual workers. In this thesis, we address these limitation by proposing the CAWS (Context-Aware Worker Selection) method which operates in two phases: in an offline phase, the correlations between the worker declarative profiles and the task types are learned. Then, in an online phase, the learned profile models are used to select the most reliable online workers for the incoming tasks depending on their types. Using declarative profiles helps eliminate any probing process, which reduces the time and the budget while maintaining the crowdsourcing quality. In order to evaluate CAWS, we introduce an information-rich dataset called CrowdED (Crowdsourcing Evaluation Dataset). The generation of CrowdED relies on a constrained sampling approach that allows to produce a dataset which respects the requester budget and type constraints. Through its generality and richness, CrowdED helps also in plugging the benchmarking gap present in the crowdsourcing community. Using CrowdED, we evaluate the performance of CAWS in terms of the quality, the time and the budget gain. Results shows that automatic grouping is able to achieve a learning quality similar to job-based grouping, and that CAWS is able to outperform the state-of-the-art profile-based worker selection when it comes to quality, especially when strong budget ant time constraints exist. Finally, we propose CREX (CReate Enrich eXtend) which provides the tools to select and sample input tasks and to automatically generate custom crowdsourcing campaign sites in order to extend and enrich CrowdED
Книги з теми "Profilage des données"
Statistics Canada. Employment Equity Program., ed. Profile of visible minorities and aboriginal people: 1986 Census--20% sample data = Profils des minorités et des autochtones : Recensement de 1986--données, échantillon, (20%). [Ottawa]: Statistics Canada, Employment Equity Program = Statistique Canada, Programme d'équité en matière d'emploi, 1990.
Знайти повний текст джерелаOntario. Ministry of Treasury and Economics. Sectoral and Regional Policy Branch. Profile of Ontario's provincial electoral districts (Bill 77, 1986 boundaries) based on 1981 census data =: Profil des circonscriptions électorales provinciales de l'Ontario (limites établies en 1986 en vertu du Projet de loi 77) d'après les données du recensement de 1981. Toronto, Ont: Ministry of Treasury and Economics = Ministère du trésor et de l'économie, 1987.
Знайти повний текст джерелаPorter, Marion R. A Profile of post-secondary students in Canada : the 1983-1984 national post-secondary student survey ; summary national data =: Profil des étudiants du niveau postsecondaire au Canada : l'enquête nationale de 1983-1984 auprès des étudiants du niveau postsecondaire : abrégé des données pour l'ensemble du Canada. Ottawa, Ont: Education Support Sector, Department of the Secretary of State = Direction générale de l'aide à l'éducation, Secrétariat d'État du Canada, 1987.
Знайти повний текст джерелаLau, Dorothy Wai Sim. Chinese Stardom in Participatory Cyberculture. Edinburgh University Press, 2018. http://dx.doi.org/10.3366/edinburgh/9781474430333.001.0001.
Повний текст джерела