Dissertationen zum Thema „Génération de données synthétiques“
Geben Sie eine Quelle nach APA, MLA, Chicago, Harvard und anderen Zitierweisen an
Machen Sie sich mit Top-50 Dissertationen für die Forschung zum Thema "Génération de données synthétiques" bekannt.
Neben jedem Werk im Literaturverzeichnis ist die Option "Zur Bibliographie hinzufügen" verfügbar. Nutzen Sie sie, wird Ihre bibliographische Angabe des gewählten Werkes nach der nötigen Zitierweise (APA, MLA, Harvard, Chicago, Vancouver usw.) automatisch gestaltet.
Sie können auch den vollen Text der wissenschaftlichen Publikation im PDF-Format herunterladen und eine Online-Annotation der Arbeit lesen, wenn die relevanten Parameter in den Metadaten verfügbar sind.
Sehen Sie die Dissertationen für verschiedene Spezialgebieten durch und erstellen Sie Ihre Bibliographie auf korrekte Weise.
Kieu, Van Cuong. „Modèle de dégradation d’images de documents anciens pour la génération de données semi-synthétiques“. Thesis, La Rochelle, 2014. http://www.theses.fr/2014LAROS029/document.
Der volle Inhalt der QuelleIn the last two decades, the increase in document image digitization projects results in scientific effervescence for conceiving document image processing and analysis algorithms (handwritten recognition, structure document analysis, spotting and indexing / retrieval graphical elements, etc.). A number of successful algorithms are based on learning (supervised, semi-supervised or unsupervised). In order to train such algorithms and to compare their performances, the scientific community on document image analysis needs many publicly available annotated document image databases. Their contents must be exhaustive enough to be representative of the possible variations in the documents to process / analyze. To create real document image databases, one needs an automatic or a manual annotation process. The performance of an automatic annotation process is proportional to the quality and completeness of these databases, and therefore annotation remains largely manual. Regarding the manual process, it is complicated, subjective, and tedious. To overcome such difficulties, several crowd-sourcing initiatives have been proposed, and some of them being modelled as a game to be more attractive. Such processes reduce significantly the price andsubjectivity of annotation, but difficulties still exist. For example, transcription and textline alignment have to be carried out manually. Since the 1990s, alternative document image generation approaches have been proposed including in generating semi-synthetic document images mimicking real ones. Semi-synthetic document image generation allows creating rapidly and cheaply benchmarking databases for evaluating the performances and trainingdocument processing and analysis algorithms. In the context of the project DIGIDOC (Document Image diGitisation with Interactive DescriptiOn Capability) funded by ANR (Agence Nationale de la Recherche), we focus on semi-synthetic document image generation adapted to ancient documents. First, we investigate new degradation models or adapt existing degradation models to ancient documents such as bleed-through model, distortion model, character degradation model, etc. Second, we apply such degradation models to generate semi-synthetic document image databases for performance evaluation (e.g the competition ICDAR2013, GREC2013) or for performance improvement (by re-training a handwritten recognition system, a segmentation system, and a binarisation system). This research work raises many collaboration opportunities with other researchers to share our experimental results with our scientific community. This collaborative work also helps us to validate our degradation models and to prove the efficiency of semi-synthetic document images for performance evaluation and re-training
Desbois-Bédard, Laurence. „Génération de données synthétiques pour des variables continues : étude de différentes méthodes utilisant les copules“. Master's thesis, Université Laval, 2017. http://hdl.handle.net/20.500.11794/27748.
Der volle Inhalt der QuelleStatistical agencies face a growing demand for releasing microdata to the public. To this end, many techniques have been proposed for publishing microdata while providing confidentiality : synthetic data generation in particular. This thesis focuses on such technique by presenting two existing methods, GAPD and C-GADP, as well as suggesting one based on vine copula models. GADP assumes that the variables of original and synthetic data are normally distributed, while C-GADP assumes that they have a normal copula distribution. Vine copula models are proposed due to their flexibility. These three methods are then assessed according to utility and risk. Data utility depends on maintaining certain similarities between the original and confidential data, while risk can be observed in two types : reidentification and inference. This work will focus on the utility examined with different analysis-specific measures, a global measure based on propensity scores and the risk of inference evaluated with a distance-based prediction.
Uzan, Kathy. „Les vaccins synthétiques : données récentes“. Paris 5, 1989. http://www.theses.fr/1989PA05P188.
Der volle Inhalt der QuelleBarrère, Killian. „Architectures de Transformer légères pour la reconnaissance de textes manuscrits anciens“. Electronic Thesis or Diss., Rennes, INSA, 2023. http://www.theses.fr/2023ISAR0017.
Der volle Inhalt der QuelleTransformer architectures deliver low error rates but are challenging to train due to limited annotated data in handwritten text recognition. We propose lightweight Transformer architectures to adapt to the limited amounts of annotated handwritten text available. We introduce a fast Transformer architecture with an encoder, processing up to 60 pages per second. We also present architectures using a Transformer decoder to incorporate language modeling into character recognition. To effectively train our architectures, we offer algorithms for generating synthetic data adapted to the visual style of modern and historical documents. Finally, we propose strategies for learning with limited data and reducing prediction errors. Our architectures, combined with synthetic data and these strategies, achieve competitive error rates on lines of text from modern documents. For historical documents, they train effectively with minimal annotated data, surpassing state-ofthe- art approaches. Remarkably, just 500 annotated lines are sufficient for character error rates close to 5%
Ruiz, Paredes Javier Antonio. „Génération d'accélérogrammes synthétiques large-bande par modélisation cinématique de la rupture sismique“. Paris, Institut de physique du globe, 2007. http://www.theses.fr/2007GLOB0009.
Der volle Inhalt der QuelleIn order to make the broadband kinematic rupture modeling more realistic with respect to dynamic modeling, physical constraints are added to the rupture parameters. To improve the slip velocity function (SVF) modeling, an evolution of the k-2 source model is proposed, which consists to decompose the slip as a sum of sub-events by band of k. This model yields to SVFclose to the solution proposed by Kostrov for a crack, while preserving the spectral characteristics of the radiated wavefield, i. E. A w2 model with spectral amplitudes at high frequency scaled to the coefficient of directivity Cd. To better control the directivity effects, acomposite source description is combined with a scaling law defining the extent of the nucleation area for each sub-event. The resulting model allows to reduce the apparent coefficient of directivity to a fraction of Cd, as well as to reproduce the standard deviation of the new empirical attenuation relationships proposed for Japan. To make source models more realistic, avariable rupture velocity in agreement with the physics of the rupture must be considered. The followed approach that is based on an analytical relation between the fracture energy, the slip and the rupture velocity, leads to higher values of the peak ground acceleration in the vicinity ofthe fault. Finally, to better account for the interaction of the wavefield with the geological medium, a semi-empirical methodology is developed combining a composite source model with empirical Green functions, and is applied to the Yamaguchi, Mw 5. 9 earthquake. The modeled synthetics reproduce satisfactorily well the observed main characteristics of ground motions
Pazat, Jean-Louis. „Génération de code réparti par distribution de données“. Habilitation à diriger des recherches, Université Rennes 1, 1997. http://tel.archives-ouvertes.fr/tel-00170867.
Der volle Inhalt der QuelleBaez, miranda Belen. „Génération de récits à partir de données ambiantes“. Thesis, Université Grenoble Alpes (ComUE), 2018. http://www.theses.fr/2018GREAM049/document.
Der volle Inhalt der QuelleStories are a communication tool that allow people to make sense of the world around them. It represents a platform to understand and share their culture, knowledge and identity. Stories carry a series of real or imaginary events, causing a feeling, a reaction or even trigger an action. For this reason, it has become a subject of interest for different fields beyond Literature (Education, Marketing, Psychology, etc.) that seek to achieve a particular goal through it (Persuade, Reflect, Learn, etc.).However, stories remain underdeveloped in Computer Science. There are works that focus on its analysis and automatic production. However, those algorithms and implementations remain constrained to imitate the creative process behind literary texts from textual sources. Thus, there are no approaches that produce automatically stories whose 1) the source consists of raw material that passed in real life and 2) and the content projects a perspective that seeks to convey a particular message. Working with raw data becomes relevant today as it increase exponentially each day through the use of connected devices.Given the context of Big Data, we present an approach to automatically generate stories from ambient data. The objective of this work is to bring out the lived experience of a person from the data produced during a human activity. Any areas that use such raw data could benefit from this work, for example, Education or Health. It is an interdisciplinary effort that includes Automatic Language Processing, Narratology, Cognitive Science and Human-Computer Interaction.This approach is based on corpora and models and includes the formalization of what we call the activity récit as well as an adapted generation approach. It consists of 4 stages: the formalization of the activity récit, corpus constitution, construction of models of activity and the récit, and the generation of text. Each one has been designed to overcome constraints related to the scientific questions asked in view of the nature of the objective: manipulation of uncertain and incomplete data, valid abstraction according to the activity, construction of models from which it is possible the Transposition of the reality collected though the data to a subjective perspective and rendered in natural language. We used the activity narrative as a case study, as practitioners use connected devices, so they need to share their experience. The results obtained are encouraging and give leads that open up many prospects for research
Morisse, Pierre. „Correction de données de séquençage de troisième génération“. Thesis, Normandie, 2019. http://www.theses.fr/2019NORMR043/document.
Der volle Inhalt der QuelleThe aims of this thesis are part of the vast problematic of high-throughput sequencing data analysis. More specifically, this thesis deals with long reads from third-generation sequencing technologies. The aspects tackled in this topic mainly focus on error correction, and on its impact on downstream analyses such a de novo assembly. As a first step, one of the objectives of this thesis is to evaluate and compare the quality of the error correction provided by the state-of-the-art tools, whether they employ a hybrid (using complementary short reads) or a self-correction (relying only on the information contained in the long reads sequences) strategy. Such an evaluation allows to easily identify which method is best tailored for a given case, according to the genome complexity, the sequencing depth, or the error rate of the reads. Moreover, developpers can thus identify the limiting factors of the existing methods, in order to guide their work and propose new solutions allowing to overcome these limitations. A new evaluation tool, providing a wide variety of metrics, compared to the only tool previously available, was thus developped. This tool combines a multiple sequence alignment approach and a segmentation strategy, thus allowing to drastically reduce the evaluation runtime. With the help of this tool, we present a benchmark of all the state-of-the-art error correction methods, on various datasets from several organisms, spanning from the A. baylyi bacteria to the human. This benchmark allowed to spot two major limiting factors of the existing tools: the reads displaying error rates above 30%, and the reads reaching more than 50 000 base pairs. The second objective of this thesis is thus the error correction of highly noisy long reads. To this aim, a hybrid error correction tool, combining different strategies from the state-of-the-art, was developped, in order to overcome the limiting factors of existing methods. More precisely, this tool combines a short reads alignmentstrategy to the use of a variable-order de Bruijn graph. This graph is used in order to link the aligned short reads, and thus correct the uncovered regions of the long reads. This method allows to process reads displaying error rates as high as 44%, and scales better to larger genomes, while allowing to reduce the runtime of the error correction, compared to the most efficient state-of-the-art tools.Finally, the third objectif of this thesis is the error correction of extremely long reads. To this aim, aself-correction tool was developed, by combining, once again, different methologies from the state-of-the-art. More precisely, an overlapping strategy, and a two phases error correction process, using multiple sequence alignement and local de Bruijn graphs, are used. In order to allow this method to scale to extremely long reads, the aforementioned segmentation strategy was generalized. This self-correction methods allows to process reads reaching up to 340 000 base pairs, and manages to scale very well to complex organisms such as the human genome
Fontin, Mickaël. „Contribution à la génération de séries synthétiques de pluies, de débits et de températures“. Toulouse, INPT, 1987. http://www.theses.fr/1987INPT117H.
Der volle Inhalt der QuelleKhalili, Malika. „Nouvelle approche de génération multi-site des données climatiques“. Mémoire, École de technologie supérieure, 2007. http://espace.etsmtl.ca/580/1/KHALILI_Malika.pdf.
Der volle Inhalt der QuelleGenestier, Richard. „Vérification formelle de programmes de génération de données structurées“. Thesis, Besançon, 2016. http://www.theses.fr/2016BESA2041/document.
Der volle Inhalt der QuelleThe general problem of proving properties of imperative programs is undecidable. Some subproblems– restricting the languages of programs and properties – are known to be decidable. Inpractice, thanks to heuristics, program proving tools sometimes automate proofs for programs andproperties living outside of the theoretical framework of known decidability results. We illustrate thisfact by building a catalog of proofs, for similar programs and properties of increasing complexity. Mostof these programs are combinatorial map generators.Thus, this work contributes to the research fields of enumerative combinatorics and softwareengineering. We distribute a C library of bounded exhaustive generators of structured arrays, formallyspecified in ACSL and verified with the WP plugin of the Frama-C analysis platform. We also proposea testing-based methodology to assist interactive proof in Coq, an original formal study of maps, andnew results in enumerative combinatorics
Caron, Maxime. „Données confidentielles : génération de jeux de données synthétisés par forêts aléatoires pour des variables catégoriques“. Master's thesis, Université Laval, 2015. http://hdl.handle.net/20.500.11794/25935.
Der volle Inhalt der QuelleConfidential data are very common in statistics nowadays. One way to treat them is to create partially synthetic datasets for data sharing. We will present an algorithm based on random forest to generate such datasets for categorical variables. We are interested by the formula used to make inference from multiple synthetic dataset. We show that the order of the synthesis has an impact on the estimation of the variance with the formula. We propose a variant of the algorithm inspired by differential privacy, and show that we are then not able to estimate a regression coefficient nor its variance. We show the impact of synthetic datasets on structural equations modeling. One conclusion is that the synthetic dataset does not really affect the coefficients between latent variables and measured variables.
Salama, Paul. „De la génération de soufre diatomique à partir de dérivés dithiocarbonylés, et de ses applications synthétiques“. Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1997. http://www.collectionscanada.ca/obj/s4/f2/dsk3/ftp04/nq26732.pdf.
Der volle Inhalt der QuelleFoundotos, Laetitia. „Génération d'accélérogrammes synthétiques large-bande. Contribution à l'estimation de l'aléa sismique par validation d'approches en aveugle“. Phd thesis, Université Nice Sophia Antipolis, 2013. http://tel.archives-ouvertes.fr/tel-00918349.
Der volle Inhalt der QuelleHonoré-Foundotos, Laëtitia. „Génération d'accélérogrammes synthétiques large-bande : contribution à l'estimation de l'aléa sismique par validation d'approches en aveugle“. Phd thesis, Université Nice Sophia Antipolis, 2013. http://tel.archives-ouvertes.fr/tel-00920208.
Der volle Inhalt der QuelleHonoré-Foundotos, Laëtitia. „Génération d'accélérogrammes synthétiques large-bande : contribution à l’estimation de l’aléa sismique par validation d’approches en aveugle“. Thesis, Nice, 2013. http://www.theses.fr/2013NICE4054/document.
Der volle Inhalt der QuelleOne of the major scientific problems in seismology is to estimate the ground motions expected at a given site from a future earthquake. The aim of this thesis is to test and validate two different methods of ground motions simulation based on the empirical Green’s function approach and to provide elements that can help to develop a blind simulation methodology. In a first part, a simulation method based on a stochastic point source approach is validated on the real data of recent earthquakes well instrumented : the Les Saintes earthquake Mw6.4 and the L’Aquila earthquake Mw6.3. We have developed a blind simulation approach by taking into account an uncertainty on the parameter of stress drop ratio C. This approach allows to generate a set of synthetic accelerograms of a target earthquake varied enough to be representative of a large number of possible source scenario and taking into account in a statistical sense potential directivity effects. This approach is also applied to the simulation of an historical Pyrenean earthquake Mw6.1. In a second part, we use a more complex extended source model, combining kinematic models of fractal composite sources with EGF approach. The potential of the method is tested on an application to L’Aquila earthquake. This has produced very satisfying results on all ground motion parameters analyzed. This simulation method appears to be very promising for the implementation of a blind simulation methodology, even if the main difficulty lies in the need to define the variability of many poorly known input parameters in the simulation of a future earthquake
Effantin, dit Toussaint Brice. „Colorations de graphes et génération exhaustive d'arbres“. Dijon, 2003. http://www.theses.fr/2003DIJOS021.
Der volle Inhalt der QuelleLagrange, Jean-Philippe. „Ogre : un système expert pour la génération de requêtes relationnelles“. Paris 9, 1992. https://portail.bu.dauphine.fr/fileviewer/index.php?doc=1992PA090035.
Der volle Inhalt der QuelleBounar, Boualem. „Génération automatique de programmes sur une base de données en réseau : couplage PROLOG-Base de données en réseau“. Lyon 1, 1986. http://www.theses.fr/1986LYO11703.
Der volle Inhalt der QuelleLeroux, (zinovieva) Elena. „Méthodes symboliques pour la génération de tests desystèmes réactifs comportant des données“. Phd thesis, Université Rennes 1, 2004. http://tel.archives-ouvertes.fr/tel-00142441.
Der volle Inhalt der Quellede transitions ne permet pas de le faire. Ceci oblige à énumérer les valeurs des données avant de construire le modèle de système de transitions d'un système, ce qui peut provoquer le problème de l'explosion de l'espace d'états. Cette énumération a également pour effet d'obtenir des cas de test où toutes les données sont instanciées. Or, cela contredit la pratique industrielle où les cas de test sont de vrais programmes avec des variables et des paramètres. La génération de tels
cas de test exige de nouveaux modèles et techniques. Dans cette thèse, nous atteignons deux objectifs. D'une part, nous introduisons un modèle appelé système symbolique de transitions à entrée/sortie qui inclut explicitement toutes les données d'un système réactif. D'autre part, nous proposons et implémentons une nouvelle technique de génération de test qui traite symboliquement les données d'un système en combinant l'approche de génération de test proposée auparavant par notre groupe de recherche avec des techniques d'interprétation abstraite. Les cas de test générés automatiquement par notre technique satisfont des propriétés de correction: ils émettent toujours un verdict correct.
Uribe, Lobello Ricardo. „Génération de maillages adaptatifs à partir de données volumiques de grande taille“. Thesis, Lyon 2, 2013. http://www.theses.fr/2013LYO22024.
Der volle Inhalt der QuelleIn this document, we have been interested in the surface extraction from the volumetric representation of an object. With this objective in mind, we have studied the spatial subdivision surface extraction algorithms. This approaches divide the volume in order to build a piecewise approximation of the surface. The general idea is to combine local and simple approximations to extract a complete representation of the object's surface.The methods based on the Marching Cubes (MC) algorithm have problems to produce good quality and to handle adaptive surfaces. Even if a lot of improvements to MC have been proposed, these approaches solved one or two problems but they don't offer a complete solution to all the MC drawbacks. Dual methods are more adapted to use adaptive sampling over volumes. These methods generate surfaces that are dual to those generated by the Marching Cubes algorithm or dual grids in order to use MC methods. These solutions build adaptive meshes that represent well the features of the object. In addition, recent improvements guarantee that the produced meshes have good geometrical and topological properties.In this dissertation, we have studied the main topological and geometrical properties of volumetric objects. In a first stage, we have explored the state of the art on spatial subdivision surface extraction methods in order to identify theirs advantages, theirs drawbacks and the implications of theirs application on volumetric objects. We have concluded that a dual approach is the best option to obtain a good compromise between mesh quality and geometrical approximation. In a second stage, we have developed a general pipeline for surface extraction based on a combination of dual methods and connected components extraction to better capture the topology and geometry of the original object. In a third stage, we have presented an out-of-core extension of our surface extraction pipeline in order to extract adaptive meshes from huge volumes. Volumes are divided in smaller sub-volumes that are processed independently to produce surface patches that are later combined in an unique and topologically correct surface. This approach can be implemented in parallel to speed up its performance. Test realized in a vast set of volumes have confirmed our results and the features of our solution
Xue, Xiaohui. „Génération et adaptation automatiques de mappings pour des sources de données XML“. Phd thesis, Université de Versailles-Saint Quentin en Yvelines, 2006. http://tel.archives-ouvertes.fr/tel-00324429.
Der volle Inhalt der QuelleNous proposons une approche de génération de mappings en trois phases : (i) la décomposition du schéma cible en sous-arbres, (ii) la recherche de mappings partiels pour chacun de ces sous-arbres et enfin (iii) la génération de mappings pour l'ensemble du schéma cible à partir de ces mappings partiels. Le résultat de notre approche est un ensemble de mappings, chacun ayant une sémantique propre. Dans le cas où l'information requise par le schéma cible n'est pas présente dans les sources, aucun mapping ne sera produit. Dans ce cas, nous proposons de relaxer certaines contraintes définies sur le schéma cible pour permettre de générer des mappings. Nous avons développé un outil pour supporter notre approche. Nous avons également proposé une approche d'adaptation des mappings existants en cas de changement survenant dans les sources ou dans le schéma cible.
Xue, Xiaohui. „Génération et adaptation automatiques de mappings pour des sources de données XML“. Phd thesis, Versailles-St Quentin en Yvelines, 2006. http://www.theses.fr/2006VERS0019.
Der volle Inhalt der QuelleThe integration of information originating from multiple heterogeneous data sources is required by many modern information systems. In this context, the applications’ needs are described by a target schema and the way in-stances of the target schema are derived from the data sources is expressed through mappings. In this thesis, we address the problem of mapping generation for multiple XML data sources and the adaptation of these mappings when the target schema or the sources evolve. We propose an automatic generation approach that first decom-poses the target schema into subtrees, then defines mappings, called partial mappings, for each of these subtrees, and finally combines these partial mappings to generate the mappings for the whole target schema. We also propose a mapping adaptation approach to keep existing mappings current if some changes occur in the target schema or in one of the sources. We have developed a prototype implementation of a tool to support these proc-esses
Fabregue, Mickael. „Extraction d'informations synthétiques à partir de données séquentielles : application à l'évaluation de la qualité des rivières“. Thesis, Strasbourg, 2014. http://www.theses.fr/2014STRAD016/document.
Der volle Inhalt der QuelleExploring temporal databases with suitable data mining methods have been the subject of several studies. However, it often leads to an excessive volume of extracted information and the analysis is difficult for the user. We addressed this issue and we specically focused on methods that synthesize and filter extracted information. The objective is to provide interpretable results for humans. Thus, we relied on the notion of partially ordered sequence and we proposed (1) an algorithm that extracts the set of closed partially ordered patterns ; (2) a post-processing to filter some interesting patterns for the user and (3) an approach that extracts a partially ordered consensus as an alternative to pattern extraction. The proposed methods were applied for validation on hydrobiological data from the Fresqueau ANR project. In addition, they have been implemented in a visualization tool designed for hydrobiologists for water course quality analysis
Zinovieva-Leroux, Eléna. „Méthodes symboliques pour la génération de tests de systèmes réactifs comportant des données“. Rennes 1, 2004. https://tel.archives-ouvertes.fr/tel-00142441.
Der volle Inhalt der QuellePinard, Hugo. „Imagerie électromagnétique 2D par inversion des formes d'ondes complètes : Approche multiparamètres sur cas synthétiques et données réelles“. Thesis, Université Grenoble Alpes (ComUE), 2017. http://www.theses.fr/2017GREAU041/document.
Der volle Inhalt der QuelleGround Penetrating Radar (GPR) is a geophysical investigation method based on electromagnetic waves propagation in the underground. With frequencies ranging from 5 MHz to a few GHz and a high sensitivity to electrical properties, GPR provides reflectivity images in a wide variety of contexts and scales: civil engineering, geology, hydrogeology, glaciology, archeology. However, in some cases, a better understanding of some subsurface processes requires a quantification of the physical parameters of the subsoil. For this purpose, inversion of full waveforms, a method initially developed for seismic exploration that exploits all the recorded signals, could prove effective. In this thesis, I propose methodological developments using a multiparameter inversion approach (dielectric permittivity and conductivity), for two-dimensional transmission configurations. These developments are then applied to a real data set acquired between boreholes.In a first part, I present the numerical method used to model the propagation of electromagnetic waves in a heterogeneous 2D environment, a much-needed element to carry out the process of imaging. Then, I introduce and study the potential of standard local optimization methods (nonlinear conjugate gradient, l-BFGS, Newton truncated in its Gauss-Newton and Exact-Newton versions) to fight the trade-off effects related to the dielectric permittivity and to the electrical conductivity. In particular, I show that effective decoupling is possible only with a sufficiently accurate initial model and the most sophisticated method (truncated Newton). As in the general case, this initial model is not available, it is necessary to introduce a scaling factor which distributes the relative weight of each parameter class in the inversion. In a realistic medium and for a cross-hole acquisition configuration, I show that the different optimization methods give similar results in terms of parameters decoupling. It is eventually the l-BFGS method that is used for the application to the real data, because of lower computation costs.In a second part, I applied the developed Full waveform inversion methodology to a set of real data acquired between two boreholes located in carbonate formations, in Rustrel (France, 84). This inversion is carried out together with a synthetic approach using a model representative of the studied site and with a similar acquisition configuration. This approach enables us to monitor and validate the observations and conclusions derived from data inversion. It shows that reconstruction of dielectrical permittivity is very robust. Conversely, conductivity estimation suffers from two major couplings: the permittivity and the amplitude of the estimated source. The derived results are successfully compared with independent data (surface geophysics and rock analysis on plugs) and provides a high resolution image of the geological formation. On the other hand, a 3D analysis confirms that 3D structures presenting high properties contrasts, such as the buried gallery present in our site, would require a 3D approach, notably to better explain the observed amplitudes
Benalia, Akram Djellal. „HELPDraw : un environnement visuel pour la génération automatique de programmes à parallélisme de données“. Lille 1, 1995. http://www.theses.fr/1995LIL10095.
Der volle Inhalt der QuelleNesvijevskaia, Anna. „Phénomène Big Data en entreprise : processus projet, génération de valeur et Médiation Homme-Données“. Thesis, Paris, CNAM, 2019. http://www.theses.fr/2019CNAM1247.
Der volle Inhalt der QuelleBig Data, a sociotechnical phenomenon carrying myths, is reflected in companies by the implementation of first projects, especially Data Science projects. However, they do not seem to generate the expected value. The action-research carried out over the course of 3 years in the field, through an in-depth qualitative study of multiple cases, points to key factors that limit this generation of value, including overly self-contained project process models. The result is (1) an open data project model (Brizo_DS), orientated on the usage, including knowledge capitalization, intended to reduce the uncertainties inherent in these exploratory projects, and transferable to the scale of portfolio management of corporate data projects. It is completed with (2) a tool for documenting the quality of the processed data, the Databook, and (3) a Human-Data Mediation device, which guarantee the alignment of the actors towards an optimal result
Nemecek, Daniel. „Une nouvelle génération d'oligonucléotides synthétiques utilisables en stratégie antisens : propriétés d'hybridation et stabilité des complexes formés avec leurs compléments d'ADN naturel“. Paris 6, 2005. http://www.theses.fr/2005PA066012.
Der volle Inhalt der QuelleThiéblin, Elodie. „Génération automatique d'alignements complexes d'ontologies“. Thesis, Toulouse 3, 2019. http://www.theses.fr/2019TOU30135.
Der volle Inhalt der QuelleThe Linked Open Data (LOD) cloud is composed of data repositories. The data in the repositories are described by vocabularies also called ontologies. Each ontology has its own terminology and model. This leads to heterogeneity between them. To make the ontologies and the data they describe interoperable, ontology alignments establish correspondences, or links between their entities. There are many ontology matching systems which generate simple alignments, i.e., they link an entity to another. However, to overcome the ontology heterogeneity, more expressive correspondences are sometimes needed. Finding this kind of correspondence is a fastidious task that can be automated. In this thesis, an automatic complex matching approach based on a user's knowledge needs and common instances is proposed. The complex alignment field is still growing and little work address the evaluation of such alignments. To palliate this lack, we propose an automatic complex alignment evaluation system. This system is based on instances. A famous alignment evaluation dataset has been extended for this evaluation
Bonnel, Nicolas. „Génération dynamique de présentations interactives en multimédia 3D, de données, pour les applications en ligne“. Phd thesis, Université Rennes 1, 2006. http://tel.archives-ouvertes.fr/tel-00532641.
Der volle Inhalt der QuelleAbdelmoula, Mariem. „Génération automatique de jeux de tests avec analyse symbolique des données pour les systèmes embarqués“. Thesis, Nice, 2014. http://www.theses.fr/2014NICE4149/document.
Der volle Inhalt der QuelleOne of the biggest challenges in hardware and software design is to ensure that a system is error-free. Small errors in reactive embedded systems can have disastrous and costly consequences for a project. Preventing such errors by identifying the most probable cases of erratic system behavior is quite challenging. Indeed, tests in industry are overall non-exhaustive, while formal verification in scientific research often suffers from combinatorial explosion problem. We present in this context a new approach for generating exhaustive test sets that combines the underlying principles of the industrial test technique and the academic-based formal verification approach. Our approach builds a generic model of the system under test according to the synchronous approach. The goal is to identify the optimal preconditions for restricting the state space of the model such that test generation can take place on significant subspaces only. So, all the possible test sets are generated from the extracted subspace preconditions. Our approach exhibits a simpler and efficient quasi-flattening algorithm compared with existing techniques and a useful compiled internal description to check security properties and reduce the state space combinatorial explosion problem. It also provides a symbolic processing technique of numeric data that provides a more expressive and concrete test of the system. We have implemented our approach on a tool called GAJE. To illustrate our work, this tool was applied to verify an industrial project on contactless smart cards security
Genevaux, Jean-David. „Représentation, modélisation et génération procédurale de terrains“. Thesis, Lyon 2, 2015. http://www.theses.fr/2015LYO22013/document.
Der volle Inhalt der QuelleThis PhD (entitled "Representation, modelisation and procedural generation of terrains") is related to movie and videogames digital content creation, especially natural scenes.Our work is dedicated to handle and to generate landscapes efficently. We propose a new model based on a construction tree inside which the user can handle parts of the terrain intuitively. We also present techniques to efficently visualize such model. Finally, we present a new algorithm for generating large-scale terrains exhibiting hierarchical structures based on their hydrographic networks: elevation is generated in a broad compliance to water-tansport principles without having to resort on costly hydraulic simulations
Kou, Huaizhong. „Génération d'adaptateurs web intelligents à l'aide de techniques de fouilles de texte“. Versailles-St Quentin en Yvelines, 2003. http://www.theses.fr/2003VERS0011.
Der volle Inhalt der QuelleThis thesis defines a system framework of semantically integrating Web information, called SEWISE. It can integrate text information from various Web sources belonging to an application domain into common domain-specific concept ontology. In SEWISE, Web wrappers are built around different Web sites to automatically extract interesting information from. Text mining technologies are then used to discover the semantics Web documents talk about. SEWISE can ease topic-oriented information researches over the Web. Three problems related to the document categorization are studied. Firstly, we investigate the approaches to feature selection and proposed two approaches CBA and IBA to select features. To estimate statistic term associations and integrate them within document similarity model, a mathematical model is proposed. Finally, the category score calculation algorithms used by k-NN classifiers are studied. Two weighted algorithms CBW and IBW to calculate category score are proposed
Raschia, Guillaume. „SaintEtiq : une approche floue pour la génération de résumés à partir de bases de données relationnelles“. Nantes, 2001. http://www.theses.fr/2001NANT2099.
Der volle Inhalt der QuellePlatzer, Auriane. „Mécanique numérique en grandes transformations pilotée par les données : De la génération de données sur mesure à une stratégie adaptative de calcul multiéchelle“. Thesis, Ecole centrale de Nantes, 2020. http://www.theses.fr/2020ECDN0041.
Der volle Inhalt der QuelleComputational mechanics is a field in which a large amount of data is both consumed and produced. On the one hand, the recent developments of experimental measurement techniques have provided rich data for the identification process of constitutive models used in finite element simulations. On the other hand, multiscale analysis produces a huge amount of discrete values of displacements, strains and stresses from which knowledge is extracted on the overall material behavior. The constitutive model then acts as a bottleneck between upstream and downstream material data. In contrast, Kirchdoerfer and Ortiz (Computer Methods in Applied Mechanics and Engineering, 304, 81-101) proposed a model-free computing paradigm, called data-driven computational mechanics. The material response is then only represented by a database of raw material data (strain-stress pairs). The boundary value problem is thus reformulated as a constrained distance minimization between (i) the mechanical strain-stress state of the body, and (ii) the material database. In this thesis, we investigate the question of material data coverage, especially in the finite strain framework. The data-driven approach is first extended to a geometrically nonlinear setting: two alternative formulations are considered and a finite element solver is proposed for both. Second, we explore the generation of tailored databases using a mechanically meaningful sampling method. The approach is assessed by means of finite element analyses of complex structures exhibiting large deformations. Finally, we propose a prototype multiscale data-driven solver, in which the material database is adaptively enriched
Zaher, Noufal Issam al. „Outils de CAO pour la génération d'opérateurs arithmétiques auto-contrôlables“. Grenoble INPG, 2001. http://www.theses.fr/2001INPG0028.
Der volle Inhalt der QuelleThiessard, Frantz. „Détection des effets indésirables des médicaments par un système de génération automatisée du signal adapté à la base nationale française de pharmacovigilance“. Bordeaux 2, 2004. http://www.theses.fr/2004BOR21184.
Der volle Inhalt der QuelleEvaluation and improvement of drugs risk/benefit ratio in population implies their adverse reactions surveillance after marketing. Pharmacovigilance main objective is to detect drugs adverse reactions relied mainly on spontaneous notifications. The French pharmacovigilance is faced to a very large data flow while no automatic method is available to edit a list of potentially suspected drug/adverse drug reaction associations. Eight methods were studied : Proportional Reporting Ratio (PRR), Reporting Odds Ratio (ROR), Uule's Q, Sequential Probability Ratio Test (SPRT2), Poisson's probabilities, X2, Information Component (IC), and Empirical Baye's Method (EBAM). Signals obtained with each method were compared through simulated data, then through real data from the French pharmacovigilance database
Tourancheau, Alan. „Analyse bio-informatique de données de séquençage de nouvelle génération pour l'étude transcriptomique d'enzymes du métabolisme“. Doctoral thesis, Université Laval, 2016. http://hdl.handle.net/20.500.11794/27295.
Der volle Inhalt der QuelleUDP-glucuronosyltransferases (UGT) catalyze the reaction of glucuronidation. These enzymes are involved in the metabolism of many endogenous (e.g. bilirubin and steroid hormones) and exogenous substrates (e.g. many anticancer agents and drugs of other classes). They are expressed, among others, in the tissues of drug metabolism of such as the liver, kidneys and gastrointestinal tract tissues. A comprehensive and detailed view of the human UGT transcriptome emerges as a key condition for the establishment of the metabolic signature of an individual. As part of my PhD research project, we uncover the transcriptome landscape of the 10 human UGT gene loci in normal and tumoral metabolic tissues by targeted RNA next-generation sequencing (Capture-Seq). For this, liver tissues, kidney, small intestine and colon as well as endometrial tissues, breast and prostate were analyzed. Alignment on the human hg19 reference genome identifies 234 novel exon-exon junctions. We recover all previously known UGT1 and UGT2 enzyme-coding transcripts and identify over 130 structurally and functionally diverse novel UGT variants. Our work establish for the first time that all UGT genes are subject to alternative splicing. We further expose a revised genomic structure of UGT loci and provide a comprehensive repertoire of transcripts for each UGT gene. Finally, the entire transcriptome of UGT genes was quantified in the major drugs metabolism tissues (liver, kidney and intestine). The results indicate that alternative transcripts represent a significant part of the UGT transcriptome varying from 6-100% of UGT gene expression. Data also uncover a remodelling of the UGT transcriptome occurring in a tissue- and tumor-specific manner. The complex alternative splicing program regulating UGT expression and protein functions is likely critical in determining detoxification capacity of an organ and stress-related responses, with significant impact on drug responses and diseases.
Gombert, Baptiste. „Vers une nouvelle génération de modèles de glissements co-sismiques : analyse stochastique et approche multi-données“. Thesis, Strasbourg, 2018. http://www.theses.fr/2018STRAH016/document.
Der volle Inhalt der QuelleThe explosion in the amount and variety of available geodetic, tsunami, and seismological observations offers an outstanding opportunity to develop new seismic source models. But these data are sensitive to different sources of uncertainty and provide heterogeneous information, which makes the solution of the inverse problem non-unique.In this thesis, we use a Bayesian sampling method to propose new slip models, which benefit from an objective weighting of the various datasets by combining observational and modelling errors. These models are less affected by data overfit and allow a realistic assessment of posterior uncertainties. We apply this method to the study of slip processes occurring in three different tectonic contexts: the Landers earthquake (1992, Mw=7.3), the Ecuador-Colombia subduction zone which hosted the Pedernales earthquake (2016, Mw=7.8), and the intraslab Tehuantepec earthquake (2017, Mw=8.2). Through these analyses, we demonstrate how the study of the seismic cycle can benefit from rigorous uncertainty estimates and Bayesian sampling
Soukane, Assia. „Génération automatique des requêtes de médiation dans un environnement hétérogène“. Versailles-St Quentin en Yvelines, 2005. http://www.theses.fr/2005VERS0038.
Der volle Inhalt der QuelleNowadays, mediation systems are widely used. However, their implementation raises several problems, especially, the definition of queries when there is a high number of sources, and a great amount of meta data. Furthermore, the problem is particularly important when sources are heterogeneous. We propose in this thesis an approach to automatically generate the mediation queries, in a relational context. Our algorithm produce candidate set of queries since heterogeneous and distributed sources. We develop a tool to automatically generate mediation queries in the heterogeneous context. Our principal goal is to provide an appropriate tool for different sizes of systems; we achieve tests series to perform scalability
Moyse, Gilles. „Résumés linguistiques de données numériques : interprétabilité et périodicité de séries“. Thesis, Paris 6, 2016. http://www.theses.fr/2016PA066526/document.
Der volle Inhalt der QuelleOur research is in the field of fuzzy linguistic summaries (FLS) that allow to generate natural language sentences to describe very large amounts of numerical data, providing concise and intelligible views of these data. We first focus on the interpretability of FLS, crucial to provide end-users with an easily understandable text, but hard to achieve due to its linguistic form. Beyond existing works on that topic, based on the basic components of FLS, we propose a general approach for the interpretability of summaries, considering them globally as groups of sentences. We focus more specifically on their consistency. In order to guarantee it in the framework of standard fuzzy logic, we introduce a new model of oppositions between increasingly complex sentences. The model allows us to show that these consistency properties can be satisfied by selecting a specific negation approach. Moreover, based on this model, we design a 4-dimensional cube displaying all the possible oppositions between sentences in a FLS and show that it generalises several existing logical opposition structures. We then consider the case of data in the form of numerical series and focus on linguistic summaries about their periodicity: the sentences we propose indicate the extent to which the series are periodic and offer an appropriate linguistic expression of their periods. The proposed extraction method, called DPE, standing for Detection of Periodic Events, splits the data in an adaptive manner and without any prior information, using tools from mathematical morphology. The segments are then exploited to compute the period and the periodicity, measuring the quality of the estimation and the extent to which the series is periodic. Lastly, DPE returns descriptive sentences of the form ``Approximately every 2 hours, the customer arrival is important''. Experiments with artificial and real data show the relevance of the proposed DPE method. From an algorithmic point of view, we propose an incremental and efficient implementation of DPE, based on established update formulas. This implementation makes DPE scalable and allows it to process real-time streams of data. We also present an extension of DPE based on the local periodicity concept, allowing the identification of local periodic subsequences in a numerical series, using an original statistical test. The method validated on artificial and real data returns natural language sentences that extract information of the form ``Every two weeks during the first semester of the year, sales are high''
Archier, Pascal. „Contribution à l'amélioration des données nucléaires neutroniques du sodium pour le calcul des réacteurs de génération IV“. Phd thesis, Université de Grenoble, 2011. http://tel.archives-ouvertes.fr/tel-00680781.
Der volle Inhalt der QuelleTata, Zafiarifety Christian. „Simulation et traitement des données d’un imageur à rayons Gamma pour une nouvelle génération de caméras Compton“. Electronic Thesis or Diss., Troyes, 2019. http://www.theses.fr/2019TROY0028.
Der volle Inhalt der QuelleThe localization of radioactivity is a crucial step in the dismantling of nuclear power plants. For this purpose, several detection systems have been developed, such as the pinhole camera, using lead or tungsten collimators, but having as main disadvantage a low detection efficiency. The Compton camera uses the kinematics of the Compton broadcast. It represents a very promising alternative compared to conventional systems because it has several advantages such as: high detection efficiency, reconstruction of radioactive source images with high spatial resolution and wide field of view, and the ability to perform spectroscopy with good energy resolution. So, in this work we developed a new Compton camera based on the use of two monolithic crystals from Cebr3 equipped with Philips DPC3200 photodetector and assembled with materials and processes developed by Damavan for obtain the detection heads of optimal quality and adapted to the constraints of the Compton camera. We have thus set up a procedure for the calibration of the time and energy of the detection heads. We also implemented a new position calculation algorithm based on the use of a new model simulated by Monte Carlo. Finally, we carried out a global evaluation of the camera’s performance, once the basic concepts in its development were tested: time, energy and position
Benna, Mehdi. „Génération et inversion de données de propagation d'ondes radio à travers un noyau cométaire (simulation de l'expérience CONSERT)“. Toulouse 3, 2002. http://www.theses.fr/2002TOU30116.
Der volle Inhalt der QuelleBanerji, Ranajoy. „Optimisation d’une mission spatiale CMB de 4eme génération“. Thesis, Sorbonne Paris Cité, 2017. http://www.theses.fr/2017USPCC199/document.
Der volle Inhalt der QuelleThe Cosmic Microwave Background radiation is a rich and clean source of Cosmological information. Study of the CMB over the past few decades has led to the establishment of a “Standard Model” for Cosmology and constrained many of its principal parameters. It hasalso transformed the field into a highly data-driven domain.Currently, Inflation is the leading paradigm describing the earliest moments of our Universe. It predicts the generation of primordial matter density fluctuations and gravitational waves. The CMB polarisation carries the signature of these gravitational waves in the form of primordial “B-modes”. A future generation of CMB polarisation space mission is well suited to observe this signature of Inflation.This thesis focuses on optimising a future CMB space mission that will observe the B-modesignal for reaching a sensitivity of r = 0.001. Specifically, I study the optimisation of the scanning strategy and the impact of systematics on the quality of polarisation measurement
Rialle, Stéphanie. „Méthodologie et outils bioinformatiques d'aide à la conception de systèmes biologiques synthétiques pour de nouveaux diagnostics en santé humaine“. Thesis, Montpellier 2, 2010. http://www.theses.fr/2010MON20091/document.
Der volle Inhalt der QuelleSynthetic biology is a growing discipline which aims to design and construct biological systems with functions that do not exist in nature. It is based on engineering principles to rationalize the design such systems. The CompuBioTic project aims at the development of a new system for the diagnosis of the colorectal cancer, based on a synthetic biology approach. A strategic choice has been done and consists in wanting to develop a non-living system, which does not require a host cell and which is based on the use of protein rather than genetic networks. Very few methodologies and tools have been developed to facilitate the design of such systems. This thesis proposes a methodology in three steps: design, simulation and experimental validation, as well as two bioinformatics tools, developed to assist the design of synthetic biochemical networks. Firstly, CompuBioTicDB is a database that registers and annotates functional devices and molecules carrying processes (proteins and small molecules) that can be exploited in a context of synthetic biology. Secondly, BioNetCAD is a tool for designing a biochemical network composed of real molecules from an abstract network. BioNetCAD also facilitates spatiotemporal simulation of the designed system with a link to the HSim software. Molecular logic gates and a device for detecting glucose have been designed, modeled and then validated experimentally. The principles of a system for the diagnosis of colorectal cancer are also proposed
Mehdi, Benna. „Génération et inversion de données de propagation d'ondes radio à travers un noyau comètaire (Experience CONSERT - Mission Spatiale ROSETTA)“. Phd thesis, Université Paul Sabatier - Toulouse III, 2002. http://tel.archives-ouvertes.fr/tel-00007990.
Der volle Inhalt der QuelleKhamakhem, Wassim. „Etude de l'évolution du combustible dans des réacteurs rapides de quatrième génération : impact des données nucléaires sur leur performance“. Paris 11, 2010. http://www.theses.fr/2010PA112173.
Der volle Inhalt der QuelleThe objective of this PhD topic is to contribute to the understanding of the variations of the core neutronic characteristics of the 4th generation reactors (Sodium Cooled Fast Reactors (SFR) and Gas Cooled Fast Reactors (GFR)) during fuel depletion. The neutron characteristics of interest are of course the burn up reactivity swing and the breeding gain but also the Doppler effect and the coolant void effect. Fuel depletion leads to a degradation of the core safety parameters. The study of these variations and their associated uncertainties contributes to justify 4th generation reactor core designs as envisaged in their last developments. These last developments concerned Sodium Cooled Fast Reactors (SFR) and Gas Cooled Fast Reactors (GFR) which were reshaped in order to meet Generation IV goals on economics, safety and reliability, sustainability and proliferation resistance. They exhibit very innovative characteristics compared to the European Fast Reactor (EFR) whose design was very much in line with those of Phenix and Super Phenix. Recent CEA studies had led to large 3600 MWth SFR cores using oxide fuel and to large 2400 MWth GFR cores using carbide fuel. Since the designs have to balance between positive breeding gain and safety characteristics such as rather low void reactivity effects (SFR) or rather sm ail core pressure drop (GFR), scoping studies for breakthrough SFR cores were performed using dense fuels either carbide (already taken as a reference for the GFR core) or metal. These preliminary breakthrough SFR images are characterized by high power density and highly positive breeding gain (Breeding Gain = 0. 17). As a first step towards the development of GFR plants, a low power experimental GFR called ALLEGRO is being envisaged and has been studied for its peculiar characteristics. To study the main neutronic characteristics of these cores, one can use analyses based on the sensitivity methods of the deterministic computer code ERANOS (neutronic code system). These methods are available in statics without the possibility of taking into account fuel depletion. Ln order to mitigate this insufficiency, a subsequent part of the thesis consisted in developing the depletion perturbation theory which requires to couple Boltzmann and Bateman equations and allows a more precise understanding of the behaviour of the previous cores. The method is now able to calculate the sensitivity of the actinides and fission products concentrations and of neutron characteristics of interest such as breeding gain, Doppler reactivity effect and the coolant void reactivity coefficient effect. Ln order to illustrate these sensitivity developments, uncertainties of the neutron characteristics have been calculated using a preliminary variance covariance matrix called BOLNA. The uncertainty analyses highlight the contribution of each isotope to the neutron characteristics of the various core designs. This determination has given relatively small uncertainty variations with burn up when possible modifications of nuclear data are applied. The in-depth study performed on sodium nuclear data evaluations (ENDFB-VII, JEFF-3. 1, JENDL-3. 3) highlight the difficulty of creating accurate enough nuclear data and their associated covariance matrix. It appears hence that although the feasibility of these core designs are not questioned (relatively optimistic values being calculated are within the target value of 700 pcm for the reactivity swing and 7% for the reactivity coefficient), their performance will require integral experiments both to confirm what has been evaluated with nuclear data covariance matrices and to reduce nuclear data uncertainties. Lastly, the sensitivity methods are used to explain the peculiar behaviour of integral characteristics Iike the void effect or the Doppler effect with depletion in the GFR and ALLEGRO cores. One reason was track back to the difference in size of the two cores but also to the different structural materials being used. Furthermore, the building up of Pu239 fission products and the change in Pu239 and Pu241 isotopes being different induce divergent behaviour of both Doppler with time. For the SFR, the distribution of the void effect in the various core zones which present different fuel depletion histories is finally analyzed to be compared to that of the power distribution and finally to that of the breeding gain. It appears that the SFR core design with a rather flat internai breeding gain has, as a consequence, a rather flat void effect which is another nice feature. One concludes on the advantages resulting from the last core designs as weil as their degree of performance from the view point of computational tools very dependent at first on the nuclear data knowledge
Fellahi, Mohammed. „Des réseaux de processus cyclo-statiques à la génération de code pour le pipeline multi-dimensionnel“. Phd thesis, Université Paris Sud - Paris XI, 2011. http://tel.archives-ouvertes.fr/tel-00683224.
Der volle Inhalt der Quelle