Dissertations / Theses on the topic 'Plongements de documents'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 23 dissertations / theses for your research on the topic 'Plongements de documents.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Mazoyer, Béatrice. "Social Media Stories. Event detection in heterogeneous streams of documents applied to the study of information spreading across social and news media." Thesis, université Paris-Saclay, 2020. http://www.theses.fr/2020UPASC009.
Full textSocial Media, and Twitter in particular, has become a privileged source of information for journalists in recent years. Most of them monitor Twitter, in the search for newsworthy stories. This thesis aims to investigate and to quantify the effect of this technological change on editorial decisions. Does the popularity of a story affects the way it is covered by traditional news media, regardless of its intrinsic interest?To highlight this relationship, we take a multidisciplinary approach at the crossroads of computer science and economics: first, we design a novel approach to collect a representative sample of 70% of all French tweets emitted during an entire year. Second, we study different types of algorithms to automatically discover tweets that relate to the same stories. We test several vector representations of tweets, looking at both text and text-image representations, Third, we design a new method to group together Twitter events and media events. Finally, we design an econometric instrument to identify a causal effect of the popularity of an event on Twitter on its coverage by traditional media. We show that the popularity of a story on Twitter does have an effect on the number of articles devoted to it by traditional media, with an increase of about 1 article per 1000 additional tweets
Morbieu, Stanislas. "Leveraging textual embeddings for unsupervised learning." Electronic Thesis or Diss., Université Paris Cité, 2020. http://www.theses.fr/2020UNIP5191.
Full textTextual data is ubiquitous and is a useful information pool for many companies. In particular, the web provides an almost inexhaustible source of textual data that can be used for recommendation systems, business or technological watch, information retrieval, etc. Recent advances in natural language processing have made possible to capture the meaning of words in their context in order to improve automatic translation systems, text summary, or even the classification of documents according to predefined categories. However, the majority of these applications often rely on a significant human intervention to annotate corpora: This annotation consists, for example in the context of supervised classification, in providing algorithms with examples of assigning categories to documents. The algorithm therefore learns to reproduce human judgment in order to apply it for new documents. The object of this thesis is to take advantage of these latest advances which capture the semantic of the text and use it in an unsupervised framework. The contributions of this thesis revolve around three main axes. First, we propose a method to transfer the information captured by a neural network for co-clustering of documents and words. Co-clustering consists in partitioning the two dimensions of a data matrix simultaneously, thus forming both groups of similar documents and groups of coherent words. This facilitates the interpretation of a large corpus of documents since it is possible to characterize groups of documents by groups of words, thus summarizing a large corpus of text. More precisely, we train the Paragraph Vectors algorithm on an augmented dataset by varying the different hyperparameters, classify the documents from the different vector representations and apply a consensus algorithm on the different partitions. A constrained co-clustering of the co-occurrence matrix between terms and documents is then applied to maintain the consensus partitioning. This method is found to result in significantly better quality of document partitioning on various document corpora and provides the advantage of the interpretation offered by the co-clustering. Secondly, we present a method for evaluating co-clustering algorithms by exploiting vector representations of words called word embeddings. Word embeddings are vectors constructed using large volumes of text, one major characteristic of which is that two semantically close words have word embeddings close by a cosine distance. Our method makes it possible to measure the matching between the partition of the documents and the partition of the words, thus offering in a totally unsupervised setting a measure of the quality of the co-clustering. Thirdly, we are interested in recommending classified ads. We present a system that allows to recommend similar classified ads when consulting one. The descriptions of classified ads are often short, syntactically incorrect, and the use of synonyms makes it difficult for traditional systems to accurately measure semantic similarity. In addition, the high renewal rate of classified ads that are still valid (product not sold) implies choices that make it possible to have low computation time. Our method, simple to implement, responds to this use case and is again based on word embeddings. The use of these has advantages but also involves some difficulties: the creation of such vectors requires choosing the values of some parameters, and the difference between the corpus on which the word embeddings were built upstream. and the one on which they are used raises the problem of out-of-vocabulary words, which have no vector representation. To overcome these problems, we present an analysis of the impact of the different parameters on word embeddings as well as a study of the methods allowing to deal with the problem of out-of-vocabulary words
Liu, Guogang. "Sur les lacets positifs des plongements legendriens lâches." Thesis, Nantes, 2016. http://www.theses.fr/2016NANT4045/document.
Full textIn the thesis, we have studied the problem of positive Lengendrian isotopies. That is to say, the isotopies preservepo the contact structure and the hamiltonnian functions of the isotopies are positive. We have proved that for a loose Legendrian there exists a positive loop of Legendrian embeddings based in it. We treated this result in two cases. In lower dimensions cases, we constructed positive loops by hand. In higher dimensions cases, we applied the advanced h-principle techniques. Given a loose Legendrian embedding, firstly, by the holonomic approximation, we constructed a loop of Legendrian embeddings based in it which is positive away from a finite number of disks. Secondly, we deformed it to a positive loop by the idea of convex integration. The result has two immediate applications. Firstly, we reprove the theorem that the spaces of contact elements are tight without holomorphic curves techniques. Secondly, we proved the contact product of an overtwisted contact manifold is overtwisted and the diagonal is loose, furthermore, the diagonal is in positive loop. In the end, we have defined a partial order on the universal cover of the contactomorphism group by positive Legendrian isotopies in the contact product. It will help us to study the properties of contactomorphism via positive Legendrian isotopies
Gaillard, Loïc. "Espaces de Müntz, plongements de Carleson, et opérateurs de Cesàro." Thesis, Artois, 2017. http://www.theses.fr/2017ARTO0406/document.
Full textFor a sequence ⋀ = (λn) satisfying the Müntz condition Σn 1/λn < +∞ and for p ∈ [1,+∞), we define the Müntz space Mp⋀ as the closed subspace of Lp([0, 1]) spanned by the monomials yn : t ↦ tλn. The space M∞⋀ is defined in the same way as a subspace of C([0, 1]). When the sequence (λn + 1/p)n is lacunary with a large ratio, we prove that the sequence of normalized Müntz monomials (gn) in Lp is (1 + ε)-isometric to the canonical basis of lp. In the case p = +∞, the monomials (yn) form a sequence which is (1 + ε)-isometric to the summing basis of c. These results are asymptotic refinements of a well known theorem for the lacunary sequences. On the other hand, for p ∈ [1, +∞), we investigate the Carleson measures for Müntz spaces, which are defined as the Borel measures μ on [0; 1) such that the embedding operator Jμ,p : Mp⋀ ⊂ Lp(μ) is bounded. When ⋀ is lacunary, we prove that if the (gn) are uniformly bounded in Lp(μ), then for any q > p, the measure μ is a Carleson measure for Mq⋀. These questions are closely related to the behaviour of μ in the neighborhood of 1. Wealso find some geometric conditions about the behaviour of μ near the point 1 that ensure the compactness of Jμ,p, or its membership to some thiner operator ideals. More precisely, we estimate the approximation numbers of Jμ,p in the lacunary case and we even obtain some equivalents for particular lacunary sequences ⋀. At last, we show that the essentialnorm of the Cesàro-mean operator Γp : Lp → Lp coincides with its norm, which is p'. This result is also valid for the Cesàro sequence operator. We introduce some Müntz subspaces of the Cesàro function spaces Cesp, for p ∈ [1, +∞]. We show that the value of the essential norm of the multiplication operator TΨ is ∥Ψ∥∞ in the Cesàaro spaces. In the Müntz-Cesàrospaces, the essential norm of TΨ is equal to |Ψ(1)|
Catusse, Nicolas. "Spanners pour des réseaux géométriques et plongements dans le plan." Thesis, Aix-Marseille 2, 2011. http://www.theses.fr/2011AIX22119/document.
Full textIn this thesis, we study several problems related to the design of geometric networks and isometric embeddings into the plane.We start by considering the generalization of the classical Minimum Manhattan Network problem to all normed planes. We search the minimum network that connects each pair of terminals by a shortest path in this norm. We propose a factor 2.5 approximation algorithm in time O(mn^3), where n is the number of terminals and m is the number of directions of the unit ball.The second problem presented is an oriented version of the minumum Manhattan Network problem, we want to obtain a minimum oriented network such that for each pair u, v of terminals, there is a shortest rectilinear path from u to v and another path from v to u.We describe a factor 2 approximation algorithm with complexity O(n^3) where n is the number of terminals for this problem.Then we study the problem of finding a planar spanner (a subgraph which approximates the distances) of the Unit Disk Graph (UDG) which is used to modelize wireless ad hoc networks. We present an algorithm for computing a constant hop stretch factor planar spanner for all UDG. This algorithm uses only local properties and it can be implemented in distributed manner.Finally, we study the problem of recognizing metric spaces that can be isometrically embbed into the rectilinear plane and we provide an optimal time O(n^2) algorithm to solve this problem. We also study the generalization of this problem to all normed planes whose unit ball is a centrally symmetric convex polygon
Netillard, François. "Plongements grossièrement Lipschitz et presque Lipschitz dans les espaces de Banach." Thesis, Bourgogne Franche-Comté, 2019. http://www.theses.fr/2019UBFCD020/document.
Full textThe central theme of this thesis is the study of embeddings of metric spaces into Banach spaces.The first study focuses on the coarse Lipschitz embeddings between James Spaces Jp for p≻1 and p finite. We obtain that, for p,q different, Jq does not coarse Lipschitz embed into Jp. We also obtain, in the case where q≺p, that the compression exponent of Jq in Jp is lower or equal to q/p. Another natural question is to know whether we have similar results for the dual spaces of James spaces. We obtain that, for p,q different, Jp* does not coarse Lipschitz embed into Jq*. Further to this work, we establish a more general result about the coarse Lipschitz embeddability of a Banach space which has a q-AUS norm into a Banach space which has a p-AMUC norm for p≺q. With the help of a renorming theorem, we deduce also a result about the Szlenk index. Moreover, after defining the quasi-Lipschitz embeddability, which is slightly different to the almost Lipschitz embeddability, we obtain the following result: For two Banach spaces X, if X is crudely finitely representable with constant C (where C≻1) in any subspace of Y of finite codimension, then every proper subset M of X quasi-Lipschitz embeds into Y. To conclude, we obtain the following corollary: Let X be a locally minimal Banach space, and Y be a Banach space which is crudely finitely representable in X. Then, for M a proper subspace of Y, M quasi-Lipschitz embeds into X
Dutailly, Bruno. "Plongement de surfaces continues dans des surfaces discrètes épaisses." Thesis, Bordeaux, 2016. http://www.theses.fr/2016BORD0444/document.
Full textIn the context of archaeological sciences, 3D images produced by Computer Tomography scanners are segmented into regions of interest corresponding to virtual objects in order to make some scientific analysis. These virtual objects are often used for the purpose of performing accurate measurements. Some of these analysis require extracting the surface of the regions of interest. This PhD falls within this framework and aims to improve the accuracy of surface extraction. We present in this document our contributions : first of all, the weighted HMH algorithm whose objective is to position precisely a point at the interface between two materials. But, applied to surface extraction, this method often leads to topology problems on the resulting surface. So we proposed two other methods : The discrete HMH method which allows to refine the 3D object segmentation, and the surface HMH method which allows a constrained surface extraction ensuring a topologically correct surface. It is possible to link these two methods on a pre-segmented 3D image in order to obtain a precise surface extraction of the objects of interest These methods were evaluated on simulated CT-scan acquisitions of synthetic objects and real acquisitions of archaeological artefacts
Boroş, Emanuela. "Neural Methods for Event Extraction." Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLS302/document.
Full textWith the increasing amount of data and the exploding number data sources, the extraction of information about events, whether from the perspective of acquiring knowledge or from a more directly operational perspective, becomes a more and more obvious need. This extraction nevertheless comes up against a recurring difficulty: most of the information is present in documents in a textual form, thus unstructured and difficult to be grasped by the machine. From the point of view of Natural Language Processing (NLP), the extraction of events from texts is the most complex form of Information Extraction (IE) techniques, which more generally encompasses the extraction of named entities and relationships that bind them in the texts. The event extraction task can be represented as a complex combination of relations linked to a set of empirical observations from texts. Compared to relations involving only two entities, there is, therefore, a new dimension that often requires going beyond the scope of the sentence, which constitutes an additional difficulty. In practice, an event is described by a trigger and a set of participants in that event whose values are text excerpts. While IE research has benefited significantly from manually annotated datasets to learn patterns for text analysis, the availability of these resources remains a significant problem. These datasets are often obtained through the sustained efforts of research communities, potentially complemented by crowdsourcing. In addition, many machine learning-based IE approaches rely on the ability to extract large sets of manually defined features from text using sophisticated NLP tools. As a result, adaptation to a new domain is an additional challenge. This thesis presents several strategies for improving the performance of an Event Extraction (EE) system using neural-based approaches exploiting morphological, syntactic, and semantic properties of word embeddings. These have the advantage of not requiring a priori modeling domain knowledge and automatically generate a much larger set of features to learn a model. More specifically, we proposed different deep learning models for two sub-tasks related to EE: event detection and argument detection and classification. Event Detection (ED) is considered an important subtask of event extraction since the detection of arguments is very directly dependent on its outcome. ED specifically involves identifying instances of events in texts and classifying them into specific event types. Classically, the same event may appear as different expressions and these expressions may themselves represent different events in different contexts, hence the difficulty of the task. The detection of the arguments is based on the detection of the expression considered as triggering the event and ensures the recognition of the participants of the event. Among the difficulties to take into account, it should be noted that an argument can be common to several events and that it does not necessarily identify with an easily recognizable named entity. As a preliminary to the introduction of our proposed models, we begin by presenting in detail a state-of-the-art model which constitutes the baseline. In-depth experiments are conducted on the use of different types of word embeddings and the influence of the different hyperparameters of the model using the ACE 2005 evaluation framework, a standard evaluation for this task. We then propose two new models to improve an event detection system. One allows increasing the context taken into account when predicting an event instance by using a sentential context, while the other exploits the internal structure of words by taking advantage of seemingly less obvious but essentially important morphological knowledge. We also reconsider the detection of arguments as a high-order relation extraction and we analyze the dependence of arguments on the ED task
Bérard, Alexandre. "Neural machine translation architectures and applications." Thesis, Lille 1, 2018. http://www.theses.fr/2018LIL1I022/document.
Full textThis thesis is centered on two main objectives: adaptation of Neural Machine Translation techniques to new tasks and research replication. Our efforts towards research replication have led to the production of two resources: MultiVec, a framework that facilitates the use of several techniques related to word embeddings (Word2vec, Bivec and Paragraph Vector); and a framework for Neural Machine Translation that implements several architectures and can be used for regular MT, Automatic Post-Editing, and Speech Recognition or Translation. These two resources are publicly available and now extensively used by the research community. We extend our NMT framework to work on three related tasks: Machine Translation (MT), Automatic Speech Translation (AST) and Automatic Post-Editing (APE). For the machine translation task, we replicate pioneer neural-based work, and do a case study on TED talks where we advance the state-of-the-art. Automatic speech translation consists in translating speech from one language to text in another language. In this thesis, we focus on the unexplored problem of end-to-end speech translation, which does not use an intermediate source-language text transcription. We propose the first model for end-to-end AST and apply it on two benchmarks: translation of audiobooks and of basic travel expressions. Our final task is automatic post-editing, which consists in automatically correcting the outputs of an MT system in a black-box scenario, by training on data that was produced by human post-editors. We replicate and extend published results on the WMT 2016 and 2017 tasks, and propose new neural architectures for low-resource automatic post-editing
Mabrouki, Mbarka. "Etude de la préservation des propriétés temporelles des réseaux de régulation génétique au travers du plongement : vers une caractérisation des systèmes complexes par l'émergence de propriétés." Thesis, Evry-Val d'Essonne, 2010. http://www.theses.fr/2010EVRY0039/document.
Full textThe thesis proposes a generic framework to denote specifications of basic system components and to characterize the notion of complex system by the presence of emergent property, that are either in conflict with the properties attached to the subsystems the constituent, either are directly due to the cooperation of the subsystems. The framework is declined for the cases of the relative systems and genetic regulatory network
Cassagnes, Cyril. "Architecture autonome et distribuée d’adressage et de routage pour la flexibilité des communications dans l’internet." Thesis, Bordeaux 1, 2012. http://www.theses.fr/2012BOR14600/document.
Full textLocal routing schemes based on virtual coordinates taken from the hyperbolic plane have attracted considerable interest in recent years.However, solutions have been applied to ad-hoc and sensor networks having a random topology and a limited number of nodes. In other hand, some research has focused on the creation of network topology models based on hyperbolic geometric laws. In this case, it has been shown that these graphs have an Internet-like topology and that local hyperbolic routing achieves a near perfect efficiency. However, with these graphs, routing success is not guaranteed even if no failures happen. In this thesis, we aim at building a scalable system for creating overlay networks on top of the Internet that would provide reliable addressing and routing service to its members in a dynamic environment.Next, we investigate how well P2PTV networks would support a growing number of users. In this thesis, we try to address this question by studying scalability and efficiency factors in a typical P2P based live streaming network. Through the use of the data provided by Zattoo a production P2PTV network, we carry out simulations whose results show that there are still hurdles to overcome before P2P based live streaming could depend uniquely of their users
Binois, Mickaël. "Uncertainty quantification on pareto fronts and high-dimensional strategies in bayesian optimization, with applications in multi-objective automotive design." Thesis, Saint-Etienne, EMSE, 2015. http://www.theses.fr/2015EMSE0805/document.
Full textThis dissertation deals with optimizing expensive or time-consuming black-box functionsto obtain the set of all optimal compromise solutions, i.e. the Pareto front. In automotivedesign, the evaluation budget is severely limited by numerical simulation times of the considered physical phenomena. In this context, it is common to resort to “metamodels” (models of models) of the numerical simulators, especially using Gaussian processes. They enable adding sequentially new observations while balancing local search and exploration. Complementing existing multi-objective Expected Improvement criteria, we propose to estimate the position of the whole Pareto front along with a quantification of the associated uncertainty, from conditional simulations of Gaussian processes. A second contribution addresses this problem from a different angle, using copulas to model the multi-variate cumulative distribution function. To cope with a possibly high number of variables, we adopt the REMBO algorithm. From a randomly selected direction, defined by a matrix, it allows a fast optimization when only a few number of variables are actually influential, but unknown. Several improvements are proposed, such as a dedicated covariance kernel, a selection procedure for the low dimensional domain and of the random directions, as well as an extension to the multi-objective setup. Finally, an industrial application in car crash-worthiness demonstrates significant benefits in terms of performance and number of simulations required. It has also been used to test the R package GPareto developed during this thesis
Ferré, Arnaud. "Représentations vectorielles et apprentissage automatique pour l’alignement d’entités textuelles et de concepts d’ontologie : application à la biologie." Thesis, Université Paris-Saclay (ComUE), 2019. http://www.theses.fr/2019SACLS117/document.
Full textThe impressive increase in the quantity of textual data makes it difficult today to analyze them without the assistance of tools. However, a text written in natural language is unstructured data, i.e. it cannot be interpreted by a specialized computer program, without which the information in the texts remains largely under-exploited. Among the tools for automatic extraction of information from text, we are interested in automatic text interpretation methods for the entity normalization task that consists in automatically matching text entitiy mentions to concepts in a reference terminology. To accomplish this task, we propose a new approach by aligning two types of vector representations of entities that capture part of their meanings: word embeddings for text mentions and concept embeddings for concepts, designed specifically for this work. The alignment between the two is done through supervised learning. The developed methods have been evaluated on a reference dataset from the biological domain and they now represent the state of the art for this dataset. These methods are integrated into a natural language processing software suite and the codes are freely shared
Alam, Ihab Al. "Géométrie des espaces de Müntz et opérateurs de composition à poids." Thesis, Lille 1, 2008. http://www.theses.fr/2008LIL10068/document.
Full textThe main subject of this PHD thesis is the study of sorne geometric aspects of Müntz spaces (M'A and M~) in C([O, 1]) and LP([O, 1]),1 ::; p < 00. This work is composed offour chapters. The first chapter is devoted to preliminary. ln the second chapter, we prove sever al basic properties of Müntz spaces, these properties explain the geometric nature of these spaces. There is also a new generalization of Müntz spaces by considering the Müntz polynomials with coefficient in any Banach space X. The aim of the third one is to construct a Müntz space having no complement in LI ([0,1]). As an application of this work, we obtain sorne results that were recently obtained in the monograph of Vladimir I. Gurariy and Wolfgang Lusky, but with a method completely different. We also provide an explicit Schauder basis equivalent to the canonical base in gl for sorne Müntz spaces MX, with A not lacunary. ln a second part of this chapter, we study the case LP([O, 1]), 1 ::; p < 00, we will see that sorne phenomena still true in the case 1 < p < 00. Finally, in the fourth chapter, we discuss the problem of compactness for weighted composition operators T'ljJoC
Bogso, Antoine Marie. "Étude de peacocks sous l'hypothèse de monotonie conditionnelle et de positivité totale." Thesis, Université de Lorraine, 2012. http://www.theses.fr/2012LORR0152/document.
Full textThis thesis deals with real valued stochastic processes which increase in the convex order. We call them peacocks. A remarkable result due to Kellerer states that a real valued process is a peacock if and only if it has the same one-dimensional marginals as a martingale. Such a martingale is said to be associated to this process. But in his article, Kellerer provides neither an example of peacock nor a concrete idea to construct an associated martingale to a given peacock. Hence, as other investigations on peacocks, our study has two purposes. We first exhibit new families of peacocks and then, we contruct associated martingales to certain of them. In the first three chapters, we exhibit several classes of peacocks using successively the notions of conditional monotonicity, very strong peacock and total positivity of order 2. In particular, we provide many extensions of Carr-Ewald-Xiao result which states that the arithmetic mean of geometric Brownian motion, also called "Asian option" is a peacock. The purpose of the last chapter is to construct associated martingales to certain peacocks. To this end, we use Azéma-Yor and Bertoin-Le Jan embedding algorithms. The originality of this chapter is the use of total positivity of order 2 in the study of Azéma-Yor embedding algorithm
Simonovsky, Martin. "Deep learning on attributed graphs." Thesis, Paris Est, 2018. http://www.theses.fr/2018PESC1133/document.
Full textGraph is a powerful concept for representation of relations between pairs of entities. Data with underlying graph structure can be found across many disciplines, describing chemical compounds, surfaces of three-dimensional models, social interactions, or knowledge bases, to name only a few. There is a natural desire for understanding such data better. Deep learning (DL) has achieved significant breakthroughs in a variety of machine learning tasks in recent years, especially where data is structured on a grid, such as in text, speech, or image understanding. However, surprisingly little has been done to explore the applicability of DL on graph-structured data directly.The goal of this thesis is to investigate architectures for DL on graphs and study how to transfer, adapt or generalize concepts working well on sequential and image data to this domain. We concentrate on two important primitives: embedding graphs or their nodes into a continuous vector space representation (encoding) and, conversely, generating graphs from such vectors back (decoding). To that end, we make the following contributions.First, we introduce Edge-Conditioned Convolutions (ECC), a convolution-like operation on graphs performed in the spatial domain where filters are dynamically generated based on edge attributes. The method is used to encode graphs with arbitrary and varying structure.Second, we propose SuperPoint Graph, an intermediate point cloud representation with rich edge attributes encoding the contextual relationship between object parts. Based on this representation, ECC is employed to segment large-scale point clouds without major sacrifice in fine details.Third, we present GraphVAE, a graph generator allowing to decode graphs with variable but upper-bounded number of nodes making use of approximate graph matching for aligning the predictions of an autoencoder with its inputs. The method is applied to the task of molecule generation
Tuong, Frédéric. "Constructing Semantically Sound Object-Logics for UML/OCL Based Domain-Specific Languages." Thesis, Université Paris-Saclay (ComUE), 2016. http://www.theses.fr/2016SACLS085/document.
Full textObject-based and object-oriented specification languages (likeUML/OCL, JML, Spec#, or Eiffel) allow for the creation and destruction, casting and test for dynamic types of statically typed objects. On this basis, class invariants and operation contracts can be expressed; the latter represent the key elements of object-oriented specifications. A formal semantics of object-oriented data structures is complex: imprecise descriptions can often imply different interpretations in resulting tools. In this thesis we demonstrate how to turn a modern proof environment into a meta-tool for definition and analysis of formal semantics of object-oriented specification languages. Given a representation of a particular language embedded in Isabelle/HOL, we build for this language an extended Isabelle environment by using a particular method of code generation, which actually involves several variants of code generation. The result supports the asynchronous editing, type-checking, and formal deduction activities, all "inherited" from Isabelle. Following this method, we obtain an object-oriented modelling tool for textual UML/OCL. We also integrate certain idioms not necessarily present in UML/OCL --- in other words, we develop support for domain-specific dialects of UML/OCL. As a meta construction, we define a meta-model of a part of UML/OCL in HOL, a meta-model of a part of the Isabelle API in HOL, and a translation function between both in HOL. The meta-tool will then exploit two kinds of code generation to produce either fairly efficient code, or fairly readable code. Thus, this provides two animation modes to inspect in more detail the semantics of a language being embedded: by loading at a native speed its semantics, or just delay at another "meta"-level the previous experimentation for another type-checking time in Isabelle, be it for performance, testing or prototyping reasons. Note that generating "fairly efficient code", and "fairly readable code" include the generation of tactic code that proves a collection of theorems forming an object-oriented datatype theory from a denotational model: given a UML/OCL class model, the proof of the relevant properties for casts, type-tests, constructors and selectors are automatically processed. This functionality is similar to the datatype theory packages in other provers of the HOL family, except that some motivations have conducted the present work to program high-level tactics in HOL itself. This work takes into account the most recent developments of the UML/OCL 2.5 standard. Therefore, all UML/OCL types including the logic types distinguish two different exception elements: invalid (exception) and null (non-existing element). This has far-reaching consequences on both the logical and algebraic properties of object-oriented data structures resulting from class models. Since our construction is reduced to a sequence of conservative theory extensions, the approach can guarantee logical soundness for the entire considered language, and provides a methodology to soundly extend domain-specific languages
Arène, Christophe. "Géométrie et arithmétique explicites des variétés abéliennes et applications à la cryptographie." Thesis, Aix-Marseille 2, 2011. http://www.theses.fr/2011AIX22069/document.
Full textThe main objects we study in this PhD thesis are the equations describing the group morphism on an abelian variety, embedded in a projective space, and their applications in cryptograhy. We denote by g its dimension and k its field of definition. This thesis is built in two parts. The first one is concerned by the study of Edwards curves, a model for elliptic curves having a cyclic subgroup of k-rational points of order 4, known in cryptography for the efficiency of their addition law and the fact that it can be defined for any couple of k-rational points (k-complete addition law). We give the corresponding geometric interpretation and deduce explicit formulae to calculate the reduced Tate pairing on twisted Edwards curves, whose efficiency compete with currently used elliptic models. The part ends with the generation, specific to pairing computation, of Edwards curves with today's cryptographic standard sizes. In the second part, we are interested in the notion of completeness introduced above. This property is cryptographically significant, indeed it permits to avoid physical attacks as side channel attacks, on elliptic -- or hyperelliptic -- curves cryptosystems. A preceeding work of Lange and Ruppert, based on cohomology of line bundles, brings a theoretic approach of addition laws. We present three important results: first of all we generalize a result of Bosma and Lenstra by proving that the group morphism can not be described by less than g+1 addition laws on the algebraic closure of k. Next, we prove that if the absolute Galois group of k is infinite, then any abelian variety can be projectively embedded together with a k-complete addition law. Moreover, a cryptographic use of abelian varieties restricting us to the dimension one and two cases, we prove that such a law exists for their classical projective embedding. Finally, we develop an algorithm, based on the theory of theta functions, computing this addition law in P^15 on the Jacobian of a genus two curve given in Rosenhain form. It is now included in AVIsogenies, a Magma package
Guo, Gaoyue. "Continuous-time Martingale Optimal Transport and Optimal Skorokhod Embedding." Thesis, Université Paris-Saclay (ComUE), 2016. http://www.theses.fr/2016SACLX038/document.
Full textThis PhD dissertation presents three research topics, the first two being independent and the last one relating the first two issues in a concrete case.In the first part we focus on the martingale optimal transport problem on the Skorokhod space, which aims at studying systematically the tightness of martingale transport plans. Using the S-topology introduced by Jakubowski, we obtain the desired tightness which yields the upper semicontinuity of the primal problem with respect to the marginal distributions, and further the first duality. Then, we provide also two dual formulations that are related to the robust superhedging in financial mathematics, and we establish the corresponding dualities by adapting the dynamic programming principle and the discretization argument initiated by Dolinsky and Soner.The second part of this dissertation addresses the optimal Skorokhod embedding problem under finitely-many marginal constraints. We formulate first this optimization problem by means of probability measures on an enlarged space as well as its dual problems. Using the classical convex duality approach together with the optimal stopping theory, we obtain the duality results. We also relate these results to the martingale optimal transport on the space of continuous functions, where the corresponding dualities are derived for a special class of reward functions. Next, We provide an alternative proof of the monotonicity principle established in Beiglbock, Cox and Huesmann, which characterizes the optimizers by their geometric support. Finally, we show a stability result that is twofold: the stability of the optimization problem with respect to target marginals and the relation with another optimal embedding problem.The last part concerns the application of stochastic control to the martingale optimal transport with a payoff depending on the local time, and the Skorokhod embedding problem. For the one-marginal case, we recover the optimizers for both primal and dual problems through Vallois' solutions, and show further the optimality of Vallois' solutions, which relates the martingale optimal transport and the optimal Skorokhod embedding. As for the two-marginal case, we obtain a generalization of Vallois' solution. Finally, a special multi-marginal case is studied, where the stopping times given by Vallois are well ordered
Khalil, Maya. "Classes de Steinitz, codes cycliques de Hamming et classes galoisiennes réalisables d'extensions non abéliennes de degré p³." Thesis, Valenciennes, 2016. http://www.theses.fr/2016VALE0012/document.
Full textXiong, Xiao. "Espaces de fonctions sur les tores quantiques." Thesis, Besançon, 2015. http://www.theses.fr/2015BESA2029/document.
Full textThis thesis gives a systematic study of Sobolev, Besov and Triebel-Lizorkin spaces on a noncommutative d-torus. We prove, arnong other basic properties, the lifting theorem for all these spaces and a Poincaré type inequality for Sobolev spaces. We establish the embedding inequalities of all these spaces, including the l3esov and Sobolev embedding theorems. We obtain Littlewood-Paley type characterizations for Besov and 'friebel-Lizorki spaces in a general way, as well as the concrete ones internas of the Poisson, heat semigroups and differences. Some of them are new even in the commutative case, for instance, oui Poisson semigroup characterization of Besov and Triebel-Lizorkin spaces improves the classical ones. As a consequence of the characterization of the Besov spaces by differences, we extend to the quantum setting the recent results of Bourgain-Brézis -Mironescu and Maz'ya-Shaposhnikova on the limits of l3esov florins. We investigate the interpolation of all these spaces, in particular, deterrnine explicitly the K-functional of the couple of Lp space and Sobolev space, winch is the quantum analogue of a classical result due to Johnen and Scherer Finally, we show that the completely bounded Fourier multipliers on all these spaces coincide with those on the corresponding spaces on the usuel d-torus. We also give a quite simple description of (completely) bounded Fourier multipliers on the Besov spaces in ternis of their behavior on the Lp-components in the Littlevvood-Paley decomposition
Bucher, Maxime. "Apprentissage et exploitation de représentations sémantiques pour la classification et la recherche d'images." Thesis, Normandie, 2018. http://www.theses.fr/2018NORMC250/document.
Full textIn this thesis, we examine some practical difficulties of deep learning models.Indeed, despite the promising results in computer vision, implementing them in some situations raises some questions. For example, in classification tasks where thousands of categories have to be recognised, it is sometimes difficult to gather enough training data for each category.We propose two new approaches for this learning scenario, called <>. We use semantic information to model classes which allows us to define models by description, as opposed to modelling from a set of examples.In the first chapter we propose to optimize a metric in order to transform the distribution of the original data and to obtain an optimal attribute distribution. In the following chapter, unlike the standard approaches of the literature that rely on the learning of a common integration space, we propose to generate visual features from a conditional generator. The artificial examples can be used in addition to real data for learning a discriminant classifier. In the second part of this thesis, we address the question of computational intelligibility for computer vision tasks. Due to the many and complex transformations of deep learning algorithms, it is difficult for a user to interpret the returned prediction. Our proposition is to introduce what we call a <> in the processing pipeline, which is a crossing point in which the representation of the image is entirely expressed with natural language, while retaining the efficiency of numerical representations. This semantic bottleneck allows to detect failure cases in the prediction process so as to accept or reject the decision
Trouillon, Théo. "Modèles d'embeddings à valeurs complexes pour les graphes de connaissances." Thesis, Université Grenoble Alpes (ComUE), 2017. http://www.theses.fr/2017GREAM048/document.
Full textThe explosion of widely available relational datain the form of knowledge graphsenabled many applications, including automated personalagents, recommender systems and enhanced web search results.The very large size and notorious incompleteness of these data basescalls for automatic knowledge graph completion methods to make these applicationsviable. Knowledge graph completion, also known as link-prediction,deals with automatically understandingthe structure of large knowledge graphs---labeled directed graphs---topredict missing entries---labeled edges. An increasinglypopular approach consists in representing knowledge graphs as third-order tensors,and using tensor factorization methods to predict their missing entries.State-of-the-art factorization models propose different trade-offs between modelingexpressiveness, and time and space complexity. We introduce a newmodel, ComplEx---for Complex Embeddings---to reconcile both expressivenessand complexity through the use of complex-valued factorization, and exploreits link with unitary diagonalization.We corroborate our approach theoretically and show that all possibleknowledge graphs can be exactly decomposed by the proposed model.Our approach based on complex embeddings is arguably simple,as it only involves a complex-valued trilinear product,whereas other methods resort to more and more complicated compositionfunctions to increase their expressiveness. The proposed ComplEx model isscalable to large data sets as it remains linear in both space and time, whileconsistently outperforming alternative approaches on standardlink-prediction benchmarks. We also demonstrateits ability to learn useful vectorial representations for other tasks,by enhancing word embeddings that improve performanceson the natural language problem of entailment recognitionbetween pair of sentences.In the last part of this thesis, we explore factorization models abilityto learn relational patterns from observed data.By their vectorial nature, it is not only hard to interpretwhy this class of models works so well,but also to understand where they fail andhow they might be improved. We conduct an experimentalsurvey of state-of-the-art models, not towardsa purely comparative end, but as a means to get insightabout their inductive abilities.To assess the strengths and weaknesses of each model, we create simple tasksthat exhibit first, atomic properties of knowledge graph relations,and then, common inter-relational inference through synthetic genealogies.Based on these experimental results, we propose new researchdirections to improve on existing models, including ComplEx