Dissertations / Theses on the topic 'Systèmes de données échantillonnées'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Systèmes de données échantillonnées.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Postoyan, Romain. "Commande et construction d’observateurs pour des systèmes non linéaires incertains à données échantillonnées et en réseau." Paris 11, 2009. http://www.theses.fr/2009PA112163.
Full textThe rise of digital technologies has promoted the development of new controllers implementations that have many advantages compared to traditional control structures. Indeed, digital controllers have become very popular due to their low cost and great flexibility in comparison with analogical controllers. The implementation of control structures via a network also offers a new point of view. They are generally easier to use and to maintain than point-by-point wiring, they allow one to significantly reduce data exchanges and, as a consequence, the energy cost. However, induced communication constraints can have a significant impact on system dynamical behaviour. In this thesis, we first propose adaptive and robust stabilisation methods for classes of nonlinear sampled-data systems affected by uncertainties; the main objective is to improve closed-loop performance compared to the emulation of a continuous-time control law. When data exchanges are sampled and time-scheduled via a network, we have developed a framework for the observer design by emulation. It is shown that various observer designs (linear, high gain, circle criterion) and various network configurations fit our framework
Mattioni, Mattia. "Stabilisation des systèmes échantillonnés en cascade et avec retards." Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLS118/document.
Full textOver the last decades the methodologies of dynamical systems and control theory have been playing an increasingly relevant role in a lot of situations of practical interest. Though, a lot of theoretical problem still remain unsolved. Among all, the ones concerning stability and stabilization are of paramount importance. In order to stabilize a physical (or not) system, it is necessary to acquire and interpret heterogeneous information on its behavior in order to correctly intervene on it. In general, those information are not available through a continuous flow but are provided in a synchronous or asynchronous way. This issue has to be unavoidably taken into account for the design of the control action. In a very natural way, all those heterogeneities define an hybrid system characterized by both continuous and discrete dynamics. This thesis is contextualized in this framework and aimed at proposing new methodologies for the stabilization of sampled-data nonlinear systems with focus toward the stabilization of cascade dynamics. In doing so, we shall propose a small number of tools for constructing sampled-data feedback laws stabilizing the origin of sampled-data nonlinear systems admitting cascade interconnection representations. To this end, we shall investigate on the effect of sampling on the properties of the continuous-time system while enhancing design procedures requiring no extra assumptions over the sampled-data equivalent model.Finally, we shall show the way sampling positively affects nonlinear retarded dynamics affected by a fixed and known time-delay over the input signal by enforcing on the implicit cascade representation the sampling process induces onto the retarded system. Academic examples will illustrate the computational aspects together with their performances throughout the whole manuscript
Chen, Fengwei. "Contributions à l'identification de modèles à temps continu à partir de données échantillonnées à pas variable." Thesis, Université de Lorraine, 2014. http://www.theses.fr/2014LORR0149/document.
Full textThe output of a system is always corrupted by additive noise, therefore it is more practical to develop estimation algorithms that are capable of handling noisy data. The effect of white additive noise has been widely studied, while a colored additive noise attracts less attention, especially for a continuous-time (CT) noise. Sampling issues of CT stochastic processes are reviewed in this thesis, several sampling schemes are presented. Estimation of a CT stochastic process is studied. An expectation-maximization-based (EM) method to CT autoregressive/autoregressive moving average model is developed, which gives accurate estimation over a large range of sampling interval. Estimation of CT Box-Jenkins models is also considered in this thesis, in which the noise part is modeled to improve the performance of plant model estimation. The proposed method for CT Box-Jenkins model identification is in a two-step and iterative framework. Two-step means the plant and noise models are estimated in a separate and alternate way, where in estimating each of them, the other is assumed to be fixed. More specifically, the plant is estimated by refined instrumental variable (RIV) method while the noise is estimated by EM algorithm. Iterative means that the proposed method repeats the estimation procedure several times until a optimal estimate is found. Many practical systems have inherent time-delay. The problem of identifying delayed systems are of great importance for analysis, prediction or control design. The presence of a unknown time-delay greatly complicates the parameter estimation problem, essentially because the model are not linear with respect to the time-delay. An approach to continuous-time model identification of time-delay systems, combining a numerical search algorithm for the delay with the RIV method for the dynamic has been developed in this thesis. In the proposed algorithm, the system parameters and time-delay are estimated reciprocally in a bootstrap manner. The time-delay is estimated by an adaptive gradient-based method, whereas the system parameters are estimated by the RIV method. Since numerical method is used in this algorithm, the bootstrap method is likely to converge to local optima, therefore a low-pass filter has been used to enlarge the convergence region for the time-delay. The performance of the proposed algorithms are evaluated by numerical examples
Ricciardi, Celsi Lorenzo. "Commande non linéaire multi-agents : applications aux systèmes en réseau." Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLS017/document.
Full textThe objective of this PhD thesis is (i) to investigate and develop methods for the analysis and design of linear and nonlinear networked control systems and (ii) to show the potential of such approaches in relevant complex applications. In this respect, multi-agent systems theory, algebraic graph theory and consensus are the most interesting methodological tools, and specific attention is paid to the characterization of the relationships between, on the one hand, the topology of the communication graph that underlies the evolution of the considered multiagent system and, on the other hand, the spectral properties of the Laplacian matrix associated with the graph itself. The control of a group of autonomous agents is investigated from different perspectives. The main control objective is to make sure that the agents work together in a cooperative fashion, where cooperation accounts for the close relationship among all agents in the team, with information sharing playing an important role. In particular, various problems regarding consensus/agreement/synchronization/rendezvous are investigated with the specific aim of driving a group of agents to some common state. Consensus is investigated in a discrete-time setting due to the fact that the system dynamics is normally continuous while the measurements and control inputs might only be made in a sampled-data setting. Moreover, game theory is relied upon in order to cope with distributed multi-agent coordination problems, with application to Software Defined Networks. In this respect, it can be shown that, under properly designed protocols, the players converge to a unique Wardrop equilibrium. We focus on distributed control, since this approach shows obvious benefits over centralization, such as scalability and robustness. Yet, it also has its own drawbacks: among all, one drawback is that each agent cannot effectively predict the overall group behaviour based on only local information. Some attention is also devoted to the need for securing power grids against the danger of cyber-physical attacks through the development of distributed intelligence technologies accompanied by appropriate security enforcements. In this respect, based on realistic power network topologies, we briefly present the design of a protection scheme against closed-loop single-point and multi-point dynamic load altering attacks. This is done by formulating and solving a non-convex optimization problem subject to a Lyapunov stability constraint for the autonomous multiagent representation of a power system obtained after linearization and application of the attack and frequency control laws. Eventually, we show some other results achieved in terms of the exact steeering of finite sampled nonlinear dynamics with input delays, of sampled-data stabilization and quasi-halo orbit following around the L₂ translunar libration point, and of heuristic algorithms based on multi-agent reinforcement learning methods capable of performing optimal adaptive Quality of Service/Quality of Experience control in model-free scenarios
Louis, Julien. "Discrétisation des systèmes de Lur'e : stabilisation et consistance." Thesis, Université de Lorraine, 2015. http://www.theses.fr/2015LORR0080/document.
Full textRecent studies dealing with discrete-time (switched) Lur’e systems involve an adapted Lur’e type function exhibiting possibly non-convex and disconnected level sets. These properties raise fundamental issues in the case of discrete-time Lur’e system obtained by the sampling of a continuous time one. This PhD thesis aims at answering these questions. The first contribution is to avoid the discrete-time disconnected level sets by a decreasing sequence of bounded and connected sets that converges to the origin and that contain the future of the continuous-time trajectory. The second contribution deals with the joint stabilization of a sampled-data Lur’e system with non-uniform sampling. When the sampling period belongs to a finite set of values, this problem is reformulated as the joint stabilization of a discrete-time Lur’e switched system with norm-bounded uncertain parameters. Futhermore, if a quadratic criterion is associated with each mode, a min-switching strategy combined with LMI constraints allow to provide a solution to this problem. Finally the property of consistency for discrete-time switched Lur’e systems is investigated. It is shown that the min-switching strategy is consistent with respect to quadratic upper bounds of the performances. This result is applied on the stabilization of Lur’e systems with non-uniform sampling
Mizere, Dominique. "Contributions à la modélisation et à l'analyse statistique des données de dénombrement." Pau, 2006. http://www.theses.fr/2006PAUU3001.
Full textThe Poisson model is a benchmark model for the statistical analysis of the count data. Sometimes count data exhibit overdispersion or underdispersion resulting in the lack of the Poisson model. The aim of this work is to present an overview of the probability models that can provide alternative framework for the statistical analysis of count data, among others, Katz's model and the weighted Poisson model. Some weighted Poisson models will be fitted on the data collected in Republic of Congo and related to a sample of spiraling whitefly Aleurodicus dispersus Russell (Homoptera : Aleyrodidae) described by the preimaginal span (from eggs to adult stage) measured on days, the total number of eggs and the laying span. The linear regression issue between these three variables will be discussed as well. We must afterwards evaluate the performances of some tests procedures devoted to valid the fitness of count data by Poisson distribution against general alternatives of overdispersion or underdispersion. Thus, we compare the Chisquare test of Pearson to tests constructed from statistics which are obtained by the BoxCox transformation of Fisher dispersion index and of its inverse
Dhar, Gaurav. "Contributions en théorie du contrôle échantillonné optimal avec contraintes d’état et données non lisses." Thesis, Limoges, 2020. http://www.theses.fr/2020LIMO0050.
Full textThis dissertation is concerned with first-order necessary optimality conditions in the form of a Pontryagin maximum principle (in short, PMP) for optimal sampled-data control problems with free sampling times, running inequality state constraints and nonsmooth Mayer cost functions.Chapter 1 is devoted to notations and basic framework needed to describe the optimal sampled-data control problems to be encountered in the manuscript. In Chapter 2, considering that the sampling times can be freely chosen, we obtain an additional necessary optimality condition in the PMP called the Hamiltonian continuity condition. Recall that the Hamiltonian function, which describes the evolution of the Hamiltonian taking values of the optimal trajectory and of theoptimal sampled-data control, is in general discontinuous when the sampling times are fixed. Our result proves that the continuity of the Hamiltonian function is recovered in the case of optimal sampled-data controls with optimal sampling times. Finally we implement a shooting method based on the Hamiltonian continuity condition in order to numerically determine the optimal sampling times in two linear-quadratic examples.In Chapter 3, we obtain a PMP for optimal sampled-data control problems with running inequality state constraints. In particular we obtain that the adjoint vectors are solutions to Cauchy-Stieltjes problems defined by Borel measures associated to functions of bounded variation. Moreover, we find that, under certain general hypotheses, any admissible trajectory (associated to a sampled-data control) necessarily bounces on the runningine quality state constraints. Taking advantage of this bouncing trajectory phenomen on, we are able to use thePMP to implement an indirect numerical method which we use to numerically solve some simple examples of optimal sampled-data control problems with running inequality state constraints. In Chapter 4, we obtain a PMP for optimal sampled-data control problems with nonsmooth Mayer cost functions. Our proof directly follows from the tools of nonsmooth analysis and does not involve any regularization technique. We determine the existence of a selection in the subdifferential of the nonsmooth Mayer cost function by establishing a more general result asserting the existence a universal separating vector for a given compact convex set. From the application of this result, which is called universal separating vector theorem, we obtain a PMP for optimal sampled-data control problems with nonsmooth Mayer cost functions where the transversality conditon on the adjoint vector is given by an inclusion in the subdifferential of the nonsmooth Mayer cost function.To obtain the optimality conditions in the form of a PMP, we use different techniques of perturbations of theoptimal control. In order to handle the state constraints, we penalize the distance to them in a corresponding cost functional and then apply the Ekeland variational principle. In particular, we invoke some results on renorming Banach spaces in order to ensure the regularity of distance functions in the infinite-dimensional context. Finally we use standard notions in nonsmooth analysis such as the Clarke generalized directional derivative and theClarke subdifferential to study optimal sampled-data control problems with nonsmooth Mayer cost functions
Sahuguède, Alexandre. "Un algorithme de découverte de chroniques pertinentes pour le diagnostic par identification et reconstitution." Thesis, Toulouse 3, 2020. http://www.theses.fr/2020TOU30166.
Full textChronicles are temporal patterns well-suited for an abstract representation of complex dynamic systems. Chronicle recognition algorithms allow the identification of chronicles in an on-line stream of data to be done and take adequate action in an quick and efficient manner. Chronicles are used in a vast array of applications such as medical field, internet networks, or industrial applications. Nevertheless, designing chronicles is not an easy task due to the sophistication and the increase of data generation capacity of modern systems. The chronicle discovery process try and tackle this problem by an automatic design of chronicles from data directly generated by the studied system. In this thesis, an innovative approach to the problem of chronicle discovery is introduced. This new approach lies of the identification of elementary chronicles and a reconstitution of complex chronicles from them. The algorithm introduced, called CDIRe (Chronicle Discovery by Identification and Reconstitution), allows the discovery of chronicles with few knowledge from the underlying system to be done
Dewalle, Anne-Sophie. "Nouvelles approches de traitement du signal et de l'image en IRM fonctionnelle d'activation cérébrale." Lille 1, 2006. https://ori-nuxeo.univ-lille1.fr/nuxeo/site/esupversions/c3ff9e5f-e824-4cc2-a6ea-a22cb4383bc2.
Full textTreangle, Clement. "Observateurs grand gain pour des systèmes non linéaires à sorties échantillonnées et retardées." Thesis, Normandie, 2018. http://www.theses.fr/2018NORMC251/document.
Full textThis manuscript deals with the synthesis of high gain observers for nonlinear systems with sampled and delayed outputs. Three contributions are proposed for consideration in this manuscript. The first contribution, for a class of Multi-input / Multi-output systems whose outputs are grouped into a single block, involves the problem of the acquisition process of output measurements (continuous, sampled, delayed or not) and proposes a common framework for all possible cases. The second contribution proposes a filtered high gain observer on this same class of systems in order to reduce the sensitivity to measurement noise, in the case where the output is continuous and then in the case where the latter is sampled. The last contribution aims to extend the standard high gain synthesis for a large class of uniformly observable Multi-input / Multi-output systems with continuous output measurements. For each of these contributions, it has been shown that the observation error of each of the proposed observers converges exponentially towards zero in the absence of uncertainties in the system. All these contributions have been illustrated through several examples from different fields of study
Yacoub, Aznam. "Une approche de vérification formelle et de simulation pour les systèmes à événements : application à PROMELA." Thesis, Aix-Marseille, 2016. http://www.theses.fr/2016AIXM4373/document.
Full textNowadays, making reliable software and systems is become harder. New technologies imply more and more interactions between complex components, whose the analysis and the understanding are become arduous.To overcome this problem, the domains of verification and validation have known a significant progress, with the emergence of new automatic methods that ensure reliability of systems. Among all these techniques, we can find two great families of tools : the formal methods and the simulation. For a long time, these two families have been considered as opposite to each other. However, recent work tries to reduce the border between them. In this context, this thesis proposes a new approach in order to integrate discrete-event simulation in formal methods. The main objective is to improve existing model-checking tools by combining them with simulation, in order to allow them detecting errors that they were not previously able to find, and especially on timed systems. This approach led us to develop a new formal language, called DEv-PROMELA. This new language, which relies on the PROMELA and on the DEVS formalism, is like both a verifiable specifications language and a simulation formalism. By combining a traditional model-checking and a discrete-event simulation on models expressed in DEv-PROMELA, it is therefore possible to detect and to understand dysfunctions which could not be found by using only a formal checking or only a simulation. This result is illustrated through the different examples which are treated in this work
Jawad, Mohamed. "Confidentialité de données dans les systèmes P2P." Phd thesis, Université de Nantes, 2011. http://tel.archives-ouvertes.fr/tel-00638721.
Full textJanyene, Abderrahmane. "Validation de données des systèmes dynamiques linéaires." Nancy 1, 1987. http://www.theses.fr/1987NAN10190.
Full textAbdali, Abdelkebir. "Systèmes experts et analyse de données industrielles." Lyon, INSA, 1992. http://www.theses.fr/1992ISAL0032.
Full textTo analyses industrial process behavio, many kinds of information are needed. As tye ar mostly numerical, statistical and data analysis methods are well-suited to this activity. Their results must be interpreted with other knowledge about analysis prcess. Our work falls within the framework of the application of the techniques of the Artificial Intelligence to the Statistics. Its aim is to study the feasibility and the development of statistical expert systems in an industrial process field. The prototype ALADIN is a knowledge-base system designed to be an intelligent assistant to help a non-specialist user analyze data collected on industrial processes, written in Turbo-Prolong, it is coupled with the statistical package MODULAD. The architecture of this system is flexible and combing knowledge with general plants, the studied process and statistical methods. Its validation is performed on continuous manufacturing processes (cement and cast iron processes). At present time, we have limited to principal Components analysis problems
Tos, Uras. "Réplication de données dans les systèmes de gestion de données à grande échelle." Thesis, Toulouse 3, 2017. http://www.theses.fr/2017TOU30066/document.
Full textIn recent years, growing popularity of large-scale applications, e.g. scientific experiments, Internet of things and social networking, led to generation of large volumes of data. The management of this data presents a significant challenge as the data is heterogeneous and distributed on a large scale. In traditional systems including distributed and parallel systems, peer-to-peer systems and grid systems, meeting objectives such as achieving acceptable performance while ensuring good availability of data are major challenges for service providers, especially when the data is distributed around the world. In this context, data replication, as a well-known technique, allows: (i) increased data availability, (ii) reduced data access costs, and (iii) improved fault-tolerance. However, replicating data on all nodes is an unrealistic solution as it generates significant bandwidth consumption in addition to exhausting limited storage space. Defining good replication strategies is a solution to these problems. The data replication strategies that have been proposed for the traditional systems mentioned above are intended to improve performance for the user. They are difficult to adapt to cloud systems. Indeed, cloud providers aim to generate a profit in addition to meeting tenant requirements. Meeting the performance expectations of the tenants without sacrificing the provider's profit, as well as managing resource elasticities with a pay-as-you-go pricing model, are the fundamentals of cloud systems. In this thesis, we propose a data replication strategy that satisfies the requirements of the tenant, such as performance, while guaranteeing the economic profit of the provider. Based on a cost model, we estimate the response time required to execute a distributed database query. Data replication is only considered if, for any query, the estimated response time exceeds a threshold previously set in the contract between the provider and the tenant. Then, the planned replication must also be economically beneficial to the provider. In this context, we propose an economic model that takes into account both the expenditures and the revenues of the provider during the execution of any particular database query. Once the data replication is decided to go through, a heuristic placement approach is used to find the placement for new replicas in order to reduce the access time. In addition, a dynamic adjustment of the number of replicas is adopted to allow elastic management of resources. Proposed strategy is validated in an experimental evaluation carried out in a simulation environment. Compared with another data replication strategy proposed in the cloud systems, the analysis of the obtained results shows that the two compared strategies respond to the performance objective for the tenant. Nevertheless, a replica of data is created, with our strategy, only if this replication is profitable for the provider
Voisard, Agnès. "Bases de données géographiques : du modèle de données à l'interface utilisateur." Paris 11, 1992. http://www.theses.fr/1992PA112354.
Full textJaff, Luaï. "Structures de Données dynamiques pour les Systèmes Complèxes." Phd thesis, Université du Havre, 2007. http://tel.archives-ouvertes.fr/tel-00167104.
Full textla porte vers des applications en économie via les systèmes complexes.
Les structures de données que nous avons étudiées sont les permutations qui ne contiennent pas de sous-suite croissante de longueur plus que deux, les tableaux de Young standards rectangles à deux lignes, les mots de Dyck et les codes qui lient ces structures de données.
Nous avons proposé un modèle économique qui modélise le bénéfice d'un compte bancaire dont l'énumération des configurations possible se fait à l'aide d'un code adapté. Une seconde application
concerne l'évolution de populations d'automate génétique . Ces populations sont étudiées par analyse spectrale et des expérimentations sont données sur des automates probabilistes dont l'évolution conduit à contrôler la dissipation par auto-régulation.
L'ensemble de ce travail a pour ambition de donner quelques outils calculatoires liés à la dynamique de structures de données pour analyser la complexité des systèmes.
Delot, Thierry. "Accès aux données dans les systèmes d'information pervasifs." Habilitation à diriger des recherches, Université de Valenciennes et du Hainaut-Cambresis, 2009. http://tel.archives-ouvertes.fr/tel-00443664.
Full textHeraud, Nicolas. "Validation de données et observabilité des systèmes multilinéairesé." Vandoeuvre-les-Nancy, INPL, 1991. http://www.theses.fr/1991INPL082N.
Full textThe aim of this study is to investigate data validation and observability of miltilinear systems to diagnose instrumentation in a process. Data validation and observability in linear systems are first reviewed and these notions are extended to multilinear systems. Differents methods such as hierarchical computation, constraint linearization and penalization functions, are presented to estimate true values when some values are lacking. After comparing the different methods, a recurrent calculus of estimates using constraint linearization and penalization functions is developed. An observable system is required in order to perform data validation. Thus, we developed an original method, based on arborescent diagrams. The technique of data validation has been successfully applied to a complex uranium processing plant owned by the French company Total Compagnie Minière France. On this partially instrumented process, measurements for volumic flow, density and uranium in both solid and liquid phase are available. The analysis allows first to obtain coherent date. Furthemore, it can be used to detect sensors faults
Meyer, Michel. "Validation de données sur des systèmes incomplètement observés." Toulouse, INPT, 1990. http://www.theses.fr/1990INPT032G.
Full textLiroz, Miguel. "Partitionnement dans les systèmes de gestion de données parallèles." Phd thesis, Université Montpellier II - Sciences et Techniques du Languedoc, 2013. http://tel.archives-ouvertes.fr/tel-01023039.
Full textPetit, Loïc. "Gestion de flux de données pour l'observation de systèmes." Phd thesis, Université de Grenoble, 2012. http://tel.archives-ouvertes.fr/tel-00849106.
Full textLiroz-Gistau, Miguel. "Partitionnement dans les Systèmes de Gestion de Données Parallèles." Phd thesis, Université Montpellier II - Sciences et Techniques du Languedoc, 2013. http://tel.archives-ouvertes.fr/tel-00920615.
Full textMichel, François. "Validation de systèmes répartis : symétries d'architecture et de données." Toulouse, INPT, 1996. http://www.theses.fr/1996INPT099H.
Full textRobin, Jean-Marc. "L'estimation des systèmes de demandes sur données individuelles d'enquêtes." Paris 1, 1988. http://www.theses.fr/1988PA010043.
Full textThe fact that all households do not purchase all commodities during short periods of time is a source of trouble when estimating demand systems from household survey data. To avoid possible selection biases when selecting only households whe did purchase during the recording period, we have to model explicitely purchasing behaviors. The various models proposed are analysed then tested on the data of the french "ecquete consommation alimentaire en 1981
Villamil, Giraldo María del Pilar. "Service de localisation de données pour les systèmes P2P." Grenoble INPG, 2006. http://www.theses.fr/2006INPG0052.
Full textThis thesis is oriented to the interrogation in massively distributed systems. It proposes a data location service for peer to peer systems based on distributed hash tables. These systems are characterized by a high degree of distribution, a large set of heterogeneous peers, a very dynamic configuration and a "blind" distribution of the data. These characteristics make it difficult to provide an efficient data management. Ln fact, it is almost impossible to have a coherent view regarding the global state of the system. Moreover, the location, using declarative queries, of data shared in the system is becoming very problematic. The objective of the location service proposed is to provide a query management adapted to the peer to peer context. As a result, the service uses distributed indexation techniques, query evaluation models, cache and materialized queries. The query evaluation models permitted theoretical performance analysis wich using prototype experiments in a large scale system (1300 peers were deployed). The behaviour observed shows good properties particularly regarding the solution scalability according to the number of participant sites. This is one of the critical issues for providing successful massively distributed systems
Liroz, Gistau Miguel. "Partitionnement dans les systèmes de gestion de données parallèles." Thesis, Montpellier 2, 2013. http://www.theses.fr/2013MON20117/document.
Full textDuring the last years, the volume of data that is captured and generated has exploded. Advances in computer technologies, which provide cheap storage and increased computing capabilities, have allowed organizations to perform complex analysis on this data and to extract valuable knowledge from it. This trend has been very important not only for industry, but has also had a significant impact on science, where enhanced instruments and more complex simulations call for an efficient management of huge quantities of data.Parallel computing is a fundamental technique in the management of large quantities of data as it leverages on the concurrent utilization of multiple computing resources. To take advantage of parallel computing, we need efficient data partitioning techniques which are in charge of dividing the whole data and assigning the partitions to the processing nodes. Data partitioning is a complex problem, as it has to consider different and often contradicting issues, such as data locality, load balancing and maximizing parallelism.In this thesis, we study the problem of data partitioning, particularly in scientific parallel databases that are continuously growing and in the MapReduce framework.In the case of scientific databases, we consider data partitioning in very large databases in which new data is appended continuously to the database, e.g. astronomical applications. Existing approaches are limited since the complexity of the workload and continuous appends restrict the applicability of traditional approaches. We propose two partitioning algorithms that dynamically partition new data elements by a technique based on data affinity. Our algorithms enable us to obtain very good data partitions in a low execution time compared to traditional approaches.We also study how to improve the performance of MapReduce framework using data partitioning techniques. In particular, we are interested in efficient data partitioning of the input datasets to reduce the amount of data that has to be transferred in the shuffle phase. We design and implement a strategy which, by capturing the relationships between input tuples and intermediate keys, obtains an efficient partitioning that can be used to reduce significantly the MapReduce's communication overhead
Madera, Cedrine. "L’évolution des systèmes et architectures d’information sous l’influence des données massives : les lacs de données." Thesis, Montpellier, 2018. http://www.theses.fr/2018MONTS071/document.
Full textData is on the heart of the digital transformation.The consequence is anacceleration of the information system evolution , which must adapt. The Big data phenomenonplays the role of catalyst of this evolution.Under its influence appears a new component of the information system: the data lake.Far from replacing the decision support systems that make up the information system, data lakes comecomplete information systems’s architecture.First, we focus on the factors that influence the evolution of information systemssuch as new software and middleware, new infrastructure technologies, but also the decision support system usage itself.Under the big data influence we study the impact that this entails especially with the appearance ofnew technologies such as Apache Hadoop as well as the current limits of the decision support system .The limits encountered by the current decision support system force a change to the information system which mustadapt and that gives birth to a new component: the data lake.In a second time we study in detail this new component, formalize our definition, giveour point of view on its positioning in the information system as well as with regard to the decision support system .In addition, we highlight a factor influencing the architecture of data lakes: data gravity, doing an analogy with the law of gravity and focusing on the factors that mayinfluence the data-processing relationship. We highlight, through a use case, that takingaccount of the data gravity can influence the design of a data lake.We complete this work by adapting the software product line approach to boot a methodof formalizations and modeling of data lakes. This method allows us:- to establish a minimum list of components to be put in place to operate a data lake without transforming it into a data swamp,- to evaluate the maturity of an existing data lake,- to quickly diagnose the missing components of an existing data lake that would have become a dataswamp- to conceptualize the creation of data lakes by being "software agnostic “
Barbier, Sébastien. "Visualisation distance temps-réel de grands volumes de données." Grenoble 1, 2009. http://www.theses.fr/2009GRE10155.
Full textNumerical simulations produce huger and huger meshes that can reach dozens of million tetrahedra. These datasets must be visually analyzed to understand the physical simulated phenomenon and draw conclusions. The computational power for scientific visualization of such datasets is often smaller than for numerical simulation. As a consequence, interactive exploration of massive meshes is barely achieved. In this document, we propose a new interactive method to interactively explore massive tetrahedral meshes with over forty million tetrahedra. This method is fully integrated into the simulation process, based on two meshes at different resolutions , one fine mesh and one coarse mesh , of the same simulation. A partition of the fine vertices is computed guided by the coarse mesh. It allows the on-the-fly extraction of a mesh, called \textit{biresolution}, mixed of the two initial resolutions as in usual multiresolution approaches. The extraction of such meshes is carried out into the main memory (CPU), the last generation of graphics cards (GPU) and with an out-of-core algorithm. They guarantee extraction rates never reached in previous work. To visualize the biresolution meshes, a new direct volume rendering (DVR) algorithm is fully implemented into graphics cards. Approximations can be performed and are evaluated in order to guarantee an interactive rendering of any biresolution meshes
Peerbocus, Mohamed Ally. "Gestion de l'évolution spatiotemporelle dans une base de données géographiques." Paris 9, 2001. https://portail.bu.dauphine.fr/fileviewer/index.php?doc=2001PA090055.
Full textAllouti, Faryel. "Visualisation dans les systèmes informatiques coopératifs." Paris 5, 2011. http://www.theses.fr/2011PA05S003.
Full textClustering techniques and visualization tools of complex data are two recurring themes in the community of Mining and Knowledge Management. At the intersection of these two themes there are the visualization methods such as multidimensional scaling or the Self-Organizing Maps (SOM). The SOM is constructed using K-means algorithm to which is added the notion of neighborhood allowing in this way the preservation of the topo-logy of the data. Thus, the learning moves closer, in the space of data, the centers that are neighbors on a two dimensions grid generally, to form a discrete surface which is a representation of the distribution of the cloud to explore. In this thesis, we are interested in the visualization in a cooperative context, where co-operation is established via an asynchronous communication and the media is the e-mail. This tool has emerged with the advent of information technology and communication. It is widely used in organizations, it allows an immediate and fast distribution of the in-formation to several persons at the same time, without worrying about their presence. Our objective consisted in proposing a tool of visual exploration of textual data which are files attached to the electronic messages. In order to do this, we combined clustering and visualization methods. We investigated the mixture approach, which is a very useful contribution for classification. In our context, we used the multinomial mixture model (Go-vaert and Nadif, 2007) to determine the classes of files. In addition, we studied the aspect of visualization of the obtained classes and documents using the multidimensional scaling and DC (Difference of Convex functions) and Self-Organizing Maps of Kohonen
Lassoued, Yassine. "Médiation de qualité dans les systèmes d'information géographique." Aix-Marseille 1, 2005. http://www.theses.fr/2005AIX11027.
Full textFilali, Rania Linda. "Sur la synchronisation et le cryptage de systèmes chaotiques à temps discret utilisant les techniques d'agrégation et la représentation en flèche des matrices." Phd thesis, Ecole Centrale de Lille, 2013. http://tel.archives-ouvertes.fr/tel-00858272.
Full textFlores, Ignacio. "On the empirical measurement of inequality." Thesis, Paris 1, 2019. http://www.theses.fr/2019PA01E003/document.
Full textThe 1st chapter presents historical series of Chilean top income shares over a period of half a century, mostly using data from tax statistics and national accounts. The study contradicts evidence based on survey data, according to which inequality has fallen constantly over the past 25 years. Rather, it changes direction, increasing from around the year 2000. Chile ranks as one of the most unequal countries among both OECD and Latin American countries over the whole period of study. The 2nd chapter measures the underestimation of factor income in distributive data. I find that households receive only half of national gross capital income,as opposed to corporations. Due to heterogeneous non-response and misreporting, Surveys only capture 20% of it, vs. 70% of labor income. This understates inequality estimates, which become insensitive to the capital share and its distribution. I formalize this system based on accounting identities. I then compute marginal effects and contributions to changes in fractile shares. The 3rd chapter, presents a method to adjust surveys. These generally fail to capturethe top of the income distribution. It has several advantages over previous ones: it is consistent with standard survey calibration methods; it has explicit probabilistic foundations and preserves the continuity of density functions; it provides an option to overcome the limitations of bounded survey-supports; and it preserves the microdata structure of the survey
Garnerin, Mahault. "Des données aux systèmes : étude des liens entre données d’apprentissage et biais de performance genrés dans les systèmes de reconnaissance automatique de la parole." Thesis, Université Grenoble Alpes, 2022. http://www.theses.fr/2022GRALL006.
Full textMachine learning systems contribute to the reproduction of social inequalities, because of the data they use and for lack of critical approches, thys feeding a discourse on the ``biases of artificial intelligence''. This thesis aims at contributing to collective thinking on the biases of automatic systems by investigating the existence of gender biases in automatic speech recognition (ASR) systems.Critically thinking about the impact of systems requires taking into account both the notion of bias (linked with the architecture, or the system and its data) and that of discrimination, defined at the level of each country's legislation. A system is considered discriminatory when it makes a difference in treatment on the basis of criteria defined as breaking the social contract. In France, sex and gender identity are among the 23 criteria protected by law.Based on theoretical considerations on the notions of bias, and in particular on the predictive (or performance) bias and the selection bias, we propose a set of experiments to try to understand the links between selection bias in training data and predictive bias of the system. We base our work on the study of an HMM-DNN system trained on French media corpus, and an end-to-end system trained on audio books in English. We observe that a significant gender selection bias in the training data contributes only partially to the predictive bias of the ASR system, but that the latter emerges nevertheless when the speech data contain different utterance situations and speaker roles. This work has also led us to question the representation of women in speech data, and more generally to rethink the links between theoretical conceptions of gender and ASR systems
Hajji, Hicham. "Gestion des risques naturels : une approche fondée sur l'intégration des données." Lyon, INSA, 2005. http://theses.insa-lyon.fr/publication/2005ISAL0039/these.pdf.
Full textThere is a huge geographic data available with many organizations collecting geographic data for centuries, but some of that is still in the form of paper maps or in traditional files or databases, and with the emergence of latest technologies in the field of software and data storage some has been digitized and is stored in latest GIS systems. However, too often their reuse for new applications is a nightmare, due to diversity of data sets, heterogeneity of existing systems in terms of data modeling concepts, data encoding techniques, obscure semantics of data,storage structures, access functionality, etc. Such difficulties are more common in natural hazards information systems. In order to support advanced natural hazards management based on heterogeneous data, this thesis develops a new approach to the integration of semantically heterogeneous geographic information which is capable of addressing the spatial and thematic aspects of geographic information. The approach is based on OpenGIS standard. It uses it as a common model for data integration. The proposed methodology takes into consideration a large number of the aspects involved in the construction and the modelling of natural hazards management information system. Another issue has been addressed in this thesis, which is the design of an ontology for natural hazards. The ontology design has been extensively studied in recent years, we have tried throughout this work to propose an ontology to deal with semantic heterogeneity existing between different actors and to model existing knowledge present for this issue. The ontology contains the main concepts and relationships between these concepts using OWL Language
Kaplan, Stéphane. "Spécification algébrique de types de données à accès concurrent." Paris 11, 1987. http://www.theses.fr/1987PA112335.
Full textSaidi, Selma. "Optimisation des transferts de données sur systèmes multiprocesseurs sur puce." Phd thesis, Université de Grenoble, 2012. http://tel.archives-ouvertes.fr/tel-00875582.
Full textSiriopoulos, Costas-Panou. "Essai sur les systèmes experts et l'analyse confirmatoire des données." Aix-Marseille 3, 1988. http://www.theses.fr/1988AIX32010.
Full textA) artificial intelligence in statistical analysis: the key use of a. I. In statistics has been to enable statisticians to study strategies of data analysis. The existence, today, of a large number of statistical packages poses problems for potential users. Incorporating expert guidance in statistical software is technically challenging but worthwhile undertaking. Different systems and projects have been suggested. Recently, m. Egea and j. P. Marciano have proposed the a. I. D. A. Project in c. D. A. We consider the autocorrelation problem in a multiple linear model and we propose a typology to study the required statistical knowledge. We also propose a corpus of 30 rules of thumb and 5 meta-rules. We conclude in a possible strategy, for the detection and correction of the problem, in the form of an hierarchical tree in which each node is represented by a frame. B) statistics in artificial intelligence: expert systems work with propositions that may be uncertain. Accepting a probabilistic nature of uncertainty, we have to make a crucial assumption, namely, the additivity axiome. Once this assumption is dropped, we have other ways of characterizing uncertainty, in particular, possibilistic uncertainty based on fuzzy sets
Pradel, Bruno. "Evaluation des systèmes de recommandation à partir d'historiques de données." Paris 6, 2013. http://www.theses.fr/2013PA066263.
Full textThis thesis presents various experimental protocols leading to abetter offline estimation of errors in recommender systems. As a first contribution, results form a case study of a recommendersystem based on purchased data will be presented. Recommending itemsis a complex task that has been mainly studied considering solelyratings data. In this study, we put the stress on predicting thepurchase a customer will make rather than the rating he will assign toan item. While ratings data are not available for many industries andpurchases data widely used, very few studies considered purchasesdata. In that setting, we compare the performances of variouscollaborative filtering models from the litterature. We notably showthat some changes the training and testing phases, and theintroduction of contextual information lead to major changes of therelative perfomances of algorithms. The following contributions will focus on the study of ratings data. Asecond contribution will present our participation to the Challenge onContext-Aware Movie Recommendation. This challenge provides two majorchanges in the standard ratings prediction protocol: models areevaluated conisdering ratings metrics and tested on two specificsperiod of the year: Christmas and Oscars. We provides personnalizedrecommendation modeling the short-term evolution of the popularitiesof movies. Finally, we study the impact of the observation process of ratings onranking evaluation metrics. Users choose the items they want to rateand, as a result, ratings on items are not observed at random. First,some items receive a lot more ratings than others and secondly, highratings are more likely to be oberved than poor ones because usersmainly rate the items they likes. We propose a formal analysis ofthese effects on evaluation metrics and experiments on the Yahoo!Musicdataset, gathering standard and randomly collected ratings. We showthat considering missing ratings as negative during training phaseleads to good performances on the TopK task, but these performancescan be misleading favoring methods modeling the popularities of itemsmore than the real tastes of users
Akbarinia, Reza. "Techniques d'accès aux données dans des systèmes pair-à-pair." Nantes, 2007. http://www.theses.fr/2007NANT2060.
Full textThe goal of this thesis is to contribute to the development of new data access techniques for query processing services in P2P environments. We focus on novel techniques for two important kinds of queries: queries with currency guarantees and top-k queries. To improve data availability, most P2P systems rely on data replication, but without currency guarantees. However, for many applications which could take advantage of a P2P system (e. G. Agenda management), the ability to get the current data is very important. To support these applications, the query processing service must be able to efficiently detect and retrieve a current, i. E. Up-to-date, replica in response to a user requesting a data. The second problem which we address is supporting top-k queries which are very useful in large scale P2P systems, e. G. They can reduce the network traffic significantly. However, efficient execution of these queries is very difficult in P2P systems because of their special characteristics, in particular in DHTs. In this thesis, we first survey the techniques which have been proposed for query processing in P2P systems. We give an overview of the existing P2P networks, and compare their properties from the perspective of query processing. Second, we propose a complete solution to the problem of current data retrieval in DHTs. We propose a service called Update Management Service (UMS) which deals with updating replicated data and efficient retrieval of current replicas based on timestamping. Third, we propose novel solutions for top-k query processing in structured, i. E. DHTs, and unstructured P2P systems. We also propose new algorithms for top-k query processing over sorted lists which is a general model for top-k queries in many centralized, distributed and P2P systems, especially in super-peer networks. We validated our solutions through a combination of implementation and simulation and the results show very good performance, in terms of communication and response time
Alustwani, Husam. "Interactivité et disponibilité des données dans les systèmes multimédias distribués." Besançon, 2009. http://www.theses.fr/2009BESA2041.
Full textThe works in this thesis have been guided by two problems: (a) how to efficiently support fast browsing interactions in streamed multimedia presentations and (b) how to enhance data availability in pure P2P streaming systems ? In order to enable quick browsing within streamed multimedia presentations, we proposed an approach that takes full advantage of object multiplicity in a multimedia presentation. Our approach allows, among other features, to preserve the semantic on the presentation, when a fast browsing interaction occurs. In a second time, we studied the performances of our approach through the proposal of a Content-Based Prefetching Strategy, called CPS. Our strategy enables to considerably reduce the new interaction’s latency, that is to reduce the response time of a fast browsing action. Data availability in P2P streaming systems differs fundamentally from that observed in classical systems, in the sense that the use data are time-dependent. Thus, this problem arises in terms of the opportunity for a peer (consumer) to entirely receive a video content, that is able to watch the content to its end. However, spontaneous PSP systems are characterised, mainly, by the volatility of the peers. The unpredictable departure of peers poses the problem of the availability of peers that are sources for streaming. We have studied this problem by setting-up, a centralised caching mechanism to reduce the effects of peer’s departure and by only replicating the suffixes (last parts) of the videos that are being accessed. In a second step, we extended our approach towards a distributed virtual cache. The simulation results showed the relevance of the proposed approaches. Finally, we described the design and implementation of a prototype, that demonstrates the feasibility of a spontaneous P2P streaming system
Saint-Joan, Denis. "Données géographiques et raisonnement : le système GEODES." Toulouse 3, 1995. http://www.theses.fr/1995TOU30179.
Full textPetitdemange, Eva. "SAMUFLUX : une démarche outillée de diagnostic et d'amélioration à base de doubles numériques : application aux centres d'appels d'urgence de trois SAMU." Thesis, Ecole nationale des Mines d'Albi-Carmaux, 2020. http://www.theses.fr/2020EMAC0012.
Full textThe demand for emergency medical services has been significant and increasing over the last decade. In a constrained medico-economic context, the maintenance of operational capacities represents a strategic strake in front of the risk of congestion and insufficient accessibility for the population. Recent events such as the COVID-19 pandemic show the limits of the current system to face crisis situations. Reinforcement in human resources cannot be the only solution in front of this observation and it becomes unavoidable to build new organizational models while aiming at a quality of service allowing to answer 99% of the incoming calls in less than 60 seconds (90% in 15s and 99% in 30s MARCUS report and HAS recommendation October 2020). However, these models must take into account the great heterogeneity of EMS and their operation. In the light of these findings, the research work presented in this manuscript aims to evaluate the organizational effiectiveness and resilience of EMS in managing the flow of emergency telephone calls to deal with daily life and crisis situations. This evaluation allows us to propose and test new organizational schemes in order to make recommendations adpated to the particularities of emergency call centers. In a first part, we propose a methodology equipped for the diagnosis and improvement of emergency call centers. It can be broken down into two main parts: the study of data from emergency call centers, and then the design and use of a dual digital system. For each step of this methodology, we propose an associated tool. In a second part, we apply the first part of the methodology to our partner EMS data. The aim is to be able to extract information and knowledge from the telephony data as well as from the business processes for handling emergency calls. The knowledge thus extracted makes it possible to design a digital duplicate that is close to the real behavior of the EMS. Finally, in a third part, we use the material produced previously to model and parameterize a digital duplicate deployed on a discrete event simulation engine. It allows us to test several scenarios by playing on different call management organizations. Thanks to this, we make recommendations on the types of organizations to adopt in order to improve the performance of call centers
Bazin, Cyril. "Tatouage de données géographiques et généralisation aux données devant préserver des contraintes." Caen, 2010. http://www.theses.fr/2010CAEN2006.
Full textDigital watermaking is a fundamental process for intellectual property protection. It consists in inserting a mark into a digital document by slightly modifications. The presence of this mark allows the owner of a document to prove the priority of his rights. The originality of our work is twofold. In one hand, we use a local approach to ensure a priori that the quality of constrained documents is preserved during the watermark insertion. On the other hand, we propose a generic watermarking scheme. The manuscript is divided in three parts. Firstly, we introduce the basic concepts of digital watermarking for constrainted data and the state of the art of geographical data watermarking. Secondly, we present our watermarking scheme for digital vectorial maps often used in geographic information systems. This scheme preserves some topological and metric qualities of the document. The watermark is robust, it is resilient against geometric transformations and cropping. We give an efficient implementation that is validated by many experiments. Finally, we propose a generalization of the scheme for constrainted data. This generic scheme will facilitate the design of watermarking schemes for new data type. We give a particular example of application of a generic schema for relational databases. In order to prove that it is possible to work directly on the generic scheme, we propose two detection protocols straightly applicable on any implementation of generic scheme
Fénié, Patrick. "Graico : méthode de modélisation et de conception de systèmes d'exploitation de systèmes de production." Bordeaux 1, 1994. http://www.theses.fr/1994BOR10622.
Full textLanglet, Thomas. "Étude de convergences de séries aléatoires échantillonnées, mesures de Markov quasi-Bernoulli ou quasi-Bernoulli faible et temps de retour uniforme dans les systèmes exponentiellement mélangeants." Amiens, 2009. http://www.theses.fr/2009AMIE0122.
Full textIn the first part of this thesis, we study the convergence of two different random series. We give some sufficient conditions on the sequence (ak)k, on the random independent variables (Xk)k, such that for almost every ω, to have the uniform convergence or almost everywhere convergence of Σk ak f(x(X1+. . . +Xk)(ω)) for a certain class of functions f. We also find some sufficient conditions on (ak)k and on the random independent variables (Xk)k to have a certain class of function g ϵ L²(μ) and a certain class of operator {Tt,t ϵ G} such that the series Σk ak T^{(X1+. . . +Xk)(ω)}(g)(x) converge in L²(μ) or for μ-almost every x. In the second part, we prove some necessary and sufficient condition for an inhomogeneous Markov measure to be a quasi-Bernoulli measure or a weak quasi-Bernoulli measure. We apply this condition on the Bernoulli measure. We proved that a Bernoulli measure which only has some non-negative probability is a weak quasi-Bernoulli measure, then it is a quasi-Bernoulli measure. The last part is devoted to the study of a uniform hitting problem in an exponentially mixing dynamical system. This is a random covering problem driven by an exponentially mixing stationary process. For such a covering, among others, we obtain a satisfactory estimation on the covering numbers
Tahir, Hassane. "Aide à la contextualisation de l’administration de base de données." Paris 6, 2013. http://www.theses.fr/2013PA066789.
Full textThe complexity of database administration tasks requires the development of tools for supporting database experts. When problems occur, the database administrator (DBA) is frequently the first person blamed. Most DBAs work in a fire-fighting mode and have little opportunity to be proactive. They must be constantly ready to analyze and correct failures based on a large set of procedures. In addition, they are continually readjusting these procedures and developing practices to manage a multitude of specific situations that differ from the generic situation by some few contextual elements. These practices have to deal with these contextual elements in order to solve the problem at hand. This thesis aims to use Contextual Graphs formalism in order to improve existing procedures used in database administration. The thesis shows also the benefits of using Contextual Graphs to capture user practices in order to be reused in the working contexts. Up to now, this improvement is achieved by a DBA through practices that adapt procedures to the context in which tasks should be performed and the incidents appear. This work will be the basis for designing and implementing a Context-Based Intelligent Assistant System (CBIAS) for supporting DBAs
Heba, Nurja Ines. "Contributions à l'analyse statistique et économétrique des données géoréférencées." Toulouse 1, 2005. http://www.theses.fr/2005TOU10048.
Full textSpatial analysis is a research topic that might develop the exploration capacity of goegraphical information systems. We study different aspects of georeferenced data modelling. 1) We build a tool-box called GEOXP (using Matlab) organised with statistical functions offering an exploratory analysis of georeferenced data with spatial dimension. These functions use statistical tools adapted for spatial data. 2) We study the theoretical context of a spatial analysis of real estate data in urban environment to prepare an empirical work. This study bring us to a new research on weight matrix choice in spatial regression models and we offer a new type of weight matrix, built on location density. 3) We build a new methodology to classify data flows between geographical units by using two matrix, one describing the flows themselves and the othe one describing their neighbourhood relations
Coupaye, Thierry. "Un modèle d'exécution paramétrique pour systèmes de bases de données actifs." Phd thesis, Université Joseph Fourier (Grenoble), 1996. http://tel.archives-ouvertes.fr/tel-00004983.
Full text