Dissertations / Theses on the topic 'Informatique appliquée à la biodiversité'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Informatique appliquée à la biodiversité.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Saliba, Elie Mario. "Understanding and managing Zoological Nomenclature in the era of Big Data and Open Science." Electronic Thesis or Diss., Sorbonne université, 2023. http://www.theses.fr/2023SORUS460.
Full textNomenclature is the discipline of taxonomy responsible for creating and managing the scientific names assigned to categories of living beings. It ensures continuity in the transmission of all kinds of accumulated data and knowledge about taxa. To this end, zoologists refer to the International Code of Zoological Nomenclature. The Code contains the rules for understanding and correctly applying this discipline.Nomenclature has become increasingly complex over the centuries, to keep pace with the evolution of scientific and technological knowledge. It currently plays, through scientific names, a crucial role in biodiversity databases. However, databases have their limitations in terms of structure when it comes to nomenclatural endeavors. The role of nomenclature in databases is explored in the thesis.While nomenclatural data is largely present in online databases, the discipline itself has remained relatively impervious to computerization. An in-depth analysis of the Code enabled the creation of a list of formal objects and their properties, which are needed to model the rules of this discipline. Moreover, the structure of the Code's rules is intended to be logical and unambiguous, which makes it ideal for translating into a series of algorithms. This hypothesis led to the creation of a web application called Lognom. Lognom is an algorithm-based software that supports decision-making in zoological nomenclature. The application does not rely on pre-existing databases, but provides an answer based on data entered by the user. The software aims to support taxonomists in their day-to-day nomenclature management, by determining whether a name or work is available and whether spelling rules have been correctly applied. It can also verify whether all requirements preceding the publication of a new name or work have been met. Additionally, Lognom allows the user to establish which name is the valid name among several candidates, and the list of synonyms that results from this decision. It also includes several tools for answering various nomenclatural questions, such as the determination of the grammatical gender of a genus name. However, it has not been possible to integrate all the rules of the International Code of Zoological Nomenclature into the application. Some rules are based on semantic interpretation, which is very complex to automate. Moreover, Lognom is highly sensitive to the quality of the data supplied by its users, even if it does provide a few controls. A proposed classification of the Code’s rules is included, to better identify the strengths and weaknesses of the Code in terms of its possible complete computerization. Recommendations for the optimization of its logical and unambiguous character are also mentioned. Similarly, various issues relating to nomenclature and its computer applications are explored, as well as a brief analysis of the social difficulties that might impede the improvement of these rules. There are still many possible future applications for the algorithms developed for Lognom. These include the possibility of working collaboratively on projects concerning the nomenclature of q given taxonomic group. This could lead to the creation of dynamic lists of names: Furthermore, the algorithms should be able to simulate the impact of changes in the rules of the Code on existing names and works. In the long term, a tool such as Lognom could enable the possibility of modeling nomenclature in its entirety, opening the door to more efficient and coordinated management of this centuries-old discipline
Blondeau, Sylvain. "Développement d'une plateforme informatique hétérogène appliquée à la pharmacognosie." Thesis, Orléans, 2011. http://www.theses.fr/2011ORLE2004.
Full textPharmacognosy, a multidisciplinary science which studies natural products, has greatly evolved in parallel with other disciplines particularly analytical chemistry and information technologies. New complementary approaches like reverse pharmacognosy, from molecules to source organisms (in contrary of pharmacognosy), have been developed. Data necessary for these studies are overwhelming and heterogeneous. Molecules, activities, targets, organisms, traditional uses, are some of essential information for pharmacognosy. The main goal of this thesis was to develop an information platform, including a large database, to collect the diverse data, along with a specific web interface and also mining and cheminformatics tools. A constant enrichment of this database has also been realized. This platform allows to query and browse data, but also to cross them in order to discover new knowledge. Truely optimised for pharmacognosy, it makes easier and enhances the study of natural products, due to a multi-level approach (ethnopharmacology, botany, chemistry). It offers new possibilities of interaction between data from different domains, such as organism hopping, a new concept of organism comparisons according to their chemical profiles
Saraydaryan, Jacques. "Détection d'anomalies comportementales appliquée à la vision globale." Lyon, INSA, 2008. http://theses.insa-lyon.fr/publication/2008ISAL0132/these.pdf.
Full textLn light of the increase in new threads and attacks, security components (Firewall, IDS) are becoming inadequate. Lndeed, complex attack scenarios tend to be confused with normal system behaviors in arder to by-pass local security components. From this perspective, we provided a new method of behavioral anomaly detection based on a global view of the system throughout our work. By taking into account the observation constraints of the entire IS (heterogeneity, high data volume), we built a statistical profile of the system and developed an anomaly detection method that showed that the continuous update of this profile allows us to follow the evolution of legitima te user behaviors and reduces false alarms. Thus, by focusing on the attacker's strategy, our works determined the observation perimeter of system behaviors to detect behavioral anomalies
Gamoudi, Oussama. "Optimisation adaptative appliquée au préchargement de données." Paris 6, 2012. http://www.theses.fr/2012PA066192.
Full textData prefetching is an effective way to bridge the increasing performance gap between processor and memory. Prefetching can improve performance but it has some side effects which may lead to no performance improvement while increasing memory pressure or to performance degradation. Adaptive prefetching aims at reducing negative effects of prefetching while keeping its advantages. This paper proposes an adaptive prefetching method based on runtime activity, which corresponds to the processor and memory activities retrieved by hardware counters, to predict the prefetch efficiency. Our approach highlights and relies on the correlation between the prefetch effects and runtime activity. Our method learns all along the execution this correlation to predict the prefetch efficiency in order to filter out predicted inefficient prefetches. Experimental results show that the proposed filter is able to cancel thenegative impact of prefetching when it is unprofitable while keeping the performance improvement due to prefetching when it is beneficial. Our filter works similarly well when several threads are running simultane-ously which shows that runtime activity enables an efficient adaptation of prefetch by providing information on running-applications behaviors and interactions
Yesse, Thierry. "Etude du parallélisme massif appliquée à la visualisation interactive de grandes bases de données scientifiques." Rouen, 1998. http://www.theses.fr/1998ROUES086.
Full textDeguet, Joris. "Intégration de l'émergence au sein des systèmes multi-agent : une étude appliquée à la recherche heuristique." Grenoble 1, 2008. http://www.theses.fr/2008GRE10051.
Full textEmergence and multi-agent systems are two domains with share similar problems. Through the study of emergence in the context of multi-agent systems, this work is centred on their principal common point; the search for a collective advantage gained through the interaction among the agents, when the global result is due to interaction. This situation is often summarized by a whole that is more than the sum of its parts. One contribution of this thesis is identifying interactions whose impact can be evaluated through the side-by-side comparison of systems including these interactions or not. Such collective benefits are defined for problem-solving by a heuristic search with a hierarchical multi-agent model. This work includes a multi-agent model of that kind of search that identifies collective benefits, which are called synergies. Three kinds of synergies are considered: the synergy between heuristics searching the same problem, the one between similar problems, and the synergy between a human user and the artificial search system. These possible synergies are included in the discussion about emergence. In this framework, some heuristic populations match the idea of a ``true composite'' that is a composed system whose result is due to interactions between components and not to its composition. The interaction between local and global levels fits in the levels induced by the model's hierarchical organisation. This way, a contribution to the study of emergence is given, by the definition possibilities that a multi-agent model allows for emergence
Terrissa, Saddek Labib. "Vision bio-fondée appliquée à l'analyse d'images." Le Havre, 2006. http://www.theses.fr/2006LEHA0001.
Full textThe environment recognition goes inevitably through the retina of alive beings. This component is a part of the central nervous system witch performs many processing, especially in motion detection and features extraction of a scene. These aspects are not reach the development level compared with static processing in artificial vision. For these reasons, we are developed a generic prototype of artificial retina inspired from biology witch we are titled Retina. This system is based on artificial neural network organised to simulate the radial variation of the acuity. The information is encoded through the implementation of analogue and impulse neurons. These retinal signal impulses are analysed in order to detect the variation of grey level in dynamic images. Moreover, we reconstitute these images using neural networks techniques and signal processing
Shwartz, Assaf. "Les interactions entre la biodiversité et les citadins au cœur d'une métropole." Paris, Muséum national d'histoire naturelle, 2012. http://www.theses.fr/2012MNHN0005.
Full textUrban conservation is emerging as an integrative discipline that explores cities, biodiversity, people and their environment. However, the role that biodiversity plays in people’s life has rarely been investigated. In this thesis, I explored the people-biodiversity interaction in small public gardens in Paris (France). I found that gardens can host rich biodiversity, which can be enhanced by certain management practices. However, while people were highly interested in discovering this diversity and related it to their well-being, they demonstrated poor capabilities to experience this diversity and in fact did not notice it. This highlights a people-biodiversity paradox, which could be related to the extinction of experience. The separation of people from nature in cities modifies people’s relation with biodiversity by limiting abilities to experience the quality of nature. Participating in conservation education programs can help avert this extinction of experience locally
Chenouard, Raphaël. "Résolution par satisfaction de contraintes appliquée à l'aide à la décision en conception architecturale." Paris, ENSAM, 2007. http://www.theses.fr/2007ENAM0041.
Full textThe embodiment and conceptual design phases of mechanical and energetic systems are concerned with preliminary design. The embodiment design phase aims at investigating the feasibility of concepts emerging from conceptual design. Embodiment achieves to make a decision concerning the main features of the architecture of a product, which will be detailed in the next design phase. The main characteristics of the product are relating to dimensions, choices of components, shapes, topologies or materials. Decisions can better be supported by carrying out models which mainly take into consideration functional specifications and physical behaviors, namely a mathematical definition of the designer knowledge. These models may be processed using a numerical Constraint Satisfaction Problem (CSP) solver. CSP solvers are based on generic algorithms suitable for the numerical treatment of problems regarded as sets of constraints, variables and domains. Through these models, designers aim at investigating design architectures, i. E. They have to match up values with the main variables of the problem (design variables and decision variables). We propose several new heuristics and algorithms taking into account some specific difficulties relating to preliminary design problems. This approach concerns the classification and relative precision of variables and piecewise constraints. The feasibility studies of several mechanical and energetic systems have been investigated. These new algorithms drastically decrease the calculation time of the numerical treatment, and also improves the quality of the computational results
Salinesi, Camille. "L'Ingénierie des Exigences appliquée aux Systèmes d'Information." Habilitation à diriger des recherches, Université Panthéon-Sorbonne - Paris I, 2010. http://tel.archives-ouvertes.fr/tel-01003149.
Full textLavaud, Gael. "Optimisation robuste appliquée au crash automobile." Ecully, Ecole centrale de Lyon, 2007. http://bibli.ec-lyon.fr/exl-doc/TH_T2092_glavaud.pdf.
Full textThis PhD thesis deals with robust design of complex systems with numerical simulations. The analysis of RENAULT design strategy suggests us concrete improvements based on the FIRST DESIGN methodology. This strategy relies on Engineering System ant the concept of robustness applied to all design step of the product. It uses all new design tools, as well statistical as modeling. This strategy allows designers to avoid design resource consuming and time demanding iterations. We use two concrete case study to illustrate our work. These cases take place in the context of costly simulations, that is why we will have to save them. First case study is about shape optimization of a pedestrian hood panel. We develop an industrial optimization tool and use it to find a satisfying shape. To save simulations, we also develop a special parameterization and a progressive validation of new shapes. Second case study is about robust design of a frontal crash. As traditional approaches could not fit our problem, we develop an original strategy based on crash scenarios. This method allows designers to identify unstable structural elements and to stabilize their behavior during the crash in the best conditions. All the tools developed along this work tend to take the best advantage of both new simulations tools and designers experience. Improvements of performance and design time on both cases study prove the interest of modifying the traditional design strategy to incorporate new numerical design tools
Burneau, Jean-Christophe. "Une méthode de conception orientée objets, appliquée aux systèmes de maintenance d'équipements de radiodiffusion." Nancy 1, 1991. http://docnum.univ-lorraine.fr/public/SCD_T_1991_0262_BURNEAU.pdf.
Full textCaelen, Olivier. "Sélection séquentielle en environnement aléatoire appliquée à l'apprentissage supervisé." Doctoral thesis, Universite Libre de Bruxelles, 2009. http://hdl.handle.net/2013/ULB-DIPOT:oai:dipot.ulb.ac.be:2013/210265.
Full textDans un premier cas, les tests viseront à maximiser la somme des gains collectés. Un juste compromis doit alors être trouvé entre l'exploitation et l'exploration. Ce problème est couramment dénommé dans la littérature scientifique "multi-armed bandit problem".
Dans un second cas, un nombre de sélections maximal est imposé et l'objectif consistera à répartir ces sélections de façon à augmenter les chances de trouver l'alternative présentant le gain moyen le plus élevé. Ce deuxième problème est couramment repris dans la littérature scientifique sous l'appellation "selecting the best".
La sélection de type gloutonne joue un rôle important dans la résolution de ces problèmes de décision et opère en choisissant l'alternative qui s'est jusqu'ici montrée optimale. Or, la nature généralement aléatoire de l'environnement rend incertains les résultats d'une telle sélection.
Dans cette thèse, nous introduisons une nouvelle quantité, appelée le "gain espéré d'une action gloutonne". Sur base de quelques propriétés de cette quantité, de nouveaux algorithmes permettant de résoudre les deux problèmes décisionnels précités seront proposés.
Une attention particulière sera ici prêtée à l'application des techniques présentées au domaine de la sélection de modèles en l'apprentissage artificiel supervisé.
La collaboration avec le service d'anesthésie de l'Hôpital Erasme nous a permis d'appliquer les algorithmes proposés à des données réelles, provenant du milieu médical. Nous avons également développé un système d'aide à la décision dont un prototype a déjà été testé en conditions réelles sur un échantillon restreint de patients.
Doctorat en Sciences
info:eu-repo/semantics/nonPublished
Shen, Yimin. "Bioinformatique appliquée à la prédiction de structure de peptides." Paris 7, 2012. http://www.theses.fr/2012PA077070.
Full textIn these years, peptides have become an alternative more and more credible than small chemical compounds. A key step of peptide functional characterization is the characterization of their 3D structure. A first part of this thesis is related to PEP-FOLD optimization. PEP-FOLD is a fast and accurate approach for the de novo peptide structure prediction, based on the concept of structural alphabet. We show it is possible to bais the conformational search to as few as 20% of the full conformational space still having the native conformation. This biais results from an a posteriori analysis of the structural alphabet trajectories used for PEP-FOLD 3D generation. Based on such bias, we show PEP-FOLD2 can efficiently generate models for peptides up to 50 amino acids, at a precision of 2. 6 A, which is slightly better than a reference approach such as Rosetta. It is also a major improvement compared to PEP-FOLD1 that could generate 3D models for peptides up to 30 amino acids only, at a precision of 2-3 A. The second part focused on candidate fragment identification from sequence. I have developed an approach for structural alphabet profile comparison, to identify fragments of know structure at low sequence identity. Results show that, mining a non redundant collection of structures at low sequence identity (25%), the procedure is able to identify fragments of size between 6 and 27 amino acids covering 90% of the structures. Besides, a comparison with a recent amino acid profile-profile approach - Hhfrag - shows that the local structure information predicted using our structural alphabet allows a significant gain in accuracy - 27%, with a slightly better coverage
Liebgott, Pierre-Pol. "Etude de la biodiversité bactérienne d'un biotope oléicole hypersalé et des voies biochimiques impliquées dans la transformation du tyrosol." Aix-Marseille 1, 2007. http://theses.univ-amu.fr.lama.univ-amu.fr/2007AIX11031.pdf.
Full textThe liquid effluents which are composed of a mixture of olive mill wastewater and brine waters, are rejected by olive industries, and contribute considerably to the environmental pollution of the Mediterranean area. They are characterized by a high DCO and significant quantities of recalcitrant phenolic compounds, of which the most representative is the tyrosol. The objective of this work was thus to understand the degradation pathways of the tyrosol, via the use of endogenous microorganisms isolated from these effluents. By using the technique ' Single Strand Conformation Polymorphism' (SSCP), the study of the biodiversity showed a profile of salted biotope. Various halophilous, aerobic and anaerobic bacterial isolates were then obtained through a targeted traditional insolation. Two of them, TYRC17 an HTB24, belonging to the Halomonas genus, were studied for their capacity to transform the tyrosol and p - hydroxyphenylacetate d (HPA), into orthodiphenolic molecules (3,4-dihydroxyphenylacetate or DHPA and homogentisate HGA), known as potential antioxydants. This study enabled us to highlight two degradation pathways of the tyrosol, never yet described to date. Thus, the tyrosol is oxidized via an aryl-dehydrogenase in HPA. This last is hydroxylated either in HGA via a HPA 1-monooxygenase from strain TYRC17, or in DHPA via a HPA 3-monooxygenase from strain HTB24. These ortho-diphenolic compounds are then cleaved by catechol-dioxygenases, and finally degraded. In the presence of transition metals, contained in the Widdel oligo-element solution, the HGA is transformed not biologically, into 2,5-dihydroxybenzaldehyde (gentisaldehyde) which is involved in the pyomelanin formation. Moreover, strain HTB24 is able to produce hydroxytyrosol by aspecific tyrosol hydroxylation, via the same HPA 3-monooxygenase, the latter being induced by the presence of HPA in the medium. Moreover, the study of other bacteria having this enzyme, such as Pseudomonas aeruginosa or Escherichia coli confirmed that they are also able to transform the tyrosol into hydroxytyrosol. These original bioconversions thus open great prospects for the bioremediation for these strongly salted effluents, but also for a promising production of antioxydant molecules with high added value
Meygret, Aimé. "Vision automatique appliquée à la détection d'obstacles dans un environnement routier." Nice, 1990. http://www.theses.fr/1990NICE4411.
Full textKhouaja, Youssef. "Une méthode de contrôle par le réseau appliquée à la gestion hiérarchique de la mobilité IPv6." Rennes 1, 2002. http://www.theses.fr/2002REN10114.
Full textSaint, Michel Thierry. "Filtrage non linéaire en vue d'une segmentation semi supervisée appliquée à l'imagerie médicale." Lille 1, 1997. http://www.theses.fr/1997LIL10110.
Full textGennero, Marie-Claude. "Contribution à l'étude théorique et appliquée des codes correcteurs d'erreurs." Toulouse 3, 1990. http://www.theses.fr/1990TOU30243.
Full textOrset, Jean-Marie. "Une architecture de test passif appliquée à la détection des attaques dans les réseaux ad hoc." Evry, Institut national des télécommunications, 2007. http://www.theses.fr/2007TELE0002.
Full textThomassey, Sébastien. "Méthodologie de la prévision des ventes appliquée à la distribution textile." Lille 1, 2002. https://pepite-depot.univ-lille.fr/LIBRE/Th_Num/2002/50376-2002-239-240.pdf.
Full textCependant, ils sont généralement inadaptés au contexte textile. En effet, leurs capacités d'apprentissage et de modélisation sont souvent limitées sur les historiques courts et perturbés des ventes textiles. L'interprétation et l'intervention de l'utlisateur sont également souvent compliquées avec les modèles classiques. Ainsi, nous proposons un système de prévision, constitué de plusieurs modèles qui abordent des prévisions sur divers horizons et à différents niveaux d'agrégation des ventes. Ce système est basé sur des techniques issues du "soft computing" telles la logique floue, les réseaux de neurones ou les procédures évolutionnistes, autorisant le traitement de données incertaines. Les performances de nos modèles sont ensuite évaluées et analysées sur un jeu de données réelles provenant d'un grand distributeur textile. Enfin, dans le cadre d'un projet nommé AIDE financé par le ministère de l'économie, des finances et de l'industrie, ce système de prévision s'intègre dans un outil d'aide à la décision à destination de chacun des acteurs de la filière textile
Guibert, Stéphane. "La thermographie infrarouge à détection synchrone appliquée aux matériaux composites." Thesis, Université Laval, 2007. http://www.theses.ulaval.ca/2007/24372/24372.pdf.
Full textInthasone, Somsack. "Techniques d'extraction de connaissances en biodiversité." Thesis, Nice, 2015. http://www.theses.fr/2015NICE4013/document.
Full textBiodiversity data are generally stored in different formats. This makes it difficult for biologists to combine and integrate them in order to retrieve useful information and discover novel knowledge for the purpose of, for example, efficiently classifying specimens. In this work, we present the BioKET data warehouse which is a consolidation of heterogeneous data stored in different formats and originating from different sources. For the time being, the scope of BioKET is botanical. Its construction required, among others things, to identify and analyze existing botanical ontologies, to standardize and relate terms in BioKET. We also developed a methodology for mapping and defining taxonomic terminologies, that are controlled vocabularies with hierarchical structures from authoritative plant ontologies, Google Maps, and OpenStreetMap geospatial information system. Data from four major biodiversity and botanical data providers and from the two previously mentioned geospatial information systems were then integrated in BioKET. The usefulness of such a data warehouse was demonstrated by applying classical knowledge pattern extraction methods, based on the classical Apriori and Galois closure based approaches, to several datasets generated from BioKET extracts. Using these methods, association rules and conceptual bi-clusters were extracted to analyze the risk status of plants endemic to Laos and Southeast Asia. Besides, BioKET is interfaced with other applications and resources, like the GeoCAT Geospatial Conservation Assessment Tool, to provide a powerful analysis tool for biodiversity data
Noulard, Eric. "Programmation parallèle orientée objet et réutilisabilité appliquée à l'algèbre linéaire." Phd thesis, Université de Versailles-Saint Quentin en Yvelines, 2000. http://tel.archives-ouvertes.fr/tel-00378738.
Full textAprès une revue des moyens de programmation parallèle et des
concepts objets, la conception et la réalisation d'une bibliothèque parallèle d'algèbre linéaire orientée-objet sont présentées. Nous étudions deux moyens de programmation parallèle, le premier, C++//, est un LAO parallèle à objets actifs dérivé de C++, le second est l'utilisation de MPI au travers d'une surcouche objet minimale.
Ces deux approches objets posent des problèmes soit de performances soit de réutilisabilité séquentielle/parallèle qui sont présentés et résolus.
Nous proposons notamment un mécanisme simple de partage en lecture pour les modèles à objets actifs, en montrant son utilité en terme de performances de nos applications. Suite à la seconde approche nous définissons les notions de formes de matrices et de matrices avec forme qui permettent d'atteindre nos objectifs de réutilisabilité séquentielle/parallèle.
Au final, la conception et la réalisation permettent d'instancier, à partir du même code [séquentiel] d'algèbre linéaire, une version séquentielle et parallèle offrant des performances satisfaisantes.
Sautot, Lucile. "Conception et implémentation semi-automatique des entrepôts de données : application aux données écologiques." Thesis, Dijon, 2015. http://www.theses.fr/2015DIJOS055/document.
Full textThis thesis concerns the semi-automatic design of data warehouses and the associated OLAP cubes analyzing ecological data.The biological sciences, including ecology and agronomy, generate data that require an important collection effort: several years are often required to obtain a complete data set. Moreover, objects and phenomena studied by these sciences are complex and require many parameter recording to be understood. Finally, the collection of complex data over a long time results in an increased risk of inconsistency. Thus, these sciences generate numerous and heterogeneous data, which can be inconsistent. It is interesting to offer to scientists, who work in life sciences, information systems able to store and restore their data, particularly when those data have a significant volume. Among the existing tools, business intelligence tools, including online analytical systems (On-Line Analytical processing: OLAP), particularly caught our attention because it is data analysis process working on large historical collections (i.e. a data warehouse) to provide support to the decision making. The business intelligence offers tools that allow users to explore large volumes of data, in order to discover patterns and knowledge within the data, and possibly confirm their hypotheses.However, OLAP systems are complex information systems whose implementation requires advanced skills in business intelligence. Thus, although they have interesting features to manage and analyze multidimensional data, their complexity makes them difficult to manage by potential users, who would not be computer scientists.In the literature, several studies have examined the automatic multidimensional design, but the examples provided by theses works were traditional data. Moreover, other articles address the multidimensional modeling adapted to complex data (inconsistency, heterogeneous data, spatial objects, texts, images within a warehouse ...) but the proposed methods are rarely automatic. The aim of this thesis is to provide an automatic design method of data warehouse and OLAP cubes. This method must be able to take into account the inherent complexity of biological data. To test the prototypes, that we proposed in this thesis, we have prepared a data set concerning bird abundance along the Loire. This data set is structured as follows: (1) we have the census of 213 bird species (described with a set of qualitative factors, such as diet) in 198 points along the river for 4 census campaigns; (2) each of the 198 points is described by a set of environmental variables from different sources (land surveys, satellite images, GIS). These environmental variables address the most important issue in terms of multidimensional modeling. These data come from different sources, sometimes independent of bird census campaigns, and are inconsistent in time and space. Moreover, these data are heterogeneous: they can be qualitative factors, quantitative varaibles or spatial objects. Finally, these environmental data include a large number of attributes (158 selected variables) (...)
Peixoto, Joana. "Les formateurs parlent : les représentations des formateurs d'enseignants dans le domaine de l'informatique appliquée à l'éducation." Paris 8, 2005. http://www.theses.fr/2005PA082491.
Full textWhy teachers prepared to work with Educational Informatics (EI) do not transfer to their practice what they learned during their formation? Our hypothesis is that there is, among EI formers, a certain representation of the formation model and a conception on the beneficial aspects of digital technologies that would be automatically transferred to the teachers' practice. The concept of social representation (Moscovici) was adopted to identify on the formers discourse how they conceive the formation and bring it to the practice. The investigation was done with the 21 teachers of the discipline "Informatics and Education" from the Catholic University of Goiás, in Brazil, using the techniques of words associations, semi-structured interview and focus group. The meanings of EI are made evident to those teachers through the many metaphors that constitute their imaginary: from the teaching machine to the bottomless well, and also a partner, a mirror, a miracle. .
Bigand, Michel. "Générateur automatique de géométries admissibles en optimisation de forme." Paris 6, 1988. http://www.theses.fr/1988PA066082.
Full textCaudy, Nathalie. "Analyse de la conformité appliquée à la détection des interactions de services dans le réseau intelligent." Versailles-St Quentin en Yvelines, 1996. http://www.theses.fr/1996VERS0007.
Full textSirugue, Léa. "Conception et développement d’une méthode de comparaison de surfaces appliquée aux protéines." Thesis, Paris, HESAM, 2020. http://www.theses.fr/2020HESAC042.
Full textProtein interactions play a crucial role in the living processes such as cell communication, immunity, cell growth, proliferation and death. These interactions occur through the surface of proteins and the disruption of their interactions is the start of many disease processes. It is therefore necessary to understand and characterize the surface of proteins and their interactions to better understand living processes. Different methods of protein surfaces comparison have been developed in the recent years but none are powerful enough to handle all the structures currently available in databases. The PhD project is to develop rapid methods of surface comparison and apply them to the surface of macromolecules
Lalanne, Vincent. "Gestion des risques appliquée aux systèmes d’information distribués." Thesis, Pau, 2013. http://www.theses.fr/2013PAUU3052/document.
Full textIn this thesis we discuss the application of risk management to distributed information systems. We handle problems of interoperability and securisation of the exchanges within DRM systems and we propose the implementation of this system for the company: it needs to permit the distribution of self-protected contents. We then present the (our) participation in the creation of an innovative company which emphasizes on the security of information, in particular the management of risks through the ISO/IEC 27005:2011 standard. We present risks related to the use of services, highlighting in particular the ones which are not technological: we approach inheritent risks in clouds (provider failure, etc ...) but also the more insidious aspects of espionage and intrusion in personal data (Case PRISM in June 2013). In the last section, we present a concept of a DRM company which uses metadata to deploy settings in usage control models. We propose a draft formalization of metadata necessary for the implementation of a security policy and guarantee respect of regulations and legislation
Duquenoy, Mickaël. "Analyse Temps-Fréquence appliquée à l'imagerie SAR polarimétrique." Phd thesis, Université Rennes 1, 2009. http://tel.archives-ouvertes.fr/tel-00480262.
Full textBeye, Mamadou. "Génomique en temps réel appliquée aux isolats bactériens cliniques atypiques." Thesis, Aix-Marseille, 2017. http://www.theses.fr/2017AIXM0558.
Full textRapid and accurate diagnosis, characterization and identification of pathogens are essential to guide treatment and detect transmission events or treatments failures. However, the biomedical field is confronted with emerging and re-emerging pathogens. Some of these clinical bacterial strains exhibit specificities concerning the virulence, contagiousness and / or resistance to antibiotics. High-throughput sequencing and comparative analysis of bacterial genomes is a reliable strategy enabling the rapid study of the characteristics of these emerging pathogens. In a short period, not exceeding 20 years, genomics has known a considerable revolution. In effect the introduction of the new high-throughput sequencingtechnologies and the increased concern of the scientist into this field, led to an exponential increase of number of available sequenced bacterial genomes in public databases. Real-time genomics is a strategy consisting on rapid analysis of the genome of a clinical bacterial strain in order to identify the genetic determinants justifying its unusual phenotypic characteristics. Thus, the objectives of this thesis project were: to rapidly exploit whole-genome sequencing data for identification of the virulence or resistance repertoire; to compare genomes from atypical clinical bacteria to those of other bacteria of the same species in order to identify their specific features; to use genomes as a taxonomic tool to rapidly describe the new bacterial species isolated in the laboratory by culturomics approach
Gagliolo, Matteo. "Online Dynamic Algorithm Portfolios: Minimizing the computational cost of problem solving." Doctoral thesis, Università della Svizzera italiana, Lugano, Switzerland, 2010. http://hdl.handle.net/2013/ULB-DIPOT:oai:dipot.ulb.ac.be:2013/250787.
Full textPermanent URL: http://doc.rero.ch/record/20245
info:eu-repo/semantics/nonPublished
Buyssens, Pierre. "Fusion de différents modes de capture pour la reconnaissance de visages appliquée aux e_transactions." Caen, 2011. http://www.theses.fr/2011CAEN2007.
Full textIn this thesis, we describe the problem of automatic face recognition in visible and long--wave infrared lights. The state of the art methods are described, and we study, in a first time, a method based on convolutional neural networks. Applied to both modalities, the proposed fusion method, based on a weighted sum of scores, yields a substantial increasing of the recognition rates. In a second time, a pretraining of the network with sparse methods is studied for automatic facial recognition. Finally, we propose an approach based on a sparse decomposition of faces, coupled with a classification scheme involving in a l1 minimization. This last approach gives good identification rates on the well known Notre-Dame database
Bame, Ndiouma. "Gestion de donnée complexes pour la modélisation de niche écologique." Thesis, Paris 6, 2015. http://www.theses.fr/2015PA066125/document.
Full textThis thesis concerns large scale biodiversity data management. Its objectives are to optimize queries for researchers who have free access to biodiversity worldwide data. These data which are shared by worldwide research laboratories are federated in GBIF data warehouse. GBIF makes accessible its data to researchers, policy makers and general public. With a significant amount of data and a rapid growth of data and users that express new needs, the GBIF portal is facing a double problem of expressiveness of queries and of efficiency. Thus, we propose a decentralized solution for biodiversity data interrogation. Our solution combines the resources of several of remote and limited machines to provide the needed computing and storage power to ensure system responsiveness for users. It also provides high-level query interface which is more expressive for users. Then, we propose a dynamic data distribution on demand approach. This approach which is based on data properties and characteristics of users analysis queries adapts dynamically machines capacities to users demands. Then, we propose a queries optimization approach that adapts dynamically data placement and machines loads according to performances in order to process users queries within deadlines. We experimentally validated our solution with real GBIF data concerning 100 million observation data
Haydar, Mahmoud. "Interaction en Réalité Mixte appliquée à l'Archéologie Sous-Marine." Phd thesis, Université d'Evry-Val d'Essonne, 2011. http://tel.archives-ouvertes.fr/tel-00677294.
Full textEl, Maadi Amar. "Vision infrarouge appliquée à la vidéosurveillance extérieure automatisée : reconnaissance, suivi, classification et détection d'événements." Thesis, Université Laval, 2006. http://www.theses.ulaval.ca/2006/23614/23614.pdf.
Full textNguyen, Le-Vinh. "Technique de programmation par contraintes pour la vérification bornée des programmes : une stratégie non-séquentielle appliquée à un système embarqué critique." Nice, 2011. http://www.theses.fr/2011NICE4042.
Full textThis thesis is devoted to program verification using the constraint programming technique. In particular, it focus on the incremental exploration strategy of executable paths of a program for verification and automatic counterexample generation using constraint solvers such as CP, LP, MIP, SMT. The context of this work is the Bounded Model Checking (BMC), an incomplete formal verification method, which only deals with paths of a bounded length in programs. In this thesis, we propose DPVS (Dynamic Postcondition-Variables driven Strategies), a new strategy based on the dynamic generation of a system of constraints in during the exploration of the control flow graph of the program. DPVS uses a backward search technique guided by the variables related to the property to prove. This strategy was evaluated on academic examples and real applications. Experiments on an industrial controller which manages the flashing lights of a car show that our system is more efficient than CPBPV, our previous approach, and than CBMC, a-state-of-the-art bounded model checker. We have developed a prototype in COMET that uses the DPVS strategy for program verification and automatic generation of counterexamples. This prototype uses many classical techniques to simplify the control flow graph such as calculating bounds of variables, slicing, propagation of constants. DPVS was successful in finding a counter–example of a real application, the Flasher Manager, that was provided by Gensoft, an industrial partner of the research project TESTEC
Camarda, Federico. "Fusion de données multi-capteurs pour la détection des bords de voie appliquée au véhicule autonome." Thesis, Compiègne, 2022. http://www.theses.fr/2022COMP2673.
Full textPerception and correct understanding of the road scene is crucial for any application of assisted and automated driving. In order to guarantee safety of the passenger and other road users, planning and navigation must be made on the basis of a reliable environment representation. Multi-sensor data and prior information is used to build this representation which incorporates identification of road users and road structure. For the latter, the focus is put on the drivablespace and lane repartition. On highways, urban streets and generally all over the road network, the drivable space is organized in oriented corridors which enablesafer and predictable navigation for everyone. In the development of intelligentvehicles, identifying the lane repartition and building an accurate road representation outlines the lane boundaries detection task. Depending on the specifics of the target automated system, car manufacturers integrate in currently commercialized vehicles ready-to-use lane detection solutions from Tier-1 suppliers generally featuring single and vision-based smart sensors. Such solutions may not be adequate in highly automated systems where the driver is allowed to divert their attention from the driving task and become passenger. This thesis addresses the problem of lane boundaries identification relying on multi-sensor fusion of smart camera data (specifically, frontal and AVM cameras) and HD-maps. In the first part, an appropriate modeling for smart sensor measurements which is independent from the sensor nature is proposed. Uncertain detections of markings, barriers and other road elements contribute to the tracking of the surrounding lane boundaries using a novel clothoid-spline model. The second part focuses on the integration of prior information coming from digital maps. Similarly to the modeling of smart sensors, the involved uncertainties in the usage of map-providers have been taken into account to support the lane boundaries estimation. For the testing of the proposed approaches, a custom dataset of road data has been recorded using both off-the-shelf smart sensors and live streamed HD-maps. Validated and tuned tracking solutions are then integrated in close-loop experimentations on Renault prototype vehicle of SAE Level 3 of automation
La percezione e la corretta comprensione della scena stradale e fondamentale per qualsiasi applicazione di guida assistita e automatizzata. Per garantire la sicurezza del passeggero e degli altri utenti della strada, la pianificazione e la navigazione devono essere effettuate sulla base di una rappresentazione affidabile dell’ambiente. Dati di origine multi-sensore e informazioni disponibili a priori sono utilizzati per costruire questa rappresentazione che incorpora l’identificazione degli utenti della strada e la struttura della strada stessa. Per quest’ultima, l’attenzione e posta sullo spazio percorribile e sulla ripartizione in corsie. Sulle autostrade, le strade urbane e in generale su tutta la rete stradale, lo spazio percorribile e organizzato in corridoi orientati che permettono una navigazione piu sicura e prevedibile per tutti. Nello sviluppo di veicoli intelligenti, l’identificazione della ripartizione in corsie e la costruzione di una rappresentazione accurata della strada delinea il compito di rilevamento dei confini delle corsie o lane boundaries detection. A seconda delle specifiche del sistema automatizzato di destinazione, le case automobilistiche integrano nei veicoli attualmente commercializzati soluzioni di rilevamento di corsia pronte all’uso da fornitori Tier-1, generalmente composte di singoli sensori intelligenti e basate sulla visione computerizzata. Tali soluzioni potrebbero non essere adeguate in sistemi altamente automatizzati dove al guidatore e permesso di distogliere l’attenzione dal compito di guida e di diventare passeggero. Questa tesi di dottorato affronta il problema dell’identificazione dei limiti di corsia basandosi sulla fusione multi-sensore di dati provenienti da telecamere intelligenti (in particolare, telecamere frontali e AVM) e mappe HD. Nella prima parte, viene proposta una modellazione appropriata per le misure dei sensori intelligenti che e indipendente dalla natura del sensore. I rilevamenti incerti di marcature, barriere e altri elementi stradali contribuiscono alla stima dei limiti delle corsie circostanti utilizzando un nuovo modello di spline di clotoidi. La seconda parte si concentra sull’integrazione di informazioni provenienti da mappe digitali. Analogamente alla modellazione dei sensori intelligenti, le incertezze coinvolte nell’uso di map-providers sono state prese in considerazione per supportare l’identificazione dei limiti di corsia. Per testare gli approcci proposti, e stato registrato un dataset personalizzato di dati stradali utilizzando sia sensori intelligenti off-the-shelf che mappe HD in live streaming. Le soluzioni di tracking convalidate e correttamente regolate sono poi integrate in sperimentazioni a circuito chiuso su un veicolo prototipo Renault di livello 3 di automazione SAE
Aljundi, Ahmad Chadi. "Une méthodologie multi-critères pour l'évaluation de performance appliquée aux architectures de réseaux d'interconnexion multi-étages." Lille 1, 2004. https://pepite-depot.univ-lille.fr/LIBRE/Th_Num/2004/50376-2004-Aljundi.pdf.
Full textHerradi, Noura. "Représentation sémantique multilingue, multiculturelle et temporelle des relations interpersonnelles, appliquée à une prothèse de mémoire." Thesis, Paris, CNAM, 2018. http://www.theses.fr/2018CNAM1205/document.
Full textIn this thesis, we propose a knowledge base for a "smart" memory prosthesis, called CaptainMemo, which aims to help Alzheimer's patients to overcome their memory impairments. This knowledge base is built over the temporal, multicultural and multilingual PersonLink ontology. This ontology gives the memory prosthesis a rigorous, multilingual and temporal semantic representation of interpersonal relationships. The PersonLink ontology is dereferenceable and available in the Linked Data. Multilingualism and temporal representation are two major research topics in computer science and in the Semantic Web in particular.Multilingualism applied to the representation of interpersonal relationships requires specific treatment because it is linked to multiculturalism. In addition, the transition from one culture / language to another is a major research problem. Indeed, literal translation is not always allowed, especially when it comes to interpersonal relationships, because they are culturally dependent. In this context, we propose an approach allowing the representation of ontologies in several cultures / languages. This approach, based on a translation algorithm, allows the transition from one culture / language to another by making a cultural translation rather than a literal one. Thus, by adopting this approach, our PersonLink ontology allows an exact representation of interpersonal relationships, because it takes into consideration the cultural aspect for the definition of each relationship, and assigns the appropriate term according to the language related to this culture. Interpersonal relationships are governed by rules and constraints that define them according to each culture, these constraints are represented semantically in the PersonLink ontology using OWL2. However, it is difficult to consider these constraints when introducing the temporal dimension to represent the time intervals of these interpersonal relationships, especially when these are diachronic and their time intervals are qualitative. Indeed, the legacy models and solutions make it possible to make a temporal representation of the time intervals (e.g. 4D-Fluents), and to link between these time intervals (e.g. Allen Relations), but do not take into account the semantics constraints of interpersonal relationships. In this context, we propose an approach that allows a semantic representation, based on OWL2 constraints, for the representation of qualitative time intervals. Finally, to deal with the intelligence of the CaptainMemo memory prosthesis, we propose an approach for reasoning over time intervals. In this approach we introduce a set of SWRL rules to assert inferred temporal Allen relationships, allowing reasoners, such as Pellet that supports DL-Safe rules, to be used for the inference and the verification of consistency over the temporal relationships between different time intervals. Thus, the table of compositions of the relations between time intervals has been considerably reduced, since it is based on a tractable set of these relations, and, consequently, the processing time of the reasoning becomes shorter
Brito, Gabrielle de. "Analyse ergonomique du suivi de procédures écrites dans les environnements dynamiques (speed) appliquée à l'aéronautique." Paris 5, 2000. http://www.theses.fr/2000PA05H057.
Full textGuilleminot, Christian. "Décomposition adaptative du signal de parole appliquée au cas de l'arabe standard et dialectal." Besançon, 2008. http://www.theses.fr/2008BESA1030.
Full textThe present work introduces in phonetics, the atomic decomposition of the signal also known as the Matching Pursuit and treats a group of atoms by compression without losses and finally measures the distance of the list of atoms compressed using the Kolmogorov's algorithms. The calibration is based on an initial classical analysis of the co-articulation of sound sequences of VCV and CV, or V ∈ {[i] [u] [a]} and C ∈ {[t] [d] [s] [δ]}∪ [tʕ] [dʕ] [sʕ [δʕ]} the excerpts culled from a corpus made up of four arabic speaking areas. The locus equation of CV vs CʕV, makes it possible to differentiate the varieties of the language. In the second analysis, an algorithm of atomic adaptative decomposition or Matching Pursuit is applied to the sequences VCV and VCʕV still on the same corpus. The atomic sequences representing VCV et VCʕV are then compressed without losses and the distances between them are searched for by Kolmogorov's algorithms. The classification of phonetic recordings obtained from these arabic speaking areas is equivalent to that of the first method. The findings of the study show how the introduction of Matching Pursuit's in phonetics works, the great robustness of the use of algorithms and suggesting important possibilities of automation of processes put in place, while opening new grounds for further investigations
Ménigot, Sébastien. "Commande optimale appliquée aux systèmes d'imagerie ultrasonore." Phd thesis, Université François Rabelais - Tours, 2011. http://tel.archives-ouvertes.fr/tel-00657980.
Full textMenet, Ludovic. "Formalisation d'une approche d'Ingénierie Dirigée par les Modèles appliquée au domaine de la gestion des données de référence." Paris 8, 2010. http://www.theses.fr/2010PA083184.
Full textOur research work is in line with the problematic of data models definition in the framework of Master Data Management. Indeed, Model Driven Engineering (MDE) is a theme in great expansion in the academic world as well as in the industrial world. It brings an important change in the conception of applications taking in account the durability of savoir-faire and of gains of productivity, and taking profits of platforms advantages without suffering of secondary effects. The MDE architecture is based on the transformation of models to come to a technical solution on a chosen platform from independent business models of any platform. In this thesis, a conceptual and technical thought process of the MDE approach is applied to the definition of pivot data models, which are the base of Master Data Management (MDM). Thus, we use Unified Modeling Language (UML) as formalism to describe the independent aspects of the platform (business model), and we propose a meta-model, in the form of an UML profile, to describe the dependent aspects of the MDM platform. Then, we present our approach to move from a business model to a platform model to be able to generate the physical pivot model. The inputs of the thesis are : the study of a MDE approach in the MDM context, the definition of UML transformations towards a MDM model (based on a XML Schema structure), besides we bring a new aspect to MDE applied to MDM, that is to say the definition of a method for incremental model validation allowing the optimization of validation stages during model conception
Saddem, Ramla. "Diagnosticabilité modulaire appliquée au Diagnostic en ligne des Systèmes Embarqués Logiques." Phd thesis, Ecole Centrale de Lille, 2012. http://tel.archives-ouvertes.fr/tel-00861200.
Full textDe, Mezzo Benoît. "Reconnaissance d'objets par la génération d'hypothèses de modèles de forme appliquée à l'extraction des feuilles de plantes dans des scènes naturelles complexes." Montpellier 2, 2004. http://www.theses.fr/2004MON20153.
Full textDepierre, Amélie. "La Composition en anglais scientifique : application au domaine de la physico-chimie." Littoral, 2001. http://www.theses.fr/2001DUNK0070.
Full textThis study stems from teaching experience and comes as a response to students' questions and errors. It has two main objectives : on the one hand a critical review of existing research on compounds and other complex nominals in English, and on the other hand the application of the theoretical data available in order to help French-speaking students and researchers, who need to communicate (understand, speak and write) in English, namely in the domain of physico-chemical sciences. After a survey of the characteristics of compounds, and an investigation of axamples taken from several corpora, a new morphological classification of binominal compounds is suggested, then the semantic relations between their elements are pointed out. The results are extended to longer nominals, with a particular attention paid to words built from Greek and Latin bases, and acronyms, which abound in the field of sciences. These competing structures : N₂ N₁, N₂'s N₁ and N₁ of N₂ are contrasted, as they are a major source of error among French speakers, especially at the encoding stage. Lastly, an electronic glossary in Excel format including some of the terms of the domain, together with examples of attested sentences, has been developed. One of the future objectives is to transfer the glossary in HTML format in order to make it more available for use in various domains
Hartmann, Chris. "Automatisation de la synthèse d’architectures appliquée aux aéronefs à voilure tournante." Thesis, Ecole centrale de Nantes, 2018. http://www.theses.fr/2018ECDN0002/document.
Full textThe research work presented in this thesis is related to the System Engineering field and the computer aided synthesis field. A methodology realized by a newsoftware is developed and introduced. The synthesis process is semi-automated and is devided into three phases: the need extraction and its translation into system requirements, a logical architecture synthesis and a physical architecture analysis. The need extraction and its translation into system requirements are highly inspired from previous work from the System Engineering field. Nevertheless, the objective, at this step, is to provide the software and the user with a unique model understandable to both. Stakeholders, life situations, needs, requirements and interfaces with the environment are modelized. The logical architecture synthesis, or logical architecture generation, is in accordance with the models we build previoulsy. That is to say that all logical architectures are instantiations of the system requirements expressed before. A C++ code translates this model into mathematical equations solved by an integer CSP solver. The result of this part of the methodology is a set of graphs, ranked by family. These graphs are views of the logical architectures. They express all the possible links between the sub-systems of the architectures. The physical architecture analysis step is an automated equation writer. The equations are non-linear and non differential and they are written for each logical architecture generated at the previous step. They are used for a first step of pre-sizing. These systems are then solved by a CSP solver for real numbers through an optimization. At the end, all the feasible architectures are ranked according to a unique state variable that iscommon to all possible solutions
Alvarez, Aldana José Alfredo. "Une méthode de test fonctionnel en-ligne basée sur une approche de monitorage distribuée continue appliquée aux systèmes communicants." Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLL005/document.
Full textMANETs represent a significant area of network research due to the many opportunities derived from the problematics and applications. The most recurring problematics are the mobility, the availability and also the limited resources. A well-known interest in networks and therefore in MANETs is to monitor properties of the network and nodes. The problematics of the MANETs can have a significant impact on the monitoring efforts. Mobility and availability can create incomplete results for the monitoring. The usual properties discussed in monitoring are simple ones, e.g., average CPU consumption, average bandwidth and so on. Moreover, the evolution of networks has led to an increasing need to examine more complex, dependent and intertwined behaviors. The literature states that accuracy of the approaches is not reliable and difficult to achieve due to the dynamic properties of the MANET. Therefore, we propose a decentralized and distributed monitoring architecture that rely on multiple points of observation. The decentralized approach combines gossip and hierarchical algorithms to provide an effective monitoring approach. Through extensive experimentation, we concluded that although we were able to achieve exceptional performance, network fragmentation still has a harsh impact on the approach. Trying to improve our approach, we proposed a distributed approach, relying on stronger bedrock to enhance the overall efficiency and accuracy. It provides a consensus mechanism that allows it to aggregate and provides a more meaningful and accurate result. We support our proposal with numerous mathematical definition that models local results for a single node and global results for the network. Our experiments were evaluated with an emulator built in-house that relies on Amazon Web Services, NS-3, Docker and GoLang with varying number of nodes, network size, network density, speed, mobility algorithms and timeouts. Through this emulator, we were able to analyze multiple aspects of the approaches by providing a repeatable, documented and accessible test beds. We obtained promising results for both approaches, but for the distributed approach, especially regarding accuracy