Contents
Academic literature on the topic 'Informatique périphérie'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Informatique périphérie.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Dissertations / Theses on the topic "Informatique périphérie"
Mazouzi, Houssemeddine. "Algorithmes pour le déchargement de tâches sur serveurs de périphérie." Thesis, Paris 13, 2019. http://www.theses.fr/2019PA131076.
Full textComputation offloading is one of the most promising paradigm to overcome the lack of computational resources in mobile devices. Basically, it allows the execution of part orall of a mobile application in the cloud. The main objective is to reduce both execution time and energy consumption for the mobile terminals. Unfortunately, even if clouds have rich computing and storage resources, they are usually geographically far from mobile applications and may suffer from large delays, which is particularly problematic for mobile applications with small response time requirements. To reduce this long delay, one of the emerging approach is to push the cloud to the network edge. This proximity gives the opportunity to mobile users to offload their tasks to “local” cloud for processing. An Edge Cloud can be seen as small data center acting as a shadow image of larger data centers. This geographical proximity between mobile applications and edge cloud means that the access delay can be greatly reduced, but affects also higher throughput, improved responsiveness and better scalability. In this thesis, we focus on computation offloading in mobile environment (Mobile Edge Computing - MEC), composed of several edge servers. Our goal is to explore new and effective offloading strategies to improve applications performances in both execution time and energy consumption, while ensuring application requirements. Our first contribution is a new offloading strategy in the case of multiple edge servers. Thenwe extend this strategy to include the Cloud. Both strategies have been evaluated theoretically and experimentally by the implementation of an offloading middleware. Finally, we propose a new elastic approach in the case of multitasking applications characterized by a graph of dependencies between tasks
Santi, Nina. "Prédiction des besoins pour la gestion de serveurs mobiles en périphérie." Electronic Thesis or Diss., Université de Lille (2022-....), 2023. http://www.theses.fr/2023ULILB050.
Full textMulti-access Edge computing is an emerging paradigm within the Internet of Things (IoT) that complements Cloud computing. This paradigm proposes the implementation of computing servers located close to users, reducing the pressure and costs of local network infrastructure. This proximity to users is giving rise to new use cases, such as the deployment of mobile servers mounted on drones or robots, offering a cheaper, more energy-efficient and flexible alternative to fixed infrastructures for one-off or exceptional events. However, this approach also raises new challenges for the deployment and allocation of resources in terms of time and space, which are often battery-dependent.In this thesis, we propose predictive tools and algorithms for making decisions about the allocation of fixed and mobile resources, in terms of both time and space, within dynamic environments. We provide rich and reproducible datasets that reflect the heterogeneity inherent in Internet of Things (IoT) applications, while exhibiting a high rate of contention and interference. To achieve this, we are using the FIT-IoT Lab, an open testbed dedicated to the IoT, and we are making all the code available in an open manner. In addition, we have developed a tool for generating IoT traces in an automated and reproducible way. We use these datasets to train machine learning algorithms based on regression techniques to evaluate their ability to predict the throughput of IoT applications. In a similar approach, we have also trained and analysed a neural network of the temporal transformer type to predict several Quality of Service (QoS) metrics. In order to take into account the mobility of resources, we are generating IoT traces integrating mobile access points embedded in TurtleBot robots. These traces, which incorporate mobility, are used to validate and test a federated learning framework based on parsimonious temporal transformers. Finally, we propose a decentralised algorithm for predicting human population density by region, based on the use of a particle filter. We test and validate this algorithm using the Webots simulator in the context of servers embedded in robots, and the ns-3 simulator for the network part
Ntumba, wa Ntumba Patient. "Ordonnancement d'opérateurs continus pour l'analyse de flux de données à la périphérie de l'Internet des Objets." Electronic Thesis or Diss., Sorbonne université, 2022. http://www.theses.fr/2022SORUS183.
Full textData stream processing and analytics (DSPA) applications are widely used to process the ever increasing amounts of data streams produced by highly geographically distributed data sources, such as fixed and mobile IoT devices, in order to extract valuable information in a timely manner for actuation. DSPA applications are typically deployed in the Cloud to benefit from practically unlimited computational resources on demand. However, such centralized and distant computing solutions may suffer from limited network bandwidth and high network delay. Additionally, data propagation to the Cloud may compromise the privacy of sensitive data. To effectively handle this volume of data streams, the emerging Edge/Fog computing paradigm is used as the middle-tier between the Cloud and the IoT devices to process data streams closer to their sources and to reduce the network resource usage and network delay to reach the Cloud. However, Edge/Fog computing comes with limited computational resource capacities and requires deciding which part of the DSPA application should be performed in the Edge/Fog layers while satisfying the application response time constraint for timely actuation. Furthermore, the computational and network resources across the Edge-Fog-Cloud architecture can be shareable among multiple DSPA (and other) applications, which calls for efficient resource usage. In this PhD research, we propose a new model for assessing the usage cost of resources across the Edge-Fog-Cloud architecture. Our model addresses both computational and network resources and enables dealing with the trade-offs that are inherent to their joint usage. It precisely characterizes the usage cost of resources by distinguishing between abundant and constrained resources as well as by considering their dynamic availability, hence covering both resources dedicated to a single DSPA application and shareable resources. We complement our system modeling with a response time model for DSPA applications that takes into account their windowing characteristics. Leveraging these models, we formulate the problem of scheduling streaming operators over a hierarchical Edge-Fog-Cloud resource architecture. Our target problem presents two distinctive features. First, it aims at jointly optimizing the resource usage cost for computational and network resources, while few existing approaches have taken computational resources into account in their optimization goals. More precisely, our aim is to schedule a DSPA application in a way that it uses available resources in the most efficient manner. This enables saving valuable resources for other DSPA (and non DSPA) applications that share the same resource architecture. Second, it is subject to a response time constraint, while few works have dealt with such a constraint; most approaches for scheduling time-critical (DSPA) applications include the response time in their optimization goals. To solve our formulated problem, we introduce several heuristic algorithms that deal with different versions of the problem: static resource-aware scheduling that each time calculates a new system deployment from the outset, time-aware and resource-aware scheduling, dynamic scheduling that takes into account the current deployment. Finally, we extensively and comparatively evaluate our algorithms with realistic simulations against several baselines that either we introduce or that originate / are inspired from the existing literature. Our results demonstrate that our solutions advance the current state of the art in scheduling DSPA applications
Antony, Geo Johns. "Cheops reloaded, further steps in decoupling geo-distribution from application business logic : a focus on externalised sharding collaboration, consistency and dependency." Electronic Thesis or Diss., Ecole nationale supérieure Mines-Télécom Atlantique Bretagne Pays de la Loire, 2024. http://www.theses.fr/2024IMTA0441.
Full textThe shift from centralized cloud computing to geo-distributed applications is critical for meeting modern demands for low-latency, highly available, and resilient services. However, existing geo-distribution solutions often require intrusive modifications to application code. My thesis extends the Cheops framework, a middleware that decouples geodistribution from application logic, offering a non-intrusive and generic solution for deploying an application across geographically distributed instances. Building on the Cheops principles of "local-first" and "collaborative-then," my research introduces Cross, a shard collaboration mechanism for partitioning resources across sites, and a new approach to decoupling consistency from the application logic, ensuring synchronization between instances. Additionally, dependency management guarantees that operations performed on one instance are reproducible across geo-distributed instances, maintaining the illusion of a unified, singlesite application. Cheops uses Scope-lang, a Domain-Specific Language (DSL), to facilitate this without altering application logic. This extension of Cheops, further enhances the separation of geo-distribution from the application business logic
Aguiari, Davide. "Exploring Computing Continuum in IoT Systems : sensing, communicating and processing at the Network Edge." Electronic Thesis or Diss., Sorbonne université, 2021. http://www.theses.fr/2021SORUS131.
Full textAs Internet of Things (IoT), originally comprising of only a few simple sensing devices, reaches 34 billion units by the end of 2020, they cannot be defined as merely monitoring sensors anymore. IoT capabilities have been improved in recent years as relatively large internal computation and storage capacity are becoming a commodity. In the early days of IoT, processing and storage were typically performed in cloud. New IoT architectures are able to perform complex tasks directly on-device, thus enabling the concept of an extended computational continuum. Real-time critical scenarios e.g. autonomous vehicles sensing, area surveying or disaster rescue and recovery require all the actors involved to be coordinated and collaborate without human interaction to a common goal, sharing data and resources, even in intermittent networks covered areas. This poses new problems in distributed systems, resource management, device orchestration,as well as data processing. This work proposes a new orchestration and communication framework, namely CContinuum, designed to manage resources in heterogeneous IoT architectures across multiple application scenarios. This work focuses on two key sustainability macroscenarios: (a) environmental sensing and awareness, and (b) electric mobility support. In the first case a mechanism to measure air quality over a long period of time for different applications at global scale (3 continents 4 countries) is introduced. The system has been developed in-house from the sensor design to the mist-computing operations performed by the nodes. In the second scenario, a technique to transmit large amounts of fine-time granularity battery data from a moving vehicle to a control center is proposed jointly with the ability of allocating tasks on demand within the computing continuum
Yu, Shuai. "Multi-user computation offloading in mobile edge computing." Electronic Thesis or Diss., Sorbonne université, 2018. http://www.theses.fr/2018SORUS462.
Full textMobile Edge Computing (MEC) is an emerging computing model that extends the cloud and its services to the edge of the network. Consider the execution of emerging resource-intensive applications in MEC network, computation offloading is a proven successful paradigm for enabling resource-intensive applications on mobile devices. Moreover, in view of emerging mobile collaborative application (MCA), the offloaded tasks can be duplicated when multiple users are in the same proximity. This motivates us to design a collaborative computation offloading scheme for multi-user MEC network. In this context, we separately study the collaborative computation offloading schemes for the scenarios of MEC offloading, device-to-device (D2D) offloading and hybrid offloading, respectively. In the MEC offloading scenario, we assume that multiple mobile users offload duplicated computation tasks to the network edge servers, and share the computation results among them. Our goal is to develop the optimal fine-grained collaborative offloading strategies with caching enhancements to minimize the overall execution delay at the mobile terminal side. To this end, we propose an optimal offloading with caching-enhancement scheme (OOCS) for femto-cloud scenario and mobile edge computing scenario, respectively. Simulation results show that compared to six alternative solutions in literature, our single-user OOCS can reduce execution delay up to 42.83% and 33.28% for single-user femto-cloud and single-user mobile edge computing, respectively. On the other hand, our multi-user OOCS can further reduce 11.71% delay compared to single-user OOCS through users' cooperation. In the D2D offloading scenario, we assume that where duplicated computation tasks are processed on specific mobile users and computation results are shared through Device-to-Device (D2D) multicast channel. Our goal here is to find an optimal network partition for D2D multicast offloading, in order to minimize the overall energy consumption at the mobile terminal side. To this end, we first propose a D2D multicast-based computation offloading framework where the problem is modelled as a combinatorial optimization problem, and then solved using the concepts of from maximum weighted bipartite matching and coalitional game. Note that our proposal considers the delay constraint for each mobile user as well as the battery level to guarantee fairness. To gauge the effectiveness of our proposal, we simulate three typical interactive components. Simulation results show that our algorithm can significantly reduce the energy consumption, and guarantee the battery fairness among multiple users at the same time. We then extend the D2D offloading to hybrid offloading with social relationship consideration. In this context, we propose a hybrid multicast-based task execution framework for mobile edge computing, where a crowd of mobile devices at the network edge leverage network-assisted D2D collaboration for wireless distributed computing and outcome sharing. The framework is social-aware in order to build effective D2D links [...]
Jouni, Zalfa. "Analog spike-based neuromorphic computing for low-power smart IoT applications." Electronic Thesis or Diss., université Paris-Saclay, 2024. http://www.theses.fr/2024UPAST114.
Full textAs the Internet of Things (IoT) expands with more connected devices and complex communications, the demand for precise, energy-efficient localization technologies has intensified. Traditional machine learning and artificial intelligence (AI) techniques provide high accuracy in radio-frequency (RF) localization but often at the cost of greater complexity and power usage. To address these challenges, this thesis explores the potential of neuromorphic computing, inspired by brain functionality, to enable energy-efficient AI-based RF localization. It introduces an end-to-end analog spike-based neuromorphic system (RF NeuroAS), with a simplified version fully implemented in BiCMOS 55 nm technology. RF NeuroAS is designed to identify source positions within a 360-degree range on a two-dimensional plane, maintaining high resolution (10 or 1 degree) even in noisy conditions. The core of this system, an analog-based spiking neural network (A-SNN), was trained and tested on a simulated dataset (SimLocRF) from MATLAB and an experimental dataset (MeasLocRF) from anechoic chamber measurements, both developed in this thesis.The learning algorithms for A-SNN were developed through two approaches: software-based deep learning (DL) and bio-plausible spike-timing-dependent plasticity (STDP). RF NeuroAS achieves a localization accuracy of 97.1% with SimLocRF and 90.7% with MeasLoc at a 10-degree resolution, maintaining high performance with low power consumption in the nanowatt range. The simplified RF NeuroAS consumes just over 1.1 nW and operates within a 30 dB dynamic range. A-SNN learning, via DL and STDP, demonstrated performance on XOR and MNIST problems. DL depends on the non-linearity of post-layout transfer functions of A-SNN's neurons and synapses, while STDP depends on the random noise in analog neuron circuits. These findings highlight advancements in energy-efficient IoT through neuromorphic computing, promising low-power smart edge IoT breakthroughs inspired by brain mechanisms
Foroughi, Parisa. "Towards network automation : planning and monitoring." Electronic Thesis or Diss., Institut polytechnique de Paris, 2022. http://www.theses.fr/2022IPPAT038.
Full textNetwork management is undergoing drastic changes due to the high expectations of the infrastructure to support new services. The diverse requirements of these services, call for the integration of new enabler technologies that complicate the network monitoring and planning process. Therefore, to alleviate the burden and increase the monitoring and planning accuracy, more automated solutions on the element/device level are required. In this thesis, we propose a semi-automated framework called AI-driven telemetry (ADT) for collecting, processing, and assessing the state of routers using streaming telemetry data. ADT consists of 4 building blocks: collector, detector, explainer, and exporter. We concentrate on the detection block in ADT and propose a multi-variate online change detection technique called DESTIN. Our study on the explainer block of ADT is limited to exploring the potential of the input data and showcasing the possibility of the automated event description. Then, we tackle the problem of planning and dimensioning in radio access networks equipped with distributed edge servers. We propose a model that satisfies the service requirements and makes use of novel enabler technologies, i.e. network slicing and virtualization techniques. We showcase the advantages of using our holistic model to automate RAN planning by utilizing simulated annealing and greedy methods
De, Souza Felipe Rodrigo. "Scheduling Solutions for Data Stream Processing Applications on Cloud-Edge Infrastructure." Thesis, Lyon, 2020. http://www.theses.fr/2020LYSEN082.
Full textTechnology has evolved to a point where applications and devicesare highly connected and produce ever-increasing amounts of dataused by organizations and individuals to make daily decisions. Forthe collected data to become information that can be used indecision making, it requires processing. The speed at whichinformation is extracted from data generated by a monitored systemTechnology has evolved to a point where applications and devicesare highly connected and produce ever-increasing amounts of dataused by organizations and individuals to make daily decisions. Forthe collected data to become information that can be used indecision making, it requires processing. The speed at whichinformation is extracted from data generated by a monitored systemor environment affects how fast organizations and individuals canreact to changes. One way to process the data under short delays isthrough Data Stream Processing (DSP) applications. DSPapplications can be structured as directed graphs, where the vertexesare data sources, operators, and data sinks, and the edges arestreams of data that flow throughout the graph. A data source is anapplication component responsible for data ingestion. Operatorsreceive a data stream, apply some transformation or user-definedfunction over the data stream and produce a new output stream,until the latter reaches a data sink, where the data is stored,visualized or provided to another application
Rasse, Alban. "Une Approche Orientée Modèles pour la Spécification, la Vérification et l’Implantation des Systèmes Logiciels Critiques." Mulhouse, 2006. https://www.learning-center.uha.fr/opac/resource/une-approche-orientee-modeles-pour-la-specification-la-verification-et-limplantation-des-systemes-lo/BUS3944436.
Full textBooks on the topic "Informatique périphérie"
Hazzah, Karen. Writing Windows VxDs and device drivers. Lawrence, Kan: R & D Publications, 1995.
Find full textWriting Windows VxDs and Device Drivers. 2nd ed. Lawrence: R&D Books, 1997.
Find full text