Índice
Literatura académica sobre el tema "Informatique dans les nuages – Économies d'énergie"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte las listas temáticas de artículos, libros, tesis, actas de conferencias y otras fuentes académicas sobre el tema "Informatique dans les nuages – Économies d'énergie".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Tesis sobre el tema "Informatique dans les nuages – Économies d'énergie"
Borgetto, Damien. "Allocation et réallocation de services pour les économies d'énergie dans les clusters et les clouds". Toulouse 3, 2013. http://thesesups.ups-tlse.fr/2100/.
Texto completoCloud computing has become over the last years an important paradigm in the computing landscape. Its principle is to provide decentralized services and allows client to consume resources on a pay-as-you-go model. The increasing need for this type of service brings the service providers to increase the size of their infrastructures, to the extent that energy consumptions as well as operating costs are becoming important. Each cloud service provider has to provide for different types of requests. Infrastructure manager then have to host all the types of services together. That's why during this thesis, we tackled energy efficient resource management in the clouds. In order to do so, we first modeled and studied the initial service allocation problem, by computing approximated solutions given by heuristics, then comparing it to the optimal solution computed with a linear program solver. We then extended the model of resources to allow us to have a more global approach, by integrating the inherent heterogeneity of clusters and the cooling infrastructures. We then validated our model via simulation. Usually, the services must face different stages of workload, as well as utilization spikes. That's why we extended the model to include dynamicity of requests and resource usage, as well as the concept of powering on or off servers, or the cost of migrating a service from one host to another. We implemented a simulated cloud infrastructure, aiming at controlling the execution of the services as well as their placement. Thus, our approach enables the reduction of the global energy consumption of the infrastructure, and limits as much as possible degrading the performances
Jacquet, Pierre. "Enhancing IaaS Consolidation with resource oversubscription". Electronic Thesis or Diss., Université de Lille (2022-....), 2024. http://www.theses.fr/2024ULILB015.
Texto completoBy migrating its workload to larger Data Centers (DCs), the digital realm has been able to improve its energy efficiency. The consumption due to the increase in usage has thus been mitigated by significant improvements in shared infrastructure (commonly referred to as Cloud Computing), which is evident through indicators such as Power Usage Efficiency (PUE). However, infrastructure is not the sole point of optimization. The server itself, and the tasks it executes, remain an important focus of research. Usage rate, in particular, is closely studied because its relatively low value also represents a considerable potential gain. Thus, from both an energy (consumption) and material (environmental and financial cost) standpoint, the use of a server loaded at 100% is preferable to that of 3 servers loaded at 30%. I propose to examine these usage rates along four complementary contributions:1. The creation of realistic controlled experiments in an Infrastructure-as-a-Service (IAAS) context. While platforms supporting Cloud infrastructures are extensively studied, generating realistic workloads is crucial. As each Cloud Provider has its characteristics (distribution of VM sizes, individual usage rates), we propose a tool to generate these workloads.2. The calculation of individual server oversubscription ratio. By considering the individual stability of servers, it is possible to fine-tune the calculation of this ratio without causing additional violations.3. The introduction of a new oversubscription paradigm. By first demonstrating that Virtual Machine (VM) vCPUs are not uniformly used in a real-world context, we expose to VMs cores of different powers (by oversubscribing them to different amounts) and demonstrate that this paradigm can improve overall performance.4. The complementarity of oversubscription techniques to reduce unallocated resources. Comparing so-called premium VMs and oversubscribed VMs identifies that they tend to saturate their hosts' resources differently. By hosting them on the same servers, it is thus possible to benefit from synergies and reduce the number of servers by up to 9.6%
Ghribi, Chaima. "Energy efficient resource allocation in cloud computing environments". Thesis, Evry, Institut national des télécommunications, 2014. http://www.theses.fr/2014TELE0035/document.
Texto completoCloud computing has rapidly emerged as a successful paradigm for providing IT infrastructure, resources and services on a pay-per-use basis over the past few years. As, the wider adoption of Cloud and virtualization technologies has led to the establishment of large scale data centers that consume excessive energy and have significant carbon footprints, energy efficiency is becoming increasingly important for data centers and Cloud. Today data centers energy consumption represents 3 percent of all global electricity production and is estimated to further rise in the future. This thesis presents new models and algorithms for energy efficient resource allocation in Cloud data centers. The first goal of this work is to propose, develop and evaluate optimization algorithms of resource allocation for traditional Infrastructutre as a Service (IaaS) architectures. The approach is Virtual Machine (VM) based and enables on-demand and dynamic resource scheduling while reducing power consumption of the data center. This initial objective is extended to deal with the new trends in Cloud services through a new model and optimization algorithms of energy efficient resource allocation for hybrid IaaS-PaaS Cloud providers. The solution is generic enough to support different type of virtualization technologies, enables both on-demand and advanced resource provisioning to deal with dynamic resource scheduling and fill the gap between IaaS and PaaS services and create a single continuum of services for Cloud users. Consequently, in the thesis, we first present a survey of the state of the art on energy efficient resource allocation in cloud environments. Next, we propose a bin packing based approach for energy efficient resource allocation for classical IaaS. We formulate the problem of energy efficient resource allocation as a bin-packing model and propose an exact energy aware algorithm based on integer linear program (ILP) for initial resource allocation. To deal with dynamic resource consolidation, an exact ILP algorithm for dynamic VM reallocation is also proposed. This algorithm is based on VM migration and aims at constantly optimizing energy efficiency at service departures. A heuristic method based on the best-fit algorithm has also been adapted to the problem. Finally, we present a graph-coloring based approach for energy efficient resource allocation in the hybrid IaaS-PaaS providers context. This approach relies on a new graph coloring based model that supports both VM and container virtualization and provides on-demand as well as advanced resource reservation. We propose and develop an exact Pre-coloring algorithm for initial/static resource allocation while maximizing energy efficiency. A heuristic Pre-coloring algorithm for initial resource allocation is also proposed to scale with problem size. To adapt reservations over time and improve further energy efficiency, we introduce two heuristic Re-coloring algorithms for dynamic resource reallocation. Our solutions are generic, robust and flexible and the experimental evaluation shows that both proposed approaches lead to significant energy savings while meeting the users' requirements
Cuadrado-Cordero, Ismael. "Microclouds : an approach for a network-aware energy-efficient decentralised cloud". Thesis, Rennes 1, 2017. http://www.theses.fr/2017REN1S003/document.
Texto completoThe current datacenter-centralized architecture limits the cloud to the location of the datacenters, generally far from the user. This architecture collides with the latest trend of ubiquity of Cloud computing. Also, current estimated energy usage of data centers and core networks adds up to 3% of the global energy production, while according to latest estimations only 42,3% of the population is connected. In the current work, we focused on two drawbacks of datacenter-centralized Clouds: Energy consumption and poor quality of service. On the one hand, due to its centralized nature, energy consumption in networks is affected by the centralized vision of the Cloud. That is, backbone networks increase their energy consumption in order to connect the clients to the datacenters. On the other hand, distance leads to increased utilization of the broadband Wide Area Network and poor user experience, especially for interactive applications. A distributed approach can provide a better Quality of Experience (QoE) in large urban populations in mobile cloud networks. To do so, the cloud should confine local traffic close to the user, running on the users and network devices. In this work, we propose a novel distributed cloud architecture based on microclouds. Microclouds are dynamically created and allow users to contribute resources from their computers, mobile and network devices to the cloud. This way, they provide a dynamic and scalable system without the need of an extra investment in infrastructure. We also provide a description of a realistic mobile cloud use case, and the adaptation of microclouds on it. Through simulations, we show an overall saving up to 75% of energy consumed in standard centralized clouds with our approach. Also, our results indicate that this architecture is scalable with the number of mobile devices and provide a significantly lower latency than regular datacenter-centralized approaches. Finally, we analyze the use of incentives for Mobile Clouds, and propose a new auction system adapted to the high dynamism and heterogeneity of these systems. We compare our solution to other existing auctions systems in a Mobile Cloud use case, and show the suitability of our solution
Villebonnet, Violaine. "Scheduling and Dynamic Provisioning for Energy Proportional Heterogeneous Infrastructures". Thesis, Lyon, 2016. http://www.theses.fr/2016LYSEN057/document.
Texto completoThe increasing number of data centers raises serious concerns regarding their energy consumption. These infrastructures are often over-provisioned and contain servers that are not fully utilized. The problem is that inactive servers can consume as high as 50% of their peak power consumption.This thesis proposes a novel approach for building data centers so that their energy consumption is proportional to the actual load. We propose an original infrastructure named BML for "Big, Medium, Little", composed of heterogeneous computing resources : from low power processors to classical servers. The idea is to take advantage of their different characteristics in terms of energy consumption, performance, and switch on reactivity to adjust the composition of the infrastructure according to the load evolutions. We define a generic methodology to compute the most energy proportional combinations of machines based on hardware profiling data.We focus on web applications whose load varies over time and design a scheduler that dynamically reconfigures the infrastructure, with application migrations and machines switch on and off, to minimize the infrastructure energy consumption according to the current application requirements.We have developed two different dynamic provisioning algorithms which take into account the time and energy overheads of the different reconfiguration actions in the decision process. We demonstrate through simulations based on experimentally acquired hardware profiles that we achieve important energy savings compared to classical data center infrastructures and management
Ghribi, Chaima. "Energy efficient resource allocation in cloud computing environments". Electronic Thesis or Diss., Evry, Institut national des télécommunications, 2014. http://www.theses.fr/2014TELE0035.
Texto completoCloud computing has rapidly emerged as a successful paradigm for providing IT infrastructure, resources and services on a pay-per-use basis over the past few years. As, the wider adoption of Cloud and virtualization technologies has led to the establishment of large scale data centers that consume excessive energy and have significant carbon footprints, energy efficiency is becoming increasingly important for data centers and Cloud. Today data centers energy consumption represents 3 percent of all global electricity production and is estimated to further rise in the future. This thesis presents new models and algorithms for energy efficient resource allocation in Cloud data centers. The first goal of this work is to propose, develop and evaluate optimization algorithms of resource allocation for traditional Infrastructutre as a Service (IaaS) architectures. The approach is Virtual Machine (VM) based and enables on-demand and dynamic resource scheduling while reducing power consumption of the data center. This initial objective is extended to deal with the new trends in Cloud services through a new model and optimization algorithms of energy efficient resource allocation for hybrid IaaS-PaaS Cloud providers. The solution is generic enough to support different type of virtualization technologies, enables both on-demand and advanced resource provisioning to deal with dynamic resource scheduling and fill the gap between IaaS and PaaS services and create a single continuum of services for Cloud users. Consequently, in the thesis, we first present a survey of the state of the art on energy efficient resource allocation in cloud environments. Next, we propose a bin packing based approach for energy efficient resource allocation for classical IaaS. We formulate the problem of energy efficient resource allocation as a bin-packing model and propose an exact energy aware algorithm based on integer linear program (ILP) for initial resource allocation. To deal with dynamic resource consolidation, an exact ILP algorithm for dynamic VM reallocation is also proposed. This algorithm is based on VM migration and aims at constantly optimizing energy efficiency at service departures. A heuristic method based on the best-fit algorithm has also been adapted to the problem. Finally, we present a graph-coloring based approach for energy efficient resource allocation in the hybrid IaaS-PaaS providers context. This approach relies on a new graph coloring based model that supports both VM and container virtualization and provides on-demand as well as advanced resource reservation. We propose and develop an exact Pre-coloring algorithm for initial/static resource allocation while maximizing energy efficiency. A heuristic Pre-coloring algorithm for initial resource allocation is also proposed to scale with problem size. To adapt reservations over time and improve further energy efficiency, we introduce two heuristic Re-coloring algorithms for dynamic resource reallocation. Our solutions are generic, robust and flexible and the experimental evaluation shows that both proposed approaches lead to significant energy savings while meeting the users' requirements
Gbaguidi, Fréjus A. Roméo. "Approche prédictive de l'efficacité énergétique dans les Clouds Datacenters". Electronic Thesis or Diss., Paris, CNAM, 2017. http://www.theses.fr/2017CNAM1163.
Texto completoWith the democratization of digital technologies, the construction of a globalized cyberspace insidiously transforms our lifestyle. Connect more than 4 billion people at high speed, requires the invention of new concept of service provision and trafic management that are capable to face the challenges. For that purpose, Cloud Computing have been set up to enable Datacenters to provide part or total IT components needed by companies for timely services delivering with performance that meets the requirements of their clients. Consequently, the proliferation of Datacenters around the world has brought to light the worrying question about the amount of energy needed for their function and the resulting difficulty for the humanity, whose current reserves are not extensible indefinitely. It was therefore necessary to develop techniques that reduce the power consumption of Datacenters by minimizing the energy losses orchestrated on servers where each wasted watt results in a chain effect on a substantial increase in the overall bill of Datacenters. Our work consisted first in making a review of the literature on the subject and then testing the ability of some prediction tools to improve the anticipation of the risks of energy loss caused by the misallocation of virtual equipment on servers. This study focused particularly on the ARMA tools and neural networks which in the literature have produced interesting results in related fields. After this step, it appeared to us that ARMA tools, although having less performance than neural networks in our context, runs faster and are best suited to be implemented in cloud computing environments. Thus, we used the results of this method to improve the decision-making process, notably for the proactive re-allocation of virtual equipment before it leads to under-consumption of resources on physical servers or over-consumption inducing breaches of SLAs. Based on our simulations, this approach enabled us to reduce energy consumption on a firm of 800 servers over a period of one day by more than 5Kwh. This gain could be significant when considering the enormous size of modern data centers and projected over a relatively long period of time. It would be even more interesting to deepen this research in order to generalize the integration of this predictive approach into existing techniques in order to significantly optimize the energy consumption within Datacenters while preserving performance and quality of service which are key requirements in the concept of Cloud Computing
Haderer, Nicolas. "APISENSE® : une plate-forme répartie pour la conception, le déploiement et l’exécution de campagnes de collecte de données sur des terminaux intelligents". Thesis, Lille 1, 2014. http://www.theses.fr/2014LIL10118/document.
Texto completoMobile crowdsensing is a new form of data collection that takes advantage of millions smart devices already deployed throughout the world to collect massively environmental or behavioral data from a population. Recently, this type of data collection has attracted interest from a large number of industrials and academic players in many areas, such as the study of urban mobility, environmental monitoring, health or the study of sociocultural attitudes. However, mobile crowdsensing is in its early stages of development, and many challenges remain to be addressed to take full advantage of its potential. These challenges include privacy, limited energy resources of devices, development of reward and recruitment models to select appropriates mobile users and dealing with heterogeneity of mobile platforms available. In this thesis, we aim to reconsider the architectural design of current mobile crowdsensing systems to provide a simple and effective way to design, deploy and manage data collection campaigns.The main contributions of this thesis are organize around APISENSE, the resulting platform of this research. APISENSE has been used to carry out a data collection campaign deployed over hundred of users in a sociological study and evaluated through experiments demonstrating the validity, effectiveness and scalability of our solution
Gbaguidi, Fréjus A. Roméo. "Approche prédictive de l'efficacité énergétique dans les Clouds Datacenters". Thesis, Paris, CNAM, 2017. http://www.theses.fr/2017CNAM1163/document.
Texto completoWith the democratization of digital technologies, the construction of a globalized cyberspace insidiously transforms our lifestyle. Connect more than 4 billion people at high speed, requires the invention of new concept of service provision and trafic management that are capable to face the challenges. For that purpose, Cloud Computing have been set up to enable Datacenters to provide part or total IT components needed by companies for timely services delivering with performance that meets the requirements of their clients. Consequently, the proliferation of Datacenters around the world has brought to light the worrying question about the amount of energy needed for their function and the resulting difficulty for the humanity, whose current reserves are not extensible indefinitely. It was therefore necessary to develop techniques that reduce the power consumption of Datacenters by minimizing the energy losses orchestrated on servers where each wasted watt results in a chain effect on a substantial increase in the overall bill of Datacenters. Our work consisted first in making a review of the literature on the subject and then testing the ability of some prediction tools to improve the anticipation of the risks of energy loss caused by the misallocation of virtual equipment on servers. This study focused particularly on the ARMA tools and neural networks which in the literature have produced interesting results in related fields. After this step, it appeared to us that ARMA tools, although having less performance than neural networks in our context, runs faster and are best suited to be implemented in cloud computing environments. Thus, we used the results of this method to improve the decision-making process, notably for the proactive re-allocation of virtual equipment before it leads to under-consumption of resources on physical servers or over-consumption inducing breaches of SLAs. Based on our simulations, this approach enabled us to reduce energy consumption on a firm of 800 servers over a period of one day by more than 5Kwh. This gain could be significant when considering the enormous size of modern data centers and projected over a relatively long period of time. It would be even more interesting to deepen this research in order to generalize the integration of this predictive approach into existing techniques in order to significantly optimize the energy consumption within Datacenters while preserving performance and quality of service which are key requirements in the concept of Cloud Computing
Politaki, Dimitra. "Vers la modélisation de clusters de centres de données vertes". Thesis, Université Côte d'Azur (ComUE), 2019. http://www.theses.fr/2019AZUR4116.
Texto completoData center clusters energy consumption is rapidly increasing making them the fastest-growing consumers of electricity worldwide. Renewable electricity sources and especially solar energy as a clean and abundant energy can be used, in many locations, to cover their electricity needs and make them "green" namely fed by photovoltaics. This potential can be explored by predicting solar irradiance and assessing the capacity provision for data center clusters. In this thesis we develop stochastic models for solar energy; one at the surface of the Earth and a second one which models the photovoltaic output current. We then compare them to the state of the art on-off model and validate them against real data. We conclude that the solar irradiance model can better capture the multiscales correlations and is suitable for small scale cases. We then propose a new job life-cycle of a complex and real cluster system and a model for data center clusters that supports batch job submissions and cons iders both impatient and persistent customer behavior. To understand the essential computer cluster characteristics, we analyze in detail two different workload type traces; the first one is the published complex Google trace and the second, simpler one, which serves scientific purposes, is from the Nef cluster located at the research center Inria Sophia Antipolis. We then implement the marmoteCore-Q, a tool for the simulation of a family of queueing models based on our multi-server model for data center clusters with abandonments and resubmissions