Dissertations / Theses on the topic 'Cloud infrastructure'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Cloud infrastructure.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Ha, Kiryong. "System Infrastructure for Mobile-Cloud Convergence." Research Showcase @ CMU, 2016. http://repository.cmu.edu/dissertations/704.
Full textBaig, Viñas Roger. "Development and management of collective network and cloud computing infrastructures." Doctoral thesis, Universitat Politècnica de Catalunya, 2019. http://hdl.handle.net/10803/667952.
Full textEn la recerca i desenvolupament de models més participatius per al desenvolupament i gestió d'infraestructura, en aquesta tesi investiguem sobre models per al finançament, desplegament i operació d'infraestructures de xarxa i de computació al núvol. La nostra preocupació principal és fer front a l’exclusió inherent dels models dominants actualment pel que fa a la participació en els processos de desenvolupament i gestió i, també, als drets d’us. El nostre treball comença amb un estudi detallat del model de Guifi.net, un cas d'èxit d'iniciativa ciutadana en la construcció d'infraestructura de xarxa, iniciatives que es coneixen com a xarxes comunitàries. En fer-ho, parem una atenció especial al sistema de governança i a l’organització econòmica perquè pensem que són els dos elements claus de l'èxit d'aquesta iniciativa. Tot seguit passem a analitzar d'altres xarxes comunitàries per abundar en la comprensió dels factors determinants per a la seva sostenibilitat i escalabilitat. Després ampliem el nostre estudi analitzant la capacitat i el comportament del model de Guifi.net en el camp de les infraestructures de computació al núvol. A resultes d'aquests estudis, proposem l'atribut extensible per a descriure aquelles infraestructures que són relativament fàcil d'ampliar i gestionar, en contraposició a les que o bé estan limitades de forma natural o be són difícils d'ampliar, com ara els recursos naturals o els sistemes artificials avançats o complexos. Finalitzem aquest treball fent una proposta de model genèric que pensem que és d'aplicabilitat, com a mínim, a tot tipus d'infraestructura extensible. El model de Guifi.net està fortament vinculat als bens comuns. És per això que la recerca en aquest àmbit, en general, i els treballs de Elinor Ostrom en particular, han deixat una forta empremta en el nostre treball. Els resultats que hem obtingut mostren que el model Guifi.net s'ajusta molt bé als principis que segons Ostrom han de complir els béns comuns per ser sostenibles. Aquest treball s'ha desenvolupat com a doctorat industrial. Com a tal, combina la investigació acadèmica amb elements de practica i persegueix una transferència efectiva de coneixement entre l'àmbit acadèmic i el sector privat. Ates que el soci del sector privat és una organització sense ànim de lucre, l’esforç per crear valor social ha prevalgut en l’ambició d’avançar en el desenvolupament d'un producte industrial específic o d'una tecnologia particular
Suriano, Francesco. "Cloud Computing: From Containerization to Infrastructure Provisioning." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2019. http://amslaurea.unibo.it/18504/.
Full textTran, Khanh-Toan. "Efficient complex service deployment in cloud infrastructure." Thesis, Evry-Val d'Essonne, 2013. http://www.theses.fr/2012EVRY0038/document.
Full textThe purpose of the work in this thesis is to provide the Service Provider a solution which is capable of deploying complex services in a cloud automatically and cost-effectively. The first contribution allows the Service Provider to construct complex services requested by the clients from basic services at his disposal. The construction must be efficient in terms of execution time and operation cost while respecting the client’s constraints. We present an analytic model for this problem and propose a heuristic solution which performance is 20-30% better than other approaches. The second contribution solves the problem of deploying the services while considering end-users’ demands. To ensure the quality of services provided to end-users, not only one instance but a set of service replicas is deployed in the network. How the service is duplicated and distributed depends on the demands of its end-users that change constantly in quantity as well as distribution, which complicates the problem. Thus the provisioning mechanism has to be capable of adapt to the change in the network, including the change in end-users’ demands. Our third contribution is a system using OpenStack which allows Service Provider to deploy complex services in a cloud that spans over different locations (multi-site cloud). Given a client’s request, the system automatically calculates the optimal provisioning plan and deploys it while respecting the client’s constraints
Su, Xueyuan. "Efficient Fault-Tolerant Infrastructure for Cloud Computing." Thesis, Yale University, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=3578459.
Full textCloud computing is playing a vital role for processing big data. The infrastructure is built on top of large-scale clusters of commodity machines. It is very challenging to properly manage the hardware resources in order to utilize them effectively and to cope with the inevitable failures that will arise with such a large collection of hardware. In this thesis, task assignment and checkpoint placement for cloud computing infrastructure are studied.
As data locality is critical in determining the cost of running a task on a specific machine, how tasks are assigned to machines has a big impact on job completion time. An idealized abstract model is presented for a popular cloud computing platform called Hadoop. Although Hadoop task assignment (HTA) is [special characters omitted]-hard, an algorithm is presented with only an additive approximation gap. Connection is established between the HTA problem and the minimum makespan scheduling problem under the restricted assignment model. A new competitive ratio bound for the online GREEDY algorithm is obtained.
Checkpoints allow recovery of long-running jobs from failures. Checkpoints themselves might fail. The effect of checkpoint failures on job completion time is investigated. The sum of task success probability and checkpoint reliability greatly affects job completion time. When possible checkpoint placements are constrained, retaining only the most recent Ω(log n) possible checkpoints has at most a constant factor penalty. When task failures follow the Poisson distribution, two symmetries for non-equidistant placements are proved and a first order approximation to optimum placement interval is generalized.
Pech, David. "Cloud Framework on Infrastructure as a Service." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2013. http://www.nusl.cz/ntk/nusl-236185.
Full textFehse, Carsten. "Infrastructure suitability assessment modeling for cloud computing solutions." Thesis, Monterey, California. Naval Postgraduate School, 2011. http://hdl.handle.net/10945/5580.
Full textMaturing virtualization in information technology systems has enabled increased implementations of the cloud com-puting paradigm, dissolving the need to co-locate user and computing power by providing desired services through the network. This thesis researches the support that current network modeling and simulation applications can provide to IT projects in planning, implementing and maintaining networks for cloud solutions. A problem-appropriate do-main model and subsequent requirements are developed for the assessment of several network modeling and simula-tion tools, which leads to the identification of a capability gap precluding the use of such tools in early stages of cloud computing projects. Consequently, a practical, modular designed methodology is proposed to measure the essential properties necessary for developing appropriate cloud computing network traffic models. The conducted proof-of-concept experiment applied to a virtual desktop environment finds the proposed methodology suitable and problem-appropriate, and results in recommended steps to close the identified capability gap.
Younis, Y. A. "Securing access to cloud computing for critical infrastructure." Thesis, Liverpool John Moores University, 2015. http://researchonline.ljmu.ac.uk/4453/.
Full textMechtri, Marouen. "Virtual networked infrastructure provisioning in distributed cloud environments." Thesis, Evry, Institut national des télécommunications, 2014. http://www.theses.fr/2014TELE0028/document.
Full textCloud computing emerged as a new paradigm for on-demand provisioning of IT resources and for infrastructure externalization and is rapidly and fundamentally revolutionizing the way IT is delivered and managed. The resulting incremental Cloud adoption is fostering to some extent cloud providers cooperation and increasing the needs of tenants and the complexity of their demands. Tenants need to network their distributed and geographically spread cloud resources and services. They also want to easily accomplish their deployments and instantiations across heterogeneous cloud platforms. Traditional cloud providers focus on compute resources provisioning and offer mostly virtual machines to tenants and cloud services consumers who actually expect full-fledged (complete) networking of their virtual and dedicated resources. They not only want to control and manage their applications but also control connectivity to easily deploy complex network functions and services in their dedicated virtual infrastructures. The needs of users are thus growing beyond the simple provisioning of virtual machines to the acquisition of complex, flexible, elastic and intelligent virtual resources and services. The goal of this thesis is to enable the provisioning and instantiation of this type of more complex resources while empowering tenants with control and management capabilities and to enable the convergence of cloud and network services. To reach these goals, the thesis proposes mapping algorithms for optimized in-data center and in-network resources hosting according to the tenants' virtual infrastructures requests. In parallel to the apparition of cloud services, traditional networks are being extended and enhanced with software networks relying on the virtualization of network resources and functions especially through network resources and functions virtualization. Software Defined Networks are especially relevant as they decouple network control and data forwarding and provide the needed network programmability and system and network management capabilities. In such a context, the first part proposes optimal (exact) and heuristic placement algorithms to find the best mapping between the tenants' requests and the hosting infrastructures while respecting the objectives expressed in the demands. This includes localization constraints to place some of the virtual resources and services in the same host and to distribute other resources in distinct hosts. The proposed algorithms achieve simultaneous node (host) and link (connection) mappings. A heuristic algorithm is proposed to address the poor scalability and high complexity of the exact solution(s). The heuristic scales much better and is several orders of magnitude more efficient in terms of convergence time towards near optimal and optimal solutions. This is achieved by reducing complexity of the mapping process using topological patterns to map virtual graph requests to physical graphs representing respectively the tenants' requests and the providers' physical infrastructures. The proposed approach relies on graph decomposition into topology patterns and bipartite graphs matching techniques. The third part propose an open source Cloud Networking framework to achieve cloud and network resources provisioning and instantiation in order to respectively host and activate the tenants' virtual resources and services. This framework enables and facilitates dynamic networking of distributed cloud services and applications. This solution relies on a Cloud Network Gateway Manager and gateways to establish dynamic connectivity between cloud and network resources. The CNG-Manager provides the application networking control and supports the deployment of the needed underlying network functions in the tenant desired infrastructure (or slice since the physical infrastructure is shared by multiple tenants with each tenant receiving a dedicated and isolated portion/share of the physical resources)
Liberman, García Ari. "The evolution of the Cloud : the work, progress and outlook of cloud infrastructure." Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/100311.
Full textThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 59-61).
Cloud computing has enabled the deployment of systems at scale without requiring deep expertise in infrastructure management or highly specialized personnel. In just a few years, cloud computing has become one of the fastest growing technology segments in the Information Technology industry and it has transformed how applications are created and how companies they manage their growth. The cloud market has quickly become one of the most competitive industries with companies committing their efforts to the creation of cloud platforms and aggressive pricing strategies in an attempt to gain market dominance. This work shows the origins of the Infrastructure-as-a-Service industry and an analysis of the market dynamics by looking at the portfolios and strategies of the top competitors in this space. Also, this report shows what are the developments that will drive the innovation in the cloud industry years to come.
by Ari Liberman García.
S.M. in Engineering and Management
AlJabban, Tarek. "Distributed database storage management for a cloud computing infrastructure." Thesis, McGill University, 2013. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=114556.
Full textLes applications Internet ont récemment connu une croissance considérable en termes de taille et de complexité. Afin de satisfaire la forte demande pour les ressources informatiques et les espaces de stockage, les technologies en distribution ont commencé à devenir plus impliquées dans les applications à grande échelle. Le Cloud Computing est l'une de ces nombreuses technologies qui ont émergé pour aider à atteindre les objectifs de ces applications, telles que la haute disponibilité, les performances et l'évolutivité.Platform as a Service (PaaS) est un type de service qui peut être fourni par les solutions de Cloud Computing. Ces systèmes suivent souvent une architecture multi-niveaux qui se compose principalement d'un niveau de présentation, un niveau d'application et d'un niveau de base de données. Les volumes de données échangées entre l'application et la base de données deviennent énormes en particulier pour les applications de niveau entreprise. En conséquence, la conception de la base de données dans les systèmes de Cloud Computing doit prendre en compte le challenge de l'évolution des quantités énormes de données. Dans cette mémoire, nous proposons une approche de distribution des données qui peuvent être utilisées pour améliorer l'évolutivité des bases de données. Nous proposons deux techniques qui peuvent être appliquées à un serveur de base de données unique traditionnelle.Ces techniques fonctionnent en remplaçant le paradigme traditionnel utilisant une seule machine de stockage avec un paradigme de stockage distribué. Les techniques proposées maintiennent les caractéristiques qui existaient à l'origine dans le système de base de données, et en plus fournissent les caractéristiques de la distribution et de la réplication. Ces deux fonctionnalités supplémentaires aident à améliorer le système de tolérance aux pannes, car ils diminuent la possibilité d'avoir une défaillance au niveau du serveur de base de données. La distribution du stockage permet de résoudre les problèmes de performances spécifiques, tels que la réduction de l'utilisation des entrées/sorties et consécutivement de diminuer la possibilité de saturation des entrées/sorties.Par ailleurs, cela produit d'autres défis de performances qui doivent être pris en compte. Pour prouver la faisabilité de nos techniques, nous les avons implémentées comme des extensions du module de gestion de stockage de la base de données PostgreSQL.
Talevi, Iacopo. "Big Data Analytics and Application Deployment on Cloud Infrastructure." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2017. http://amslaurea.unibo.it/14408/.
Full textWang, Tony. "A Service for Provisioning Compute Infrastructure in the Cloud." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-393349.
Full textZhang, Bo. "Self-optimization of infrastructure and platform resources in cloud computing." Thesis, Lille 1, 2016. http://www.theses.fr/2016LIL10207/document.
Full textElasticity is considered as an important solution to handle the performance issues in scalable distributed system. However, most of the researches of elasticity only concern the provisioning and de-provisioning resources in automatic ways, but always ignore the resource utilization of provisioned resources. This might lead to resource leaks while provisioning redundant resources, thereby causing unnecessary expenditure. To avoid the resource leaks and redundant resources, my research therefore focus on how to maximize resource utilization by self resource management. In this thesis, relevant to diverse problems of resource usage and allocation in different layers, I propose two resource management approaches corresponding to infrastructure and platform, respectively. To overcome infrastructure limitation, I propose CloudGC as middleware service which aims to free occupied resources by recycling idle VMs. In platform-layer, a self-balancing approach is introduced to adjust Hadoop configuration at runtime, thereby avoiding memory loss and dynamically optimizing Hadoop performance. Finally, this thesis concerns rapid deployment of service which is also an issue of elasticity. A new tool, named "hadoop-benchmark", applies docker to accelerate the installation of Hadoop cluster and to provide a set of docker images which contain several well-known Hadoop benchmarks.The assessments show that these approaches and tool can well achieve resource management and self-optimization in various layers, and then facilitate the elasticity of infrastructure and platform in scalable platform, such as Cloud computing
Wicaksana, Arief. "Infrastructure portable pour un système hétérogène reconfigurable dans un environnement de cloud-FPGA." Thesis, Université Grenoble Alpes (ComUE), 2018. http://www.theses.fr/2018GREAT088/document.
Full textField-Programmable Gate Arrays (FPGAs) have been gaining popularity as hardware accelerators in heterogeneous architectures thanks to their high performance and low energy consumption. This argument has been supported by the recent integration of FPGA devices in cloud services and data centers. The potential offered by the reconfigurable architectures can still be optimized by treating FPGAs as virtualizable resources and offering them multitasking capability. The solution to preempt a hardware task on an FPGA with the objective of context switching it has been in research for many years. The previous works mainly proposed the strategy to extract the context of a running task from the FPGA to provide the possibility of its resumption at a later time. The communication during the process, on the contrary, has not been receiving much attention.In this work, we study the communication management of a hardware task whileit is being context switched. This communication management is necessary to ensure the consistency in the communication of a task with context switch capability in a reconfigurable system. Otherwise, a hardware context switch can only be allowed under restrictive constraints which may lead to a considerable penalty in performance; context switching a task is possible after the communication flows finish and the input/output data have been consumed. Furthermore, certain techniques demand homogeneity in the platform for a hardware context switch can take place.We present a mechanism which preserves the communication consistency during ahardware context switch in a reconfigurable architecture. The input/output communication data are managed together with the task context to ensure their integrity. The overall management of the hardware task context and communication data follows a dedicated protocol developed for heterogeneous reconfigurable architectures. This protocol thus allows a hardware context switch to take place while the task still has ongoing communication flows on Reconfigurable System-on-Chips (RSoCs). From the experiments, we discover that the overhead due to managing the communication data becomes negligible since our mechanism provides the necessary high responsiveness for preemptive scheduling, besides the consistency in communication. Finally, the applications of the proposed solution are presented in a task migration prototyping and in a hypervisor-based system
Okonski, Aleksander. "Implementing Security Rules, Safeguards, and IPS tools for Private Cloud Infrastructures : GROOT: Infrastructure Security as a Service(ISaaS)." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-345658.
Full textBianchi, Stefano. "Design and Implementation of a Cloud Infrastructure for Distributed Scientific Calculation." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2016.
Find full textCroubois, Hadrien. "Toward an autonomic engine for scientific workflows and elastic Cloud infrastructure." Thesis, Lyon, 2018. http://www.theses.fr/2018LYSEN061/document.
Full textThe constant development of scientific and industrial computation infrastructures requires the concurrent development of scheduling and deployment mechanisms to manage such infrastructures. Throughout the last decade, the emergence of the Cloud paradigm raised many hopes, but achieving full platformautonomicity is still an ongoing challenge. Work undertaken during this PhD aimed at building a workflow engine that integrated the logic needed to manage workflow execution and Cloud deployment on its own. More precisely, we focus on Cloud solutions with a dedicated Data as a Service (DaaS) data management component. Our objective was to automate the execution of workflows submitted by many users on elastic Cloud resources.This contribution proposes a modular middleware infrastructure and details the implementation of the underlying modules:• A workflow clustering algorithm that optimises data locality in the context of DaaS-centeredcommunications;• A dynamic scheduler that executes clustered workflows on Cloud resources;• A deployment manager that handles the allocation and deallocation of Cloud resources accordingto the workload characteristics and users’ requirements. All these modules have been implemented in a simulator to analyse their behaviour and measure their effectiveness when running both synthetic and real scientific workflows. We also implemented these modules in the Diet middleware to give it new features and prove the versatility of this approach.Simulation running the WASABI workflow (waves analysis based inference, a framework for the reconstruction of gene regulatory networks) showed that our approach can decrease the deployment cost byup to 44% while meeting the required deadlines
De, Souza Felipe Rodrigo. "Scheduling Solutions for Data Stream Processing Applications on Cloud-Edge Infrastructure." Thesis, Lyon, 2020. http://www.theses.fr/2020LYSEN082.
Full textTechnology has evolved to a point where applications and devicesare highly connected and produce ever-increasing amounts of dataused by organizations and individuals to make daily decisions. Forthe collected data to become information that can be used indecision making, it requires processing. The speed at whichinformation is extracted from data generated by a monitored systemTechnology has evolved to a point where applications and devicesare highly connected and produce ever-increasing amounts of dataused by organizations and individuals to make daily decisions. Forthe collected data to become information that can be used indecision making, it requires processing. The speed at whichinformation is extracted from data generated by a monitored systemor environment affects how fast organizations and individuals canreact to changes. One way to process the data under short delays isthrough Data Stream Processing (DSP) applications. DSPapplications can be structured as directed graphs, where the vertexesare data sources, operators, and data sinks, and the edges arestreams of data that flow throughout the graph. A data source is anapplication component responsible for data ingestion. Operatorsreceive a data stream, apply some transformation or user-definedfunction over the data stream and produce a new output stream,until the latter reaches a data sink, where the data is stored,visualized or provided to another application
Soares, João Monteiro. "Integration of the cloud computing paradigm with the opeerator network’s infrastructure." Doctoral thesis, Universidade de Aveiro, 2015. http://hdl.handle.net/10773/14854.
Full textThe proliferation of Internet access allows that users have the possibility to use services available directly through the Internet, which translates in a change of the paradigm of using applications and in the way of communicating, popularizing in this way the so-called cloud computing paradigm. Cloud computing brings with it requirements at two different levels: at the cloud level, usually relying in centralized data centers, where information technology and network resources must be able to guarantee the demand of such services; and at the access level, i.e., depending on the service being consumed, different quality of service is required in the access network, which is a Network Operator (NO) domain. In summary, there is an obvious network dependency. However, the network has been playing a relatively minor role, mostly as a provider of (best-effort) connectivity within the cloud and in the access network. The work developed in this Thesis enables for the effective integration of cloud and NO domains, allowing the required network support for cloud. We propose a framework and a set of associated mechanisms for the integrated management and control of cloud computing and NO domains to provide endto- end services. Moreover, we elaborate a thorough study on the embedding of virtual resources in this integrated environment. The study focuses on maximizing the host of virtual resources on the physical infrastructure through optimal embedding strategies (considering the initial allocation of resources as well as adaptations through time), while at the same time minimizing the costs associated to energy consumption, in single and multiple domains. Furthermore, we explore how the NO can take advantage of the integrated environment to host traditional network functions. In this sense, we study how virtual network Service Functions (SFs) should be modelled and managed in a cloud environment and enhance the framework accordingly. A thorough evaluation of the proposed solutions was performed in the scope of this Thesis, assessing their benefits. We implemented proof of concepts to prove the added value, feasibility and easy deployment characteristics of the proposed framework. Furthermore, the embedding strategies evaluation has been performed through simulation and Integer Linear Programming (ILP) solving tools, and it showed that it is possible to reduce the physical infrastructure energy consumption without jeopardizing the virtual resources acceptance. This fact can be further increased by allowing virtual resource adaptation through time. However, one should have in mind the costs associated to adaptation processes. The costs can be minimized, but the virtual resource acceptance can be also reduced. This tradeoff has also been subject of the work in this Thesis.
A proliferação do acesso à Internet permite aos utilizadores usar serviços disponibilizados diretamente através da Internet, o que se traduz numa mudança de paradigma na forma de usar aplicações e na forma de comunicar, popularizando desta forma o conceito denominado de cloud computing. Cloud computing traz consigo requisitos a dois níveis: ao nível da própria cloud, geralmente dependente de centros de dados centralizados, onde as tecnologias de informação e recursos de rede têm que ser capazes de garantir as exigências destes serviços; e ao nível do acesso, ou seja, dependendo do serviço que esteja a ser consumido, são necessários diferentes níveis de qualidade de serviço na rede de acesso, um domínio do operador de rede. Em síntese, existe uma clara dependência da cloud na rede. No entanto, o papel que a rede tem vindo a desempenhar neste âmbito é reduzido, sendo principalmente um fornecedor de conectividade (best-effort) tanto no dominio da cloud como no da rede de acesso. O trabalho desenvolvido nesta Tese permite uma integração efetiva dos domínios de cloud e operador de rede, dando assim à cloud o efetivo suporte da rede. Para tal, apresentamos uma plataforma e um conjunto de mecanismos associados para gestão e controlo integrado de domínios cloud computing e operador de rede por forma a fornecer serviços fim-a-fim. Além disso, elaboramos um estudo aprofundado sobre o mapeamento de recursos virtuais neste ambiente integrado. O estudo centra-se na maximização da incorporação de recursos virtuais na infraestrutura física por meio de estratégias de mapeamento ótimas (considerando a alocação inicial de recursos, bem como adaptações ao longo do tempo), enquanto que se minimizam os custos associados ao consumo de energia. Este estudo é feito para cenários de apenas um domínio e para cenários com múltiplos domínios. Além disso, exploramos como o operador de rede pode aproveitar o referido ambiente integrado para suportar funções de rede tradicionais. Neste sentido, estudamos como as funções de rede virtualizadas devem ser modeladas e geridas num ambiente cloud e estendemos a plataforma de acordo com este conceito. No âmbito desta Tese foi feita uma avaliação extensa das soluções propostas, avaliando os seus benefícios. Implementámos provas de conceito por forma a demonstrar as mais-valias, viabilidade e fácil implantação das soluções propostas. Além disso, a avaliação das estratégias de mapeamento foi realizada através de ferramentas de simulação e de programação linear inteira, mostrando que é possível reduzir o consumo de energia da infraestrutura física, sem comprometer a aceitação de recursos virtuais. Este aspeto pode ser melhorado através da adaptação de recursos virtuais ao longo do tempo. No entanto, deve-se ter em mente os custos associados aos processos de adaptação. Os custos podem ser minimizados, mas isso implica uma redução na aceitação de recursos virtuais. Esta compensação foi também um tema abordado nesta Tese.
Retana, Solano German F. "Technology support and demand for cloud infrastructure services: the role of service providers." Diss., Georgia Institute of Technology, 2013. http://hdl.handle.net/1853/50236.
Full textAmarasinghe, Heli. "Network Resource Management in Infrastructure-as-a-Service Clouds." Thesis, Université d'Ottawa / University of Ottawa, 2019. http://hdl.handle.net/10393/39141.
Full textCroker, Trevor D. "Formation of the Cloud: History, Metaphor, and Materiality." Diss., Virginia Tech, 2020. http://hdl.handle.net/10919/96439.
Full textDoctor of Philosophy
This dissertation tells the story of cloud computing by looking at the history of the cloud and then discussing the social and political implications of this history. I start by arguing that the cloud is connected to earlier visions of computing (specifically, utility computing and ubiquitous computing). By referencing these older histories, I argue that much of what we currently understand as cloud computing is actually connected to earlier debates and efforts to shape a computing future. Using the history of computing, I demonstrate the role that metaphor plays in the development of a technology. Using these earlier histories, I explain how cloud computing was coined in the 1990s and eventually became a dominant vision of computing in the late 2000s. Much of the research addresses how the metaphor of the cloud is used, the initial reaction to the idea of the cloud, and how the creation of the cloud did (or did not) borrow from older visions of computing. This research looks at which people use the cloud, how the cloud is marketed to different groups, and the challenges of conceptualizing this new distributed computing network. This dissertation gives particular weight to the materiality of the cloud. My research focuses on the cloud's impact on data centers and submarine communication data cables. Additionally, I look at the impact of the cloud on a local community (Los Angeles, CA). Throughout this research, I argue that the metaphor of the cloud often hides deeper complexities. By looking at the material impact of the cloud, I demonstrate how larger economic, social, and political realities are entangled in the story and metaphor of the cloud.
Božić, Nikola. "Blockchain technologies and their application to secure virtualized infrastructure control." Electronic Thesis or Diss., Sorbonne université, 2019. http://www.theses.fr/2019SORUS596.
Full textBlockchain is a technology making the shared registry concept from distributed systems a reality for a number of application domains, from the cryptocurrency one to potentially any industrial system requiring decentralized, robust, trusted and automated decision making in a multi-stakeholder situation. Nevertheless, the actual advantages in using blockchain instead of any other traditional solution (such as centralized databases) are not completely understood to date, or at least there is a strong need for a vademecum guiding designers toward the right decision about when to adopt blockchain or not, which kind of blockchain better meets use-case requirements, and how to use it. At first, we aim at providing the community with such a vademecum, while giving a general presentation of blockchain that goes beyond its usage in Bitcoin and surveying a selection of the vast literature that emerged in the last few years. We draw the key requirements and their evolution when passing from permissionless to permissioned blockchains, presenting the differences between proposed and experimented consensus mechanisms, and describing existing blockchain platforms. Furthermore, we present the B-VMOA blockchain to secure virtual machine orchestration operations for cloud computing and network functions virtualization systems applying the proposed vademecum logic. Using tutorial examples, we describe our design choices and draw implementation plans. We further develop the vademecum logic applied to cloud orchestration and how it can lead to precise platform specifications. We capture the key system operations and complex interactions between them. We focus on the last release of Hyperledger Fabric platform as a way to develop B-VMOA system. Besides, Hyperledger Fabric optimizes conceived B-VMOA network performance, security, and scalability by way of workload separation across: (i) transaction execution and validation peers, and (ii) transaction ordering nodes. We study and use a distributed execute-order-validate architecture which differentiates our conceived B-VMOA system from legacy distributed systems that follow a traditional state-machine replication architecture. We parameterize and validate our model with data collected from a realistic testbed, presenting an empirical study to characterize system performance and identify potential performance bottlenecks. Furthermore, we present the tools we used, the network setup and the discussion on empirical observations from the data collection. We examine the impact of various configurable parameters to conduct an in-dept study of core components and benchmark performance for common usage patterns. Namely, B-VMOA is meant to be run within data center. Different data center interconnection topologies scale differently due to communication protocols. Enormous challenges appear to efficiently design the network interconnections so that the deployment and maintenance of the infrastructure is cost-effective. We analyze the structural properties of several DCN topologies and also present some comparison among these network architectures with the aim to reduce B-VMOA overhead costs. From our analysis, we recommend the hypercube topology as a solution to address the performance bottleneck in the B-VMOA control plane caused by gossip dissemination protocol along with an estimate of performance improvement
Meinel, Christoph, Christian Willems, Sebastian Roschke, and Maxim Schnjakin. "Virtualisierung und Cloud Computing : Konzepte, Technologiestudie, Marktübersicht." Universität Potsdam, 2011. http://opus.kobv.de/ubp/volltexte/2011/4970/.
Full textVirtualization and Cloud Computing belong to the most important issues for operators of large ICT infrastructures today. There are a large number of various technologies, products, and business models for entirely different application scenarios. The study at hand gives a detailed overview on latest developments in concepts and technologies of virtualization – beginning with classic server virtualization, continuing with infrastructures for virtual workplaces, through to application virtualization and makes an attempt to classify all these variants of virtualization. When investigating on the concepts of Cloud Computing, the report introduces basic principles as well as different types of architecture and use cases. The extensive analysis of benefits of Cloud Computing and possible reservations when using cloud resources within an enterprise context is evidence that Cloud Computing offers great opportunities, but is not worth considering for any kind of application scenario, legal framework or business scenario. The subsequent market study on virtualization technology shows that each of the major manufacturers – Citrix, Microsoft, and VMware – offer products for any variant of virtualization and highlights the important differences between the products and the respective strengths of the vendors. For example, the Citrix solution on Virtual Desktop Infrastructures comes up very well-engineered, while Microsoft can only rely on standard technology in this field. VMware, the market leader in virtualization technology, has gained the biggest popularity in data centers and offers the only product implementing real fault tolerance. On the other hand, Microsoft is able to score with seamless integration of their virtualization products into existing Windows-based infrastructures. In the area of Cloud Computing systems, there are some open source software projects that are very possibly suitable for the productive operation of so called private clouds.
Salazar, Javier. "Resource allocation optimization algorithms for infrastructure as a service in cloud computing." Thesis, Sorbonne Paris Cité, 2016. http://www.theses.fr/2016USPCB074.
Full textThe cloud architecture offers on-demand computing, storage and applications. Within this structure, Cloud Providers (CPs) not only administer infrastructure resources but also directly benefit from leasing them. In this thesis, we propose three optimization models to assist CPs reduce the costs incurred in the resource allocation process when serving users’ demands. Implementing the proposed models will not only increase the CP’s revenue but will also enhance the quality of the services offered, benefiting all parties. We focus on Infrastructure as a Service (IaaS) resources which constitute the physical infrastructure of the cloud and are contained in datacenters (DCs). Following existing research in DC design and cloud computing applications, we propose the implementation of smaller DCs (Edge DCs) be located close to end users as an alternative to large centralized DCs. Lastly, we use the Column Generation optimization technique to handle large scale optimization models efficiently. The proposed formulation optimizes both the communications and information technology resources in a single phase to serve IaaS requests. Based on this formulation, we also propose a second model that includes QoS guarantees under the same Infrastructure as a Service resource allocation perspective, to provide different solutions to diverse aspects of the resource allocation problem such as cost and delay reduction while providing different levels of service. Additionally, we consider the multimedia cloud computing scenario. When Edge DCs architecture is applied to this scenario it results in the creation of the Multimedia Edge Cloud (MEC) architecture. In this context we propose a resource allocation approach to help with the placement of these DCs to reduce communication related problems such as jitter and latency. We also propose the implementation of optical fiber network technologies to enhance communication between DCs. Several studies can be found proposing new methods to improve data transmission and performance. For this study, we decided to implement Wavelength Division Multiplexing (WDM) to strengthen the link usage and light-paths and, by doing so, group different signals over the same wavelength. Using a realistic simulation environment, we evaluate the efficiency of the approaches proposed in this thesis using a scenario specifically designed for the DCs, comparing them with different benchmarks and also simulating the effect of the optical formulation on the network performance. The numerical results obtained show that by using the proposed models, a CP can efficiently reduce allocation costs while maintaining satisfactory request acceptance and QoS ratios
Margaris, Nikos. "Cloud Computing v českém prostředí." Master's thesis, Vysoká škola ekonomická v Praze, 2011. http://www.nusl.cz/ntk/nusl-72541.
Full textSvärd, Petter. "Dynamic Cloud Resource Management : Scheduling, Migration and Server Disaggregation." Doctoral thesis, Umeå universitet, Institutionen för datavetenskap, 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-87904.
Full textYampolsky, Vincent. "An evaluation of the power consumption and carbon footprint of a cloud infrastructure." Thesis, Edinburgh Napier University, 2010. http://researchrepository.napier.ac.uk/Output/3973.
Full textZia, Umar. "Enabling Context Awareness in Ambient Environments using Cloud Infrastructures." Thesis, Mittuniversitetet, Institutionen för informationsteknologi och medier, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:miun:diva-15712.
Full textAhmadi, Mehri Vida. "An Investigation of CPU utilization relationship between host and guests in a Cloud infrastructure." Thesis, Blekinge Tekniska Högskola, Institutionen för kommunikationssystem, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-10809.
Full textManco, Filipe José Jesus. "Network infrastructure control for virtual campus." Master's thesis, Universidade de Aveiro, 2013. http://hdl.handle.net/10773/12725.
Full textAn evolution of the current university’s networking infrastructure and service models has been shown to be necessary to enable the provisioning of innovative services that are able to respond to today’s needs. On this work a network orchestration tool is proposed that, integrated with the OpenStack cloud framework, is able to virtualize any network deployment in a non-disruptive manner, providing a clean upgrade path from the traditional networking to the world of virtualization. The framework is able to extend virtual networks created on the datacenter by OpenStack or other cloud frameworks, to the outside campus. It does so by directly reconfiguring the network devices according to the needs, independently of the device vendor, the type of device or its specificities, and independently of the specific physical network topology. This service is provided to the end user using a cloud like service model, much more flexible than the current one, properly integrated with the cloud framework services. The project was developed with the Aveiro University use cases in mind, but the final result can be applied in many academic or business environments. The framework is presented both from a conceptual perspective, by describing the abstractions and mechanisms created, and from an implementation perspective, provinding the reader the necessary understanding about the framework operation and the integration with OpenStack. The integration of this work with a broader vision for the future of the university’s services is left as future work.
A evolução da actual infraestrutura de rede e modelos de serviço da universidade tem-se mostrado necessária para permitir o fornecimento de serviços inovadores capazes de responder às necessidades do mundo actual. Neste trabalho é proposta uma ferramenta de orquestração de rede que, integrada com a plataforma de cloud OpenStack, é capaz de virtualizar qualquer infraestrutura de rede de forma não disruptiva, proporcionando um modelo de actualização simples dos tradicionais serviços para o novo mundo da virtualização. A framework é capaz de estender as redes virtuais criadas no datacenter pelo OpenStack ou qualquer outra plataforma de cloud, para o campus. Fá-lo reconfigurando directamente os dispositivos de rede de acordo com as necessidades, independentemente do fabricante, do tipo de dispositivo ou das suas especificidades, e independentemente da topologia da rede física. O serviço é fornecido ao utilizador usando um modelo de cloud, muito mais flexivel que o modelo actual, devidamente integrado com os serviços da plataforma de cloud. O projecto foi desenvolvido com os casos de uso da Universidade de Aveiro em mente, mas o resultado final pode ser aplicado em muitos outros ambientes académicos ou empresariais. A framework é apresentada tanto do ponto de vista conceptual, descrevendo as abstrações e mecanismos criados, como do ponto de vista de implementação, dando ao leitor o entendimento necessário acerca da operação da ferramenta e da sua integração com o OpenStack. A integração deste trabalho com uma visão mais abrangente para o futuro dos serviços da universidade é deixada como trabalho futuro.
Hasan, MD Sabbir. "Smart management of renewable energy in clouds : from infrastructure to application." Thesis, Rennes, INSA, 2017. http://www.theses.fr/2017ISAR0010/document.
Full textWith the advent of cloud enabling technologies and adoption of cloud computing, enterprise and academic institutions are moving their IT workload to the cloud. Although this prolific advancement and easy to access model have greatly impacted our scientific and industrial community in terms of reducing complexity and increasing revenue, data centers are consuming enormous amount of energy, which translates into higher carbon emission. In response, varieties of research work have focused on environmental sustainability for Cloud Computing paradigm through energy consumption reduction by devising energy efficient strategies. However, energy efficiency in cloud infrastructure alone is not going to be enough to boost carbon footprint reduction. Therefore, it is imperative to envision of smartly using green energy at infrastructure and application level for further reduction of carbon footprint. In recent years, some cloud providers are powering their data centers with renewable energy. The characteristics of renewable energy sources are highly intermittent which creates several challenges to manage them efficiently. To overcome the problem, we investigate the options and challenges to integrate different renewable energy sources in a realistic way and propose a Cloud energy broker, which can adjust the availability and price combination to buy Green energy dynamically from the energy market in advance to make a data center partially green. Later, we introduce the concept of Virtualization of Green Energy, which can be seen as an alternative to energy storage used in data center to eliminate the intermittency problem to some extent. With the adoption of virtualization concept, we maximize the usage of green energy contrary to energy storage which induces energy losses, while introduce Green Service Level Agreement based on green energy for service provider and end users. By •using realistic traces and extensive simulation and analysis, we show that, the proposal can provide an efficient, robust and cost-effective energy management scheme for data center. While an efficient energy management in the presence of intermittent green energy is necessary, how modern Cloud applications can take advantage of the presence/absence of green energy has not been studied with requisite effort. Unlike Batch applications, Interactive Cloud applications have to be always accessible and car not be scheduled in advance to match with green energy profile. Therefore, this thesis proposes an energy adaptive autoscaling solution to exploit applications internal to create green energy awareness in the application, while respecting traditional QoS properties. To elaborate, we design green energy aware application controller that takes advantage of green energy availability to perform opportunistic adaptation in an application along with performance aware application controller. Experiment is performed with real life application at Grid5000 and results show significant reduction of energy consumption while respecting traditional QoS attributes compared to performance aware approach
Lakew, Ewnetu Bayuh. "Autonomous cloud resource provisioning : accounting, allocation, and performance control." Doctoral thesis, Umeå universitet, Institutionen för datavetenskap, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-107955.
Full textHans, Ronny [Verfasser], Ralf [Akademischer Betreuer] Steinmetz, and Schahram [Akademischer Betreuer] Dustdar. "QoS-aware Cloud Infrastructure Provisioning in Heterogeneous Environments / Ronny Hans ; Ralf Steinmetz, Schahram Dustdar." Darmstadt : Universitäts- und Landesbibliothek Darmstadt, 2018. http://d-nb.info/1165227606/34.
Full textFilipsson, Emma. "Modelling and simulating Identity and Access Management based lateral movement in a cloud infrastructure." Thesis, KTH, Fysik, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-252836.
Full textCarreño, Emmanuell Diaz. "Migration and evaluation of a numerical weather prediction application in a cloud computing infrastructure." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2015. http://hdl.handle.net/10183/127446.
Full textThe usage of clusters and grids has benefited for years the High Performance Computing (HPC) community. These kind of systems have allowed scientists to use bigger datasets and to perform more intensive computations, helping them to achieve results in less time but has also increased the upfront costs associated with this area of science. As some e-Science projects are carried out also in highly distributed network environments or using immense data sets that sometimes require grid computing, they are good candidates for cloud computing initiatives. The Cloud Computing paradigm has emerged as a practical solution to perform large-scale scientific computing. The elasticity of the cloud and its pay-as-you-go model presents an attractive opportunity for applications commonly executed in clusters or supercomputers. In this context, the user does not need to buy infrastructure, the resources can be rented from a provider and used for a period of time. This thesis presents the challenges and solutions of migrating a numerical weather prediction (NWP) application to a cloud computing infrastructure. We performed the migration of this HPC application and evaluated its performance in a local cluster and the cloud using different instance sizes. We analyzed the main characteristics of the application running in the cloud. The experiments demonstrate that, although processing and networking create a limiting factor, storing input and output datasets in the cloud presents an attractive option to share results and ease the deployment of a test-bed for a weather research platform. Results show that cloud infrastructure can be used as a viable HPC alternative for numerical weather prediction software.
Pastor, Jonathan. "Contributions à la mise en place d'une infrastructure de Cloud Computing à large échelle." Thesis, Nantes, Ecole des Mines, 2016. http://www.theses.fr/2016EMNA0240/document.
Full textThe continuous increase of computing power needs has favored the triumph of the Cloud Computing model. Customers asking for computing power will receive supplies via Internet resources hosted by providers of Cloud Computing infrastructures. To make economies of scale, Cloud Computing that are increasingly large and concentrated in few attractive places, leading to problems such energy supply, fault tolerance and the fact that these infrastructures are far from most of their end users. During this thesis we studied the implementation of an fully distributed and decentralized IaaS system operating a network of micros data-centers deployed in the Internet backbone, using a modified version of OpenStack that leverages non relational databases. A prototype has been experimentally validated onGrid’5000, showing interesting results, however limited by the fact that OpenStack doesn’t take advantage of a geographically distributed functioning. Thus, we focused on adding the support of network locality to improve performance of Cloud Computing services by favoring collaborations between close nodes. A prototype of the DVMS algorithm, working with an unstructured topology based on the Vivaldi algorithm, has been validated on Grid’5000. This prototype got the first prize at the large scale challenge of the Grid’5000 spring school in 2014. Finally, the work made with DVMS enabled us to participate at the development of the VMPlaceS simulator
Cardoso, Igor Duarte. "Network infrastructure control for virtual campuses." Master's thesis, Universidade de Aveiro, 2014. http://hdl.handle.net/10773/14707.
Full textThis dissertation provides a way to merge Cloud Computing infrastructures with traditional or legacy network deployments, leveraging the best in both worlds and enabling a logically centralized control for it. A design/architecture is proposed to extend existing Cloud Computing software stacks so they are able to manage networks outside the Cloud Computing infrastructure, by extending the internal, virtualized network segments. This is useful in a variety of use cases such as incremental Legacy to Cloud network migration, hybrid virtual/ traditional networking, centralized control of existing networks, bare metal provisioning and even offloading of advanced services from typical home gateways into the operator, improving control and reducing maintenance costs. An implementation is presented and tested on top of OpenStack, the principal Open-Source Cloud Computing software stack available. It includes changes to the API, command line interface and existing mechanisms which previously only supported homogeneous vendor equipment, such that they support any hardware and be able to automate their provisioning. By using what is called External Drivers, any organization (an hardware vendor, a Cloud provider or even a telecommunications operator) can develop their own driver to support new, specific networking equipment. Besides this ease of development and extensibility, two drivers are already developed in the context of this work: one for OpenWrt switches/routers and one for Cisco EtherSwitch IOS switching modules. Test results indicate that there are low penalties on latency and throughput, and that provisioning times (for setting up or tearing down networks) are reduced in comparison with similar maintenance operations on traditional computer networks.
Esta dissertação apresenta uma forma de juntar infrastruturas Cloud Computing com redes tradicionais ou legadas, trazendo o melhor de ambos os mundos e possibilitando um controlo logicamente centralizado. Uma arquitetura é proposta com o intuito de extender implementações de Cloud Computing para que possam gerir também redes fora da infrastrutura de Cloud Computing, extendendo os segmentos de rede internos, virtualizados. Isto é útil para um variado conjunto de casos de uso, tais como migração incremental de redes legadas para a Cloud, redes híbridas virtuais/tradicionais, controlo centralizado de redes já existentes, aprovisionamento de bare metal e até mesmo a passagem de serviços tipicamente fornecidos por um home gateway para o lado do operador, melhorando o controlo e reduzindo custos de manutenção. Uma implementação da solução é apresentada e testada em cima do OpenStack, a principal solução Open-Source de Cloud Computing disponível. A implementação inclui alterações à API, à interface de linha de comandos e aos mecanismos já existentes, que apenas suportam implementações homogéneas, para que possam suportar qualquer equipamento e automatizar o aprovisionamento dos mesmos. Através daquilo que se chamam drivers externos, qualquer organização (seja um fabricante de equipamentos de rede, um fornecedor de Cloud ou uma operadora de telecomunicações) pode desenvolver o seu próprio drivers para suportar novos, específicos equipamentos de hardware. Para além da facilidade de desenvolvimento e extensibilidade, dois drivers são também fruto deste trabalho: um para switches/routers OpenWrt e outro para os módulos de switching Cisco EtherSwitch, sistema operativo IOS. Testes efetuados indicam que há baixas penalizações na latência e largura de banda, e ainda que os tempos de aprovisionamento são reduzidos em comparação com semelhantes operações de manutenção em redes informáticas tradicionais.
Metwally, Khaled. "A Resource Management Framework for IaaS in Cloud Computing Environment." Thesis, Université d'Ottawa / University of Ottawa, 2016. http://hdl.handle.net/10393/34951.
Full textLukášová, Pavlína. "Cloud Computing jako nástroj BCM." Master's thesis, Vysoká škola ekonomická v Praze, 2010. http://www.nusl.cz/ntk/nusl-75556.
Full textEkanayake, Mudiyanselage Wijaya Dheeshakthi. "An SDN-based Framework for QoSaware Mobile Cloud Computing." Thesis, Université d'Ottawa / University of Ottawa, 2016. http://hdl.handle.net/10393/35117.
Full textIzumo, Naoki. "Clouded space: Internet physicality." Thesis, University of Iowa, 2017. https://ir.uiowa.edu/etd/5515.
Full textCastagnoli, Carlo. "Cloud Computing: gli Hypervisor e la funzionalità di Live Migration nelle Infrastructure as a Service." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2011. http://amslaurea.unibo.it/1856/.
Full textHawasli, Ahmad. "azureLang: a probabilistic modeling and simulation language for cyber attacks in Microsoft Azure cloud infrastructure." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-240406.
Full textCyberattacksimulering är en lämplig metod som används för att bedöma säkerhetenhos nätverkssystem. En angrepsimulering går stegvis från ett visst systeminmatningspunkt för att utforska angreppsbanorna som leder till olika svagheter i modellen. Varje steg analyseras och tiden för kompromettera beräknas.Attack-simuleringar baseras huvudsakligen på attackgrafer. Graferna används för att modellera angreppssteg där noder kan representera tillgångar i systemet, och kanterna kan representera attackenstegen. För att minska kostnaden för att skapa attackgrafer för varje specifikt system används domänspecifika språk eller DSL förkortat.Den slutliga produkten av detta examensarbete är azureLang, ett probabilistisk hotmodelleringsoch attacksimuleringsspråk för analys av Microsoft Azure Cloud Infrastructure. AzureLang är en DSL som definierar en generisk attacklogik för Microsoft Azure-system. Med hjälp av azureLang kan systemadministratörer enkelt ordna specifika systemscenarier som efterliknar deras Microsoft Azure cloudsystem infrastruktur. Efter att ha skapat modellen kan attack simu-lering köras för att bedöma modellens säkerhet.
Noroozi, Hamid. "A Cloud-native Vehicular Public Key Infrastructure : Towards a Highly-available and Dynamically- scalable VPKIaaS." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-300658.
Full textAnsträngningar för standardisering av Vehicular Communication Systems har varit avgörande för användandet av Vehicular Public-Key Infrastructure (VPKI) för att etablera förtroende mellan nätverksdeltagare. Användande av VPKI i Vehicular Communication (VC) garanterar integritet och autenticitet av meddelanden. Det erbjuder ett lager av säkerhet för fordon då VPKI ger dem en mängd av icke länkbara certifikat, kallade pseudonym, som används medan de kommunicerar med andra fordon, kallat Vehicle-to-Vehicle (V2V) eller Roadside Units (RSUs) kallat Vehicle-to-Infrastructure (V2I). Varje fordon använder ett pseudonym under en begränsad tid och genom att byta till ett icke tidigare använt pseudonym kan det fortsätta kommunicera utan att riskera sin integritet. I litteratur har två metoder föreslagits för hur man ska ladda fordon med pseudonym de behöver. Den ena metoden det så kallade offline-läget, som proponerar att man för-laddar fordonen med alla pseudonym som det behöver vilket ökar kostnaden för revokering i fall de blir komprometterat. Den andra metoden föreslår ett on-demand tillvägagångssätt som erbjuder pseudonym via VPKI på fordonets begäran vid början av varje färd. Valet av på begäran metoden sätter en stor börda på tillgänglighet och motståndskraft av VPKI tjänster. I det här arbetet, möter vi problem med storskaliga driftsättningar av en på begäran VPKI som är motståndskraftig, har hög tillgänglighet och dynamiskt skalbarhet i syfte att uppnå dessa attribut genom att nyttja toppmoderna verktyg och designparadigmer. Vi har förbättrat ett VPKI system för att säkerställa att det är kapabelt att möta SLA:er av företagsklass gällande tillgänglighet och att det även kan vara kostnadseffektivt eftersom tjänster dynamiskt kan skala ut vid högre last eller skala ner vid lägre last. Detta har möjliggjorts genom att arkitekta om en existerande VPKI till en cloud-native lösning driftsatt som mikrotjänster. En av nyckelutmaningarna till att ha en pålitlig arkitektur baserad på distribuerade mikrotjänster är sybil-baserad missuppförande. Genom att utnyttja Sybil baserade attacker på VPKI, kan illvilliga fordon påverka trafik att tjäna dess egna syften. Därför är det av största vikt att förhindra Sybil attacker. Å andra sidan så dras traditionella metoder att stoppa dem med prestandakostnader. Vi föreslår en lösning för att adressera Sybilbaserade attacker genom att nyttja Redis, en in-memory data-store utan att märkbart kompromissa på systemets effektivitet och prestanda. Att köra våra VPKI tjänster på Google Cloud Platform (GCP) och genomföra diverse stresstester mot dessa har visat att storskaliga driftsättningar av VPKI as a Service (VPKIaaS) kan göras effektivt samtidigt som riktigt trafik hanteras. Vi har testat VPKIaaS under syntetisk genererat normalt trafikflöde samt flow och flash mängd scenarier. Det har visat sig att VPKIaaS klarar att utfärda 100 pseudonym per förfråga utsänt av 1000 fordon (där fordonen bad om en ny uppsättning pseudonym varje 1 till 5 sekunder), och varje fordon fått svar inom 77 millisekunder. Vi demonstrerar även att under en flashcrowd situation, där antalet fordon höjs till 50000 med en kläckningsgrad på 100. VPKIaaS dynamiskt skalar ut och tar ≈192 millisekunder att betjäna 100 pseudonymer per förfrågan gjord av fordon.
Lam, Juan Carlos. "Dynamic Analysis of Levee Infrastructure Failure Risk: A Framework for Enhanced Critical Infrastructure Management." Thesis, Virginia Tech, 2012. http://hdl.handle.net/10919/43106.
Full textMaster of Science
Nemanja, Popović. "Napredni distributivni menadžment sistem zasnovan na Cloud infrastrukturi." Phd thesis, Univerzitet u Novom Sadu, Fakultet tehničkih nauka u Novom Sadu, 2018. https://www.cris.uns.ac.rs/record.jsf?recordId=107176&source=NDLTD&language=en.
Full textThis dissertation inspects the possibility of transferring the Advanced Distribution Management System (ADMS) based on a common operating technology platform to the virtual Cloud Infrastructure Environment. First, a traditional ADMS solution based on physical computing architecture was chosen and functional blocks were identified. Furthermore, their performances were profiled to four key resources: processor, operating memory, computer network, and storage. Then, virtual cloud-based solution was proposed based on Cloud infrastructure which is verified on two imaginary power distribution networks of real size (small and large) and in two test scenarios (steady state and high activity). Finally, the test results show that ADMS can be transferred to a virtual Cloud environment without adversely affecting the functional and non-functional requirements of the ADMS solution.
Künsemöller, Jörn [Verfasser]. "Tragedy of the common cloud : game theory on the infrastructure-as-a-service market / Jörn Künsemöller." Paderborn : Universitätsbibliothek, 2014. http://d-nb.info/106464709X/34.
Full textPérennou, Loïc. "Virtual machine experience design : a predictive resource allocation approach for cloud infrastructures." Thesis, Paris, CNAM, 2019. http://www.theses.fr/2019CNAM1246/document.
Full textOne of the main challenges for cloud computing providers remains to offer trustable performance for all users, while maintaining an efficient use of hardware and energy resources. In the context of this CIFRE thesis lead with Outscale, apublic cloud provider, we perform an in-depth study aimed at making management algorithms use new sources of information. We characterize Outscale’s workload to understand the resulting stress for the orchestrator, and the contention for hardware resources. We propose models to predict the runtime of VMs based on features which are available when they start. We evaluate the sensitivity with respect to prediction error of a VM placement algorithm from the literature that requires such predictions. We do not find any advantage in coupling our prediction model and the selected algorithm, but we propose alternative ways to use predictions to optimize the placement of VMs