To see the other types of publications on this topic, follow the link: Cloud infrastructure.

Dissertations / Theses on the topic 'Cloud infrastructure'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Cloud infrastructure.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Ha, Kiryong. "System Infrastructure for Mobile-Cloud Convergence." Research Showcase @ CMU, 2016. http://repository.cmu.edu/dissertations/704.

Full text
Abstract:
The convergence of mobile computing and cloud computing enables new mobile applications that are both resource-intensive and interactive. For these applications, end-to-end network bandwidth and latency matter greatly when cloud resources are used to augment the computational power and battery life of a mobile device. This dissertation designs and implements a new architectural element called a cloudlet, that arises from the convergence of mobile computing and cloud computing. Cloudlets represent the middle tier of a 3-tier hierarchy, mobile device — cloudlet—cloud, to achieve the right balance between cloud consolidation and network responsiveness. We first present quantitative evidence that shows cloud location can affect the performance of mobile applications and cloud consolidation. We then describe an architectural solution using cloudlets that are a seamless extension of todays cloud computing infrastructure. Finally, we define minimal functionalities that cloudlets must offer above/beyond standard cloud computing, and address corresponding technical challenges.
APA, Harvard, Vancouver, ISO, and other styles
2

Baig, Viñas Roger. "Development and management of collective network and cloud computing infrastructures." Doctoral thesis, Universitat Politècnica de Catalunya, 2019. http://hdl.handle.net/10803/667952.

Full text
Abstract:
In the search and development of more participatory models for infrastructure development and management, in this dissertation, we investigate models for the financing, deployment, and operation of network and cloud computing infrastructures. Our main concern is to overcome the inherent exclusion in participation in the processes of development and management and in the right of use in the current dominant models. Our work starts by studying in detail the model of Guifi.net, a successful bottom-up initiative for building network infrastructure, generally referred to as a community networks. We pay special attention to its governance system and economic organisation because we argue that these are the key components of the success of this initiative. Then, we generalise our findings for any community network, aiming at becoming sustainable and scalable, and we explore the suitability of the Guifi.net model to the cloud computing infrastructure. As a result of both, we coin the attribute extensible to refer to infrastructure that is relatively easy to expand and maintain in contrast to those naturally limited or hard to expand, such as natural resources or highly complex or advanced artificial systems. We conclude proposing a generic model which, in our opinion, is suitable, at least, for managing extensible infrastructure. The Guifi.net model is deeply rooted in the commons; thus, the research in this field, in general, and Elinor Ostrom’s work, in particular, have left a profound imprint in our work. Our results show that the \guifinet model meets almost entirely the principles of long-enduring commons identified by E. Ostrom. This work has been developed as an industrial doctorate. As such, it combines academic research with elements of practice and pursues an effective knowledge transfer between academia and the private sector. Given that the private sector’s partner is a not-for-profit organisation, the effort to create social value has prevailed over the ambition to advance the development of a specific industrial product or particular technology.
En la recerca i desenvolupament de models més participatius per al desenvolupament i gestió d'infraestructura, en aquesta tesi investiguem sobre models per al finançament, desplegament i operació d'infraestructures de xarxa i de computació al núvol. La nostra preocupació principal és fer front a l’exclusió inherent dels models dominants actualment pel que fa a la participació en els processos de desenvolupament i gestió i, també, als drets d’us. El nostre treball comença amb un estudi detallat del model de Guifi.net, un cas d'èxit d'iniciativa ciutadana en la construcció d'infraestructura de xarxa, iniciatives que es coneixen com a xarxes comunitàries. En fer-ho, parem una atenció especial al sistema de governança i a l’organització econòmica perquè pensem que són els dos elements claus de l'èxit d'aquesta iniciativa. Tot seguit passem a analitzar d'altres xarxes comunitàries per abundar en la comprensió dels factors determinants per a la seva sostenibilitat i escalabilitat. Després ampliem el nostre estudi analitzant la capacitat i el comportament del model de Guifi.net en el camp de les infraestructures de computació al núvol. A resultes d'aquests estudis, proposem l'atribut extensible per a descriure aquelles infraestructures que són relativament fàcil d'ampliar i gestionar, en contraposició a les que o bé estan limitades de forma natural o be són difícils d'ampliar, com ara els recursos naturals o els sistemes artificials avançats o complexos. Finalitzem aquest treball fent una proposta de model genèric que pensem que és d'aplicabilitat, com a mínim, a tot tipus d'infraestructura extensible. El model de Guifi.net està fortament vinculat als bens comuns. És per això que la recerca en aquest àmbit, en general, i els treballs de Elinor Ostrom en particular, han deixat una forta empremta en el nostre treball. Els resultats que hem obtingut mostren que el model Guifi.net s'ajusta molt bé als principis que segons Ostrom han de complir els béns comuns per ser sostenibles. Aquest treball s'ha desenvolupat com a doctorat industrial. Com a tal, combina la investigació acadèmica amb elements de practica i persegueix una transferència efectiva de coneixement entre l'àmbit acadèmic i el sector privat. Ates que el soci del sector privat és una organització sense ànim de lucre, l’esforç per crear valor social ha prevalgut en l’ambició d’avançar en el desenvolupament d'un producte industrial específic o d'una tecnologia particular
APA, Harvard, Vancouver, ISO, and other styles
3

Suriano, Francesco. "Cloud Computing: From Containerization to Infrastructure Provisioning." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2019. http://amslaurea.unibo.it/18504/.

Full text
Abstract:
L'elaborato tratta argomenti relativi al cloud computing, in particolare delle tecniche di containerizzazione, orchestrazione e gestione dell'infrastruttura. Sono presentati e confrontati diversi strumenti per risolvere questi tipi di problemi. Lo studio e la scelta di un determinato strumento è stata fatta tenendo conto di esigenze reali di Kopjra, azienda in cui è stato svolto il tirocinio curricolare.
APA, Harvard, Vancouver, ISO, and other styles
4

Tran, Khanh-Toan. "Efficient complex service deployment in cloud infrastructure." Thesis, Evry-Val d'Essonne, 2013. http://www.theses.fr/2012EVRY0038/document.

Full text
Abstract:
Le but du travail réalisé dans cette thèse est de fournir aux fournisseurs de services une solution qui est capable de non seulement déployer les services dans le cloud de façon économique, automatique, mais aussi à grande échelle. La première contribution traite le problème de la construction d’un nouveau service demandé par le client à partir de services disponibles dans le cloud de manière à satisfaire les exigences en termes de qualité de service ainsi qu’en termes de coût. Nous présentons ce problème par un modèle analytique et proposons un algorithme heuristique dont la performance est améliorée de 20-30% par rapport aux autres approches. La seconde contribution est une solution pour déployer les services dans le cloud en considérant les demandes des utilisateurs finaux. Pour assurer qualité de services dans une grande échelle, le service demandé est dupliqué et distribué dans le réseau; chacun de ses réplicas servira les utilisateurs à proximité. Le plan d’approvisionnement selon lequel le service est dupliqué dépend de sa demande, ce qui ne cesse pas de changer en quantité ainsi qu’en distribution, ce qui rend le problème plus compliqué. Nous proposons une solution qui est capable de s’adapter aux changements dans le réseau, y compris ceux des demandes de ses utilisateurs. Enfin, nous proposons un système basé sur OpenStack qui permet de déployer les services complexes dans un cloud qui couvre différente locations (cloud multi-site). A partir d’une demande du client, le système automatiquement calcule le plan d’approvisionnement optimal et le déploie en respectant les contraintes du client
The purpose of the work in this thesis is to provide the Service Provider a solution which is capable of deploying complex services in a cloud automatically and cost-effectively. The first contribution allows the Service Provider to construct complex services requested by the clients from basic services at his disposal. The construction must be efficient in terms of execution time and operation cost while respecting the client’s constraints. We present an analytic model for this problem and propose a heuristic solution which performance is 20-30% better than other approaches. The second contribution solves the problem of deploying the services while considering end-users’ demands. To ensure the quality of services provided to end-users, not only one instance but a set of service replicas is deployed in the network. How the service is duplicated and distributed depends on the demands of its end-users that change constantly in quantity as well as distribution, which complicates the problem. Thus the provisioning mechanism has to be capable of adapt to the change in the network, including the change in end-users’ demands. Our third contribution is a system using OpenStack which allows Service Provider to deploy complex services in a cloud that spans over different locations (multi-site cloud). Given a client’s request, the system automatically calculates the optimal provisioning plan and deploys it while respecting the client’s constraints
APA, Harvard, Vancouver, ISO, and other styles
5

Su, Xueyuan. "Efficient Fault-Tolerant Infrastructure for Cloud Computing." Thesis, Yale University, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=3578459.

Full text
Abstract:

Cloud computing is playing a vital role for processing big data. The infrastructure is built on top of large-scale clusters of commodity machines. It is very challenging to properly manage the hardware resources in order to utilize them effectively and to cope with the inevitable failures that will arise with such a large collection of hardware. In this thesis, task assignment and checkpoint placement for cloud computing infrastructure are studied.

As data locality is critical in determining the cost of running a task on a specific machine, how tasks are assigned to machines has a big impact on job completion time. An idealized abstract model is presented for a popular cloud computing platform called Hadoop. Although Hadoop task assignment (HTA) is [special characters omitted]-hard, an algorithm is presented with only an additive approximation gap. Connection is established between the HTA problem and the minimum makespan scheduling problem under the restricted assignment model. A new competitive ratio bound for the online GREEDY algorithm is obtained.

Checkpoints allow recovery of long-running jobs from failures. Checkpoints themselves might fail. The effect of checkpoint failures on job completion time is investigated. The sum of task success probability and checkpoint reliability greatly affects job completion time. When possible checkpoint placements are constrained, retaining only the most recent Ω(log n) possible checkpoints has at most a constant factor penalty. When task failures follow the Poisson distribution, two symmetries for non-equidistant placements are proved and a first order approximation to optimum placement interval is generalized.

APA, Harvard, Vancouver, ISO, and other styles
6

Pech, David. "Cloud Framework on Infrastructure as a Service." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2013. http://www.nusl.cz/ntk/nusl-236185.

Full text
Abstract:
Práce se zabývá podrobnou analýzou požadavků na moderní aplikační rámec pro prostředí cloud. Za pomoci standardních návrhových vzorů a technik připravuje teoretický základ a pravidla, která musí uvnitř rámce platit. V práci je realizována referenční implementace a připravena demonstrační aplikace středního rozsahu, aby představila výhody plynoucí z užití frameworku.
APA, Harvard, Vancouver, ISO, and other styles
7

Fehse, Carsten. "Infrastructure suitability assessment modeling for cloud computing solutions." Thesis, Monterey, California. Naval Postgraduate School, 2011. http://hdl.handle.net/10945/5580.

Full text
Abstract:
Approved for public release; distribution is unlimited
Maturing virtualization in information technology systems has enabled increased implementations of the cloud com-puting paradigm, dissolving the need to co-locate user and computing power by providing desired services through the network. This thesis researches the support that current network modeling and simulation applications can provide to IT projects in planning, implementing and maintaining networks for cloud solutions. A problem-appropriate do-main model and subsequent requirements are developed for the assessment of several network modeling and simula-tion tools, which leads to the identification of a capability gap precluding the use of such tools in early stages of cloud computing projects. Consequently, a practical, modular designed methodology is proposed to measure the essential properties necessary for developing appropriate cloud computing network traffic models. The conducted proof-of-concept experiment applied to a virtual desktop environment finds the proposed methodology suitable and problem-appropriate, and results in recommended steps to close the identified capability gap.
APA, Harvard, Vancouver, ISO, and other styles
8

Younis, Y. A. "Securing access to cloud computing for critical infrastructure." Thesis, Liverpool John Moores University, 2015. http://researchonline.ljmu.ac.uk/4453/.

Full text
Abstract:
Cloud computing offers cost effective services on-demand which encourage critical infrastructure providers to consider migrating to the cloud. Critical infrastructures are considered as a backbone of modern societies such as power plants and water. Information in cloud computing is likely to be shared among different entities, which could have various degrees of sensitivity. This requires robust isolation and access control mechanisms. Although various access control models and policies have been developed, they cannot fulfil requirements for a cloud based access control system. The reason is that cloud computing has a diverse sets of security requirements and unique security challenges such as multi-tenant and heterogeneity of security policies, rules and domains. This thesis provides a detailed study of cloud computing security challenges and threats, which were used to identify security requirements for various critical infrastructure providers. We found that an access control system is a crucial security requirement for the surveyed critical infrastructure providers. Furthermore, the requirement analysis was used to propose a new criteria to evaluate access control systems for cloud computing. Moreover, this work presents a new cloud based access control model to meet the identified cloud access control requirements. The model does not only ensure the secure sharing of resources among potential untrusted tenants, but also has the capacity to support different access permissions for the same cloud user. Our focused in the proposed model is the lack of data isolation in lower levels (CPU caches), which could lead to bypass access control models to gain some sensitive information by using cache side-channel attacks. Therefore, the thesis investigates various real attack scenarios and the gaps in existing mitigation approaches. It presents a new Prime and Probe cache side-channel attack, which can give detailed information about addresses accessed by a virtual machine with no need for any information about cache sets accessed by the virtual machine. The design, implementation and evaluation of a proposed solution preventing cache side-channel attacks are also presented in the thesis. It is a new lightweight solution, which introduces very low overhead (less than 15,000 CPU cycles). It can be applied in any operating system and prevents cache side-channel attacks in cloud computing. The thesis also presents a new detecting cache side-channel attacks solution. It focuses on the infrastructure used to host cloud computing tenants by counting cache misses caused by a virtual machine. The detection solutions has 0% false negative and 15% false positive.
APA, Harvard, Vancouver, ISO, and other styles
9

Mechtri, Marouen. "Virtual networked infrastructure provisioning in distributed cloud environments." Thesis, Evry, Institut national des télécommunications, 2014. http://www.theses.fr/2014TELE0028/document.

Full text
Abstract:
L'informatique en nuage (Cloud Computing) a émergé comme un nouveau paradigme pour offrir des ressources informatiques à la demande et pour externaliser des infrastructures logicielles et matérielles. Le Cloud Computing est rapidement et fondamentalement en train de révolutionner la façon dont les services informatiques sont mis à disposition et gérés. Ces services peuvent être demandés à partir d’un ou plusieurs fournisseurs de Cloud d’où le besoin de la mise en réseau entre les composants des services informatiques distribués dans des emplacements géographiquement répartis. Les utilisateurs du Cloud veulent aussi déployer et instancier facilement leurs ressources entre les différentes plateformes hétérogènes de Cloud Computing. Les fournisseurs de Cloud assurent la mise à disposition des ressources de calcul sous forme des machines virtuelles à leurs utilisateurs. Par contre, ces clients veulent aussi la mise en réseau entre leurs ressources virtuelles. En plus, ils veulent non seulement contrôler et gérer leurs applications, mais aussi contrôler la connectivité réseau et déployer des fonctions et des services de réseaux complexes dans leurs infrastructures virtuelles dédiées. Les besoins des utilisateurs avaient évolué au-delà d'avoir une simple machine virtuelle à l'acquisition de ressources et de services virtuels complexes, flexibles, élastiques et intelligents. L'objectif de cette thèse est de permettre le placement et l’instanciation des ressources complexes dans des infrastructures de Cloud distribués tout en permettant aux utilisateurs le contrôle et la gestion de leurs ressources. En plus, notre objectif est d'assurer la convergence entre les services de cloud et de réseau. Pour atteindre cela, nous proposons des algorithmes de mapping d’infrastructures virtuelles dans les centres de données et dans le réseau tout en respectant les exigences des utilisateurs. Avec l'apparition du Cloud Computing, les réseaux traditionnels sont étendus et renforcés avec des réseaux logiciels reposant sur la virtualisation des ressources et des fonctions réseaux. En plus, le nouveau paradigme d'architecture réseau (Software Defined Networks) est particulièrement pertinent car il vise à offrir la programmation du réseau et à découpler, dans un équipement réseau, la partie plan de données de la partie plan de contrôle. Dans ce contexte, la première partie propose des algorithmes optimaux (exacts) et heuristiques de placement pour trouver le meilleur mapping entre les demandes des utilisateurs et les infrastructures sous-jacentes, tout en respectant les exigences exprimées dans les demandes. Cela inclut des contraintes de localisation permettant de placer une partie des ressources virtuelles dans le même nœud physique. Ces contraintes assurent aussi le placement des ressources dans des nœuds distincts. Les algorithmes proposés assurent le placement simultané des nœuds et des liens virtuels sur l’infrastructure physique. Nous avons proposé aussi un algorithme heuristique afin d’accélérer le temps de résolution et de réduire la complexité du problème. L'approche proposée se base sur la technique de décomposition des graphes et la technique de couplage des graphes bipartis. Dans la troisième partie, nous proposons un cadriciel open source (framework) permettant d’assurer la mise en réseau dynamique entre des ressources Cloud distribués et l’instanciation des fonctions réseau dans l’infrastructure virtuelle de l’utilisateur. Ce cadriciel permettra de déployer et d’activer les composants réseaux afin de mettre en place les demandes des utilisateurs. Cette solution se base sur un gestionnaire des ressources réseaux "Cloud Network Gateway Manager" et des passerelles logicielles permettant d’établir la connectivité dynamique et à la demande entre des ressources cloud et réseau. Le CNG-Manager offre le contrôle de la partie réseau et prend en charge le déploiement des fonctions réseau nécessaires dans l'infrastructure virtuelle des utilisateurs
Cloud computing emerged as a new paradigm for on-demand provisioning of IT resources and for infrastructure externalization and is rapidly and fundamentally revolutionizing the way IT is delivered and managed. The resulting incremental Cloud adoption is fostering to some extent cloud providers cooperation and increasing the needs of tenants and the complexity of their demands. Tenants need to network their distributed and geographically spread cloud resources and services. They also want to easily accomplish their deployments and instantiations across heterogeneous cloud platforms. Traditional cloud providers focus on compute resources provisioning and offer mostly virtual machines to tenants and cloud services consumers who actually expect full-fledged (complete) networking of their virtual and dedicated resources. They not only want to control and manage their applications but also control connectivity to easily deploy complex network functions and services in their dedicated virtual infrastructures. The needs of users are thus growing beyond the simple provisioning of virtual machines to the acquisition of complex, flexible, elastic and intelligent virtual resources and services. The goal of this thesis is to enable the provisioning and instantiation of this type of more complex resources while empowering tenants with control and management capabilities and to enable the convergence of cloud and network services. To reach these goals, the thesis proposes mapping algorithms for optimized in-data center and in-network resources hosting according to the tenants' virtual infrastructures requests. In parallel to the apparition of cloud services, traditional networks are being extended and enhanced with software networks relying on the virtualization of network resources and functions especially through network resources and functions virtualization. Software Defined Networks are especially relevant as they decouple network control and data forwarding and provide the needed network programmability and system and network management capabilities. In such a context, the first part proposes optimal (exact) and heuristic placement algorithms to find the best mapping between the tenants' requests and the hosting infrastructures while respecting the objectives expressed in the demands. This includes localization constraints to place some of the virtual resources and services in the same host and to distribute other resources in distinct hosts. The proposed algorithms achieve simultaneous node (host) and link (connection) mappings. A heuristic algorithm is proposed to address the poor scalability and high complexity of the exact solution(s). The heuristic scales much better and is several orders of magnitude more efficient in terms of convergence time towards near optimal and optimal solutions. This is achieved by reducing complexity of the mapping process using topological patterns to map virtual graph requests to physical graphs representing respectively the tenants' requests and the providers' physical infrastructures. The proposed approach relies on graph decomposition into topology patterns and bipartite graphs matching techniques. The third part propose an open source Cloud Networking framework to achieve cloud and network resources provisioning and instantiation in order to respectively host and activate the tenants' virtual resources and services. This framework enables and facilitates dynamic networking of distributed cloud services and applications. This solution relies on a Cloud Network Gateway Manager and gateways to establish dynamic connectivity between cloud and network resources. The CNG-Manager provides the application networking control and supports the deployment of the needed underlying network functions in the tenant desired infrastructure (or slice since the physical infrastructure is shared by multiple tenants with each tenant receiving a dedicated and isolated portion/share of the physical resources)
APA, Harvard, Vancouver, ISO, and other styles
10

Liberman, García Ari. "The evolution of the Cloud : the work, progress and outlook of cloud infrastructure." Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/100311.

Full text
Abstract:
Thesis: S.M. in Engineering and Management, Massachusetts Institute of Technology, Engineering Systems Division, System Design and Management Program, 2015.
This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 59-61).
Cloud computing has enabled the deployment of systems at scale without requiring deep expertise in infrastructure management or highly specialized personnel. In just a few years, cloud computing has become one of the fastest growing technology segments in the Information Technology industry and it has transformed how applications are created and how companies they manage their growth. The cloud market has quickly become one of the most competitive industries with companies committing their efforts to the creation of cloud platforms and aggressive pricing strategies in an attempt to gain market dominance. This work shows the origins of the Infrastructure-as-a-Service industry and an analysis of the market dynamics by looking at the portfolios and strategies of the top competitors in this space. Also, this report shows what are the developments that will drive the innovation in the cloud industry years to come.
by Ari Liberman García.
S.M. in Engineering and Management
APA, Harvard, Vancouver, ISO, and other styles
11

AlJabban, Tarek. "Distributed database storage management for a cloud computing infrastructure." Thesis, McGill University, 2013. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=114556.

Full text
Abstract:
Internet applications have recently witnessed tremendous growth in terms of both size and complexity. Cloud computing is one of the several distributed technologies that have emerged to help meeting the objectives of these applications in terms of achieving high availability, performance and scalability.Platform as a Service (PaaS) is one kind of services provided by cloud solutions. These systems often follow a multi-tier architecture consisting mainly of a presentation tier, an application tier and a database tier. The volumes of data exchanged between the application tier and the database tier become huge, especially for enterprise level applications. As a result, the design of the database tier in cloud systems has to carefully address the scalability challenges rising from the huge data volumes. In this thesis, we propose a data distribution approach to improve the scalability of the database tier. Our approach is applied to a traditional single database server. It works by replacing the traditionally used single machine storage paradigm with a distributed storage paradigm. The suggested approach maintains the features that originally exists in the database system, and additionally provides the features of distribution and replication. Distributing the data storage helps improving the system fault-tolerance as it decreases the possibility of having a failure at the database server. It also helps resolve specific performance issues such as reducing the I/O usage and consecutively decreasing the possibility of an I/O bottleneck. Yet, it produces other performance challenges that need to be addressed. To prove the feasibility of our proposed approach, we use it to implement two extensions to the storage manager module of the PostgreSQL database system, using the HDFS distributed file system, and the HBase distributed key-value store.
Les applications Internet ont récemment connu une croissance considérable en termes de taille et de complexité. Afin de satisfaire la forte demande pour les ressources informatiques et les espaces de stockage, les technologies en distribution ont commencé à devenir plus impliquées dans les applications à grande échelle. Le Cloud Computing est l'une de ces nombreuses technologies qui ont émergé pour aider à atteindre les objectifs de ces applications, telles que la haute disponibilité, les performances et l'évolutivité.Platform as a Service (PaaS) est un type de service qui peut être fourni par les solutions de Cloud Computing. Ces systèmes suivent souvent une architecture multi-niveaux qui se compose principalement d'un niveau de présentation, un niveau d'application et d'un niveau de base de données. Les volumes de données échangées entre l'application et la base de données deviennent énormes en particulier pour les applications de niveau entreprise. En conséquence, la conception de la base de données dans les systèmes de Cloud Computing doit prendre en compte le challenge de l'évolution des quantités énormes de données. Dans cette mémoire, nous proposons une approche de distribution des données qui peuvent être utilisées pour améliorer l'évolutivité des bases de données. Nous proposons deux techniques qui peuvent être appliquées à un serveur de base de données unique traditionnelle.Ces techniques fonctionnent en remplaçant le paradigme traditionnel utilisant une seule machine de stockage avec un paradigme de stockage distribué. Les techniques proposées maintiennent les caractéristiques qui existaient à l'origine dans le système de base de données, et en plus fournissent les caractéristiques de la distribution et de la réplication. Ces deux fonctionnalités supplémentaires aident à améliorer le système de tolérance aux pannes, car ils diminuent la possibilité d'avoir une défaillance au niveau du serveur de base de données. La distribution du stockage permet de résoudre les problèmes de performances spécifiques, tels que la réduction de l'utilisation des entrées/sorties et consécutivement de diminuer la possibilité de saturation des entrées/sorties.Par ailleurs, cela produit d'autres défis de performances qui doivent être pris en compte. Pour prouver la faisabilité de nos techniques, nous les avons implémentées comme des extensions du module de gestion de stockage de la base de données PostgreSQL.
APA, Harvard, Vancouver, ISO, and other styles
12

Talevi, Iacopo. "Big Data Analytics and Application Deployment on Cloud Infrastructure." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2017. http://amslaurea.unibo.it/14408/.

Full text
Abstract:
This dissertation describes a project began in October 2016. It was born from the collaboration between Mr.Alessandro Bandini and me, and has been developed under the supervision of professor Gianluigi Zavattaro. The main objective was to study, and in particular to experiment with, the cloud computing in general and its potentiality in the data elaboration field. Cloud computing is a utility-oriented and Internet-centric way of delivering IT services on demand. The first chapter is a theoretical introduction on cloud computing, analyzing the main aspects, the keywords, and the technologies behind clouds, as well as the reasons for the success of this technology and its problems. After the introduction section, I will briefly describe the three main cloud platforms in the market. During this project we developed a simple Social Network. Consequently in the third chapter I will analyze the social network development, with the initial solution realized through Amazon Web Services and the steps we took to obtain the final version using Google Cloud Platform with its charateristics. To conclude, the last section is specific for the data elaboration and contains a initial theoretical part that describes MapReduce and Hadoop followed by a description of our analysis. We used Google App Engine to execute these elaborations on a large dataset. I will explain the basic idea, the code and the problems encountered.
APA, Harvard, Vancouver, ISO, and other styles
13

Wang, Tony. "A Service for Provisioning Compute Infrastructure in the Cloud." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-393349.

Full text
Abstract:
The amount of data has grown tremendously over the last decade. Cloud computing is a solution to handle large-scale computations and immense data sets. However, cloud computing comes with a multitude of challenges that scientist who are using the data have to tackle. Provisioning and orchestration cloud infrastructure is a challenge in itself with a wide variety of applications and cloud providers that are available. This thesis explores the idea of simplifying the provisioning of computing cloud applications in the cloud. The result of this work is a service which can seamlessly provision and execute cloud computations using different applications and cloud providers.
APA, Harvard, Vancouver, ISO, and other styles
14

Zhang, Bo. "Self-optimization of infrastructure and platform resources in cloud computing." Thesis, Lille 1, 2016. http://www.theses.fr/2016LIL10207/document.

Full text
Abstract:
L’élasticité est pensée comme une solution importante pour gérer des problèmes de performance dans les systèmes répartis. Toutefois, la plupart des recherches d’élasticité ne concernent que l’approvisionnement de ressources de manière automatique, mais ignorent toujours l’utilisation des ressources provisionnées. Cela pourrait conduire à des fuites de ressources, ce qui entraîne des dépenses inutiles. Pour éviter des problèmes, mes recherches se concentrent donc sur la maximisation de l’utilisation des ressources par l’auto-gestion des ressources. Dans cette thèse, en raison de divers problèmes de l’utilisation des ressources dans les différentes couches, je propose deux approches de gestion de ressources correspondant à l’infrastructure et la plate-forme, respectivement. Pour surmonter la limitation des infrastructures, je propose CloudGC comme service de middleware qui vise à libérer les ressources occupées par machines virtuelles qui tournent au ralenti. Dans la couche de plate-forme, une approche d’auto-équilibrage est introduite pour ajuster la configuration de Hadoop lors de l’exécution, ce qui optimise dynamiquement les performances de système. Enfin, cette thèse aussi concerne le déploiement rapide de Hadoop. Un nouvel outil, appelé "hadoop-benchmark", applique docker pour accélérer l’installation de Hadoop et fournir un ensemble d’images de docker qui contiennent plusieurs Hadoop benchmarks bien connus. Les évaluations montrent que ces approches et cet outil peuvent parvenir à l’auto-gestion des ressources dans différentes couches, puis de faciliter l’élasticité de l’infrastructure et de la plate-forme dans le Cloud computing
Elasticity is considered as an important solution to handle the performance issues in scalable distributed system. However, most of the researches of elasticity only concern the provisioning and de-provisioning resources in automatic ways, but always ignore the resource utilization of provisioned resources. This might lead to resource leaks while provisioning redundant resources, thereby causing unnecessary expenditure. To avoid the resource leaks and redundant resources, my research therefore focus on how to maximize resource utilization by self resource management. In this thesis, relevant to diverse problems of resource usage and allocation in different layers, I propose two resource management approaches corresponding to infrastructure and platform, respectively. To overcome infrastructure limitation, I propose CloudGC as middleware service which aims to free occupied resources by recycling idle VMs. In platform-layer, a self-balancing approach is introduced to adjust Hadoop configuration at runtime, thereby avoiding memory loss and dynamically optimizing Hadoop performance. Finally, this thesis concerns rapid deployment of service which is also an issue of elasticity. A new tool, named "hadoop-benchmark", applies docker to accelerate the installation of Hadoop cluster and to provide a set of docker images which contain several well-known Hadoop benchmarks.The assessments show that these approaches and tool can well achieve resource management and self-optimization in various layers, and then facilitate the elasticity of infrastructure and platform in scalable platform, such as Cloud computing
APA, Harvard, Vancouver, ISO, and other styles
15

Wicaksana, Arief. "Infrastructure portable pour un système hétérogène reconfigurable dans un environnement de cloud-FPGA." Thesis, Université Grenoble Alpes (ComUE), 2018. http://www.theses.fr/2018GREAT088/document.

Full text
Abstract:
La haute performance ainsi que la basse consommation d’énergie offertes par lesField-Programmable Gate Arrays (FPGAs) contribuent à leur popularité en tant queaccélérateurs matériels. Cet argument a été soutenu par les intégrations récentes des FPGAs dans des systèmes cloud et centre de données. Toutefois, le potentiel d’une architecture reconfigurable peut être encore optimisé en traitant les FPGAs comme une ressource virtualisée et en les offrant une capacité de multitâche. La solution pour interrompre une tâche sur FPGAs à pour objectif d’effectuer un changement de contexte matériel (hardware context switch) a été un sujet de recherche depuis des nombreuses années. Les travaux précédents ont principalement proposé une stratégie pour extraire le contexte d’une tâche en cours de son exécution d’un FPGA pour offrir la possibilité de sa reprise plus tard. Cependant, la communication tout au long du processus n’a pas reçu autant d’attention.Dans cette thèse, nous étudions la gestion de communication d’une tâche matérielle durant son changement de contexte. Cette gestion de communicationest nécessaire pour garantir la cohérence de la communication d’une tâche dans un système reconfigurable avec la capacité de changement de contexte. Autrement, un changement de contexte matériel est seulement autorisé sous des contraintes restrictifs; il est possible après que les flux de communication soient fini et que toutes les données d’entrées/de sorties sont déjà consommées. De plus, certaines techniques demandent l’homogénéité au sein de la plate-forme pour qu’un changement de contexte matériel puisse se réaliser.Nous présentons içi un mécanisme qui conserve la cohérence de communication durant un changement de contexte matériel dans une architecture reconfigurable. Les données de communication sont gérées avec le contexte de tâche pour assurer leur intégrité. La gestion du contexte et les données de communication suivent un protocole spécifique pour des architectures hétérogènes reconfigurables. Ce protocole permet donc un changement de contexte matériel pendant que la tâche a encore des flux de communication. À partir des expérimentations, nous découvrons que le surcoût de la gestion de communication devient négligeable car notre mécanisme fournit une grande réactivité nécessaire pour l’allocation de tâche de façon préemptive - outre que sa consistance de communication. Enfin, les applications de solution proposée sont présentées dans un prototypage de tâche migration et dans un système utilisant un hyperviseur
Field-Programmable Gate Arrays (FPGAs) have been gaining popularity as hardware accelerators in heterogeneous architectures thanks to their high performance and low energy consumption. This argument has been supported by the recent integration of FPGA devices in cloud services and data centers. The potential offered by the reconfigurable architectures can still be optimized by treating FPGAs as virtualizable resources and offering them multitasking capability. The solution to preempt a hardware task on an FPGA with the objective of context switching it has been in research for many years. The previous works mainly proposed the strategy to extract the context of a running task from the FPGA to provide the possibility of its resumption at a later time. The communication during the process, on the contrary, has not been receiving much attention.In this work, we study the communication management of a hardware task whileit is being context switched. This communication management is necessary to ensure the consistency in the communication of a task with context switch capability in a reconfigurable system. Otherwise, a hardware context switch can only be allowed under restrictive constraints which may lead to a considerable penalty in performance; context switching a task is possible after the communication flows finish and the input/output data have been consumed. Furthermore, certain techniques demand homogeneity in the platform for a hardware context switch can take place.We present a mechanism which preserves the communication consistency during ahardware context switch in a reconfigurable architecture. The input/output communication data are managed together with the task context to ensure their integrity. The overall management of the hardware task context and communication data follows a dedicated protocol developed for heterogeneous reconfigurable architectures. This protocol thus allows a hardware context switch to take place while the task still has ongoing communication flows on Reconfigurable System-on-Chips (RSoCs). From the experiments, we discover that the overhead due to managing the communication data becomes negligible since our mechanism provides the necessary high responsiveness for preemptive scheduling, besides the consistency in communication. Finally, the applications of the proposed solution are presented in a task migration prototyping and in a hypervisor-based system
APA, Harvard, Vancouver, ISO, and other styles
16

Okonski, Aleksander. "Implementing Security Rules, Safeguards, and IPS tools for Private Cloud Infrastructures : GROOT: Infrastructure Security as a Service(ISaaS)." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-345658.

Full text
Abstract:
Clouds are a large part of todays computing space, with individuals having the ability to crate their own cloud. The Open Stack cloud suite eases the deployment and management of cloud services. However, security is one core area that is difficult to isolate and has to be addressed at each level, ranging from low-level system securityto the user-facing multi-tenant environments. There are solutions available that offer end-to-end security but most of them are proprietary and with their sophisticated licensing scheme, expertise that might be affordable for large enterprises but difficult for medium and smallscale organizations is required. The aim of this project is to design a minimalistic security service for the Open Stack environment that helps cloud administrators get first hand information regarding any activities that may cause threats to the instance or the whole tenant in the cloud infrastructure. The project created a proof of concept system that once deployed was able to detect potential misconfigurations and threats. The system was tested in a real world scenarios and proved to work, finding several machines that were launched without correct configurations.Tryckt av:
APA, Harvard, Vancouver, ISO, and other styles
17

Bianchi, Stefano. "Design and Implementation of a Cloud Infrastructure for Distributed Scientific Calculation." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2016.

Find full text
Abstract:
Cloud computing enables independent end users and applications to share data and pooled resources, possibly located in geographically distributed Data Centers, in a fully transparent way. This need is particularly felt by scientific applications to exploit distributed resources in efficient and scalable way for the processing of big amount of data. This paper proposes an open so- lution to deploy a Platform as a service (PaaS) over a set of multi- site data centers by applying open source virtualization tools to facilitate operation among virtual machines while optimizing the usage of distributed resources. An experimental testbed is set up in Openstack environment to obtain evaluations with different types of TCP sample connections to demonstrate the functionality of the proposed solution and to obtain throughput measurements in relation to relevant design parameters.
APA, Harvard, Vancouver, ISO, and other styles
18

Croubois, Hadrien. "Toward an autonomic engine for scientific workflows and elastic Cloud infrastructure." Thesis, Lyon, 2018. http://www.theses.fr/2018LYSEN061/document.

Full text
Abstract:
Les infrastructures de calcul scientifique sont en constante évolution, et l’émergence de nouvelles technologies nécessite l’évolution des mécanismes d’ordonnancement qui leur sont associé. Durant la dernière décennie, l’apparition du modèle Cloud a suscité de nombreux espoirs, mais l’idée d’un déploiement et d’une gestion entièrement automatique des plates-formes de calcul est jusque la resté un voeu pieu. Les travaux entrepris dans le cadre de ce doctorat visent a concevoir un moteur de gestion de workflow qui intègre les logiques d’ordonnancement ainsi que le déploiement automatique d’une infrastructure Cloud. Plus particulièrement, nous nous intéressons aux plates-formes Clouds disposant de système de gestion de données de type DaaS (Data as a Service). L’objectif est d’automatiser l’exécution de workflows arbitrairement complexe, soumis de manière indépendante par de nombreux utilisateurs, sur une plate-forme Cloud entièrement élastique. Ces travaux proposent une infrastructure globale, et décrivent en détail les différents composants nécessaires à la réalisation de cette infrastructure :• Un mécanisme de clustering des tâches qui prend en compte les spécificités des communications via un DaaS ;• Un moteur décentralisé permettant l’exécution des workflows découpés en clusters de tâches ;• Un système permettant l’analyse des besoins et le déploiement automatique. Ces différents composants ont fait l’objet d’un simulateur qui a permis de tester leur comportement sur des workflows synthétiques ainsi que sur des workflows scientifiques réels issues du LBMC (Laboratoire de Biologie et Modélisation de la Cellule). Ils ont ensuite été implémentés dans l’intergiciel Diet. Les travaux théoriques décrivant la conception des composants, et les résultats de simulations qui les valident, ont été publié dans des workshops et conférences de portée internationale
The constant development of scientific and industrial computation infrastructures requires the concurrent development of scheduling and deployment mechanisms to manage such infrastructures. Throughout the last decade, the emergence of the Cloud paradigm raised many hopes, but achieving full platformautonomicity is still an ongoing challenge. Work undertaken during this PhD aimed at building a workflow engine that integrated the logic needed to manage workflow execution and Cloud deployment on its own. More precisely, we focus on Cloud solutions with a dedicated Data as a Service (DaaS) data management component. Our objective was to automate the execution of workflows submitted by many users on elastic Cloud resources.This contribution proposes a modular middleware infrastructure and details the implementation of the underlying modules:• A workflow clustering algorithm that optimises data locality in the context of DaaS-centeredcommunications;• A dynamic scheduler that executes clustered workflows on Cloud resources;• A deployment manager that handles the allocation and deallocation of Cloud resources accordingto the workload characteristics and users’ requirements. All these modules have been implemented in a simulator to analyse their behaviour and measure their effectiveness when running both synthetic and real scientific workflows. We also implemented these modules in the Diet middleware to give it new features and prove the versatility of this approach.Simulation running the WASABI workflow (waves analysis based inference, a framework for the reconstruction of gene regulatory networks) showed that our approach can decrease the deployment cost byup to 44% while meeting the required deadlines
APA, Harvard, Vancouver, ISO, and other styles
19

De, Souza Felipe Rodrigo. "Scheduling Solutions for Data Stream Processing Applications on Cloud-Edge Infrastructure." Thesis, Lyon, 2020. http://www.theses.fr/2020LYSEN082.

Full text
Abstract:
L’évolution des technologies ont conduit à une forte connexion entre les applications et le matériel produisant des quantités de données en perpétuelle augmentation. Ces données sont utilisées par les entreprises, les organisations et les individus pour prendre des décisions quotidiennes. Pour que les données collectées soient réellement utiles il convient de les traiter à temps et donc suffisamment rapidement. La vitesse à laquelle les informations sont extraites depuis les données générées par un système ou un environnement surveillé a un impact sur la capacité des entités (entreprises, organisations ou individus) à réagir aux changements. Une solution pour le traitement des données dans un délais réduit consiste à utiliser des applications de traitement de flux de données.Les applications de traitement de flux de données peuvent être modélisées sous forme de graphes orientés, où les sommets sont des sources de données, des opérateurs ou des récepteurs de données(i.e., data sinks), et les arêtes représentent les flux de données entre les opérateurs. Une source de données est un composant d’application responsable de la génération des données. Les opérateurs reçoivent un flux de données, appliquent une transformation ou effectuent une fonction définie par l’utilisateur sur le flux de données entrant et produisent un nouveau flux de sortie, jusqu’à ce que ce dernier atteigne un récepteur de données,où les données sont alors stockées, visualisées ou envoyées à une autre application. Habituellement, les applications de traitement de flux de données sont conçues pour fonctionner sur des infrastructures cloud ou sur une grappe homogène de ressources (i.e., cluster) en raison du nombre de ressources que ces infrastructures peuvent fournir et de la bonne connectivité de leur réseau. Dans les scénarios où les données utilisées par l’application de traitement du flux de données sont produites dans le cloud lui-même alors le déploiement de l’ensemble de l’application sur le cloud est une approche pertinente. Cependant, à l’heure où l’Internet des objets devient de plus en plus omniprésent, il existe un nombre croissant de scénarios où les applications de traitement de flux de données consomment des flux de données générés à la périphérie du réseau (via les nombreux appareils et capteurs répartis géographiquement). Dans de tels la bonne connectivité de leur réseau. Dans les scénarios où les données utilisées par l’application de traitement du flux de données sont produites dans le cloud lui-même alors le déploiement de l’ensemble de l’application sur le cloud est une approche pertinente.Cependant, à l’heure où l’Internet des objets devient de plus en plus omniprésent, il existe un nombre croissant de scénarios où les applications de traitement de flux de données consomment des flux de données générés à la périphérie du réseau (via les nombreux appareils et capteurs répartis géographiquement). Dans de tels scénarios, l’envoi de toutes les données via Internet pour être traitées sur un cloud distant, loin de la périphérie du réseau d’où proviennent les données, conduirait à générer un trafic réseau considérable. Cela augmente ainsi de façon significative la latence de bout en bout pour l’application; c’est-à-dire, le délai entre le moment où les données sont collectées et la fin du traitement. L’informatique de périphérie (edge computing) est devenu un paradigme pour alléger les tâches de traitement du cloud vers des ressources situées plus près des sources de données. Bien que l’utilisation combinée de ces ressources soit parfois appelée fog computing, la communauté scientifique ne semble pas avoir atteint un consensus sur la terminologie. Nous appelons la combinaison de ressources cloud et de ressources périphériques une infrastructure cloud-edge
Technology has evolved to a point where applications and devicesare highly connected and produce ever-increasing amounts of dataused by organizations and individuals to make daily decisions. Forthe collected data to become information that can be used indecision making, it requires processing. The speed at whichinformation is extracted from data generated by a monitored systemTechnology has evolved to a point where applications and devicesare highly connected and produce ever-increasing amounts of dataused by organizations and individuals to make daily decisions. Forthe collected data to become information that can be used indecision making, it requires processing. The speed at whichinformation is extracted from data generated by a monitored systemor environment affects how fast organizations and individuals canreact to changes. One way to process the data under short delays isthrough Data Stream Processing (DSP) applications. DSPapplications can be structured as directed graphs, where the vertexesare data sources, operators, and data sinks, and the edges arestreams of data that flow throughout the graph. A data source is anapplication component responsible for data ingestion. Operatorsreceive a data stream, apply some transformation or user-definedfunction over the data stream and produce a new output stream,until the latter reaches a data sink, where the data is stored,visualized or provided to another application
APA, Harvard, Vancouver, ISO, and other styles
20

Soares, João Monteiro. "Integration of the cloud computing paradigm with the opeerator network’s infrastructure." Doctoral thesis, Universidade de Aveiro, 2015. http://hdl.handle.net/10773/14854.

Full text
Abstract:
Doutoramento em Engenharia Informática
The proliferation of Internet access allows that users have the possibility to use services available directly through the Internet, which translates in a change of the paradigm of using applications and in the way of communicating, popularizing in this way the so-called cloud computing paradigm. Cloud computing brings with it requirements at two different levels: at the cloud level, usually relying in centralized data centers, where information technology and network resources must be able to guarantee the demand of such services; and at the access level, i.e., depending on the service being consumed, different quality of service is required in the access network, which is a Network Operator (NO) domain. In summary, there is an obvious network dependency. However, the network has been playing a relatively minor role, mostly as a provider of (best-effort) connectivity within the cloud and in the access network. The work developed in this Thesis enables for the effective integration of cloud and NO domains, allowing the required network support for cloud. We propose a framework and a set of associated mechanisms for the integrated management and control of cloud computing and NO domains to provide endto- end services. Moreover, we elaborate a thorough study on the embedding of virtual resources in this integrated environment. The study focuses on maximizing the host of virtual resources on the physical infrastructure through optimal embedding strategies (considering the initial allocation of resources as well as adaptations through time), while at the same time minimizing the costs associated to energy consumption, in single and multiple domains. Furthermore, we explore how the NO can take advantage of the integrated environment to host traditional network functions. In this sense, we study how virtual network Service Functions (SFs) should be modelled and managed in a cloud environment and enhance the framework accordingly. A thorough evaluation of the proposed solutions was performed in the scope of this Thesis, assessing their benefits. We implemented proof of concepts to prove the added value, feasibility and easy deployment characteristics of the proposed framework. Furthermore, the embedding strategies evaluation has been performed through simulation and Integer Linear Programming (ILP) solving tools, and it showed that it is possible to reduce the physical infrastructure energy consumption without jeopardizing the virtual resources acceptance. This fact can be further increased by allowing virtual resource adaptation through time. However, one should have in mind the costs associated to adaptation processes. The costs can be minimized, but the virtual resource acceptance can be also reduced. This tradeoff has also been subject of the work in this Thesis.
A proliferação do acesso à Internet permite aos utilizadores usar serviços disponibilizados diretamente através da Internet, o que se traduz numa mudança de paradigma na forma de usar aplicações e na forma de comunicar, popularizando desta forma o conceito denominado de cloud computing. Cloud computing traz consigo requisitos a dois níveis: ao nível da própria cloud, geralmente dependente de centros de dados centralizados, onde as tecnologias de informação e recursos de rede têm que ser capazes de garantir as exigências destes serviços; e ao nível do acesso, ou seja, dependendo do serviço que esteja a ser consumido, são necessários diferentes níveis de qualidade de serviço na rede de acesso, um domínio do operador de rede. Em síntese, existe uma clara dependência da cloud na rede. No entanto, o papel que a rede tem vindo a desempenhar neste âmbito é reduzido, sendo principalmente um fornecedor de conectividade (best-effort) tanto no dominio da cloud como no da rede de acesso. O trabalho desenvolvido nesta Tese permite uma integração efetiva dos domínios de cloud e operador de rede, dando assim à cloud o efetivo suporte da rede. Para tal, apresentamos uma plataforma e um conjunto de mecanismos associados para gestão e controlo integrado de domínios cloud computing e operador de rede por forma a fornecer serviços fim-a-fim. Além disso, elaboramos um estudo aprofundado sobre o mapeamento de recursos virtuais neste ambiente integrado. O estudo centra-se na maximização da incorporação de recursos virtuais na infraestrutura física por meio de estratégias de mapeamento ótimas (considerando a alocação inicial de recursos, bem como adaptações ao longo do tempo), enquanto que se minimizam os custos associados ao consumo de energia. Este estudo é feito para cenários de apenas um domínio e para cenários com múltiplos domínios. Além disso, exploramos como o operador de rede pode aproveitar o referido ambiente integrado para suportar funções de rede tradicionais. Neste sentido, estudamos como as funções de rede virtualizadas devem ser modeladas e geridas num ambiente cloud e estendemos a plataforma de acordo com este conceito. No âmbito desta Tese foi feita uma avaliação extensa das soluções propostas, avaliando os seus benefícios. Implementámos provas de conceito por forma a demonstrar as mais-valias, viabilidade e fácil implantação das soluções propostas. Além disso, a avaliação das estratégias de mapeamento foi realizada através de ferramentas de simulação e de programação linear inteira, mostrando que é possível reduzir o consumo de energia da infraestrutura física, sem comprometer a aceitação de recursos virtuais. Este aspeto pode ser melhorado através da adaptação de recursos virtuais ao longo do tempo. No entanto, deve-se ter em mente os custos associados aos processos de adaptação. Os custos podem ser minimizados, mas isso implica uma redução na aceitação de recursos virtuais. Esta compensação foi também um tema abordado nesta Tese.
APA, Harvard, Vancouver, ISO, and other styles
21

Retana, Solano German F. "Technology support and demand for cloud infrastructure services: the role of service providers." Diss., Georgia Institute of Technology, 2013. http://hdl.handle.net/1853/50236.

Full text
Abstract:
Service providers have long recognized that their customers play a vital role in the service delivery process since they are not only recipients but also producers, or co-producers, of the service delivered. Moreover, in the particular context of self-service technology (SST) offerings, it is widely recognized that customers’ knowledge, skills and abilities in co-producing the service are key determinants of the services’ adoption and usage. However, despite the importance of customers’ capabilities, prior research has not yet paid much attention to the mechanisms by which service providers can influence them and, in turn, how the providers’ efforts affect customers’ use of the service. This dissertation addresses research questions associated with the role of a provider’s technology support and education in influencing customer use of an SST, namely public cloud computing infrastructure services. The unique datasets used to answer these research questions were collected from one of the major global providers in the cloud infrastructure services industry. This research context offers an excellent opportunity to study the role of technology support since, when adapting the standardized and commoditized components of the cloud service to their individual needs, customers may face important co-production costs that can be mitigated by the provider’s assistance. Specifically, customers must configure their computing servers and deploy their software applications on their own, relying on their own capabilities. Moreover, the cloud’s offering of on-demand computing servers through a fully pay-per-use model allows us to directly observe variation in the actual use customers make of the service. The first study of this dissertation examines how varying levels of technology support, which differ in the level of participation and assistance of the provider in customers’ service co-production process, influence the use that customers make of the service. The study matches and compares 20,179 firms that used the service between March 2009 and August 2012, and who over time accessed one of the two levels of support available: full and basic. Using fixed effects panel data models and a difference-in-difference identification strategy, we find that customers who have access to full support or accessed it in the past use (i.e., consume) more of the service than customers who have only accessed basic support. Moreover, the provider’s involvement in the co-production process is complementary with firm size in the sense that larger firms use more of the service than smaller ones if they upgrade from basic to full support. Finally, the provider’s co-participation through full support also has a positive influence on the effectiveness with which buyers make use of the service. Firms that access full support are more likely to deploy computing architectures that leverage on the cloud’s advanced features. The second study examines the value of early proactive education, which is defined as any provider-initiated effort to increase its customers’ service co-production related knowledge and skills immediately after service adoption. The study analyzes the outcome of a field experiment executed by the provider between October and November 2011, during which 366 randomly-selected customers out of 2,673 customers that adopted during the field experiment period received early proactive education treatment. The treatment consisted in a short phone call followed up by a support ticket through which the provider offered initial guidance on how to use the basic features of the service. We use survival analysis (i.e., hazard models) to compare the treatment’s effect on customer retention, and find that it reduces by half the number of customers who leave the service offering during the first week. We also use count data models to examine the treatment’s effect on customers’ demand for technology support, and find that the treated customers ask about 19.55% fewer questions during the first week of their lifetimes than the controls.
APA, Harvard, Vancouver, ISO, and other styles
22

Amarasinghe, Heli. "Network Resource Management in Infrastructure-as-a-Service Clouds." Thesis, Université d'Ottawa / University of Ottawa, 2019. http://hdl.handle.net/10393/39141.

Full text
Abstract:
Cloud Infrastructure-as-a-Service (IaaS) is a form of utility computing which has emerged with the recent innovations in the service computing and data communication technologies. Regardless of the fact that IaaS is attractive for application service providers, satisfying user requests while ensuring cloud operational objectives is a complicated task that raises several resource management challenges. Among these challenges, limited controllability over network services delivered to cloud consumers is prominent in single datacenter cloud environments. In addition, the lack of seamless service migration and optimization, poor infrastructure utilization, and unavailability of efficient fault tolerant techniques are noteworthy challenges in geographically distributed datacenter clouds. Initially in this thesis, a datacenter resource management framework is presented to address the challenge of limited controllability over cloud network traffic. The proposed framework integrates network virtualization functionalities offered by software defined networking (SDN) into cloud ecosystem. To provide rich traffic control features to IaaS consumers, control plane virtualization capabilities offered by SDN have been employed. Secondly, a quality of service (QoS) aware seamless service migration and optimization framework has been proposed in the context of geo-distributed datacenters. Focus has been given to a mobile end-user scenario where frequent cloud service migrations are required to mitigate QoS violations. Finally, an SDN-based dynamic fault restoration scheme and a shared backup-based fault protection scheme have been proposed. The fault restoration has been achieved by introducing QoS-aware reactive and shared risk link group-aware proactive path computation algorithms. Shared backup protection has been achieved by optimizing virtual and backup link embedding through a novel integer linear programming approach. The proposed solutions significantly improve bandwidth utilization in inter-datacenter networks while recovering from substrate link failures.
APA, Harvard, Vancouver, ISO, and other styles
23

Croker, Trevor D. "Formation of the Cloud: History, Metaphor, and Materiality." Diss., Virginia Tech, 2020. http://hdl.handle.net/10919/96439.

Full text
Abstract:
In this dissertation, I look at the history of cloud computing to demonstrate the entanglement of history, metaphor, and materiality. In telling this story, I argue that metaphors play a powerful role in how we imagine, construct, and maintain our technological futures. The cloud, as a metaphor in computing, works to simplify complexities in distributed networking infrastructures. The language and imagery of the cloud has been used as a tool that helps cloud providers shift public focus away from potentially important regulatory, environmental, and social questions while constructing a new computing marketplace. To address these topics, I contextualize the history of the cloud by looking back at the stories of utility computing (1960s-70s) and ubiquitous computing (1980s-1990s). These visions provide an alternative narrative about the design and regulation of new technological systems. Drawing upon these older metaphors of computing, I describe the early history of the cloud (1990-2008) in order to explore how this new vision of computing was imagined. I suggest that the metaphor of the cloud was not a historical inevitability. Rather, I argue that the social-construction of metaphors in computing can play a significant role in how the public thinks about, develops, and uses new technologies. In this research, I explore how the metaphor of the cloud underplays the impact of emerging large-scale computing infrastructures while at the same time slowly transforming traditional ownership-models in digital communications. Throughout the dissertation, I focus on the role of materiality in shaping digital technologies. I look at how the development of the cloud is tied to the establishment of cloud data centers and the deployment of global submarine data cables. Furthermore, I look at the materiality of the cloud by examining its impact on a local community (Los Angeles, CA). Throughout this research, I argue that the metaphor of the cloud often hides deeper socio-technical complexities. Both the materials and metaphor of the cloud work to make the system invisible. By looking at the material impact of the cloud, I demonstrate how these larger economic, social, and political realities are entangled in the story and metaphor of the cloud.
Doctor of Philosophy
This dissertation tells the story of cloud computing by looking at the history of the cloud and then discussing the social and political implications of this history. I start by arguing that the cloud is connected to earlier visions of computing (specifically, utility computing and ubiquitous computing). By referencing these older histories, I argue that much of what we currently understand as cloud computing is actually connected to earlier debates and efforts to shape a computing future. Using the history of computing, I demonstrate the role that metaphor plays in the development of a technology. Using these earlier histories, I explain how cloud computing was coined in the 1990s and eventually became a dominant vision of computing in the late 2000s. Much of the research addresses how the metaphor of the cloud is used, the initial reaction to the idea of the cloud, and how the creation of the cloud did (or did not) borrow from older visions of computing. This research looks at which people use the cloud, how the cloud is marketed to different groups, and the challenges of conceptualizing this new distributed computing network. This dissertation gives particular weight to the materiality of the cloud. My research focuses on the cloud's impact on data centers and submarine communication data cables. Additionally, I look at the impact of the cloud on a local community (Los Angeles, CA). Throughout this research, I argue that the metaphor of the cloud often hides deeper complexities. By looking at the material impact of the cloud, I demonstrate how larger economic, social, and political realities are entangled in the story and metaphor of the cloud.
APA, Harvard, Vancouver, ISO, and other styles
24

Božić, Nikola. "Blockchain technologies and their application to secure virtualized infrastructure control." Electronic Thesis or Diss., Sorbonne université, 2019. http://www.theses.fr/2019SORUS596.

Full text
Abstract:
Blockchain est une technologie qui fait du concept de registre partagé à partir de systèmes distribués une réalité pour un certain nombre de domaines d’application, du crypto-monnaie à potentiellement tout système industriel nécessitant une prise de décision décentralisée, robuste, fiable et automatisée dans une situation à plusieurs parties prenantes. Néanmoins, les avantages réels de l’utilisation de la blockchain au lieu de toute autre solution traditionnelle (telle que des bases de données centralisées) ne sont pas complètement compris à ce jour, ni quel type de blockchain répond le mieux aux exigences du cas d'utilisation et à son utilisation. Au début, notre objectif est de fournir une sorte de « vademecum » à la communauté, tout en donnant une présentation générale de la blockchain qui dépasse son cas d'utilisation en des crypto monnaies telle que Bitcoin, et en passant en revue une sélection de la vaste littérature qui est apparue au cours des dernières années. Nous décrivons les exigences clés et leur évolution lorsque nous passons des blockchains publics à priver, en présentant les différences entre les mécanismes de consensus proposés et expérimentés, et en décrivant les plateformes de blockchain existantes. De plus, nous présentons la blockchain B-VMOA pour sécuriser les opérations d’orchestration de machines virtuelles pour les systèmes de cloud computing et de virtualisation des fonctions réseau en appliquant la logique de vademecum proposée. À l'aide d'exemples de didacticiels, nous décrivons nos choix de conception et élaborons des plans de mise en œuvre. Nous développons plus avant la logique de vademecum appliquée à l'orchestration dans le cloud et comment elle peut conduire à des spécifications de plateforme précises. Nous capturons les opérations du système clés et les interactions complexes entre elles. Nous nous concentrons sur la dernière version de la plateforme Hyperledger Fabric en tant que moyen de développer le système B-VMOA. De plus, Hyperledger Fabric optimise les performances, la sécurité et l’évolutivité conçues pour le réseau B-VMOA en séparant la charge de travail entre (i) les homologues d’exécution et de validation de transaction et (ii) les nœuds qui sont charges pour l'ordre des transactions. Nous étudions et utilisons une architecture <> qui différencie notre système B-VMOA conçu des systèmes distribués hérités qui suivent une architecture de réplication d'état de machine traditionnelle. Nous paramétrons et validons notre modèle avec les données recueillies sur un banc d'essai réaliste, en présentant une étude empirique pour caractériser les performances du système et identifier les goulots d'étranglement potentiels. En outre, nous présentons les outils que nous avons utilisés, la configuration du réseau et la discussion sur les observations empiriques issues de la collecte de données. Nous examinons l'impact de divers paramètres configurables pour mener une étude approfondie des composants principaux et des performances de référence pour les modèles d'utilisation courants. À savoir, B-VMOA est destiné à être exécuté dans un centre de données. Différentes topologies d'interconnexion de centres de données évoluent différemment en raison des protocoles de communication. Il semble difficile de concevoir efficacement les interconnexions réseau de manière à rentabiliser le déploiement et la maintenance de l’infrastructure. Nous analysons les propriétés structurelles de plusieurs topologies DCN et présentons également une comparaison entre ces architectures de réseau dans le but de réduire les coûts indirects de la technologie B-VMOA. D'après notre analyse, nous recommandons l'hypercube topologie comme solution pour remédier au goulot d'étranglement des performances dans le plan de contrôle B-VMOA provoqué par gossip, le protocole de diffusion, ainsi qu'une estimation de l'amélioration des performances
Blockchain is a technology making the shared registry concept from distributed systems a reality for a number of application domains, from the cryptocurrency one to potentially any industrial system requiring decentralized, robust, trusted and automated decision making in a multi-stakeholder situation. Nevertheless, the actual advantages in using blockchain instead of any other traditional solution (such as centralized databases) are not completely understood to date, or at least there is a strong need for a vademecum guiding designers toward the right decision about when to adopt blockchain or not, which kind of blockchain better meets use-case requirements, and how to use it. At first, we aim at providing the community with such a vademecum, while giving a general presentation of blockchain that goes beyond its usage in Bitcoin and surveying a selection of the vast literature that emerged in the last few years. We draw the key requirements and their evolution when passing from permissionless to permissioned blockchains, presenting the differences between proposed and experimented consensus mechanisms, and describing existing blockchain platforms. Furthermore, we present the B-VMOA blockchain to secure virtual machine orchestration operations for cloud computing and network functions virtualization systems applying the proposed vademecum logic. Using tutorial examples, we describe our design choices and draw implementation plans. We further develop the vademecum logic applied to cloud orchestration and how it can lead to precise platform specifications. We capture the key system operations and complex interactions between them. We focus on the last release of Hyperledger Fabric platform as a way to develop B-VMOA system. Besides, Hyperledger Fabric optimizes conceived B-VMOA network performance, security, and scalability by way of workload separation across: (i) transaction execution and validation peers, and (ii) transaction ordering nodes. We study and use a distributed execute-order-validate architecture which differentiates our conceived B-VMOA system from legacy distributed systems that follow a traditional state-machine replication architecture. We parameterize and validate our model with data collected from a realistic testbed, presenting an empirical study to characterize system performance and identify potential performance bottlenecks. Furthermore, we present the tools we used, the network setup and the discussion on empirical observations from the data collection. We examine the impact of various configurable parameters to conduct an in-dept study of core components and benchmark performance for common usage patterns. Namely, B-VMOA is meant to be run within data center. Different data center interconnection topologies scale differently due to communication protocols. Enormous challenges appear to efficiently design the network interconnections so that the deployment and maintenance of the infrastructure is cost-effective. We analyze the structural properties of several DCN topologies and also present some comparison among these network architectures with the aim to reduce B-VMOA overhead costs. From our analysis, we recommend the hypercube topology as a solution to address the performance bottleneck in the B-VMOA control plane caused by gossip dissemination protocol along with an estimate of performance improvement
APA, Harvard, Vancouver, ISO, and other styles
25

Meinel, Christoph, Christian Willems, Sebastian Roschke, and Maxim Schnjakin. "Virtualisierung und Cloud Computing : Konzepte, Technologiestudie, Marktübersicht." Universität Potsdam, 2011. http://opus.kobv.de/ubp/volltexte/2011/4970/.

Full text
Abstract:
Virtualisierung und Cloud Computing gehören derzeit zu den wichtigsten Schlagworten für Betreiber von IT Infrastrukturen. Es gibt eine Vielzahl unterschiedlicher Technologien, Produkte und Geschäftsmodelle für vollkommen verschiedene Anwendungsszenarien. Die vorliegende Studie gibt zunächst einen detaillierten Überblick über aktuelle Entwicklungen in Konzepten und Technologien der Virtualisierungstechnologie – von klassischer Servervirtualisierung über Infrastrukturen für virtuelle Arbeitsplätze bis zur Anwendungsvirtualisierung und macht den Versuch einer Klassifikation der Virtualisierungsvarianten. Bei der Betrachtung des Cloud Computing-Konzepts werden deren Grundzüge sowie verschiedene Architekturvarianten und Anwendungsfälle eingeführt. Die ausführliche Untersuchung von Vorteilen des Cloud Computing sowie möglicher Bedenken, die bei der Nutzung von Cloud-Ressourcen im Unternehmen beachtet werden müssen, zeigt, dass Cloud Computing zwar große Chancen bietet, aber nicht für jede Anwendung und nicht für jeden rechtlichen und wirtschaftlichen Rahmen in Frage kommt.. Die anschließende Marktübersicht für Virtualisierungstechnologie zeigt, dass die großen Hersteller – Citrix, Microsoft und VMware – jeweils Produkte für fast alle Virtualisierungsvarianten anbieten und hebt entscheidende Unterschiede bzw. die Stärken der jeweiligen Anbieter heraus. So ist beispielsweise die Lösung von Citrix für Virtual Desktop Infrastructures sehr ausgereift, während Microsoft hier nur auf Standardtechnologie zurückgreifen kann. VMware hat als Marktführer die größte Verbreitung in Rechenzentren gefunden und bietet als einziger Hersteller echte Fehlertoleranz. Microsoft hingegen punktet mit der nahtlosen Integration ihrer Virtualisierungsprodukte in bestehende Windows-Infrastrukturen. Im Bereich der Cloud Computing-Systeme zeigen sich einige quelloffene Softwareprojekte, die durchaus für den produktiven Betrieb von sogenannten privaten Clouds geeignet sind.
Virtualization and Cloud Computing belong to the most important issues for operators of large ICT infrastructures today. There are a large number of various technologies, products, and business models for entirely different application scenarios. The study at hand gives a detailed overview on latest developments in concepts and technologies of virtualization – beginning with classic server virtualization, continuing with infrastructures for virtual workplaces, through to application virtualization and makes an attempt to classify all these variants of virtualization. When investigating on the concepts of Cloud Computing, the report introduces basic principles as well as different types of architecture and use cases. The extensive analysis of benefits of Cloud Computing and possible reservations when using cloud resources within an enterprise context is evidence that Cloud Computing offers great opportunities, but is not worth considering for any kind of application scenario, legal framework or business scenario. The subsequent market study on virtualization technology shows that each of the major manufacturers – Citrix, Microsoft, and VMware – offer products for any variant of virtualization and highlights the important differences between the products and the respective strengths of the vendors. For example, the Citrix solution on Virtual Desktop Infrastructures comes up very well-engineered, while Microsoft can only rely on standard technology in this field. VMware, the market leader in virtualization technology, has gained the biggest popularity in data centers and offers the only product implementing real fault tolerance. On the other hand, Microsoft is able to score with seamless integration of their virtualization products into existing Windows-based infrastructures. In the area of Cloud Computing systems, there are some open source software projects that are very possibly suitable for the productive operation of so called private clouds.
APA, Harvard, Vancouver, ISO, and other styles
26

Salazar, Javier. "Resource allocation optimization algorithms for infrastructure as a service in cloud computing." Thesis, Sorbonne Paris Cité, 2016. http://www.theses.fr/2016USPCB074.

Full text
Abstract:
L’informatique, le stockage des données et les applications à la demande font partie des services offerts par l’architecture informatique en Nuage. Dans ce cadre, les fournisseurs de nuage (FN) agissent non seulement en tant qu’administrateurs des ressources d'infrastructure mais ils profitent aussi financièrement de la location de ces ressources. Dans cette thèse, nous proposons trois modèles d'optimisation du processus d'allocation des ressources dans le nuage dans le but de réduire les coûts générés et d’accroitre la qualité du service rendu. Cela peut être accompli en fournissant au FN les outils formels nécessaires pour réduire au minimum le prix des ressources dédiées à servir les requêtes des utilisateurs. Ainsi, la mise en œuvre des modèles proposés permettra non seulement l’augmentation des revenus du FN, mais aussi l’amélioration de la qualité des services offerts, ce qui enrichira l’ensemble des interactions qui se produisent dans le nuage. A cet effet, nous nous concentrons principalement sur les ressources de l’infrastructure en tant que service (IaaS), lesquels sont contenus dans des centres de données (DCs), et constituent l'infrastructure physique du nuage. Comme une alternative aux immenses DCs centralisés, la recherche dans ce domaine comprend l’installation de petits centres de données (Edge DCs) placés à proximité des utilisateurs finaux. Dans ce contexte nous adressons le problème d’allocation des ressources et pour ce faire nous utilisons la technique d'optimisation nommée génération de colonnes. Cette technique nous permet de traiter des modèles d'optimisation à grande échelle de manière efficace. La formulation proposée comprend à la fois, et dans une seule phase, les communications et les ressources informatiques à optimiser dans le but de servir les requêtes de service d'infrastructure. Sur la base de cette formulation, nous proposons également un deuxième modèle qui comprend des garanties de qualité de service toujours sous la même perspective d'allocation des ressources d’infrastructure en tant que service. Ceci nous permet de fournir plusieurs solutions applicables à divers aspects du même problème, tels que le coût et la réduction des délais, tout en offrant différents niveaux de service. En outre, nous introduisons le scénario informatique en nuage multimédia, qui, conjointement avec l'architecture des Edge DCs, résulte en l'architecture Multimédia Edge Cloud (MEC). Dans ce cadre, nous proposons une nouvelle approche pour l'allocation des ressources dans les architectures informatique en nuage multimédia lors du positionnement de ces DCs afin de réduire les problèmes liés à la communication, tels que la latence et la gigue. Dans cette formulation, nous proposons également de mettre en œuvre des technologies optiques de réseau de fibres pour améliorer les communications entre les DCs. Plusieurs travaux ont proposé de nouvelles méthodes pour améliorer la performance et la transmission de données. Dans nos travaux, nous avons décidé de mettre en œuvre le multiplexage en longueur d'onde (WDM) pour renforcer l'utilisation des liens et les chemins optiques dans le but de grouper différents signaux sur la même longueur d'onde. Un environnement de simulation réel est également présenté pour l’évaluation des performances et de l'efficacité des approches proposées. Pour ce faire, nous utilisons le scénario spécifié pour les DCs, et nous comparons par simulation nos modèles au moyen de différents critères de performances tel que l'impact de la formulation optique sur la performance du réseau. Les résultats numériques obtenus ont montré que, en utilisant nos modèles, le FN peut efficacement réduire les coûts d'allocation en maintenant toujours un niveau satisfaisant quant à l'acceptation de requêtes et la qualité du service
The cloud architecture offers on-demand computing, storage and applications. Within this structure, Cloud Providers (CPs) not only administer infrastructure resources but also directly benefit from leasing them. In this thesis, we propose three optimization models to assist CPs reduce the costs incurred in the resource allocation process when serving users’ demands. Implementing the proposed models will not only increase the CP’s revenue but will also enhance the quality of the services offered, benefiting all parties. We focus on Infrastructure as a Service (IaaS) resources which constitute the physical infrastructure of the cloud and are contained in datacenters (DCs). Following existing research in DC design and cloud computing applications, we propose the implementation of smaller DCs (Edge DCs) be located close to end users as an alternative to large centralized DCs. Lastly, we use the Column Generation optimization technique to handle large scale optimization models efficiently. The proposed formulation optimizes both the communications and information technology resources in a single phase to serve IaaS requests. Based on this formulation, we also propose a second model that includes QoS guarantees under the same Infrastructure as a Service resource allocation perspective, to provide different solutions to diverse aspects of the resource allocation problem such as cost and delay reduction while providing different levels of service. Additionally, we consider the multimedia cloud computing scenario. When Edge DCs architecture is applied to this scenario it results in the creation of the Multimedia Edge Cloud (MEC) architecture. In this context we propose a resource allocation approach to help with the placement of these DCs to reduce communication related problems such as jitter and latency. We also propose the implementation of optical fiber network technologies to enhance communication between DCs. Several studies can be found proposing new methods to improve data transmission and performance. For this study, we decided to implement Wavelength Division Multiplexing (WDM) to strengthen the link usage and light-paths and, by doing so, group different signals over the same wavelength. Using a realistic simulation environment, we evaluate the efficiency of the approaches proposed in this thesis using a scenario specifically designed for the DCs, comparing them with different benchmarks and also simulating the effect of the optical formulation on the network performance. The numerical results obtained show that by using the proposed models, a CP can efficiently reduce allocation costs while maintaining satisfactory request acceptance and QoS ratios
APA, Harvard, Vancouver, ISO, and other styles
27

Margaris, Nikos. "Cloud Computing v českém prostředí." Master's thesis, Vysoká škola ekonomická v Praze, 2011. http://www.nusl.cz/ntk/nusl-72541.

Full text
Abstract:
Focus of this diploma thesis is on Cloud Computing -- a new delivery model of IS/ICT services for companies. Cloud Computing services are IT resources available to users via internet on pay-as-you-use basis. The aim of the thesis is define underlying Cloud Computing concepts and evaluate the current state of Cloud Computing adoption in the Czech Republic. We discuss the framework for definition of Cloud Computing concepts in the theoretical part of the thesis drawing on resources available in the literature. Cloud Computing survey is conducted using an online questionnaire and followsthe principles established in theoretical part. The main contribution of this thesis is creating working definition of Cloud Computing and identifying reasons that affect adoption rate in the Czech Republic.
APA, Harvard, Vancouver, ISO, and other styles
28

Svärd, Petter. "Dynamic Cloud Resource Management : Scheduling, Migration and Server Disaggregation." Doctoral thesis, Umeå universitet, Institutionen för datavetenskap, 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-87904.

Full text
Abstract:
A key aspect of cloud computing is the promise of infinite, scalable resources, and that cloud services should scale up and down on demand. This thesis investigates methods for dynamic resource allocation and management of services in cloud datacenters, introducing new approaches as well as improvements to established technologies.Virtualization is a key technology for cloud computing as it allows several operating system instances to run on the same Physical Machine, PM, and cloud services normally consists of a number of Virtual Machines, VMs, that are hosted on PMs. In this thesis, a novel virtualization approach is presented. Instead of running each PM isolated, resources from multiple PMs in the datacenter are disaggregated and exposed to the VMs as pools of CPU, I/O and memory resources. VMs are provisioned by using the right amount of resources from each pool, thereby enabling both larger VMs than any single PM can host as well as VMs with tailor-made specifications for their application. Another important aspect of virtualization is live migration of VMs, which is the concept moving VMs between PMs without interruption in service. Live migration allows for better PM utilization and is also useful for administrative purposes. In the thesis, two improvements to the standard live migration algorithm are presented, delta compression and page transfer reordering. The improvements can reduce migration downtime, i.e., the time that the VM is unavailable, as well as the total migration time. Postcopy migration, where the VM is resumed on the destination before the memory content is transferred is also studied. Both userspace and in-kernel postcopy algorithms are evaluated in an in-depth study of live migration principles and performance.Efficient mapping of VMs onto PMs is a key problem for cloud providers as PM utilization directly impacts revenue. When services are accepted into a datacenter, a decision is made on which PM should host the service VMs. This thesis presents a general approach for service scheduling that allows for the same scheduling software to be used across multiple cloud architectures. A number of scheduling algorithms to optimize objectives like revenue or utilization are also studied. Finally, an approach for continuous datacenter consolidation is presented. As VM workloads fluctuate and server availability varies any initial mapping is bound to become suboptimal over time. The continuous datacenter consolidation approach adjusts this VM-to-PM mapping during operation based on combinations of management actions, like suspending/resuming PMs, live migrating VMs, and suspending/resuming VMs. Proof-of-concept software and a set of algorithms that allows cloud providers to continuously optimize their server resources are presented in the thesis.
APA, Harvard, Vancouver, ISO, and other styles
29

Yampolsky, Vincent. "An evaluation of the power consumption and carbon footprint of a cloud infrastructure." Thesis, Edinburgh Napier University, 2010. http://researchrepository.napier.ac.uk/Output/3973.

Full text
Abstract:
The Information and Communication Technology (ICT) sector represent two to three percentsof the world energy consumption and about the same percentage of GreenHouse Gas(GHG) emission. Moreover the IT-related costs represent fifty per-cents of the electricity billof a company. In January 2010 the Green Touch consortium composed of sixteen leading companies and laboratories in the IT field led by Bell's lab and Alcatel-Lucent have announced that in five years the Internet could require a thousand times less energy than it requires now. Furthermore Edinburgh Napier University is committed to reduce its carbon footprint by 25% on the 2007/8 to 2012/13 period (Edinburgh Napier University Sustainability Office, 2009) and one of the objectives is to deploy innovative C&IT solutions. Therefore there is a general interest to reduce the electrical cost of the IT infrastructure, usually led by environmental concerns. One of the most prominent technologies when Green IT is discussed is Cloud Computing (Stephen Ruth, 2009). This technology allows the on-demand self service provisioning by making resources available as a service. Its elasticity allows the automatic scaling of thedemand and hardware consolidation thanks to virtualization. Therefore an increasing number of companies are moving their resources into a cloud managed by themselves or a third party. However this is known to reduce the electricity bill of a company if the cloud is managed by a third-party off-premise but this does not say to which extent is the powerconsumption is reduced. Indeed the processing resources seem to be just located somewhere else. Moreover hardware consolidation suggest that power saving is achieved only during off-peak time (Xiaobo Fan et al, 2007). Furthermore the cost of the network is never mentioned when cloud is referred as power saving and this cost might not be negligible. Indeed the network might need upgrades because what was being done locally is done remotely with cloud computing. In the same way cloud computing is supposed to enhance the capabilities of mobile devices but the impact of cloud communication on their autonomy is mentioned anywhere. Experimentations have been performed in order to evaluate the power consumption of an infrastructure relying on a cloud used for desktop virtualization and also to measure the cost of the same infrastructure without a cloud. The overall infrastructure have been split in different elements respectively the cloud infrastructure, the network infrastructure and enddevices and the power consumption of each element have been monitored separately. The experimentation have considered different severs, network equipment (switches, wireless access-points, router) and end-devices (desktops Iphone, Ipad and Sony-Ericsson Xperia running Android). The experiments have also measured the impact of a cloud communication on the battery of mobile devices. The evaluation have considered different deployment sizes and estimated the carbon emission of the technologies tested. The cloud infrastructure happened to be power saving and not only during off-peak time from a deployment size large enough (approximately 20 computers) for the same processing power. The power saving is large enough for wide deployment (500 computers) that it could overcome the cost of a network upgrade to a Gigabit access infrastructure and still reduce the carbon emission by 4 tonnes or 43.97% over a year and on Napier campuses compared to traditional deployment with a Fast-Ethernet access-network. However the impact of cloud communication on mobile-devices is important and has increase the power consumption by 57% to 169%.
APA, Harvard, Vancouver, ISO, and other styles
30

Zia, Umar. "Enabling Context Awareness in Ambient Environments using Cloud Infrastructures." Thesis, Mittuniversitetet, Institutionen för informationsteknologi och medier, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:miun:diva-15712.

Full text
Abstract:
This thesis was prepared in collaboration with Acreo (Research and Development Company) and OpenCare (IT Company). Its purpose was to design and develop a research testbed to enable context awareness in pervasive environments by modifying the MediaSense framework (EU Funded project) of Mid Sweden University. In ubiquitous environments, the proliferation of devices such as sensors, active badges, mobile phones, RFID and NfC tags enables the development of intelligent services towards new forms of pervasive applications. These intelligent services seamlessly gather context information and the benefits offered are in the provision of better services. The inspiration given by these intelligent services has meant that the focus of thesis has been on using these services in a healthcare application. The challenge is that the proposed testbed should address the intelligent delivery of health information to any host, anywhere, based on the user’s context. Further, context reasoning requires substantial computing power and smart devices have limited resources in terms of processing and memory, so, the testbed should enable smart communication to take place between these devices. The proposed solution satisfies the stated requirements by using a cloud infrastructure and a Distributed Context eXchange Protocol (DCXP). Any device on the internet that is DCXP capable may register with the architecture and share context information in an efficient way. In order to view context information, TV, smartphones, internet tablets and web interfaces have been provided for both the user and the health centre. By successfully addressing the requirements of the testbed, this thesis has enabled the creation of a pervasive healthcare application. Hence, this thesis concludes with the observation that the proposed approach for context awareness in a healthcare system has the ability to deal with the stated challenges.
APA, Harvard, Vancouver, ISO, and other styles
31

Ahmadi, Mehri Vida. "An Investigation of CPU utilization relationship between host and guests in a Cloud infrastructure." Thesis, Blekinge Tekniska Högskola, Institutionen för kommunikationssystem, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-10809.

Full text
Abstract:
Cloud computing stands as a revolution in IT world in recent years. This technology facilitates resource sharing by reducing hardware costs for business users and promises energy efficiency and better resource utilization to the service providers. CPU utilization is a key metric considered in resource management across clouds. The main goal of this thesis study is directed towards investigating CPU utilization behavior with regard to host and guest, which would help us in understanding the relationship between them. It is expected that perception of these relationships would be helpful in resource management. Working towards our goal, the methodology we adopted is experi- mental research. This involves experimental modeling, measurements and observations from the results. The experimental setup covers sev- eral complex scenarios including cloud and a standalone virtualization system. The results are further analyzed for a visual correlation. Results show that CPU utilization in cloud and virtualization sce- nario coincides. More experimental scenarios are designed based on the first observations. The obtaining results show the irregular behav- ior between PM and VM in variable workload. CPU utilization retrieved from both cloud and a standalone system is similar. 100% workload situations showed that CPU utilization is constant with no correlation co-efficient obtained. Lower workloads showed (more/less) correlation in most of the cases in our correlation analysis. It is expected that more number of iterations can possibly vary the output. Further analysis of these relationships for proper resource management techniques will be considered.
APA, Harvard, Vancouver, ISO, and other styles
32

Manco, Filipe José Jesus. "Network infrastructure control for virtual campus." Master's thesis, Universidade de Aveiro, 2013. http://hdl.handle.net/10773/12725.

Full text
Abstract:
Mestrado em Engenharia Electrónica e Telecomunicações
An evolution of the current university’s networking infrastructure and service models has been shown to be necessary to enable the provisioning of innovative services that are able to respond to today’s needs. On this work a network orchestration tool is proposed that, integrated with the OpenStack cloud framework, is able to virtualize any network deployment in a non-disruptive manner, providing a clean upgrade path from the traditional networking to the world of virtualization. The framework is able to extend virtual networks created on the datacenter by OpenStack or other cloud frameworks, to the outside campus. It does so by directly reconfiguring the network devices according to the needs, independently of the device vendor, the type of device or its specificities, and independently of the specific physical network topology. This service is provided to the end user using a cloud like service model, much more flexible than the current one, properly integrated with the cloud framework services. The project was developed with the Aveiro University use cases in mind, but the final result can be applied in many academic or business environments. The framework is presented both from a conceptual perspective, by describing the abstractions and mechanisms created, and from an implementation perspective, provinding the reader the necessary understanding about the framework operation and the integration with OpenStack. The integration of this work with a broader vision for the future of the university’s services is left as future work.
A evolução da actual infraestrutura de rede e modelos de serviço da universidade tem-se mostrado necessária para permitir o fornecimento de serviços inovadores capazes de responder às necessidades do mundo actual. Neste trabalho é proposta uma ferramenta de orquestração de rede que, integrada com a plataforma de cloud OpenStack, é capaz de virtualizar qualquer infraestrutura de rede de forma não disruptiva, proporcionando um modelo de actualização simples dos tradicionais serviços para o novo mundo da virtualização. A framework é capaz de estender as redes virtuais criadas no datacenter pelo OpenStack ou qualquer outra plataforma de cloud, para o campus. Fá-lo reconfigurando directamente os dispositivos de rede de acordo com as necessidades, independentemente do fabricante, do tipo de dispositivo ou das suas especificidades, e independentemente da topologia da rede física. O serviço é fornecido ao utilizador usando um modelo de cloud, muito mais flexivel que o modelo actual, devidamente integrado com os serviços da plataforma de cloud. O projecto foi desenvolvido com os casos de uso da Universidade de Aveiro em mente, mas o resultado final pode ser aplicado em muitos outros ambientes académicos ou empresariais. A framework é apresentada tanto do ponto de vista conceptual, descrevendo as abstrações e mecanismos criados, como do ponto de vista de implementação, dando ao leitor o entendimento necessário acerca da operação da ferramenta e da sua integração com o OpenStack. A integração deste trabalho com uma visão mais abrangente para o futuro dos serviços da universidade é deixada como trabalho futuro.
APA, Harvard, Vancouver, ISO, and other styles
33

Hasan, MD Sabbir. "Smart management of renewable energy in clouds : from infrastructure to application." Thesis, Rennes, INSA, 2017. http://www.theses.fr/2017ISAR0010/document.

Full text
Abstract:
Avec l'avènement des technologies de Cloud computing et son adoption, les entreprises et les institutions académiques transfèrent de plus en plus leurs calculs et leurs données vers le Cloud. Alors que ce progrès et ce modèle simple d'accès ont eu un impact considérable sur notre communauté scientifique et industrielle en termes de réduction de la complexité et augmentation des revenus, les centres de données consomment énormément d'énergie, ce qui se traduit par des émissions plus élevées de C02. En réponse, de nombreux travaux de recherche se sont focalisés sur les enjeux du développement durable pour le Cloud à travers la réduction de la consommation d'énergie en concevant des stratégies d'efficacité énergétiques. Cependant, l'efficacité énergétique dans l'infrastructure du C!oud ne suffira pas à stimuler la réduction de • l'empreinte carbone. Il est donc impératif d'envisager une utilisation intelligente de l'énergie verte à la fois au niveau de l'infrastructure et de l'application pour réduire davantage l'empreinte carbone. Depuis peu, certains fournisseurs de Cloud computing alimentent leurs centres de données avec de l'énergie renouvelable. Les sources d'énergie renouvelable sont très intermittentes, ce qui crée plusieurs défis pour les gérer efficacement. Pour surmonter ces défis, nous étudions les options pour intégrer les différentes sources d'énergie renouvelable de manière réaliste et proposer un Cloud energy broker qui peut ajuster la disponibilité et la combinaison de prix pour acheter de l'énergie verte dynamiquement sur le marché de l'énergie et rendre les centres de données partiellement verts. Puis, nous introduisons le concept de la virtualisation de l'énergie verte, qui peut être vu comme une alternative au stockage d'énergie utilisé dans les centres de données pour éliminer le problème d'intermittence dans une certaine mesure. Avec l'adoption du concept de virtualisation, nous maximisons l'utilisation de l'énergie verte contrairement au stockage d'énergie qui induit des pertes d'énergie, tout en introduisant des Green SLA basé sur l'énergie verte pour le fournisseur de services et les utilisateurs finaux. En utilisant des traces réalistes et une simulation et une analyse approfondie, nous montrons que la proposition peut fournir un système efficace, robuste et rentable de gestion de l'énergie pour le centre de données. Si une gestion efficace de l'énergie en présence d'énergie verte intermittente est nécessaire, la façon dont les applications Cloud modernes peuvent tirer profit de la présence ou l'absence d'énergie verte n'a pas été suffisamment étudiée. Contrairement aux applications Batch, les applications Interactive Cloud doivent toujours être accessibles et ne peuvent pas être programmées à l'avance pour correspondre au profil d'énergie verte. Par conséquent, cette thèse propose une solution d'autoscaling adaptée à l'énergie pour exploiter les caractéristiques internes des applications et créer une conscience d'énergie verte dans l'application, tout en respectant les propriétés traditionnelles de QoS. Pour cela, nous concevons un contrôleur d'application green qui profite de la disponibilité de l'énergie verte pour effectuer une adaptation opportuniste dans une application gérée par un contrôleur orienté performance. L'expérience est réalisée avec une application réelle sur Grid5000 et les résultats montrent une réduction significative de la consommation d'énergie par rapport à l'approche orientée performance, tout en respectant les attributs traditionnels de QoS
With the advent of cloud enabling technologies and adoption of cloud computing, enterprise and academic institutions are moving their IT workload to the cloud. Although this prolific advancement and easy to access model have greatly impacted our scientific and industrial community in terms of reducing complexity and increasing revenue, data centers are consuming enormous amount of energy, which translates into higher carbon emission. In response, varieties of research work have focused on environmental sustainability for Cloud Computing paradigm through energy consumption reduction by devising energy efficient strategies. However, energy efficiency in cloud infrastructure alone is not going to be enough to boost carbon footprint reduction. Therefore, it is imperative to envision of smartly using green energy at infrastructure and application level for further reduction of carbon footprint. In recent years, some cloud providers are powering their data centers with renewable energy. The characteristics of renewable energy sources are highly intermittent which creates several challenges to manage them efficiently. To overcome the problem, we investigate the options and challenges to integrate different renewable energy sources in a realistic way and propose a Cloud energy broker, which can adjust the availability and price combination to buy Green energy dynamically from the energy market in advance to make a data center partially green. Later, we introduce the concept of Virtualization of Green Energy, which can be seen as an alternative to energy storage used in data center to eliminate the intermittency problem to some extent. With the adoption of virtualization concept, we maximize the usage of green energy contrary to energy storage which induces energy losses, while introduce Green Service Level Agreement based on green energy for service provider and end users. By •using realistic traces and extensive simulation and analysis, we show that, the proposal can provide an efficient, robust and cost-effective energy management scheme for data center. While an efficient energy management in the presence of intermittent green energy is necessary, how modern Cloud applications can take advantage of the presence/absence of green energy has not been studied with requisite effort. Unlike Batch applications, Interactive Cloud applications have to be always accessible and car not be scheduled in advance to match with green energy profile. Therefore, this thesis proposes an energy adaptive autoscaling solution to exploit applications internal to create green energy awareness in the application, while respecting traditional QoS properties. To elaborate, we design green energy aware application controller that takes advantage of green energy availability to perform opportunistic adaptation in an application along with performance aware application controller. Experiment is performed with real life application at Grid5000 and results show significant reduction of energy consumption while respecting traditional QoS attributes compared to performance aware approach
APA, Harvard, Vancouver, ISO, and other styles
34

Lakew, Ewnetu Bayuh. "Autonomous cloud resource provisioning : accounting, allocation, and performance control." Doctoral thesis, Umeå universitet, Institutionen för datavetenskap, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-107955.

Full text
Abstract:
The emergence of large-scale Internet services coupled with the evolution of computing technologies such as distributed systems, parallel computing, utility computing, grid, and virtualization has fueled a movement toward a new resource provisioning paradigm called cloud computing. The main appeal of cloud computing lies in its ability to provide a shared pool of infinitely scalable computing resources for cloud services, which can be quickly provisioned and released on-demand with minimal effort. The rapidly growing interest in cloud computing from both the public and industry together with the rapid expansion in scale and complexity of cloud computing resources and the services hosted on them have made monitoring, controlling, and provisioning cloud computing resources at runtime into a very challenging and complex task. This thesis investigates algorithms, models and techniques for autonomously monitoring, controlling, and provisioning the various resources required to meet services’ performance requirements and account for their resource usage. Quota management mechanisms are essential for controlling distributed shared resources so that services do not exceed their allocated or paid-for budget. Appropriate cloud-wide monitoring and controlling of quotas must be exercised to avoid over- or under-provisioning of resources. To this end, this thesis presents new distributed algorithms that efficiently manage quotas for services running across distributed nodes. Determining the optimal amount of resources to meet services’ performance requirements is a key task in cloud computing. However, this task is extremely challenging due to multi-faceted issues such as the dynamic nature of cloud environments, the need for supporting heterogeneous services with different performance requirements, the unpredictable nature of services’ workloads, the non-triviality of mapping performance measurements into resources, and resource shortages. Models and techniques that can predict the optimal amount of resources needed to meet service performance requirements at runtime irrespective of variations in workloads are proposed. Moreover, different service differentiation schemes are proposed for managing temporary resource shortages due to, e.g., flash crowds or hardware failures. In addition, the resources used by services must be accounted for in order to properly bill customers. Thus, monitoring data for running services should be collected and aggregated to maintain a single global state of the system that can be used to generate a single bill for each customer. However, collecting and aggregating such data across geographical distributed locations is challenging because the management task itself may consume significant computing and network resources unless done with care. A consistency and synchronization mechanism that can alleviate this task is proposed.
APA, Harvard, Vancouver, ISO, and other styles
35

Hans, Ronny [Verfasser], Ralf [Akademischer Betreuer] Steinmetz, and Schahram [Akademischer Betreuer] Dustdar. "QoS-aware Cloud Infrastructure Provisioning in Heterogeneous Environments / Ronny Hans ; Ralf Steinmetz, Schahram Dustdar." Darmstadt : Universitäts- und Landesbibliothek Darmstadt, 2018. http://d-nb.info/1165227606/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
36

Filipsson, Emma. "Modelling and simulating Identity and Access Management based lateral movement in a cloud infrastructure." Thesis, KTH, Fysik, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-252836.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Carreño, Emmanuell Diaz. "Migration and evaluation of a numerical weather prediction application in a cloud computing infrastructure." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2015. http://hdl.handle.net/10183/127446.

Full text
Abstract:
O uso de clusters e grids tem beneficiado durante anos a comunidade de computação de alto desempenho (HPC). O uso deste tipo de sistemas tem permitido aos cientistas usar conjuntos de dados maiores para executar cálculos mais complexos. A computação de alto desempenho tem ajudado para obter aqueles resultados em menos tempo, mas aumentou o custo das despesas de capital nesta área da ciência. Como alguns projetos de e-science são realizados também em ambientes de rede altamente distribuídos, ou usando conjuntos de dados imensos que muitas vezes requerem computação em grade, eles são muito bons candidatos para as iniciativas de computação em nuvem. O paradigma Cloud Computing surgiu como uma solução prática com foco comercial para realizar computação científica em larga escala. A elasticidade da nuvem e o modelo pay-as-you-go apresenta uma oportunidade interessante para aplicações comumente executados em supercomputadores ou clusters. Esta tese apresenta e avalia os desafios da migração e execução da previsão numérica de tempo (NWP) numa infra-estrutura de computação em nuvem. Foi realizada a migração desta aplicação HPC e foi avaliado o desempenho em um cluster local e na nuvem utilizando diferentes tamanhos de instâncias virtuais. Analisamos as principais características da aplicação executando na nuvem. As experiências demonstram que, embora o processamento e a rede criam um fator limitante, o armazenamento dos conjuntos de dados de entrada e saída na nuvem apresentam uma opção atraente para compartilhar resultados e facilitar a implantação de um ambiente de ensaio para investigação meteorológica. Os resultados mostram que a infraestrutura de nuvem pode ser usada como uma alternativa viável de HPC para software de previsão numérica do tempo.
The usage of clusters and grids has benefited for years the High Performance Computing (HPC) community. These kind of systems have allowed scientists to use bigger datasets and to perform more intensive computations, helping them to achieve results in less time but has also increased the upfront costs associated with this area of science. As some e-Science projects are carried out also in highly distributed network environments or using immense data sets that sometimes require grid computing, they are good candidates for cloud computing initiatives. The Cloud Computing paradigm has emerged as a practical solution to perform large-scale scientific computing. The elasticity of the cloud and its pay-as-you-go model presents an attractive opportunity for applications commonly executed in clusters or supercomputers. In this context, the user does not need to buy infrastructure, the resources can be rented from a provider and used for a period of time. This thesis presents the challenges and solutions of migrating a numerical weather prediction (NWP) application to a cloud computing infrastructure. We performed the migration of this HPC application and evaluated its performance in a local cluster and the cloud using different instance sizes. We analyzed the main characteristics of the application running in the cloud. The experiments demonstrate that, although processing and networking create a limiting factor, storing input and output datasets in the cloud presents an attractive option to share results and ease the deployment of a test-bed for a weather research platform. Results show that cloud infrastructure can be used as a viable HPC alternative for numerical weather prediction software.
APA, Harvard, Vancouver, ISO, and other styles
38

Pastor, Jonathan. "Contributions à la mise en place d'une infrastructure de Cloud Computing à large échelle." Thesis, Nantes, Ecole des Mines, 2016. http://www.theses.fr/2016EMNA0240/document.

Full text
Abstract:
La croissance continue des besoins en puissance de calcul a conduit au triomphe du modèle de Cloud Computing. Des clients demandeurs en puissance de calcul vont s’approvisionner auprès de fournisseurs d’infrastructures de Cloud Computing, mises à disposition via Internet. Pour réaliser des économies d’échelles, ces infrastructures sont toujours plus grandes et concentrées en quelques endroits, conduisant à des problèmes tels que l’approvisionnement en énergie, la tolérance aux pannes et l’éloignement des utilisateurs. Cette thèse s’est intéressée à la mise en place d’un système d’IaaS massivement distribué et décentralisé exploitant un réseau de micros centres de données déployés sur la dorsale Internet, utilisant une version d’OpenStack revisitée pendant cette thèse autour du support non intrusif de bases de données non relationnelles. Des expériences sur Grid’5000 ont montré des résultats intéressants sur le plan des performances, toutefois limités par le fait qu’OpenStack ne tirait pas avantage nativement d’un fonctionnement géographiquement réparti. Nous avons étudié la prise en compte de la localité réseau pour améliorer les performances des services distribués en favorisant les collaborations proches. Un prototype de l’algorithme de placement de machines virtuelles DVMS, fonctionnant sur une topologie non structurée basée sur l’algorithme Vivaldi, a été validé sur Grid’5000. Ce prototype a fait l’objet d’un prix scientifique lors de l’école de printemps Grid’50002014. Enfin, ces travaux nous ont amenés à participer au développement du simulateur VMPlaceS
The continuous increase of computing power needs has favored the triumph of the Cloud Computing model. Customers asking for computing power will receive supplies via Internet resources hosted by providers of Cloud Computing infrastructures. To make economies of scale, Cloud Computing that are increasingly large and concentrated in few attractive places, leading to problems such energy supply, fault tolerance and the fact that these infrastructures are far from most of their end users. During this thesis we studied the implementation of an fully distributed and decentralized IaaS system operating a network of micros data-centers deployed in the Internet backbone, using a modified version of OpenStack that leverages non relational databases. A prototype has been experimentally validated onGrid’5000, showing interesting results, however limited by the fact that OpenStack doesn’t take advantage of a geographically distributed functioning. Thus, we focused on adding the support of network locality to improve performance of Cloud Computing services by favoring collaborations between close nodes. A prototype of the DVMS algorithm, working with an unstructured topology based on the Vivaldi algorithm, has been validated on Grid’5000. This prototype got the first prize at the large scale challenge of the Grid’5000 spring school in 2014. Finally, the work made with DVMS enabled us to participate at the development of the VMPlaceS simulator
APA, Harvard, Vancouver, ISO, and other styles
39

Cardoso, Igor Duarte. "Network infrastructure control for virtual campuses." Master's thesis, Universidade de Aveiro, 2014. http://hdl.handle.net/10773/14707.

Full text
Abstract:
Mestrado em Engenharia de Computadores e Telemática
This dissertation provides a way to merge Cloud Computing infrastructures with traditional or legacy network deployments, leveraging the best in both worlds and enabling a logically centralized control for it. A design/architecture is proposed to extend existing Cloud Computing software stacks so they are able to manage networks outside the Cloud Computing infrastructure, by extending the internal, virtualized network segments. This is useful in a variety of use cases such as incremental Legacy to Cloud network migration, hybrid virtual/ traditional networking, centralized control of existing networks, bare metal provisioning and even offloading of advanced services from typical home gateways into the operator, improving control and reducing maintenance costs. An implementation is presented and tested on top of OpenStack, the principal Open-Source Cloud Computing software stack available. It includes changes to the API, command line interface and existing mechanisms which previously only supported homogeneous vendor equipment, such that they support any hardware and be able to automate their provisioning. By using what is called External Drivers, any organization (an hardware vendor, a Cloud provider or even a telecommunications operator) can develop their own driver to support new, specific networking equipment. Besides this ease of development and extensibility, two drivers are already developed in the context of this work: one for OpenWrt switches/routers and one for Cisco EtherSwitch IOS switching modules. Test results indicate that there are low penalties on latency and throughput, and that provisioning times (for setting up or tearing down networks) are reduced in comparison with similar maintenance operations on traditional computer networks.
Esta dissertação apresenta uma forma de juntar infrastruturas Cloud Computing com redes tradicionais ou legadas, trazendo o melhor de ambos os mundos e possibilitando um controlo logicamente centralizado. Uma arquitetura é proposta com o intuito de extender implementações de Cloud Computing para que possam gerir também redes fora da infrastrutura de Cloud Computing, extendendo os segmentos de rede internos, virtualizados. Isto é útil para um variado conjunto de casos de uso, tais como migração incremental de redes legadas para a Cloud, redes híbridas virtuais/tradicionais, controlo centralizado de redes já existentes, aprovisionamento de bare metal e até mesmo a passagem de serviços tipicamente fornecidos por um home gateway para o lado do operador, melhorando o controlo e reduzindo custos de manutenção. Uma implementação da solução é apresentada e testada em cima do OpenStack, a principal solução Open-Source de Cloud Computing disponível. A implementação inclui alterações à API, à interface de linha de comandos e aos mecanismos já existentes, que apenas suportam implementações homogéneas, para que possam suportar qualquer equipamento e automatizar o aprovisionamento dos mesmos. Através daquilo que se chamam drivers externos, qualquer organização (seja um fabricante de equipamentos de rede, um fornecedor de Cloud ou uma operadora de telecomunicações) pode desenvolver o seu próprio drivers para suportar novos, específicos equipamentos de hardware. Para além da facilidade de desenvolvimento e extensibilidade, dois drivers são também fruto deste trabalho: um para switches/routers OpenWrt e outro para os módulos de switching Cisco EtherSwitch, sistema operativo IOS. Testes efetuados indicam que há baixas penalizações na latência e largura de banda, e ainda que os tempos de aprovisionamento são reduzidos em comparação com semelhantes operações de manutenção em redes informáticas tradicionais.
APA, Harvard, Vancouver, ISO, and other styles
40

Metwally, Khaled. "A Resource Management Framework for IaaS in Cloud Computing Environment." Thesis, Université d'Ottawa / University of Ottawa, 2016. http://hdl.handle.net/10393/34951.

Full text
Abstract:
Cloud computing Infrastructure-as-a-Service (IaaS) has gained momentum in the cloud computing research field due to its ability to provide efficient infrastructures. Cloud Service Providers (CSPs) are striving to offer Quality of Service (QoS)-guaranteed IaaS services while also improving their resource utilization and maximizing profit. In addition, CSPs are challenged by the need to manipulate diverse and heterogeneous resources, realizing multiple objectives for both customers and CSPs, and handling scalability issues. These challenges are the motivations behind this work which aims at developing a multi-layered framework for constructing and managing efficient IaaS. The fundamental layer in this framework, the Virtual Infrastructure (VI) composition layer, is dedicated to composing and delivering VIs as an IaaS service. This framework relies on a preparatory step that is defined when all the available resources in the managed space are collected in a large repository, the Virtual Resource Pool (VRP). The VRP creation process unifies the representation of all the diverse and heterogeneous resources available. Subsequently, the proposed framework performs various resource allocation approaches as working solutions through the VI composition layer. These approaches adopt efficient techniques and methodologies in performing their operations. The working solutions are initiated by designing a composition approach that relies on an ontology-based model representation. The composition approach exploits semantic similarity, closeness centrality, and random walk techniques for efficient resource allocation. As a result, it provides an efficient solution in a reasonable computational time with no guarantee for the optimality of the obtained solutions. To achieve an optimal solution, the composition approach uses a mathematical modeling formulation. In this solution, the concepts of the composition approach have been integrated into a multi-objective Mixed Integer Linear Programming (MILP) model that has been solved optimally. Despite the optimality of the resulting solution, the MILP-based model restricts IaaS resource allocation to a computational running-time challenge, and the issue of limited-size datacenters. To circumvent these issues, a cost-efficient model is proposed. The new model introduces a Column Generation (CG) formulation for the IaaS resource allocation problem in large datacenters acquainted with QoS requirements. Furthermore, this formulation is realistic, adopts large-scale optimization tools that are adequate for large datacenters, and ensures optimal solutions in a reasonable time. However, growing costs in large datacenters in accordance with the growth of recent large-scale application demands, makes large datacenters economically inefficient. Thus, we advocate a distributed framework for IaaS provisioning that guarantees affordable, scalable, and QoS-assured infrastructure for hosting large-scale applications in geo-distributed datacenters. The framework incorporates two decentralized resource allocation approaches, hierarchical and distributed, that use efficient economic models. These approaches are quite promising solutions for the scalability and computational complexity issues of existing centralized approaches. Finally, the cost-efficient model has been extended to fit the distributed infrastructure by considering additional constraints that impact CSP revenue. Simulation results showcase the effectiveness of the presented work along with the potential benefits of the proposed solutions in terms of satisfying the customers’ requirements, while achieving a better resource utilization and CSP payoffs.
APA, Harvard, Vancouver, ISO, and other styles
41

Lukášová, Pavlína. "Cloud Computing jako nástroj BCM." Master's thesis, Vysoká škola ekonomická v Praze, 2010. http://www.nusl.cz/ntk/nusl-75556.

Full text
Abstract:
This thesis deals with possible interconnections between two concepts playing a big role in contemporary business and IT world. These concepts are Business Continuity Management and Cloud Computing. In the scope of this thesis there are certain areas identified where both concepts are complement, where Cloud Computing brings new opportunities for Business Continuity Management and where could possible problems arise during particular implementation. From the BCM perspective the impact lies on IT services, from the Cloud Computing perspective the thesis deals especially with security aspects. The thesis is also aimed at the characteristics of higher education and basic differences from commercial sphere. Based on defined differences and identified interconnections between BCM and Cloud Computing, the thesis argues for usage of suitable Cloud Computing solution for higher education regarding Business Continuity improvement. The multi-criterion comparison of several Infrastructure-as-a-Service solutions stems from this analysis focusing on technical, financial, and Business Continuity aspects. The result from this comparison together with conclusions from previous chapters serve as an input for subsequent practical proposal of Cloud Computing solution and its verification against Business Continuity improvement in specific conditions on University of Economics in Prague. The proposal is also represented by strategic map.
APA, Harvard, Vancouver, ISO, and other styles
42

Ekanayake, Mudiyanselage Wijaya Dheeshakthi. "An SDN-based Framework for QoSaware Mobile Cloud Computing." Thesis, Université d'Ottawa / University of Ottawa, 2016. http://hdl.handle.net/10393/35117.

Full text
Abstract:
In mobile cloud computing (MCC), rich mobile application data is processed at the cloud infrastructure by reliving resource limited mobile devices from computationally complex tasks. However, due to the ubiquitous and mobility nature, providing time critical rich applications over remote cloud infrastructure is a challenging task for mobile application service providers. Therefore, according to the literature, close proximity placement of cloud services has been identified as a way to achieve lower end-to-end access delay and thereby provide a higher quality of experience (QoE) for rich mobile application users. However, providing a higher Quality of Service (QoS) with mobility is still a challenge within close proximity clouds. Access delay to a closely placed cloud tends to be increased over time when users move away from the cloud. However, reactive resource relocation mechanism proposed in literature does not provide a comprehensive mechanism to guarantee the QoS and as well as to minimize service provisioning cost for mobile cloud service providers. As a result, using the benefits of SDN and the data plane programmability with logically centralized controllers, a resource allocation framework was proposed for IaaS mobile clouds with regional datacenters. The user mobility problem was analyzed within SDN-enabled wireless networks and addressed the possible service level agreement violations that could occur with inter-regional mobility. The proposed framework is composed of an optimization algorithm to provide seamless cloud service during user mobility. Further a service provisioning cost minimization criteria was considered during an event of resource allocation and inter-regional user mobility.
APA, Harvard, Vancouver, ISO, and other styles
43

Izumo, Naoki. "Clouded space: Internet physicality." Thesis, University of Iowa, 2017. https://ir.uiowa.edu/etd/5515.

Full text
Abstract:
On Friday October 21st, 2016, there was a large-scale hack of an Internet domain hosting provider that took several websites including Netflix, Amazon, Reddit, and Twitter offline. Dyn, a cloud-based Internet Performance Management company, announced at 9:20AM ET that it resolved an attack that began at 7AM ET that day. However, another attack happened at 11:52AM ET. The attacks raised concern among the public and directed our attention towards Internet security. This also revealed the precariousness of Internet infrastructure. The infrastructure being used today is opaque, unregulated, and incontestable. Municipally provided public utilities are built without any transparency; thus, we do not expect failure from those systems. For instance, the Flint, Michigan water crisis raised issues of water infrastructure. Not only did the crisis spark talks about the corrosion of pipes, but also larger societal issues. Flint, a poor, largely African American community, became a victim of environmental racism—a type of discrimination where communities of color or low-income residents are forced to live in environmental dangerous areas. In order for myself and the larger public to understand this opaque system, we need to understand the infrastructure and how it works. With regards to Internet infrastructure, I focus on data centers, where there are backup servers, batteries and generators built into the architectural landscape in case of failure. There is a common held thought that overshadows the possibility of imminent technological failure—it cannot happen. This sort of thinking influences other modes of our daily lives: individuals building concrete bomb shelters underground for the apocalypse, stocking food, but not preparing for data breakdown. The consciousness of loss is further perpetuated by technology and its life expectancy. Clouded Space: Internet Physicality attempts to explore the unexceptional infrastructure of the Internet and how it exists right beneath our feet. That in itself is not very cloud-like. The work questions integrity of our infrastructure as much as environmental issues, highlighting the questionable relationship we have with data and our inclination to backup data to protect ourselves from failure. This is a relatively new topic and the challenges are not well understood. There seem to be cracks in the foundation, and though they are not yet obvious, they appear to be widening.
APA, Harvard, Vancouver, ISO, and other styles
44

Castagnoli, Carlo. "Cloud Computing: gli Hypervisor e la funzionalità di Live Migration nelle Infrastructure as a Service." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2011. http://amslaurea.unibo.it/1856/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
45

Hawasli, Ahmad. "azureLang: a probabilistic modeling and simulation language for cyber attacks in Microsoft Azure cloud infrastructure." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-240406.

Full text
Abstract:
Cyber-attack simulation is a suitable method used for assessing the security ofnetwork systems. An attack simulation advances step-wise from a certain systementry-point to explore the attack paths that lead to dierent weaknesses inthe model. Each step is analyzed, and the time to compromise is calculated.Attack simulations are primarily based on attack graphs. The graphs areemployed to model attack steps where nodes can represent assets in the system,and edges can represent the attack steps. To reduce the computational cost associatedwith building an attack graph for each specic system, domain-specicattack languages, or DSL for short, are used.The nal product of this thesis work is azureLang, a probabilistic modelingand simulation language for modeling Microsoft Azure cloud infrastructure.AzureLang is a DSL which denes a generic attack logic for MicrosoftAzure systems. Using azureLang, system administrators can easily instantiatespecic-system scenarios which emulate their Microsoft Azure cloud system infrastructure.After creating the model, attack simulation can be run to assessthe security of the model.
Cyberattacksimulering är en lämplig metod som används för att bedöma säkerhetenhos nätverkssystem. En angrepsimulering går stegvis från ett visst systeminmatningspunkt för att utforska angreppsbanorna som leder till olika svagheter i modellen. Varje steg analyseras och tiden för kompromettera beräknas.Attack-simuleringar baseras huvudsakligen på attackgrafer. Graferna används för att modellera angreppssteg där noder kan representera tillgångar i systemet, och kanterna kan representera attackenstegen. För att minska kostnaden för att skapa attackgrafer för varje specifikt system används domänspecifika språk eller DSL förkortat.Den slutliga produkten av detta examensarbete är azureLang, ett probabilistisk hotmodelleringsoch attacksimuleringsspråk för analys av Microsoft Azure Cloud Infrastructure. AzureLang är en DSL som definierar en generisk attacklogik för Microsoft Azure-system. Med hjälp av azureLang kan systemadministratörer enkelt ordna specifika systemscenarier som efterliknar deras Microsoft Azure cloudsystem infrastruktur. Efter att ha skapat modellen kan attack simu-lering köras för att bedöma modellens säkerhet.
APA, Harvard, Vancouver, ISO, and other styles
46

Noroozi, Hamid. "A Cloud-native Vehicular Public Key Infrastructure : Towards a Highly-available and Dynamically- scalable VPKIaaS." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-300658.

Full text
Abstract:
Efforts towards standardization of Vehicular Communication Systems (VCSs) have been conclusive on the use of Vehicular Public-Key Infrastructure (VPKI) for the establishment of trust among network participants. Employing VPKI in Vehicular Communication (VC) guarantees the integrity and authenticity of Cooperative Awareness Messages (CAMs) and Decentralized Environmental Notification Messages (DENMs). It also offers a level of privacy for vehicles as VPKI provides them with a set of non-linkable short-lived certificates, called pseudonyms, which are used to sign outgoing messages by vehicles while they communicate with other vehicles referred to as Vehicle-to-Vehicle (V2V) or Roadside Units (RSUs) referred to as Vehicle-to-Infrastructure (V2I). Each vehicle uses a pseudonym for its lifetime and by switching to a not- previously- used pseudonym, it continues to communicate without risking its privacy. There have been two approaches suggested by the literature on how to provide vehicles with pseudonyms. One is the so-called pre-loading mode, suggesting to pre-load vehicles with all pseudonyms they need, which increases the cost of revocation in case they are compromised. The other one is the on-demand mode, suggesting a real-time offering of pseudonyms by VPKI at vehicles request e.g., on starting each trip. Choosing the on-demand approach imposes a considerable burden of availability and resilience on VPKI services. In this work, we are confronting the problems regarding a large-scale deployment of an on-demand VPKI that is resilient, highly available, and dynamically scalable. In order to achieve that, by leveraging state-of-the-art tools and design paradigms, we have enhanced a VPKI system to ensure that it is capable of meeting enterprise-grade Service Level Agreement (SLA) in terms of availability, and it can also be cost-efficient as services can dynamically scale-out in the presence of high load, or possibly scale-in when facing less demand. That has been made possible by re-architecting and refactoring an existing VPKI into a cloud-native solution deployed as microservices. Towards having a reliable architecture based on distributed microservices, one of the key challenges to deal with is Sybil-based misbehavior. By exploiting Sybil-based attacks in VPKI, malicious vehicles can gain influential advantage in the system, e.g., one can affect the traffic to serve its own will. Therefore, preventing the occurrence of Sybil attacks is paramount. On the other hand, traditional approaches to stop them, often come with a performance penalty as they verify requests against a relational database which is a bottleneck of the operations. We propose a solution to address Sybil-based attacks, utilizing Redis, an in-memory data store, without compromising the system efficiency and performance considerably. Running our VPKI services on Google Cloud Platform (GCP) shows that a large-scale deployment of VPKI as a Service (VPKIaaS) can be done efficiently. Conducting various stress tests against the services indicates that the VPKIaaS is capable of serving real world traffic. We have tested VPKIaaS under synthetically generated normal traffic flow and flash crowd scenarios. It has been shown that VPKIaaS managed to issue 100 pseudonyms per request, submitted by 1000 vehicles where vehicles kept asking for a new set of pseudonyms every 1 to 5 seconds. Each vehicle has been served in less than 77 milliseconds. We also demonstrate that, under a flash crowd situation, with 50000 vehicles, VPKIaaS dynamically scales out, and takes ≈192 milliseconds to serve 100 pseudonyms per request submitted by vehicles.
Ansträngningar för standardisering av Vehicular Communication Systems har varit avgörande för användandet av Vehicular Public-Key Infrastructure (VPKI) för att etablera förtroende mellan nätverksdeltagare. Användande av VPKI i Vehicular Communication (VC) garanterar integritet och autenticitet av meddelanden. Det erbjuder ett lager av säkerhet för fordon då VPKI ger dem en mängd av icke länkbara certifikat, kallade pseudonym, som används medan de kommunicerar med andra fordon, kallat Vehicle-to-Vehicle (V2V) eller Roadside Units (RSUs) kallat Vehicle-to-Infrastructure (V2I). Varje fordon använder ett pseudonym under en begränsad tid och genom att byta till ett icke tidigare använt pseudonym kan det fortsätta kommunicera utan att riskera sin integritet. I litteratur har två metoder föreslagits för hur man ska ladda fordon med pseudonym de behöver. Den ena metoden det så kallade offline-läget, som proponerar att man för-laddar fordonen med alla pseudonym som det behöver vilket ökar kostnaden för revokering i fall de blir komprometterat. Den andra metoden föreslår ett on-demand tillvägagångssätt som erbjuder pseudonym via VPKI på fordonets begäran vid början av varje färd. Valet av på begäran metoden sätter en stor börda på tillgänglighet och motståndskraft av VPKI tjänster. I det här arbetet, möter vi problem med storskaliga driftsättningar av en på begäran VPKI som är motståndskraftig, har hög tillgänglighet och dynamiskt skalbarhet i syfte att uppnå dessa attribut genom att nyttja toppmoderna verktyg och designparadigmer. Vi har förbättrat ett VPKI system för att säkerställa att det är kapabelt att möta SLA:er av företagsklass gällande tillgänglighet och att det även kan vara kostnadseffektivt eftersom tjänster dynamiskt kan skala ut vid högre last eller skala ner vid lägre last. Detta har möjliggjorts genom att arkitekta om en existerande VPKI till en cloud-native lösning driftsatt som mikrotjänster. En av nyckelutmaningarna till att ha en pålitlig arkitektur baserad på distribuerade mikrotjänster är sybil-baserad missuppförande. Genom att utnyttja Sybil baserade attacker på VPKI, kan illvilliga fordon påverka trafik att tjäna dess egna syften. Därför är det av största vikt att förhindra Sybil attacker. Å andra sidan så dras traditionella metoder att stoppa dem med prestandakostnader. Vi föreslår en lösning för att adressera Sybilbaserade attacker genom att nyttja Redis, en in-memory data-store utan att märkbart kompromissa på systemets effektivitet och prestanda. Att köra våra VPKI tjänster på Google Cloud Platform (GCP) och genomföra diverse stresstester mot dessa har visat att storskaliga driftsättningar av VPKI as a Service (VPKIaaS) kan göras effektivt samtidigt som riktigt trafik hanteras. Vi har testat VPKIaaS under syntetisk genererat normalt trafikflöde samt flow och flash mängd scenarier. Det har visat sig att VPKIaaS klarar att utfärda 100 pseudonym per förfråga utsänt av 1000 fordon (där fordonen bad om en ny uppsättning pseudonym varje 1 till 5 sekunder), och varje fordon fått svar inom 77 millisekunder. Vi demonstrerar även att under en flashcrowd situation, där antalet fordon höjs till 50000 med en kläckningsgrad på 100. VPKIaaS dynamiskt skalar ut och tar ≈192 millisekunder att betjäna 100 pseudonymer per förfrågan gjord av fordon.
APA, Harvard, Vancouver, ISO, and other styles
47

Lam, Juan Carlos. "Dynamic Analysis of Levee Infrastructure Failure Risk: A Framework for Enhanced Critical Infrastructure Management." Thesis, Virginia Tech, 2012. http://hdl.handle.net/10919/43106.

Full text
Abstract:
Current models that assess infrastructure failure risk are â linear,â and therefore, only consider the direct influence attributed to each factor that defines risk. These models do not consider the undeniable relationships that exist among these parameters. In reality, factors that define risk are interdependent and influence each other in a â non-linearâ fashion through feedback effects. Current infrastructure failure risk assessment models are also static, and do not allow infrastructure managers and decision makers to evaluate the impacts over time, especially the long-term impact of risk mitigation actions. Factors that define infrastructure failure risk are in constant change. In a strategic manner, this research proposes a new risk-based infrastructure management framework and supporting system, Risk-Based Dynamic Infrastructure Management System (RiskDIMS), which moves from linear to non-linear risk assessment by applying systems engineering methods and analogs developed to address non-linear complex problems. The approach suggests dynamically integrating principal factors that define infrastructure failure risk using a unique platform that leverages Geospatial Information System services and extensions in an unprecedented manner. RiskDIMS is expected to produce results that are often counterintuitive and unexpected, but aligned to our complex reality, suggesting that the combination of geospatial and temporal analyses is required for sustainable risk-based decision making. To better illustrate the value added of temporal analysis in risk assessment, this study also develops and implements a non-linear dynamic model to simulate the behavior over time of infrastructure failure risk associated with an existing network of levees in New Orleans due to diverse infrastructure management investments. Although, the framework and RiskDIMS are discussed here in the context of levees, the concept applies to other critical infrastructure assets and systems. This research aims to become the foundation for future risk analysis system implementation.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
48

Nemanja, Popović. "Napredni distributivni menadžment sistem zasnovan na Cloud infrastrukturi." Phd thesis, Univerzitet u Novom Sadu, Fakultet tehničkih nauka u Novom Sadu, 2018. https://www.cris.uns.ac.rs/record.jsf?recordId=107176&source=NDLTD&language=en.

Full text
Abstract:
U ovoj doktorskoj disertaciji ispitana je mogućnost prebacivanja Naprednog Distributivnog Menadžment Sistema (NDMS) zasnovanog na zajedničkoj operaciono tehnološkoj platformi na virtualno Cloud infrastrukturno okruženje. Prvo je odabrano tradicionalno NDMS rešenje zasnovano na fizičkoj računarskoj arhitekturi i identifikovani su funkcionalni blokovi. Zatim su profilirane njihove performanse prema četiri ključna resursa: procesor, operativna memorija, računarska mreža i stalna memorija. Dalje predloženo je virtualno rešenje zasnovano na Cloud infrastrukturi koje je verifikovano na dve zamišljene elektrodistributivne mreže realnih veličina (maloj i velikoj) i u dva testna scenarija (stabilnog stanja i visoke aktivnosti). Na kraju predstavljeni su rezultati testiranja koji pokazaju da se NDMS može prebaciti u virtualno Cloud okruženje bez negativnog uticaja na funkcionalne i nefunkcionalne zahteve NDMS rešenja.
This dissertation inspects the possibility of transferring the Advanced Distribution Management System (ADMS) based on a common operating technology platform to the virtual Cloud Infrastructure Environment. First, a traditional ADMS solution based on physical computing architecture was chosen and functional blocks were identified. Furthermore, their performances were profiled to four key resources: processor, operating memory, computer network, and storage. Then, virtual cloud-based solution was proposed based on Cloud infrastructure which is verified on two imaginary power distribution networks of real size (small and large) and in two test scenarios (steady state and high activity). Finally, the test results show that ADMS can be transferred to a virtual Cloud environment without adversely affecting the functional and non-functional requirements of the ADMS solution.
APA, Harvard, Vancouver, ISO, and other styles
49

Künsemöller, Jörn [Verfasser]. "Tragedy of the common cloud : game theory on the infrastructure-as-a-service market / Jörn Künsemöller." Paderborn : Universitätsbibliothek, 2014. http://d-nb.info/106464709X/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
50

Pérennou, Loïc. "Virtual machine experience design : a predictive resource allocation approach for cloud infrastructures." Thesis, Paris, CNAM, 2019. http://www.theses.fr/2019CNAM1246/document.

Full text
Abstract:
L’un des principaux défis des fournisseurs de services cloud est d’offrir aux utilisateurs une performance acceptable, tout en minimisant les besoins en matériel et énergie. Dans cette thèse CIFRE menée avec Outscale, un fournisseur de cloud, nous visons à optimiser l’allocation des ressources en utilisant de nouvelles sources d’information. Nous caractérisons la charge de travail pour comprendre le stress résultant sur l’orchestrateur, et la compétition pour les ressources disponibles qui dégrade la qualité de service. Nous proposons un modèle pour prédire la durée d’exécution des VMs à partir de caractéristiques prédictives disponibles au démarrage. Enfin, nous évaluons la sensibilité aux erreurs d’un algorithme de placement des VMs de la littérature qui se base sur ces prédictions. Nous ne trouvons pas d’intérêt à coupler note système prédictif avec cet algorithme, mais nous proposons d’autres façons d’utiliser les prédictions pour optimiser le placement des VMs
One of the main challenges for cloud computing providers remains to offer trustable performance for all users, while maintaining an efficient use of hardware and energy resources. In the context of this CIFRE thesis lead with Outscale, apublic cloud provider, we perform an in-depth study aimed at making management algorithms use new sources of information. We characterize Outscale’s workload to understand the resulting stress for the orchestrator, and the contention for hardware resources. We propose models to predict the runtime of VMs based on features which are available when they start. We evaluate the sensitivity with respect to prediction error of a VM placement algorithm from the literature that requires such predictions. We do not find any advantage in coupling our prediction model and the selected algorithm, but we propose alternative ways to use predictions to optimize the placement of VMs
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography