Dissertations / Theses on the topic 'Cloud Computing Performance'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Cloud Computing Performance.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Al-Refai, Ali, and Srinivasreddy Pandiri. "Cloud Computing : Trends and Performance Issues." Thesis, Blekinge Tekniska Högskola, Sektionen för datavetenskap och kommunikation, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-3672.
Full textMani, Sindhu. "Empirical Performance Analysis of High Performance Computing Benchmarks Across Variations in Cloud Computing." UNF Digital Commons, 2012. http://digitalcommons.unf.edu/etd/418.
Full textPelletingeas, Christophe. "Performance evaluation of virtualization with cloud computing." Thesis, Edinburgh Napier University, 2010. http://researchrepository.napier.ac.uk/Output/4010.
Full textNoureddine, Moustafa. "Enterprise adoption oriented cloud computing performance optimization." Thesis, University of East London, 2014. http://roar.uel.ac.uk/4026/.
Full textPenmetsa, Jyothi Spandana. "AUTOMATION OF A CLOUD HOSTED APPLICATION : Performance, Automated Testing, Cloud Computing." Thesis, Blekinge Tekniska Högskola, Institutionen för kommunikationssystem, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-12849.
Full textAUTOMATION OF A CLOUD HOSTED APPLICATION
Roloff, Eduardo. "Viability and performance of high-performance computing in the cloud." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2013. http://hdl.handle.net/10183/79594.
Full textCloud computing is a new paradigm, where computational resources are offered as services. In this context, the user does not need to buy infrastructure, the resources can be rented from a provider and used for a period of time. Furthermore the user can easily allocate as many resources as needed, and deallocate them as well, in a totally elastic environment. The resources need to be paid only for the effective usage time. On the other hand, High-Performance Computing (HPC) requires a large amount of computational power. To acquire systems capable for HPC, large financial investments are necessary. Apart from the initial investment, the user must pay the maintenance costs, and has only limited computational resources. To overcome these issues, this thesis aims to evaluate the cloud computing paradigm as a candidate environment for HPC. We analyze the efforts and challenges for porting and deploy HPC applications to the cloud. We evaluate if this computing model can provide sufficient capacities for running HPC applications, and compare its cost efficiency to traditional HPC systems, such as clusters. The cloud computing paradigm was analyzed to identify which models have the potential to be used for HPC purposes. The identified models were then evaluated using major cloud providers, Microsoft Windows Azure, Amazon EC2 and Rackspace and compare them to a traditional HPC system. We analyzed the capabilities to create HPC environments, and evaluated their performance. For the evaluation of the cost efficiency, we developed an economic model. The results show that all the evaluated providers have the capability to create HPC environments. In terms of performance, there are some cases where cloud providers present a better performance than the traditional system. From the cost perspective, the cloud presents an interesting alternative due to the pay-per-use model. Summarizing the results, this dissertation shows that cloud computing can be used as a realistic alternative for HPC environments.
Sridharan, Suganya. "A Performance Comparison of Hypervisors for Cloud Computing." UNF Digital Commons, 2012. http://digitalcommons.unf.edu/etd/269.
Full textDanielsson, Simon, and Staffan Johansson. "Cloud Computing - A Study of Performance and Security." Thesis, Malmö högskola, Fakulteten för teknik och samhälle (TS), 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-20326.
Full textCloud Computing - the big buzz word of the IT world. It has become more and more popular in recent years but questions has arisen about it’s performance and security. How safe is it and is there any real difference in performance between a locally based server and a cloud based server? This thesis will examine these questions. A series of performance tests combined with a literature study were performed to achieve the results of this thesis.This thesis could be of use for those who have an interest in Cloud Computing and do not have much knowledge of it. The results can be used as an example for how future research in Cloud Computing can be done.
Calatrava, Arroyo Amanda. "High Performance Scientific Computing over Hybrid Cloud Platforms." Doctoral thesis, Universitat Politècnica de València, 2016. http://hdl.handle.net/10251/75265.
Full textLas aplicaciones científicas generalmente precisan grandes requisitos de cómputo, memoria y gestión de datos para su ejecución. Este tipo de aplicaciones tradicionalmente ha empleado recursos de altas prestaciones, como supercomputadores de memoria compartida, clústers de PCs de memoria distribuida, o recursos provenientes de infraestructuras Grid, sobre los que se adaptaba la aplicación para que se ejecutara satisfactoriamente. El auge que han tenido las técnicas de virtualización en los últimos años, propiciando la aparición de la computación en la nube (Cloud Computing), ha provocado un importante cambio en la forma de ejecutar este tipo de aplicaciones. Sin embargo, la gestión de la ejecución de aplicaciones científicas sobre plataformas de computación elásticas de altas prestaciones no es una tarea trivial. En esta tesis doctoral se ha desarrollado Elastic Cloud Computing Cluster (EC3), una herramienta de código abierto capaz de llevar a cabo la ejecución de aplicaciones científicas de altas prestaciones creando para ello clústers virtuales, híbridos y elásticos, autogestionados y eficientes en cuanto a costes, sobre plataformas Cloud de tipo Infraestructura como Servicio (IaaS). Estos clústers autogestionados tienen la capacidad de adaptar su tamaño, es decir, el número de nodos, a la carga de trabajo, creando así la ilusión de un clúster real sin requerir una inversión por encima del uso actual. Además, son completamente configurables y pueden ser migrados de un proveedor a otro de manera automática y transparente a los usuarios y trabajos en ejecución en el cluster. EC3 también permite desplegar clústers híbridos sobre recursos Cloud públicos y privados, donde los recursos privados son complementados con recursos Cloud públicos para acelerar el proceso de ejecución. Otras configuraciones híbridas, como el empleo de diferentes tipos de instancias y el uso de instancias puntuales combinado con instancias bajo demanda son también soportadas por EC3. Además, el uso de instancias puntuales junto con técnicas de checkpointing permite a EC3 reducir significantemente el coste total de las ejecuciones a la vez que proporciona tolerancia a fallos. EC3 está concebido para facilitar el uso de clústers virtuales a los usuarios, que, aunque no tengan un conocimiento extenso sobre este tipo de tecnologías, pueden beneficiarse fácilmente de ellas. Por ello, la herramienta ofrece dos interfaces diferentes a sus usuarios, una interfaz web donde se expone EC3 como servicio para usuarios no experimentados y una potente interfaz de línea de comandos. Además, esta tesis doctoral se adentra en el campo de la virtualización ligera, mediante el uso de contenedores como alternativa a la solución tradicional de virtualización basada en máquinas virtuales. Este estudio analiza el escenario propicio para el uso de contenedores y propone una arquitectura para el despliegue de clusters virtuales elásticos basados en esta tecnología. Finalmente, para demostrar la funcionalidad y ventajas de las herramientas desarrolladas durante esta tesis, esta memoria recoge varios casos de uso que abarcan diferentes escenarios y campos de conocimiento, como estudios estructurales de edificios, astrofísica o biodiversidad.
Les aplicacions científiques generalment precisen grans requisits de còmput, de memòria i de gestió de dades per a la seua execució. Este tipus d'aplicacions tradicionalment hi ha empleat recursos d'altes prestacions, com supercomputadors de memòria compartida, clústers de PCs de memòria distribuïda, o recursos provinents d'infraestructures Grid, sobre els quals s'adaptava l'aplicació perquè s'executara satisfactòriament. L'auge que han tingut les tècniques de virtualitzaciò en els últims anys, propiciant l'aparició de la computació en el núvol (Cloud Computing), ha provocat un important canvi en la forma d'executar este tipus d'aplicacions. No obstant això, la gestió de l'execució d'aplicacions científiques sobre plataformes de computació elàstiques d'altes prestacions no és una tasca trivial. En esta tesi doctoral s'ha desenvolupat Elastic Cloud Computing Cluster (EC3), una ferramenta de codi lliure capaç de dur a terme l'execució d'aplicacions científiques d'altes prestacions creant per a això clústers virtuals, híbrids i elàstics, autogestionats i eficients quant a costos, sobre plataformes Cloud de tipus Infraestructura com a Servici (IaaS). Estos clústers autogestionats tenen la capacitat d'adaptar la seua grandària, es dir, el nombre de nodes, a la càrrega de treball, creant així la il·lusió d'un cluster real sense requerir una inversió per damunt de l'ús actual. A més, són completament configurables i poden ser migrats d'un proveïdor a un altre de forma automàtica i transparent als usuaris i treballs en execució en el cluster. EC3 també permet desplegar clústers híbrids sobre recursos Cloud públics i privats, on els recursos privats són complementats amb recursos Cloud públics per a accelerar el procés d'execució. Altres configuracions híbrides, com l'us de diferents tipus d'instàncies i l'ús d'instàncies puntuals combinat amb instàncies baix demanda són també suportades per EC3. A més, l'ús d'instàncies puntuals junt amb tècniques de checkpointing permet a EC3 reduir significantment el cost total de les execucions al mateix temps que proporciona tolerància a fallades. EC3e stà concebut per a facilitar l'ús de clústers virtuals als usuaris, que, encara que no tinguen un coneixement extensiu sobre este tipus de tecnologies, poden beneficiar-se fàcilment d'elles. Per això, la ferramenta oferix dos interfícies diferents dels seus usuaris, una interfície web on s'exposa EC3 com a servici per a usuaris no experimentats i una potent interfície de línia d'ordres. A més, esta tesi doctoral s'endinsa en el camp de la virtualitzaciò lleugera, per mitjà de l'ús de contenidors com a alternativa a la solució tradicional de virtualitzaciò basada en màquines virtuals. Este estudi analitza l'escenari propici per a l'ús de contenidors i proposa una arquitectura per al desplegament de clusters virtuals elàstics basats en esta tecnologia. Finalment, per a demostrar la funcionalitat i avantatges de les ferramentes desenrotllades durant esta tesi, esta memòria arreplega diversos casos d'ús que comprenen diferents escenaris i camps de coneixement, com a estudis estructurals d'edificis, astrofísica o biodiversitat.
Calatrava Arroyo, A. (2016). High Performance Scientific Computing over Hybrid Cloud Platforms [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/75265
TESIS
Hutchins, Richard Chad. "Feasibility of virtual machine and cloud computing technologies for high performance computing." Thesis, Monterey, California. Naval Postgraduate School, 2013. http://hdl.handle.net/10945/42447.
Full textReissued May 2014 with additions to the acknowledgments
Knowing the future weather on the battlefield with high certainty can result in a higher advantage over the adversary. To create this advantage for the United States, the U.S. Navy utilizes the Coupled Ocean/Atmosphere Mesoscale Prediction System (COAMPS) to create high spatial resolution, regional, numerical weather prediction (NWP) forecasts. To compute a forecast, COAMPS runs on high performance computing (HPC) systems. These HPC systems are large, dedicated supercomputers with little ability to scale or move. This makes these systems vulnerable to outages without a costly, equally powerful secondary system. Recent advancements in cloud computing and virtualization technologies provide a method for high mobility and scalability without sacrificing performance. This research used standard benchmarks in order to quantitatively compare a virtual machine (VM) to a native HPC cluster. The benchmark tests showed that the VM was feasible platform for executing HPC applications. Then we ran the COAMPS NWP on a VM within a cloud infrastructure to prove the ability to run a HPC application in a virtualized environment. The VM COAMPS model run performed better than the native HPC machine model run. These results show that VM and cloud computing technologies can be used to run HPC applications for the Department of Defense
Agarwal, Dinesh. "Scientific High Performance Computing (HPC) Applications On The Azure Cloud Platform." Digital Archive @ GSU, 2013. http://digitalarchive.gsu.edu/cs_diss/75.
Full textKhan, Majid, and Muhammad Faisal Amin. "Web Server Performance Evaluation in Cloud Computing and Local Environment." Thesis, Blekinge Tekniska Högskola, Sektionen för datavetenskap och kommunikation, 2012. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-1965.
Full textBelhareth, Sonia. "Performances réseaux et système pour le cloud computing." Thesis, Nice, 2014. http://www.theses.fr/2014NICE4146/document.
Full textCloud computing enables a flexible access to computation and storage services. This requires, for the cloud service provider, mastering network and system issues. During this PhD thesis, we focused on the performance of TCP Cubic, which is the default version of TCP in Linux and thus widely used in today's data centers. Cloud environments feature low bandwidth-delay products (BPD) in the case of intra data center communications and high BDP in the case of inter data center communications. We have developed analytical models to study the performance of a Cubic connection in isolation or a set of competing Cubic connections. Our models turn out to be precise in the low BDP case but fail at capturing the synchronization of losses that ns-2 simulations reveal in the high BDP case. We have complemented our simulations studies with tests in real environments: (i) an experimental network at I3S and (ii) a cloud solution available internally at Orange: Cube. Studies performed in Cube have highlighted the high correlation that might exist between network and system performance and the complexity to analyze the performance of applications in a cloud context. Studies in the controlled environment of I3S have confirmed the existence of synchronization and enabled us to identify its condition of appearance. We further investigated two types of solution to combat synchronization: client level solutions that entail modifications of TCP and network level solutions based on queue management solutions, in particular PIE and Codel
Palopoli, Amedeo. "Containerization in Cloud Computing: performance analysis of virtualization architectures." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2017. http://amslaurea.unibo.it/14818/.
Full textKanthla, Arjun Reddy. "Network Performance Improvement for Cloud Computing using Jumbo Frames." Thesis, KTH, Radio Systems Laboratory (RS Lab), 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-143806.
Full textDen kraftiga ökningen i datormoln är på grund av dess kostnadseffektiva fördelar och den snabba skalbarhet av datorresurser, och kärnan i detta är virtualisering. Virtualiseringsteknik möjliggör att man kan köra era operativsystem på en enda fysisk maskin. Detta ökar effektiviteten av hårdvaran, vilket gör att kostnaden minskar för datormoln. Men eftersom lasten i gästoperativsystemet ökar, gör att de fysiska resurserna inte kan stödja alla program på ett effektivt sätt. In-och utgångstjänster, speciellt nätverksapplikationer, måste dela samma totala bandbredd gör att denna delning kan påverkas negativt av virtualisering. Nätverkspaket kan genomgå ytterligare behandling och måste vänta tills den virtuella maskinen är planerad av den underliggande hypervisor innan den slutliga services applikation, till exempel en webbserver. I en virtuell miljö är det inte belastningen (på grund av behandlingen av användarens data) utan nätverket overhead, som är det största problemet. Moderna nätverkskort har förbättrat nätverk virtualisering genom att hantera IP-paket mer intelligent genom TCP- segmenterings avlastning, avbrotts sammansmältning och genom en annan hårdvara som är specifik för virtualisering. Jumboramar har länge föreslagits för sina fördelar i traditionell miljö. De ökar nätverk genomströmning och minska CPU-användning. Genom att använda Jumbo frames kan Gigabit Ethernet användandet förbättras samt erbjuda stora förbättringar för virtualiserad miljö genom att utnyttja bandbredden mer effektivt samtidigt sänka processor overhead. Det här examensarbetet visar ett nätverk prestandaförbättring på 4,7% i en Xen virtualiserad miljö genom att använda jumbo frames. Dessutom undersöker det TCP prestanda i Xen och jämför Xen med samma funktion som körs på en Linux system.
Kaza, Bhagavathi. "Performance Evaluation of Data Intensive Computing In The Cloud." UNF Digital Commons, 2013. http://digitalcommons.unf.edu/etd/450.
Full textSangroya, Amit. "Towards dependability and performance benchmarking for cloud computing services." Thesis, Grenoble, 2014. http://www.theses.fr/2014GRENM016/document.
Full textCloud computing models are attractive because of various benefits such asscalability, cost and exibility to develop new software applications. However,availability, reliability, performance and security challenges are still not fullyaddressed. Dependability is an important issue for the customers of cloudcomputing who want to have guarantees in terms of reliability and availability.Many studies investigated the dependability and performance of cloud services,ranging from job scheduling to data placement and replication, adaptiveand on-demand fault-tolerance to new fault-tolerance models. However, thead-hoc and overly simplified settings used to evaluate most cloud service fault toleranceand performance improvement solutions pose significant challengesto the analysis and comparison of the effectiveness of these solutions.This thesis precisely addresses this problem and presents a benchmarkingapproach for evaluating the dependability and performance of cloud services.Designing of dependability and performance benchmarks for a cloud serviceis a particular challenge because of high complexity and the large amount ofdata processed by such service. Infrastructure as a Service (IaaS), Platform asa Service (PaaS) and Software as a Service (SaaS) are the three well definedmodels of cloud computing. In this thesis, we will focus on the PaaS modelof cloud computing. PaaS model enables operating systems and middlewareservices to be delivered from a managed source over a network. We introduce ageneric benchmarking architecture which is further used to build dependabilityand performance benchmarks for PaaS model of cloud services
Němec, Petr. "Cloud computing a jeho aplikace." Master's thesis, Vysoká škola ekonomická v Praze, 2010. http://www.nusl.cz/ntk/nusl-112946.
Full textTunc, Cihan. "Autonomic Cloud Resource Management." Diss., The University of Arizona, 2015. http://hdl.handle.net/10150/347144.
Full textAlarcon, Jean-Luc Bruno. "Emerging Capabilities and Firm Performance in the Cloud Computing Environment." Diss., Temple University Libraries, 2018. http://cdm16002.contentdm.oclc.org/cdm/ref/collection/p245801coll10/id/493895.
Full textD.B.A.
New capabilities required to succeed in the new Cloud environment represent a radical change for software companies, which have to transition from selling on-the-premises software products to providing subscription-based cloud services (aka Software-as-a-Service or SaaS). While emerging SaaS vendors have led the exponential growth of the market, the traditional software industry has been disrupted. The purpose of this dissertation is to analyze which capabilities are driving the performance of software firms in today’s cloud-computing environment by drawing upon the resource-based view (RBV) of the firm. What is the optimum spending across the primary firm capabilities (e.g., service delivery, R&D, marketing and sales) to maximize financial performance? This dissertation focuses on publicly-traded SaaS companies using publicly-available information from financial databases, corporate investor relations materials, and industry research. It is comprised of two essays. The first essay is a quantitative study based on secondary data. The second essay includes an extensive literature review, an analysis of in-depth interviews of practitioners, and mini case studies. Together, the essays contribute to RBV theory and provide useful insights to help assess the quality of execution of SaaS growth strategies and improve financial planning and performance in the software industry for the cloud computing environment. Although the results come from firms in the SaaS industry, the findings from this study could cautiously generalize to firms in other emerging technology industries. The dissertation concludes with a detailed agenda for future research.
Temple University--Theses
Farahani, Rad Ali. "Cloud computing and its implications for organizational design and performance." Thesis, Massachusetts Institute of Technology, 2013. http://hdl.handle.net/1721.1/81073.
Full textCataloged from PDF version of thesis.
Includes bibliographical references (p. 35-36).
Cloud computing has been at the center of attention for a while now. This attention is directed towards different aspects of this concept which concern different stakeholders from IT companies to cloud adopters to simple users of cloud services. Cloud computing is affecting the IT industry in terms of operations, product and service development, and strategy. It also affects the organizations that adopt cloud computing. Any new technology or new means of conducting business processes could have a potential effect on organizational structure as well as performance. The purpose of this thesis is to investigate the implications of adopting cloud computing for organizational design and performance. It focuses on how a lowering in the costs associated with IT might influence the organizational design and in particular the decision making structure of organizations. Past research conducted on the effects of the costs associated with IT on organizational structure have been used to explain the implications of adopting cloud computing.
by Ali Farahani Rad.
S.M.in Management Studies
Ibidunmoye, Olumuyiwa. "Performance problem diagnosis in cloud infrastructures." Licentiate thesis, Umeå universitet, Institutionen för datavetenskap, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-120287.
Full textCloud Control (C0590801)
Lindgren, Hans. "Performance Management for Cloud Services: Implementation and Evaluation of Schedulers for OpenStack." Thesis, KTH, Kommunikationsnät, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-124542.
Full textFör att erhålla maximal prestanda ur ett IaaS moln måste dess resurshanteringssystem kunna schemalägga resursutnyttjandet av den underliggande infrastrukturen på ett optimalt sätt. En nyttjandebaserad schemaläggare kan dra nytta av det faktum att allokerade resurser och faktiskt använda resurser ofta skiljer sig åt, för att på så sätt kunna fatta mer välgrundade beslut om var framtida förfrågningar skall placeras. Detta examensarbete presenterar såväl utformning, implementation samt utvärdering av en kontrollenhet för initial placering som till sin hjälp använder information om värdutnyttjande som indata för placering av virtuella maskiner enligt ett av ett antal stödda förvaltningssyften. Implementationen, som baseras på molnplattformen OpenStack, behandlar två sådana syften, balanserad last och energieffektivitet. I tillägg diskuteras krav för att stödja ytterligare syften. En testmiljö och demonstrationsplattform, bestående av ovan nämnda kontrollenhet, en syntetisk lastgenererare samt en övervakningsplattform byggs upp och används vid utvärdering av systemet. Resultat av utvärderingen påvisar att schemaläggaren presterar likvärdigt för båda syftena vid användande av last bestående av såväl interaktiva som applikationer av batch-typ. En diskussion om nuvarande begränsningar hos schemaläggaren och hur dessa kan överbryggas sammanfattar arbetet. Bland det som tas upp diskuteras bl.a. hur hastigheten vid vilken värdutnyttjande data samlas in påverkar prestandan hos schemaläggaren samt under vilka förhållanden dynamisk placering av virtuella maskiner bör användas som komplement till nyttjandebaserad schemaläggning för att undvika risk för överbelastning i molnet.
López, Huguet Sergio. "Elastic, Interoperable and Container-based Cloud Infrastructures for High Performance Computing." Doctoral thesis, Universitat Politècnica de València, 2021. http://hdl.handle.net/10251/172327.
Full text[ES] Las aplicaciones científicas implican generalmente una carga computacional variable y no predecible a la que las instituciones deben hacer frente variando dinámicamente la asignación de recursos en función de las distintas necesidades computacionales. Las aplicaciones científicas pueden necesitar grandes requisitos. Por ejemplo, una gran cantidad de recursos computacionales para el procesado de numerosos trabajos independientes (High Throughput Computing o HTC) o recursos de alto rendimiento para la resolución de un problema individual (High Performance Computing o HPC). Los recursos computacionales necesarios en este tipo de aplicaciones suelen acarrear un coste muy alto que puede exceder la disponibilidad de los recursos de la institución o estos pueden no adaptarse correctamente a las necesidades de las aplicaciones científicas, especialmente en el caso de infraestructuras preparadas para la ejecución de aplicaciones de HPC. De hecho, es posible que las diferentes partes de una aplicación necesiten distintos tipos de recursos computacionales. Actualmente las plataformas de servicios en la nube se han convertido en una solución eficiente para satisfacer la demanda de las aplicaciones HTC, ya que proporcionan un abanico de recursos computacionales accesibles bajo demanda. Por esta razón, se ha producido un incremento en la cantidad de clouds híbridos, los cuales son una combinación de infraestructuras alojadas en servicios en la nube y en las propias instituciones (on-premise). Dado que las aplicaciones pueden ser procesadas en distintas infraestructuras, actualmente la portabilidad de las aplicaciones se ha convertido en un aspecto clave. Probablemente, las tecnologías de contenedores son la tecnología más popular para la entrega de aplicaciones gracias a que permiten reproducibilidad, trazabilidad, versionado, aislamiento y portabilidad. El objetivo de la tesis es proporcionar una arquitectura y una serie de servicios para proveer infraestructuras elásticas híbridas de procesamiento que puedan dar respuesta a las diferentes cargas de trabajo. Para ello, se ha considerado la utilización de elasticidad vertical y horizontal desarrollando una prueba de concepto para proporcionar elasticidad vertical y se ha diseñado una arquitectura cloud elástica de procesamiento de Análisis de Datos. Después, se ha trabajo en una arquitectura cloud de recursos heterogéneos de procesamiento de imágenes médicas que proporciona distintas colas de procesamiento para trabajos con diferentes requisitos. Esta arquitectura ha estado enmarcada en una colaboración con la empresa QUIBIM. En la última parte de la tesis, se ha evolucionado esta arquitectura para diseñar e implementar un cloud elástico, multi-site y multi-tenant para el procesamiento de imágenes médicas en el marco del proyecto europeo PRIMAGE. Esta arquitectura utiliza un almacenamiento distribuido integrando servicios externos para la autenticación y la autorización basados en OpenID Connect (OIDC). Para ello, se ha desarrollado la herramienta kube-authorizer que, de manera automatizada y a partir de la información obtenida en el proceso de autenticación, proporciona el control de acceso a los recursos de la infraestructura de procesamiento mediante la creación de las políticas y roles. Finalmente, se ha desarrollado otra herramienta, hpc-connector, que permite la integración de infraestructuras de procesamiento HPC en infraestructuras cloud sin necesitar realizar cambios en la infraestructura HPC ni en la arquitectura cloud. Cabe destacar que, durante la realización de esta tesis, se han utilizado distintas tecnologías de gestión de trabajos y de contenedores de código abierto, se han desarrollado herramientas y componentes de código abierto y se han implementado recetas para la configuración automatizada de las distintas arquitecturas diseñadas desde la perspectiva DevOps.
[CA] Les aplicacions científiques impliquen generalment una càrrega computacional variable i no predictible a què les institucions han de fer front variant dinàmicament l'assignació de recursos en funció de les diferents necessitats computacionals. Les aplicacions científiques poden necessitar grans requisits. Per exemple, una gran quantitat de recursos computacionals per al processament de nombrosos treballs independents (High Throughput Computing o HTC) o recursos d'alt rendiment per a la resolució d'un problema individual (High Performance Computing o HPC). Els recursos computacionals necessaris en aquest tipus d'aplicacions solen comportar un cost molt elevat que pot excedir la disponibilitat dels recursos de la institució o aquests poden no adaptar-se correctament a les necessitats de les aplicacions científiques, especialment en el cas d'infraestructures preparades per a l'avaluació d'aplicacions d'HPC. De fet, és possible que les diferents parts d'una aplicació necessiten diferents tipus de recursos computacionals. Actualment les plataformes de servicis al núvol han esdevingut una solució eficient per satisfer la demanda de les aplicacions HTC, ja que proporcionen un ventall de recursos computacionals accessibles a demanda. Per aquest motiu, s'ha produït un increment de la quantitat de clouds híbrids, els quals són una combinació d'infraestructures allotjades a servicis en el núvol i a les mateixes institucions (on-premise). Donat que les aplicacions poden ser processades en diferents infraestructures, actualment la portabilitat de les aplicacions s'ha convertit en un aspecte clau. Probablement, les tecnologies de contenidors són la tecnologia més popular per a l'entrega d'aplicacions gràcies al fet que permeten reproductibilitat, traçabilitat, versionat, aïllament i portabilitat. L'objectiu de la tesi és proporcionar una arquitectura i una sèrie de servicis per proveir infraestructures elàstiques híbrides de processament que puguen donar resposta a les diferents càrregues de treball. Per a això, s'ha considerat la utilització d'elasticitat vertical i horitzontal desenvolupant una prova de concepte per proporcionar elasticitat vertical i s'ha dissenyat una arquitectura cloud elàstica de processament d'Anàlisi de Dades. Després, s'ha treballat en una arquitectura cloud de recursos heterogenis de processament d'imatges mèdiques que proporciona distintes cues de processament per a treballs amb diferents requisits. Aquesta arquitectura ha estat emmarcada en una col·laboració amb l'empresa QUIBIM. En l'última part de la tesi, s'ha evolucionat aquesta arquitectura per dissenyar i implementar un cloud elàstic, multi-site i multi-tenant per al processament d'imatges mèdiques en el marc del projecte europeu PRIMAGE. Aquesta arquitectura utilitza un emmagatzemament integrant servicis externs per a l'autenticació i autorització basats en OpenID Connect (OIDC). Per a això, s'ha desenvolupat la ferramenta kube-authorizer que, de manera automatitzada i a partir de la informació obtinguda en el procés d'autenticació, proporciona el control d'accés als recursos de la infraestructura de processament mitjançant la creació de les polítiques i rols. Finalment, s'ha desenvolupat una altra ferramenta, hpc-connector, que permet la integració d'infraestructures de processament HPC en infraestructures cloud sense necessitat de realitzar canvis en la infraestructura HPC ni en l'arquitectura cloud. Es pot destacar que, durant la realització d'aquesta tesi, s'han utilitzat diferents tecnologies de gestió de treballs i de contenidors de codi obert, s'han desenvolupat ferramentes i components de codi obert, i s'han implementat receptes per a la configuració automatitzada de les distintes arquitectures dissenyades des de la perspectiva DevOps.
[EN] Scientific applications generally imply a variable and an unpredictable computational workload that institutions must address by dynamically adjusting the allocation of resources to their different computational needs. Scientific applications could require a high capacity, e.g. the concurrent usage of computational resources for processing several independent jobs (High Throughput Computing or HTC) or a high capability by means of using high-performance resources for solving complex problems (High Performance Computing or HPC). The computational resources required in this type of applications usually have a very high cost that may exceed the availability of the institution's resources or they are may not be successfully adapted to the scientific applications, especially in the case of infrastructures prepared for the execution of HPC applications. Indeed, it is possible that the different parts that compose an application require different type of computational resources. Nowadays, cloud service platforms have become an efficient solution to meet the need of HTC applications as they provide a wide range of computing resources accessible on demand. For this reason, the number of hybrid computational infrastructures has increased during the last years. The hybrid computation infrastructures are the combination of infrastructures hosted in cloud platforms and the computation resources hosted in the institutions, which are named on-premise infrastructures. As scientific applications can be processed on different infrastructures, the application delivery has become a key issue. Nowadays, containers are probably the most popular technology for application delivery as they ease reproducibility, traceability, versioning, isolation, and portability. The main objective of this thesis is to provide an architecture and a set of services to build up hybrid processing infrastructures that fit the need of different workloads. Hence, the thesis considered aspects such as elasticity and federation. The use of vertical and horizontal elasticity by developing a proof of concept to provide vertical elasticity on top of an elastic cloud architecture for data analytics. Afterwards, an elastic cloud architecture comprising heterogeneous computational resources has been implemented for medical imaging processing using multiple processing queues for jobs with different requirements. The development of this architecture has been framed in a collaboration with a company called QUIBIM. In the last part of the thesis, the previous work has been evolved to design and implement an elastic, multi-site and multi-tenant cloud architecture for medical image processing has been designed in the framework of a European project PRIMAGE. This architecture uses a storage integrating external services for the authentication and authorization based on OpenID Connect (OIDC). The tool kube-authorizer has been developed to provide access control to the resources of the processing infrastructure in an automatic way from the information obtained in the authentication process, by creating policies and roles. Finally, another tool, hpc-connector, has been developed to enable the integration of HPC processing infrastructures into cloud infrastructures without requiring modifications in both infrastructures, cloud and HPC. It should be noted that, during the realization of this thesis, different contributions to open source container and job management technologies have been performed by developing open source tools and components and configuration recipes for the automated configuration of the different architectures designed from the DevOps perspective. The results obtained support the feasibility of the vertical elasticity combined with the horizontal elasticity to implement QoS policies based on a deadline, as well as the feasibility of the federated authentication model to combine public and on-premise clouds.
López Huguet, S. (2021). Elastic, Interoperable and Container-based Cloud Infrastructures for High Performance Computing [Tesis doctoral]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/172327
TESIS
Compendio
Celaya, Tracy A. "Cloud-Based Computing and human resource management performance| A Delphi study." Thesis, University of Phoenix, 2016. http://pqdtopen.proquest.com/#viewpdf?dispub=10004286.
Full textThe purpose of this qualitative study with a modified Delphi research design was to understand the reasons human resource (HR) leaders are slow to implement Cloud-based technologies and potentially identify how Cloud-Based Computing influences human resource management (HRM) and HR effectiveness, and potentially the overall performance of the organization. Business executives and HR leaders acknowledge the effect of technology on business processes and strategies, and the leader's influence on technology implementation and adoption. Cloud-Based Computing is fast becoming the standard for conducting HR processes and HR leaders must be prepared to implement the change effectively. Study findings revealed characteristics demonstrated by HR leaders successfully implementing cloud technology, best practices for successful implementation, factors championing and challenging Cloud-Based Computing adoption, and effects on HRM and organizational performance as a result of using Cloud-Based Computing. The outcomes of this study may provide the foundation of a model for implementing Cloud-Based Computing, a leadership model including characteristics of technology early adopters in HR, and identify factors impeding adoption and may assist HR leaders in creating effective change management strategies for adopting and implementing Cloud-Based Computing. Findings and recommendation from this study will enable HR professionals and leaders to make informed decisions on the adoption of Cloud-Based Computing and improve the effectiveness, efficiency, and strategic capability of HR.
Struhar, Vaclav. "Improving Soft Real-time Performance of Fog Computing." Licentiate thesis, Mälardalens högskola, Inbyggda system, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:mdh:diva-55679.
Full textZhang, Dean. "Data Parallel Application Development and Performance with Azure." The Ohio State University, 2011. http://rave.ohiolink.edu/etdc/view?acc_num=osu1308168930.
Full textDawoud, Wesam. "Scalability and performance management of internet applications in the cloud." Phd thesis, Universität Potsdam, 2013. http://opus.kobv.de/ubp/volltexte/2013/6818/.
Full textCloud computing ist ein Model fuer einen Pool von Rechenressourcen, den sie auf Anfrage zur Verfuegung stellt. Internetapplikationen in einer Cloud-Infrastruktur koennen bei einer erhoehten Auslastung schnell die Lage meistern, indem sie die durch die Cloud-Infrastruktur auf Anfrage zur Verfuegung stehenden und virtuell unbegrenzten Ressourcen in Anspruch nehmen. Allerdings sind solche Applikationen durch den Verwaltungsaufwand zur Bereitstellung der Ressourcen mit Perioden von Verschlechterung der Performanz und Ressourcenunterversorgung konfrontiert. Ausserdem ist das Management der Performanz aufgrund der Konsolidierung in einer Cloud Umgebung kompliziert. Um die Auswirkung des Mehraufwands zur Bereitstellung von Ressourcen abzuschwächen, schlagen wir in dieser Dissertation zwei Methoden vor. Die erste Methode verwendet die Kontrolltheorie, um Ressourcen vertikal zu skalieren und somit schneller mit einer erhoehten Auslastung umzugehen. Diese Methode setzt voraus, dass der Provider das Wissen und die Kontrolle über die in virtuellen Maschinen laufende Plattform hat. Der Provider ist dadurch als „Plattform als Service (PaaS)“ und als „Software als Service (SaaS)“ Provider definiert. Die zweite Methode bezieht sich auf die Clientseite und behandelt die horizontale Skalierbarkeit in einem Infrastruktur als Service (IaaS)-Model. Sie behandelt den Zielkonflikt zwischen den Kosten und der Performanz mit einer mehrzieloptimierten Loesung. Sie findet massstaebliche Schwellenwerte, die die hoechste Performanz mit der niedrigsten Steigerung der Kosten gewaehrleisten. Ausserdem ist in der zweiten Methode ein Algorithmus der Zeitreifenvorhersage verwendet, um die Applikation proaktiv zu skalieren und Perioden der nicht optimalen Ausnutzung zu vermeiden. Um die Performanz der Internetapplikation zu verbessern, haben wir zusaetzlich ein System entwickelt, das die unter Beeintraechtigung der Performanz leidenden virtuellen Maschinen findet und entfernt. Das entwickelte System ist eine leichtgewichtige Lösung, die keine Provider-Beteiligung verlangt. Um die Skalierbarkeit unserer Methoden und der entwickelten Algorithmen auszuwerten, haben wir einen Simulator namens „ScaleSim“ entwickelt. In diesem Simulator haben wir Komponenten implementiert, die als Skalierbarkeitskomponenten der Amazon EC2 agieren. Die aktuelle Skalierbarkeitsimplementierung in Amazon EC2 ist als Referenzimplementierung fuer die Messesung der Verbesserungen in der Performanz von skalierbaren Applikationen. Der Simulator wurde auf realistische Modelle der RUBiS-Benchmark angewendet, die aus einer echten Umgebung extrahiert wurden. Die Auslastung ist aus den Zugriffslogs der World Cup Website von 1998 erzeugt. Die Ergebnisse zeigen, dass die Optimierung der Schwellenwerte und der angewendeten proaktiven Skalierbarkeit den Verwaltungsaufwand zur Bereitstellung der Ressourcen bis um 88% reduziert kann, während sich die Kosten nur um 9% erhöhen.
Tamanampudi, Monica, and Mohith Kumar Reddy Sannareddy. "Performance Optimization of a Service in Virtual and Non - Virtual Environment." Thesis, Blekinge Tekniska Högskola, Institutionen för kommunikationssystem, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-18462.
Full textSilva, Jefferson de Carvalho. "A framework for building component-based applications on a cloud computing platform for high performance computing services." Universidade Federal do CearÃ, 2016. http://www.teses.ufc.br/tde_busca/arquivo.php?codArquivo=16543.
Full textDeveloping High Performance Computing applications (HPC), which optimally access the available computing resources in a higher level of abstraction, is a challenge for many scientists. To address this problem, we present a proposal of a component computing cloud called HPC Shelf, where HPC applications perform and SAFe framework, a front-end aimed to create applications in HPC Shelf and the author's main contribution. SAFe is based on Scientific Workflows Management Systems (SWMS) projects and it allows the specification of computational solutions formed by components to solve problems specified by the expert user through a high level interface. For that purpose, it implements SAFeSWL, an architectural and orchestration description language for describing scientific worflows. Compared with other SWMS alternatives, besides rid expert users from concerns about the construction of parallel and efficient computational solutions from the components offered by the cloud, SAFe integrates itself to a system of contextual contracts which is aligned to a system of dynamic discovery (resolution) of components. In addition, SAFeSWL allows explicit control of life cycle stages (resolution, deployment, instantiation and execution) of components through embedded operators, aimed at optimizing the use of cloud resources and minimize the overall execution cost of computational solutions (workflows). Montage and Map/Reduce are the case studies that have been applied for demonstration, evaluation and validation of the particular features of SAFe in building HPC applications aimed to the HPC Shelf platform.
Desenvolver aplicaÃÃes de ComputaÃÃo de Alto Desempenho (CAD), que acessem os recursos computacionais disponÃveis de forma otimizada e em um nÃvel maior de abstraÃÃo, à um desafio para cientistas de diversos domÃnios. Esta Tese apresenta a proposta de uma nuvem de componentes chamada HPC Shelf, pano de fundo onde as aplicaÃÃes CAD executam, e o arcabouÃo SAFe, Front-End para criaÃÃo de aplicaÃÃes na HPC Shelf e contribuiÃÃo principal do autor. O SAFe toma como base o projeto de sistemas gerenciadores de workflows cientÃficos (SGWC), permitindo a implementaÃÃo de soluÃÃes computacionais baseadas em componentes para resolver os problemas especificados por meio de uma interface de nÃvel de abstraÃÃo mais alto. Para isso, foi desenvolvido o SAFeSWL, uma linguagem de descriÃÃo arquitetural e orquestraÃÃo de worflows cientÃficos. Comparado com outros SGWC, alÃm de livrar usuÃrios finais de preocupaÃÃes em relaÃÃo à construÃÃo de soluÃÃes computacionais paralelas e eficientes a partir dos componentes oferecidos pela nuvem, o SAFe faz uso de um sistema de contratos contextuais integrado a um sistema de descoberta (resoluÃÃo) dinÃmica de componentes. A linguagem SAFeSWL permite o controle explÃcito das etapas do ciclo de vida de um componente em execuÃÃo (resoluÃÃo, implantaÃÃo, instanciaÃÃo e execuÃÃo), atravÃs de operadores embutidos, a fim de otimizar o uso dos recursos da nuvem e minimizar os custos de sua utilizaÃÃo. Montage e Map/Reduce constituem os estudos de caso aplicados para demonstraÃÃo e avaliaÃÃo das propriedades originais do SAFe e do SAFeSWL na construÃÃo de aplicaÃÃes de CAD.
van, Ketwich Willem. "IT Governance of Cloud Computing| Performance Measures using an IT Outsourcing Perspective." Thesis, University of Melbourne (Australia), 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=1527429.
Full textWith the advent of cloud computing and the success of the cloud computing industry, organisations are beginning to adopt this service model and technology at an increasing rate. As the rate and level of use increases, organisations are faced with how best to govern these investments and obtain maximum benefit from the services offered by providers. This includes measuring the performance of these services, the corresponding organisational performance and the associated business value generated. In investigating these areas, this study compares cloud computing and IT outsourcing. It is found that while cloud measures relate, to a great extent, to the operational level of an organisation, IT outsourcing measures are concerned more with the strategic level. This highlights that cloud computing lacks strategic measures and that measures from IT outsourcing may be adopted to fill this gap.
Coutinho, Emanuel Ferreira. "FOLE: A Conceptual Framework for Elasticity Performance Analysis in Cloud Computing Environments." Universidade Federal do CearÃ, 2014. http://www.teses.ufc.br/tde_busca/arquivo.php?codArquivo=13197.
Full textAtualmente muitos clientes e provedores estÃo utilizando recursos de ambientes de ComputaÃÃo em Nuvem, tais como processamento e armazenamento, para suas aplicaÃÃes e serviÃos. Devido à facilidade de utilizaÃÃo, baseada no modelo de pagamento por uso, à natural que a quantidade de usuÃrios e suas respectivas cargas de trabalho tambÃm cresÃam. Como consequÃncia, os provedores devem ampliar seus recursos e manter o nÃvel de qualidade acordado com os clientes, sob pena de quebras do Service Level Agreement (SLA) e consequentes multas. Com o aumento na utilizaÃÃo de recursos computacionais, uma das caracterÃsticas principais da ComputaÃÃo em Nuvem tem se tornado bastante atrativa: a elasticidade. A elasticidade pode ser definida como o quanto uma nuvem computacional se adapta a variaÃÃes na sua carga de trabalho atravÃs do provisionamento e desprovisionamento de recursos. Devido à pouca disponibilidade de informaÃÃo em relaÃÃo à configuraÃÃo dos experimentos, em geral nÃo à trivial implementar conceitos de elasticidade, muito menos aplicÃ-los em ambientes de nuvens computacionais. AlÃm disso, a maneira de se medir a elasticidade nÃo à tÃo Ãbvia, e bastante variada, nÃo havendo ainda uma padronizaÃÃo para esta tarefa, e sua avaliaÃÃo pode ser executada de diferentes maneiras devido Ãs diversas tecnologias e estratÃgias para o provimento da elasticidade. Um aspecto comum na avaliaÃÃo de desempenho da elasticidade à a utilizaÃÃo de recursos do ambiente, como CPU e memÃria, e mesmo sem ter uma mÃtrica especÃfica para a elasticidade, à possÃvel se obter uma avaliaÃÃo indireta. Nesse contexto, este trabalho propÃe o FOLE, um framework conceitual para a realizaÃÃo de anÃlise de desempenho da elasticidade em nuvens computacionais de maneira sistemÃtica, flexÃvel e reproduzÃvel. Para apoiar o framework, mÃtricas especÃficas para a elasticidade e mÃtricas para sua mediÃÃo indireta foram propostas. Para a mediÃÃo da elasticidade em ComputaÃÃo em Nuvem, propomos mÃtricas baseadas em conceitos da FÃsica, como tensÃo e estresse, e da Microeconomia, como Elasticidade do PreÃo da Demanda. Adicionalmente, mÃtricas baseadas em tempos de operaÃÃes de alocaÃÃo e desalocaÃÃo de recursos, e na utilizaÃÃo desses recursos foram propostas para apoiar a mediÃÃo da elasticidade. Para verificaÃÃo e validaÃÃo da proposta, dois estudos de caso foram realizados, um em uma nuvem privada e outro em uma nuvem hÃbrida, com experimentos projetados utilizando microbenchmarks e uma aplicaÃÃo cientÃfica clÃssica, executados sobre uma infraestrutura baseada em conceitos de ComputaÃÃo AutonÃmica. Por meio desses experimentos, o FOLE foi validado em suas atividades, permitindo a sistematizaÃÃo de uma anÃlise de desempenho da elasticidade. Os resultados mostram que à possÃvel avaliar a elasticidade de um ambiente de ComputaÃÃo em Nuvem por meio de mÃtricas especÃficas baseadas em conceitos de outras Ãreas de conhecimento, e tambÃm complementada por mÃtricas relacionadas a tempos de operaÃÃes e recursos de maneira satisfatÃria.
Nigro, Michele. "Progettazione di un cluster sul cloud con il framework HPC." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2020. http://amslaurea.unibo.it/20316/.
Full textLakew, Ewnetu Bayuh. "Autonomous cloud resource provisioning : accounting, allocation, and performance control." Doctoral thesis, Umeå universitet, Institutionen för datavetenskap, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-107955.
Full textSmith, James William. "Investigating performance and energy efficiency on a private cloud." Thesis, University of St Andrews, 2014. http://hdl.handle.net/10023/6540.
Full textJamaliannasrabadi, Saba. "High Performance Computing as a Service in the Cloud Using Software-Defined Networking." Bowling Green State University / OhioLINK, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1433963448.
Full textKehrer, Stefan [Verfasser], and Wolfgang [Akademischer Betreuer] Blochinger. "Elastic parallel systems for high performance cloud computing / Stefan Kehrer ; Betreuer: Wolfgang Blochinger." Stuttgart : Universitätsbibliothek der Universität Stuttgart, 2020. http://d-nb.info/121990578X/34.
Full textMcConnell, Aaron. "End-to-end performance monitoring and SLA-complaint resource optimisation in cloud computing." Thesis, University of Ulster, 2012. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.588494.
Full textLeite, Alessandro Ferreira. "A user-centered and autonomic multi-cloud architecture for high performance computing applications." reponame:Repositório Institucional da UnB, 2014. http://repositorio.unb.br/handle/10482/18262.
Full textSubmitted by Ana Cristina Barbosa da Silva (annabds@hotmail.com) on 2015-05-25T14:38:06Z No. of bitstreams: 1 2014_AlessandroFerreiraLeite.pdf: 9950238 bytes, checksum: 5899f0fba30e3075ce700c4440d984f9 (MD5)
Approved for entry into archive by Guimaraes Jacqueline(jacqueline.guimaraes@bce.unb.br) on 2015-05-25T15:49:14Z (GMT) No. of bitstreams: 1 2014_AlessandroFerreiraLeite.pdf: 9950238 bytes, checksum: 5899f0fba30e3075ce700c4440d984f9 (MD5)
Made available in DSpace on 2015-05-25T15:49:14Z (GMT). No. of bitstreams: 1 2014_AlessandroFerreiraLeite.pdf: 9950238 bytes, checksum: 5899f0fba30e3075ce700c4440d984f9 (MD5)
A computação em nuvem tem sido considerada como uma opção para executar aplicações de alto desempenho. Entretanto, enquanto as plataformas de alto desempenho tradicionais como grid e supercomputadores oferecem um ambiente estável quanto à falha, desempenho e número de recursos, a computação em nuvem oferece recursos sob demanda, geralmente com desempenho imprevisível à baixo custo financeiro. Além disso, em ambiente de nuvem, as falhas fazem parte da sua normal operação. No entanto, as nuvens podem ser combinadas, criando uma federação, para superar os limites de uma nuvem muitas vezes com um baixo custo para os usuários. A federação de nuvens pode ajudar tanto os provedores quanto os usuários das nuvens a atingirem diferentes objetivos tais como: reduzir o tempo de execução de uma aplicação, reduzir o custo financeiro, aumentar a disponibilidade do ambiente, reduzir o consumo de energia, entre outros. Por isso, a federação de nuvens pode ser uma solução elegante para evitar o sub-provisionamento de recursos ajudando os provedores a reduzirem os custos operacionais e a reduzir o número de recursos ativos, que outrora ficariam ociosos consumindo energia, por exemplo. No entanto, a federação de nuvens aumenta as opções de recursos disponíveis para os usuários, requerendo, em muito dos casos, conhecimento em administração de sistemas ou em computação em nuvem, bem como um tempo considerável para aprender sobre as opções disponíveis. Neste contexto, surgem algumas questões, tais como: (a) qual dentre os recursos disponíveis é apropriado para uma determinada aplicação? (b) como os usuários podem executar suas aplicações na nuvem e obter um desempenho e um custo financeiro aceitável, sem ter que modificá-las para atender as restrições do ambiente de nuvem? (c) como os usuários não especialistas em nuvem podem maximizar o uso da nuvem, sem ficar dependente de um provedor? (d) como os provedores podem utilizar a federação para reduzir o consumo de energia dos datacenters e ao mesmo tempo atender os acordos de níveis de serviços? A partir destas questões, este trabalho apresenta uma solução para consolidação de aplicações em nuvem federalizadas considerando os acordos de serviços. Nossa solução utiliza um sistema multi-agente para negociar a migração das máquinas virtuais entres as nuvens. Simulações mostram que nossa abordagem pode reduzir em até 46% o consumo de energia e atender os requisitos de qualidade. Nós também desenvolvemos e avaliamos uma solução para executar uma aplicação de bioinformática em nuvens federalizadas, a custo zero. Nesse caso, utilizando a federação, conseguimos diminuir o tempo de execução da aplicação em 22,55%, considerando o seu tempo de execução na melhor nuvem. Além disso, este trabalho apresenta uma arquitetura chamada Excalibur, que possibilita escalar a execução de aplicações comuns em nuvem. Excalibur conseguiu escalar automaticamente a execução de um conjunto de aplicações de bioinformática em até 11 máquinas virtuais, reduzindo o tempo de execução em 63% e o custo financeiro em 84% quando comparado com uma configuração definida pelos usuários. Por fim, este trabalho apresenta um método baseado em linha de produto de software para lidar com as variabilidades dos serviços oferecidos por nuvens de infraestrutura (IaaS), e um sistema que utiliza deste processo para configurar o ambiente e para lidar com falhas de forma automática. O nosso método utiliza modelo de feature estendido com atributos para descrever os recursos e para selecioná-los com base nos objetivos dos usuários. Experimentos realizados com dois provedores diferentes mostraram que utilizando o nosso processo, os usuários podem executar as suas aplicações em um ambiente de nuvem federalizada, sem conhecer as variabilidades e limitações das nuvens. _______________________________________________________________________________________ ABSTRACT
Cloud computing has been seen as an option to execute high performance computing (HPC) applications. While traditional HPC platforms such as grid and supercomputers offer a stable environment in terms of failures, performance, and number of resources, cloud computing offers on-demand resources generally with unpredictable performance at low financial cost. Furthermore, in cloud environment, failures are part of its normal operation. To overcome the limits of a single cloud, clouds can be combined, forming a cloud federation often with minimal additional costs for the users. A cloud federation can help both cloud providers and cloud users to achieve their goals such as to reduce the execution time, to achieve minimum cost, to increase availability, to reduce power consumption, among others. Hence, cloud federation can be an elegant solution to avoid over provisioning, thus reducing the operational costs in an average load situation, and removing resources that would otherwise remain idle and wasting power consumption, for instance. However, cloud federation increases the range of resources available for the users. As a result, cloud or system administration skills may be demanded from the users, as well as a considerable time to learn about the available options. In this context, some questions arise such as: (a) which cloud resource is appropriate for a given application? (b) how can the users execute their HPC applications with acceptable performance and financial costs, without needing to re-engineer the applications to fit clouds’ constraints? (c) how can non-cloud specialists maximize the features of the clouds, without being tied to a cloud provider? and (d) how can the cloud providers use the federation to reduce power consumption of the clouds, while still being able to give service-level agreement (SLA) guarantees to the users? Motivated by these questions, this thesis presents a SLA-aware application consolidation solution for cloud federation. Using a multi-agent system (MAS) to negotiate virtual machine (VM) migrations between the clouds, simulation results show that our approach could reduce up to 46% of the power consumption, while trying to meet performance requirements. Using the federation, we developed and evaluated an approach to execute a huge bioinformatics application at zero-cost. Moreover, we could decrease the execution time in 22.55% over the best single cloud execution. In addition, this thesis presents a cloud architecture called Excalibur to auto-scale cloud-unaware application. Executing a genomics workflow, Excalibur could seamlessly scale the applications up to 11 virtual machines, reducing the execution time by 63% and the cost by 84% when compared to a user’s configuration. Finally, this thesis presents a software product line engineering (SPLE) method to handle the commonality and variability of infrastructure-as-a-service (IaaS) clouds, and an autonomic multi-cloud architecture that uses this method to configure and to deal with failures autonomously. The SPLE method uses extended feature model (EFM) with attributes to describe the resources and to select them based on the users’ objectives. Experiments realized with two different cloud providers show that using the proposed method, the users could execute their application on a federated cloud environment, without needing to know the variability and constraints of the clouds. _______________________________________________________________________________________ RÉSUMÉ
Le cloud computing a été considéré comme une option pour exécuter des applications de calcul haute performance (HPC). Bien que les plateformes traditionnelles de calcul haute performance telles que les grilles et les supercalculateurs offrent un environnement stable du point de vue des défaillances, des performances, et de la taille des ressources, le cloud computing offre des ressources à la demande, généralement avec des performances imprévisibles mais à des coûts financiers abordables. En outre, dans un environnement de cloud, les défaillances sont perçues comme étant ordinaires. Pour surmonter les limites d’un cloud individuel, plusieurs clouds peuvent être combinés pour former une fédération de clouds, souvent avec des coûts supplémentaires légers pour les utilisateurs. Une fédération de clouds peut aider autant les fournisseurs que les utilisateurs à atteindre leurs objectifs tels la réduction du temps d’exécution, la minimisation des coûts, l’augmentation de la disponibilité, la réduction de la consummation d’énergie, pour ne citer que ceux-là. Ainsi, la fédération de clouds peut être une solution élégante pour éviter le sur-approvisionnement, réduisant ainsi les coûts d’exploitation en situation de charge moyenne, et en supprimant des ressources qui, autrement, resteraient inutilisées et gaspilleraient ainsi de énergie. Cependant, la fédération de clouds élargit la gamme des ressources disponibles. En conséquence, pour les utilisateurs, des compétences en cloud computing ou en administration système sont nécessaires, ainsi qu’un temps d’apprentissage considérable pour maîtrises les options disponibles. Dans ce contexte, certaines questions se posent : (a) Quelle ressource du cloud est appropriée pour une application donnée ? (b) Comment les utilisateurs peuvent-ils exécuter leurs applications HPC avec un rendement acceptable et des coûts financiers abordables, sans avoir à reconfigurer les applications pour répondre aux norms et contraintes du cloud ? (c) Comment les non-spécialistes du cloud peuvent-ils maximiser l’usage des caractéristiques du cloud, sans être liés au fournisseur du cloud ? et (d) Comment les fournisseurs de cloud peuvent-ils exploiter la fédération pour réduire la consommation électrique, tout en étant en mesure de fournir un service garantissant les normes de qualité préétablies ? À partir de ces questions, la presente thèse propose une solution de consolidation d’applications pour la fédération de clouds qui garantit le respect des normes de qualité de service. On utilise un système multi-agents (SMA) pour négocier la migration des machines virtuelles entre les clouds. Les résultats de simulations montrent que notre approche pourrait réduire jusqu’à 46% la consommation totale d’énergie, tout en respectant les exigencies de performance. En nous basant sur la fédération de clouds, nous avons développé et évalué une approche pour exécuter une énorme application de bioinformatique à coût zéro. En outre, nous avons pu réduire le temps d’exécution de 22,55% par rapport à la meilleure exécution dans un cloud individuel. Cette thèse présente aussi une architecture de cloud baptisée « Excalibur » qui permet l’adaptation automatique des applications standards pour le cloud. Dans l’exécution d’une chaîne de traitements de la génomique, Excalibur a pu parfaitement mettre à l’échelle les applications sur jusqu’à 11 machines virtuelles, ce qui a réduit le temps d’exécution de 63% et le coût de 84% par rapport à la configuration de l’utilisateur. Enfin, cette thèse présente un processus d’ingénierie des lignes de produits (PLE) pour gérer la variabilité de l’infrastructure à la demande du cloud, et une architecture multi-cloud autonome qui utilise ce processus pour configurer et faire face aux défaillances de manière indépendante. Le processus PLE utilise le modele étendu de fonction (EFM) avec des attributs pour décrire les ressources et les sélectionner en fonction dês objectifs de l’utilisateur. Les expériences réalisées avec deux fournisseurs de cloud différents montrent qu’em utilisant le modèle proposé, les utilisateurs peuvent exécuter leurs applications dans un environnement de clouds fédérés, sans avoir besoin de connaître les variabilités et contraintes du cloud.
Ferreira, Leite Alessandro. "A user-centered and autonomic multi-cloud architecture for high performance computing applications." Thesis, Paris 11, 2014. http://www.theses.fr/2014PA112355/document.
Full textCloud computing has been seen as an option to execute high performance computing (HPC) applications. While traditional HPC platforms such as grid and supercomputers offer a stable environment in terms of failures, performance, and number of resources, cloud computing offers on-Demand resources generally with unpredictable performance at low financial cost. Furthermore, in cloud environment, failures are part of its normal operation. To overcome the limits of a single cloud, clouds can be combined, forming a cloud federation often with minimal additional costs for the users. A cloud federation can help both cloud providers and cloud users to achieve their goals such as to reduce the execution time, to achieve minimum cost, to increase availability, to reduce power consumption, among others. Hence, cloud federation can be an elegant solution to avoid over provisioning, thus reducing the operational costs in an average load situation, and removing resources that would otherwise remain idle and wasting power consumption, for instance. However, cloud federation increases the range of resources available for the users. As a result, cloud or system administration skills may be demanded from the users, as well as a considerable time to learn about the available options. In this context, some questions arise such as: (a) which cloud resource is appropriate for a given application? (b) how can the users execute their HPC applications with acceptable performance and financial costs, without needing to re-Engineer the applications to fit clouds' constraints? (c) how can non-Cloud specialists maximize the features of the clouds, without being tied to a cloud provider? and (d) how can the cloud providers use the federation to reduce power consumption of the clouds, while still being able to give service-Level agreement (SLA) guarantees to the users? Motivated by these questions, this thesis presents a SLA-Aware application consolidation solution for cloud federation. Using a multi-Agent system (MAS) to negotiate virtual machine (VM) migrations between the clouds, simulation results show that our approach could reduce up to 46% of the power consumption, while trying to meet performance requirements. Using the federation, we developed and evaluated an approach to execute a huge bioinformatics application at zero-Cost. Moreover, we could decrease the execution time in 22.55% over the best single cloud execution. In addition, this thesis presents a cloud architecture called Excalibur to auto-Scale cloud-Unaware application. Executing a genomics workflow, Excalibur could seamlessly scale the applications up to 11 virtual machines, reducing the execution time by 63% and the cost by 84% when compared to a user's configuration. Finally, this thesis presents a product line engineering (PLE) process to handle the variabilities of infrastructure-As-A-Service (IaaS) clouds, and an autonomic multi-Cloud architecture that uses this process to configure and to deal with failures autonomously. The PLE process uses extended feature model (EFM) with attributes to describe the resources and to select them based on users' objectives. Experiments realized with two different cloud providers show that using the proposed model, the users could execute their application in a cloud federation environment, without needing to know the variabilities and constraints of the clouds
Izurieta, Iván Carrera. "Performance modeling of MapReduce applications for the cloud." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2014. http://hdl.handle.net/10183/99055.
Full textIn the last years, Cloud Computing has become a key technology that made possible running applications without needing to deploy a physical infrastructure with the advantage of lowering costs to the user by charging only for the computational resources used by the application. The challenge with deploying distributed applications in Cloud Computing environments is that the virtual machine infrastructure should be planned in a way that is time and cost-effective. Also, in the last years we have seen how the amount of data produced by applications has grown bigger than ever. This data contains valuable information that has to be extracted using tools like MapReduce. MapReduce is an important framework to analyze large amounts of data since it was proposed by Google, and made open source by Apache with its Hadoop implementation. The goal of this work is to show that the execution time of a distributed application, namely, a MapReduce application, in a Cloud computing environment, can be predicted using a mathematical model based on theoretical specifications. This prediction is made to help the users of the Cloud Computing environment to plan their deployments, i.e., quantify the number of virtual machines and its characteristics in order to have a lesser cost and/or time. After measuring the application execution time and varying parameters stated in the mathematical model, and after that, using a linear regression technique, the goal is achieved finding a model of the execution time which was then applied to predict the execution time of MapReduce applications with satisfying results. The experiments were conducted in several configurations: namely, private and public clusters, as well as commercial cloud infrastructures, running different MapReduce applications, and varying the number of nodes composing the cluster, as well as the amount of workload given to the application. Experiments showed a clear relation with the theoretical model, revealing that the model is in fact able to predict the execution time of MapReduce applications. The developed model is generic, meaning that it uses theoretical abstractions for the computing capacity of the environment and the computing cost of the MapReduce application. Further work in extending this approach to fit other types of distributed applications is encouraged, as well as including this mathematical model into Cloud services offering MapReduce platforms, in order to aid users plan their deployments.
Zhang, Ziming. "Adaptive Power Management for Autonomic Resource Configuration in Large-scale Computer Systems." Thesis, University of North Texas, 2015. https://digital.library.unt.edu/ark:/67531/metadc804939/.
Full textAli-Eldin, Hassan Ahmed. "Workload characterization, controller design and performance evaluation for cloud capacity autoscaling." Doctoral thesis, Umeå universitet, Institutionen för datavetenskap, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-108398.
Full textMarathe, Aniruddha Prakash. "Evaluation and Optimization of Turnaround Time and Cost of HPC Applications on the Cloud." Diss., The University of Arizona, 2014. http://hdl.handle.net/10150/332825.
Full textGuan, Shichao. "Efficient and Proactive Offloading Techniques for Sustainable and Mobility-aware Resource Management in Heterogeneous Mobile Cloud Environments." Thesis, Université d'Ottawa / University of Ottawa, 2020. http://hdl.handle.net/10393/40562.
Full textOlsson, Philip. "A Study of OpenStack Networking Performance." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-191023.
Full textMolntjänster är en snabbt växande sektor bland mjukvaruföretag. Molnplattformar tillhandahåller tjänster så som utspridning av lagring och beräkningskraft över olika geografiska områden, resursallokering på begäran och flexibla betalningsmetoder. Virtualisering är en teknik som används tillsammans med molnteknologi och erbjuder möjligheten att dela de fysiska resurserna hos en värddator mellan olika virtuella maskiner som kör på samma fysiska dator. Varje virtuell maskin kör sitt egna operativsystem vilket gör att de virtuella maskinerna blir hårdvaruoberoende. Moln och virtualiseringslagret lägger till ytterligare mjukvarulager till servermiljöer för att göra teknikerna möjliga. De extra mjukvarulagrerna orsakar ett pålägg på responstiden vilket kan vara ett problem för applikationer som kräver snabb responstid. Det primära målet i detta examensarbete är att undersöka hur de extra nätverkskomponenterna i molnplattformen OpenStack påverkar responstiden. Nätverkskomonenterna var utvärderade under olika belastningsscenarion och resultaten indikerar att den extra responstiden som orsakades av de extra nätverkskomponenterna inte har allt för stor betydelse på responstiden i den använda nätverksinstallationen. En signifikant perstandaförsämring sågs på applikationerna som körde på den virtuella maskinen vilket stod för den större delen av den ökade responstiden.
Cheng, Luwei, and 程芦伟. "Network performance isolation for virtual machines." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2011. http://hub.hku.hk/bib/B47753183.
Full textpublished_or_final_version
Computer Science
Master
Master of Philosophy
Sreenibha, Reddy Byreddy. "Performance Metrics Analysis of GamingAnywhere with GPU accelerated Nvidia CUDA." Thesis, Blekinge Tekniska Högskola, Institutionen för datalogi och datorsystemteknik, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-16846.
Full textChauhan, Maneesh. "Measurement and Analysis of Networking Performance in Virtualised Environments." Thesis, KTH, Skolan för informations- och kommunikationsteknik (ICT), 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-177327.
Full textMobil cloud computing, har anammat ideerna som beräknings avlastning, att en låg latens, höghastighetsnät för att tillfredsställa tjänsternas kvalitet och användbarhet garantier för mobila applikationer. Nätverks prestanda moln baserade på Xen och VMware virtualiseringslösningar har studerats av forskare, även om de har mestadels fokuserat på nätverksgenomströmning och bandbredd statistik. Arbetet är inriktat på mätning och analys av nätverksprestanda i virtuella maskiner i en liten, KVM baserade datacenter, betonar betydelsen av virtualiserings omkostnader i värd-VM latens och så småningom till den totala fördröjningen upplevs av fjärrklienter. Wealso presentera några användbara verktyg som Driftanalyser, VirtoCalc och Trotter som vi utvecklat för att utföra specifika mätningar och analyser. Vårt arbete visar att en ökning av en VM processor arbetsbelastning har direkta konsekvenser för nätverket Round restider. Vi visar också att Virtualiserings omkostnader (VO) har stor betydelse för början till slut latens och kan bidra med upp till 70 % av rundtrippstid mellan värd och VM. Dessutom är vi noga studera Latency grund Virtualiserings Omkostnader som en nätverksprestanda och undersöka effekterna av CPU-belastning och nätverks arbetsbelastning på den. Vi analyserar också de resursdelningsmönster och deras effekter bland virtuella maskiner i olika storlekar på samma värd. Slutligen, efter att ha observerat ett beroende mellan nätverksprestanda i ett VM och värd CPU belastning, föreslar vi att i en KVM baserad moln installation, arbetsbelastning profilering och optimal processor pinning mekanism kan anvandas effektivt för att reglera VM nätverksprestanda. Resultaten från denna forskning gäller att optimera latens orienterade VM provisione i molnet datacenter, som skulle dra störst latency känsliga mobila molnapplikationer.
Killian, Rudi. "Dynamic superscalar grid for technical debt reduction." Thesis, Cape Peninsula University of Technology, 2018. http://hdl.handle.net/20.500.11838/2726.
Full textOrganizations and the private individual, look to technology advancements to increase their ability to make informed decisions. The motivation for technology adoption by entities sprouting from an innate need for value generation. The technology currently heralded as the future platform to facilitate value addition, is popularly termed cloud computing. The move to cloud computing however, may conceivably increase the obsolescence cycle for currently retained Information Technology (IT) assets. The term obsolescence, applied as the inability to repurpose or scale an information system resource for needed functionality. The incapacity to reconfigure, grow or shrink an IT asset, be it hardware or software is a well-known narrative of technical debt. The notion of emergent technical debt realities is professed to be all but inevitable when informed by Moore’s Law, as technology must inexorably advance. Of more imminent concern however are that major accelerating factors of technical debt are deemed as non-holistic conceptualization and design conventions. Should management of IT assets fail to address technical debt continually, the technology platform would predictably require replacement. The unrealized value, functional and fiscal loss, together with the resultant e-waste generated by technical debt is meaningfully unattractive. Historically, the cloud milieu had evolved from the grid and clustering paradigms which allowed for information sourcing across multiple and often dispersed computing platforms. The parallel operations in distributed computing environments are inherently value adding, as enhanced effective use of resources and efficiency in data handling may be achieved. The predominant information processing solutions that implement parallel operations in distributed environments are abstracted constructs, styled as High Performance Computing (HPC) or High Throughput Computing (HTC). Regardless of the underlying distributed environment, the archetypes of HPC and HTC differ radically in standard implementation. The foremost contrasting factors of parallelism granularity, failover and locality in data handling have recently been the subject of greater academic discourse towards possible fusion of the two technologies. In this research paper, we uncover probable platforms of future technical debt and subsequently recommend redeployment alternatives. The suggested alternatives take the form of scalable grids, which should provide alignment with the contemporary nature of individual information processing needs. The potential of grids, as efficient and effective information sourcing solutions across geographically dispersed heterogeneous systems are envisioned to reduce or delay aspects of technical debt. As part of an experimental investigation to test plausibility of concepts, artefacts are designed to generically implement HPC and HTC. The design features exposed by the experimental artefacts, could provide insights towards amalgamation of HPC and HTC.