To see the other types of publications on this topic, follow the link: Cloud networking.

Dissertations / Theses on the topic 'Cloud networking'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Cloud networking.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Bou, Abdo Jacques. "Efficient and secure mobile cloud networking." Thesis, Paris 6, 2014. http://www.theses.fr/2014PA066551.

Full text
Abstract:
MCC (Mobile Cloud Computing) est un candidat très fort pour le NGN (Next Generation Network) qui permet aux utilisateurs mobiles d’avoir une mobilité étendue, une continuité de service et des performances supérieures. Les utilisateurs peuvent s’attendre à exécuter leurs travaux plus rapidement, avec une faible consommation de batterie et à des prix abordables ; mais ce n’est pas toujours le cas. Diverses applications mobiles ont été développées pour tirer parti de cette nouvelle technologie, mais chacune de ces applications possède ses propres exigences. Plusieurs MCA (Mobile Cloud Architectures) ont été proposées, mais aucune n'a été adaptée pour toutes les applications mobiles, ce qui a mené à une faible satisfaction du client. De plus, l'absence d'un modèle d'affaires (business model) valide pour motiver les investisseurs a empêché son déploiement à l'échelle de production. Cette thèse propose une nouvelle architecture de MCA (Mobile Cloud Architecture) qui positionne l'opérateur de téléphonie mobile au cœur de cette technologie avec un modèle d'affaires de recettes. Cette architecture, nommée OCMCA (Operator Centric Mobile Cloud Architecture), relie l'utilisateur d’un côté et le fournisseur de services Cloud (CSP) de l'autre côté, et héberge un cloud dans son réseau. La connexion OCMCA / utilisateur peut utiliser les canaux multiplex menant à un service beaucoup moins cher pour les utilisateurs, mais avec plus de revenus, et de réduire les embouteillages et les taux de rejet pour l'opérateur. La connexion OCMCA / CSP est basée sur la fédération, ainsi un utilisateur qui a été enregistré avec n’importe quel CSP, peut demander que son environnement soit déchargé de cloud hébergé par l'opérateur de téléphonie mobile afin de recevoir tous les services et les avantages de OCMCA.Les contributions de cette thèse sont multiples. Premièrement, nous proposons OCMCA et nous prouvons qu'il a un rendement supérieur à toutes les autres MCA (Mobile Cloud Architectures). Le modèle d'affaires (business model) de cette architecture se concentre sur la liberté de l'abonnement de l'utilisateur, l'utilisateur peut ainsi être abonné à un fournisseur de cloud et être toujours en mesure de se connecter via cette architecture à son environnement à l'aide du déchargement et de la fédération
Mobile cloud computing is a very strong candidate for the title "Next Generation Network" which empowers mobile users with extended mobility, service continuity and superior performance. Users can expect to execute their jobs faster, with lower battery consumption and affordable prices; however this is not always the case. Various mobile applications have been developed to take advantage of this new technology, but each application has its own requirements. Several mobile cloud architectures have been proposed but none was suitable for all mobile applications which resulted in lower customer satisfaction. In addition to that, the absence of a valid business model to motivate investors hindered its deployment on production scale. This dissertation proposes a new mobile cloud architecture which positions the mobile operator at the core of this technology equipped with a revenue-making business model. This architecture, named OCMCA (Operator Centric Mobile Cloud Architecture), connects the user from one side and the Cloud Service Provider (CSP) from the other and hosts a cloud within its network. The OCMCA/user connection can utilize multicast channels leading to a much cheaper service for the users and more revenues, lower congestion and rejection rates for the operator. The OCMCA/CSP connection is based on federation, thus a user who has been registered with any CSP, can request her environment to be offloaded to the mobile operator's hosted cloud in order to receive all OCMCA's services and benefits
APA, Harvard, Vancouver, ISO, and other styles
2

Bou, Abdo Jacques. "Efficient and secure mobile cloud networking." Electronic Thesis or Diss., Paris 6, 2014. http://www.theses.fr/2014PA066551.

Full text
Abstract:
MCC (Mobile Cloud Computing) est un candidat très fort pour le NGN (Next Generation Network) qui permet aux utilisateurs mobiles d’avoir une mobilité étendue, une continuité de service et des performances supérieures. Les utilisateurs peuvent s’attendre à exécuter leurs travaux plus rapidement, avec une faible consommation de batterie et à des prix abordables ; mais ce n’est pas toujours le cas. Diverses applications mobiles ont été développées pour tirer parti de cette nouvelle technologie, mais chacune de ces applications possède ses propres exigences. Plusieurs MCA (Mobile Cloud Architectures) ont été proposées, mais aucune n'a été adaptée pour toutes les applications mobiles, ce qui a mené à une faible satisfaction du client. De plus, l'absence d'un modèle d'affaires (business model) valide pour motiver les investisseurs a empêché son déploiement à l'échelle de production. Cette thèse propose une nouvelle architecture de MCA (Mobile Cloud Architecture) qui positionne l'opérateur de téléphonie mobile au cœur de cette technologie avec un modèle d'affaires de recettes. Cette architecture, nommée OCMCA (Operator Centric Mobile Cloud Architecture), relie l'utilisateur d’un côté et le fournisseur de services Cloud (CSP) de l'autre côté, et héberge un cloud dans son réseau. La connexion OCMCA / utilisateur peut utiliser les canaux multiplex menant à un service beaucoup moins cher pour les utilisateurs, mais avec plus de revenus, et de réduire les embouteillages et les taux de rejet pour l'opérateur. La connexion OCMCA / CSP est basée sur la fédération, ainsi un utilisateur qui a été enregistré avec n’importe quel CSP, peut demander que son environnement soit déchargé de cloud hébergé par l'opérateur de téléphonie mobile afin de recevoir tous les services et les avantages de OCMCA.Les contributions de cette thèse sont multiples. Premièrement, nous proposons OCMCA et nous prouvons qu'il a un rendement supérieur à toutes les autres MCA (Mobile Cloud Architectures). Le modèle d'affaires (business model) de cette architecture se concentre sur la liberté de l'abonnement de l'utilisateur, l'utilisateur peut ainsi être abonné à un fournisseur de cloud et être toujours en mesure de se connecter via cette architecture à son environnement à l'aide du déchargement et de la fédération
Mobile cloud computing is a very strong candidate for the title "Next Generation Network" which empowers mobile users with extended mobility, service continuity and superior performance. Users can expect to execute their jobs faster, with lower battery consumption and affordable prices; however this is not always the case. Various mobile applications have been developed to take advantage of this new technology, but each application has its own requirements. Several mobile cloud architectures have been proposed but none was suitable for all mobile applications which resulted in lower customer satisfaction. In addition to that, the absence of a valid business model to motivate investors hindered its deployment on production scale. This dissertation proposes a new mobile cloud architecture which positions the mobile operator at the core of this technology equipped with a revenue-making business model. This architecture, named OCMCA (Operator Centric Mobile Cloud Architecture), connects the user from one side and the Cloud Service Provider (CSP) from the other and hosts a cloud within its network. The OCMCA/user connection can utilize multicast channels leading to a much cheaper service for the users and more revenues, lower congestion and rejection rates for the operator. The OCMCA/CSP connection is based on federation, thus a user who has been registered with any CSP, can request her environment to be offloaded to the mobile operator's hosted cloud in order to receive all OCMCA's services and benefits
APA, Harvard, Vancouver, ISO, and other styles
3

Turull, Daniel. "Network virtualization as enabler for cloud networking." Licentiate thesis, KTH, Network Systems Laboratory (NS Lab), 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-185454.

Full text
Abstract:
The Internet has exponentially grown and now it is part of our everyday life. Internet services and applications rely on back-end servers that are deployed on local servers and data centers. With the growing use of data centers and cloud computing, the locations of these servers have been externalized and centralized, taking advantage of economies of scale. However, some applications need to define complex network topologies and require more than simple connectivity to the remote sites. Therefore, the network part of cloud computing, what is called cloud networking, needs to be improved and simplified. This thesis argues that network virtualization permits to fill the missing gap and we propose a network virtualization abstraction layer to ease the use of cloud networking for the end users. We implement a software prototype of our ideas using OpenFlow. We also evaluate our prototype with state of the art controllers that has similar functionalities for network virtualization. A second part of this thesis focuses on developing a tool for performance testing. We have improved the widely used tool pktgen with receiver functionalities. We use pktgen to generate traffic for our experiments with network virtualization.

QC 20160428

APA, Harvard, Vancouver, ISO, and other styles
4

Olsson, Philip. "A Study of OpenStack Networking Performance." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-191023.

Full text
Abstract:
Cloud computing is a fast-growing sector among software companies. Cloud platforms provide services such as spreading out storage and computational power over several geographic locations, on-demand resource allocation and flexible payment options. Virtualization is a technology used in conjunction with cloud technology and offers the possibility to share the physical resources of a host machine by hosting several virtual machines on the same physical machine. Each virtual machine runs its operating system which makes the virtual machines hardware independent. The cloud and virtualization layers add additional layers of software to the server environments to provide the services. The additional layers cause an overlay in latency which can be problematic for latency sensitive applications. The primary goal of this thesis is to investigate how the networking components impact the latency in an OpenStack cloud compared to a traditional deployment. The networking components were benchmarked under different load scenarios, and the results indicate that the additional latency added by the networking components is not too significant in the used network setup. Instead, a significant performance degradation could be seen on the applications running in the virtual machine which caused most of the added latency in the cloud environment.
Molntjänster är en snabbt växande sektor bland mjukvaruföretag. Molnplattformar tillhandahåller tjänster så som utspridning av lagring och beräkningskraft över olika geografiska områden, resursallokering på begäran och flexibla betalningsmetoder. Virtualisering är en teknik som används tillsammans med molnteknologi och erbjuder möjligheten att dela de fysiska resurserna hos en värddator mellan olika virtuella maskiner som kör på samma fysiska dator. Varje virtuell maskin kör sitt egna operativsystem vilket gör att de virtuella maskinerna blir hårdvaruoberoende. Moln och virtualiseringslagret lägger till ytterligare mjukvarulager till servermiljöer för att göra teknikerna möjliga. De extra mjukvarulagrerna orsakar ett pålägg på responstiden vilket kan vara ett problem för applikationer som kräver snabb responstid. Det primära målet i detta examensarbete är att undersöka hur de extra nätverkskomponenterna i molnplattformen OpenStack påverkar responstiden. Nätverkskomonenterna var utvärderade under olika belastningsscenarion och resultaten indikerar att den extra responstiden som orsakades av de extra nätverkskomponenterna inte har allt för stor betydelse på responstiden i den använda nätverksinstallationen. En signifikant perstandaförsämring sågs på applikationerna som körde på den virtuella maskinen vilket stod för den större delen av den ökade responstiden.
APA, Harvard, Vancouver, ISO, and other styles
5

Medhioub, Houssem. "Architectures et mécanismes de fédération dans les environnements cloud computing et cloud networking." Thesis, Evry, Institut national des télécommunications, 2015. http://www.theses.fr/2015TELE0009/document.

Full text
Abstract:
Présenté dans la littérature comme une nouvelle technologie, le Cloud Computing est devenu incontournable dans la mise en place et la fourniture des services informatiques. Cette thèse s’inscrit dans le contexte de cette nouvelle technologie qui est en mesure de transformer la mise en place, la gestion et l’utilisation des systèmes d’information. L'adoption et la vulgarisation du Cloud ont été ralenties par la jeunesse même des concepts et l'hétérogénéité des solutions existantes. Cette difficulté d'adoption se manifeste par l'absence de standard, l'hétérogénéité des architectures et des API, le Vendor Lock-In imposé par les leaders du marché et des manques qui ralentissent la fédération. La motivation principale de la thèse est de simplifier l'adoption du cloud et la migration vers ses environnements et technologies. Notre objectif est de proposer des solutions d'interopérabilité et de fédération dans le Cloud. Le travail de recherche s’est aussi articulé autour de deux grands axes. Le premier concerne le rapprochement des réseaux du futur et des Clouds. Le deuxième axe concerne l'interopérabilité et la fédération entre solutions et services cloud. Une analyse de l’état de l’art sur le Cloud Computing et le Cloud Networking, a permis de confirmer des manques pressentis et de proposer deux architectures de fédération Cloud. La première architecture permet le rapprochement entre le Cloud Computing et le Cloud Networking. La seconde architecture facilite l'interopérabilité et le courtage de services Cloud. L'étude des deux architectures a fait ressortir deux composants primordiaux et essentiels pour assurer la fédération: une interface générique et un système d'échange de messages. Ces deux composants correspondent à deux contributions centrales de la thèse et reflètent l’ensemble des contributions (quatre au total) du travail de recherche
Presented in the literature as a new technology, Cloud Computing has become essential in the development and delivery of IT services. Given the innovative potential of Cloud, our thesis was conducted in the context of this promising technology. It was clear that the Cloud would change the way we develop, manage and use information systems. However, the adoption and popularization of Cloud were slow and difficult given the youth of the concepts and heterogeneity of the existing solutions. This difficulty in adoption is reflected by the lack of standard, the presence of heterogeneous architectures and APIs, the introduction of Vendor Lock-In imposed by the market leaders and the lack of cloud federation principles and facilitators. The main motivation for our PhD is to simplify the adoption of the cloud paradigm and the migration to cloud environments and technologies. Our goal has consequently been to improve interoperability and enable federation in the cloud. The thesis focused on two main areas. The first concerns the convergence of future networks and clouds and the second the improvement of federation and interoperability between heterogeneous cloud solutions and services. Based on our work in state of the art about Cloud Computing and Cloud Networking, we defined in this thesis two architectures for Cloud federation. The first architecture enables the merging (convergence) of Cloud Computing and Cloud Networking. The second architecture addresses interoperability between services and proposes cloud-brokering solutions. The study enabled the identification of two essential components for cloud federation, namely: a generic interface and a message exchange system. These two components have been two contributions of our thesis. The proposed federation architectures and these two components summarize the four major contributions of our work
APA, Harvard, Vancouver, ISO, and other styles
6

Medhioub, Houssem. "Architectures et mécanismes de fédération dans les environnements cloud computing et cloud networking." Electronic Thesis or Diss., Evry, Institut national des télécommunications, 2015. http://www.theses.fr/2015TELE0009.

Full text
Abstract:
Présenté dans la littérature comme une nouvelle technologie, le Cloud Computing est devenu incontournable dans la mise en place et la fourniture des services informatiques. Cette thèse s’inscrit dans le contexte de cette nouvelle technologie qui est en mesure de transformer la mise en place, la gestion et l’utilisation des systèmes d’information. L'adoption et la vulgarisation du Cloud ont été ralenties par la jeunesse même des concepts et l'hétérogénéité des solutions existantes. Cette difficulté d'adoption se manifeste par l'absence de standard, l'hétérogénéité des architectures et des API, le Vendor Lock-In imposé par les leaders du marché et des manques qui ralentissent la fédération. La motivation principale de la thèse est de simplifier l'adoption du cloud et la migration vers ses environnements et technologies. Notre objectif est de proposer des solutions d'interopérabilité et de fédération dans le Cloud. Le travail de recherche s’est aussi articulé autour de deux grands axes. Le premier concerne le rapprochement des réseaux du futur et des Clouds. Le deuxième axe concerne l'interopérabilité et la fédération entre solutions et services cloud. Une analyse de l’état de l’art sur le Cloud Computing et le Cloud Networking, a permis de confirmer des manques pressentis et de proposer deux architectures de fédération Cloud. La première architecture permet le rapprochement entre le Cloud Computing et le Cloud Networking. La seconde architecture facilite l'interopérabilité et le courtage de services Cloud. L'étude des deux architectures a fait ressortir deux composants primordiaux et essentiels pour assurer la fédération: une interface générique et un système d'échange de messages. Ces deux composants correspondent à deux contributions centrales de la thèse et reflètent l’ensemble des contributions (quatre au total) du travail de recherche
Presented in the literature as a new technology, Cloud Computing has become essential in the development and delivery of IT services. Given the innovative potential of Cloud, our thesis was conducted in the context of this promising technology. It was clear that the Cloud would change the way we develop, manage and use information systems. However, the adoption and popularization of Cloud were slow and difficult given the youth of the concepts and heterogeneity of the existing solutions. This difficulty in adoption is reflected by the lack of standard, the presence of heterogeneous architectures and APIs, the introduction of Vendor Lock-In imposed by the market leaders and the lack of cloud federation principles and facilitators. The main motivation for our PhD is to simplify the adoption of the cloud paradigm and the migration to cloud environments and technologies. Our goal has consequently been to improve interoperability and enable federation in the cloud. The thesis focused on two main areas. The first concerns the convergence of future networks and clouds and the second the improvement of federation and interoperability between heterogeneous cloud solutions and services. Based on our work in state of the art about Cloud Computing and Cloud Networking, we defined in this thesis two architectures for Cloud federation. The first architecture enables the merging (convergence) of Cloud Computing and Cloud Networking. The second architecture addresses interoperability between services and proposes cloud-brokering solutions. The study enabled the identification of two essential components for cloud federation, namely: a generic interface and a message exchange system. These two components have been two contributions of our thesis. The proposed federation architectures and these two components summarize the four major contributions of our work
APA, Harvard, Vancouver, ISO, and other styles
7

Jouin, Lionel. "Network Service Mesh Solving Cloud Native IMS Networking Needs." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-425220.

Full text
Abstract:
With the growing demand for mobile networks and specially IPMultimedia subsystem (IMS), new cloud native orchestration tools providing more flexibility and efficiency start to be used within telecommunication companies in order to improve the robustness and the reliability of these systems. However, Kubernetes, the most used among cloud native orchestration tools does not fulfill completelyall the needs and use cases in terms of networking the telecommunication industry meets. Network Service Mesh (NSM), a new Cloud Native Computing Foundation (CNCF) project, aiming to address complex networking use cases in Kubernetes might solve the different issues IP multimedia subsystem face. Detailed designs and implementations using Network Service Mesh coupled with diverse networking technologies are shown in this thesis with the objectiveof solving the networking IP multimedia subsystem requirements (e.g. the NAT issue and the secondary network). In addition, an analysis and an evaluation of Network Service Mesh is given together with a presentation of the ability of this new project to bring solutions toIP Multimedia subsystem based on a cloud native technology.
APA, Harvard, Vancouver, ISO, and other styles
8

Pourvali, Mahsa. "Resilience of Cloud Networking Services for Large Scale Outages." Scholar Commons, 2017. http://scholarcommons.usf.edu/etd/6664.

Full text
Abstract:
Cloud infrastructure services are enabling organizations and enterprises to outsource a wide range of computing, storage, and networking needs to external service providers. These offerings make extensive use of underlying network virtualization, i.e., virtual network (VN) embedding, techniques to provision and interconnect customized storage/computing resource pools across large network substrates. However, as cloud-based services continue to gain traction, there is a growing need to address a range of resiliency concerns, particularly with regards to large-scale outages. These conditions can be triggered by events such as natural disasters, malicious man-made attacks, and even cascading power failures. Overall, a wide range of studies have looked at network virtualization survivability, with most efforts focusing on pre-fault protection strategies to set aside backup datacenter and network bandwidth resources. These contributions include single node/link failure schemes as well as recent studies on correlated multi-failure \disaster" recovery schemes. However, pre-fault provisioning is very resource-intensive and imposes high costs for clients. Moreover this approach cannot guarantee recovery under generalized multi-failure conditions. Although post-fault restoration (remapping) schemes have also been studied, the effectiveness of these methods is constrained by the scale of infrastructure damage. As a result there is a pressing need to investigate longer-term post-fault infrastructure repair strategies to minimize VN service disruption. However this is a largely unexplored area and requires specialized consideration as damaged infrastructures will likely be repaired in a time-staged, incremental manner, i.e., progressive recovery. Furthermore, more specialized multicast VN (MVN) services are also being used to support a range of content distribution and real-time streaming needs over cloud-based infrastructures. In general, these one-to-many services impose more challenging requirements in terms of geographic coverage, delay, delay variation, and reliability. Now some recent studies have looked at MVN embedding and survivability design. In particular, the latter contributions cover both pre-fault protection and post-fault restoration methods, and also include some multi-failure recovery techniques. Nevertheless, there are no known efforts that incorporate risk vulnerabilities into the MVN embedding process. Indeed, there is a strong need to develop such methods in order to reduce the impact of large-scale outages, and this remains an open topic area. In light of the above, this dissertation develops some novel solutions to further improve the resiliency of the network virtualization services in the presence of large outages. Foremost, new multi-stage (progressive) infrastructure repair strategies are proposed to improve the post-fault recovery of VN services. These contributions include advanced simulated annealing metaheuristics as well as more scalable polynomial-time heuristic algorithms. Furthermore, enhanced \risk-aware" mapping solutions are also developed to achieve more reliable multicast (MVN) embedding, providing a further basis to develop more specialized repair strategies in the future. The performance of these various solutions is also evaluated extensively using custom-developed simulation models.
APA, Harvard, Vancouver, ISO, and other styles
9

Andersson, Elin, and Abdel Ahmid. "OpenStack Networking Scheduler : RESOURCE MANAGEMENT AND SCHEDULINGFOR CLOUD NETWORK SERVICES." Thesis, KTH, Skolan för informations- och kommunikationsteknik (ICT), 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-154597.

Full text
Abstract:
The opportunity of vitalizing recourses is becoming increasingly attractive, and traditional machine virtualization is no longer the only thing in mind. Today, virtualization also concerns resources such as firewalls, routers and switches. For this to be possible and successful, a flexible and evolutionary scheduler is necessary. The focus of this thesis project has been to explore and develop a generic cross-service scheduler between the computing service Nova and networking service Neutron. We have designed a generic foundation for a scheduler and the aim in the future is to use this solution for other services in OpenStack such as volume, storage and image. Our starting point was the filter scheduler in Nova, which originally is designed for scheduling virtual machines (VMs) rather than scheduling network resources. A unique cross-service scheduler between Nova and Neutron is implemented and discussed in this paper. This thesis presents the design, implementation and functional verification of the Neutron scheduler and cross-service communication. The cross-service scheduling is designed so that Nova and Neutron are loosely coupled. By loosely coupled we mean that Nova and Neutron are as independent of each other as possible. The report concludes with a discussion about the project process and implementation process, and the how the solution can be developed further e.g. by implementing CLI commands for increasing the functionality of filter chains.
APA, Harvard, Vancouver, ISO, and other styles
10

Adduci, Pietro. "Software-Defined Networking: lo standard Openflow." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2014. http://amslaurea.unibo.it/7241/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
11

Peddireddy, Vidyadhar reddy. "Enhancement of Networking Capabilities in P2P OpenStack." Thesis, Blekinge Tekniska Högskola, Institutionen för datavetenskap, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-17822.

Full text
Abstract:
In recent times, there’s been a trend towards setting up smaller clouds at the edge of the network and interconnecting them across multiple sites. In these scenarios, the software used for managing the resources should be flexible enough to scale. Considering OpenStack the most widely used cloud software, It is observed that the compute service has shown performance degradation when the deployment reaches fewer hundreds of nodes. Finding out solutions to address the scalability issue in OpenStack, Ericsson has developed a new architecture that supports massive scalability of OpenStack clouds. However, the challenges with multicloud networking in P2P OpenStack remained unsolved. This thesis work as an extension to Ericsson’s P2P OpenStack project investigates various multi-cloud networking techniques and proposes two decentralized designs for cross Neutron networking in P2P OpenStack. The design-1 is based on OpenStack Tricircle project and design-2 is based on VPNaaS. This thesis work implements VPNaaS design to support the automatic interconnection of Virtual machines that belong to the same user but deployed in different OpenStack clouds. We evaluate this thesis for control plane operation under two different scenarios namely single user case and multiple users cases. In both scenarios, request-response time is chosen as an evaluating parameter. Results show that there is an increase in request-response time when users in the system are increased.
APA, Harvard, Vancouver, ISO, and other styles
12

Liu, Binghan. "Software Defined Networking and Tunneling for Mobile Networks." Thesis, KTH, Kommunikationssystem, CoS, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-118376.

Full text
Abstract:
With the deployment of Long Term Evolution (LTE) networks, mobile networks will become an important infrastructure component in the cloud ecosystem.  However, in the cloud computing era, traditional routing and switching platforms do not meet the requirements of this new trend, especially in a mobile network environment. With the recent advances in software switches and efficient virtualization using commodity servers, Software Defined Networking (SDN) has emerged as a powerful technology to meet the new requirements for supporting a new generation of cloud service. This thesis describers an experimental investigation of cloud computing, SDN, and a mobile network’s packet core. The design of a mobile network exploiting the evolution of SDN is also presented. The actual implementation consists of a GTP enabled Open vSwitch together with the transparent mode of mobile network SDN evolution. Open vSwitch is a SDN product designed for computer networks. The implementation extends Open vSwitch with an implementation of the GTP protocol. This extension enables Open vSwitch to be an excellent SDN component for mobile networks. In transparent mode, a cloud data center is deployed without making any modification to the existing mobile networks.  In the practical evaluation of the GTP-U tunnel protocol implementation, the measured metrics are UDP and TCP throughput, end-to-end latency and jitter.  Two experiments have been conducted and described in the evaluation chapter. Cloud computing has become one of the hottest Internet topics. It is attractive for the mobile network to adopt cloud computing technology in order to enjoy the benefits of cloud computing. For example, to reduce network construction cost, make the network deployment more flexible, etc. This thesis presents an potential direction for mobile network cloud computing. Since this thesis relies on open source projects, readers may use the results to explore a feasible direction for mobile network cloud computing evolution.
Med utbyggnaden av långa (LTE) Term Evolution nätverk, mobila nätverk kommer blivit en viktig infrastruktur komponent i molnet ekosystemet. Men i cloud computing eran, uppfyller traditionella routing och switching plattformar inte kraven i denna nya trend, särskilt i ett mobilnät miljö. Med de senaste framstegen i programvara växlar och effektiv virtualisering påråvaror servrar, programvarustyrd Nätverk (SDN) har utvecklats till en kraftfull teknik för att möta de nya kraven för att stödja en ny generation av molntjänst. Denna avhandling beskrivarna en försöksverksamhet inriktad undersökning av cloud computing, SDN och ett mobilnät är Packet Core. Utformningen av ett mobilnät utnyttja SDN utveckling presenteras också. Det faktiska genomförandet består av en GTP aktiverad Open Vswitch tillsammans med transparent läge av mobilnätet SDN evolution. Öppna Vswitch är en SDN-produkt avsedd för datornätverk. Genomförandet utökar Open Vswitch med en implementering av GTP-protokollet. Denna uppgradering gör Open Vswitch vara som en utmärkt SDN komponent för mobila nätverk. I transparent läge är ett moln datacenter utplacerade utan göra eventuella ändringar till befintliga mobilnät. I den praktiska utvärderingen av GTP-U tunnel protokollimplementering, de uppmätta mått är UDP och TCP genomströmning, end-to-end-latens, jitter och paketförluster.  Tvåexperiment har utförts i utvärderingen kapitlet. Cloud computing har blivit en av de hetaste av Internet. Således kan framtiden för det mobila nätet ocksåanta teknik cloud computing och dra nytta av cloud computing. Till exempel minska kostnaderna nätbyggnad, gör nätverket distribuera mer flexibla, etc. .. Denna avhandling presenterar en möjlig inriktning för mobilnät cloud computing. Eftersom denna avhandling bygger påopen source-projekt, läsarna använda resultatet av den att utforska möjliga riktning mobilnät cloud computing utveckling.
APA, Harvard, Vancouver, ISO, and other styles
13

Chauhan, Maneesh. "Measurement and Analysis of Networking Performance in Virtualised Environments." Thesis, KTH, Skolan för informations- och kommunikationsteknik (ICT), 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-177327.

Full text
Abstract:
Mobile cloud computing, having embraced the ideas like computation ooading, mandates a low latency, high speed network to satisfy the quality of service and usability assurances for mobile applications. Networking performance of clouds based on Xen and Vmware virtualisation solutions has been extensively studied by researchers, although, they have mostly been focused on network throughput and bandwidth metrics. This work focuses on the measurement and analysis of networking performance of VMs in a small, KVM based data centre, emphasising the role of virtualisation overheads in the Host-VM latency and eventually to the overall latency experienced by remote clients. We also present some useful tools such as Driftanalyser, VirtoCalc and Trotter that we developed for carrying out specific measurements and analysis. Our work proves that an increase in a VM's CPU workload has direct implications on the network Round trip times. We also show that Virtualisation Overheads (VO) have significant bearing on the end to end latency and can contribute up to 70% of the round trip time between the Host and VM. Furthermore, we thoroughly study Latency due to Virtualisation Overheads as a networking performance metric and analyse the impact of CPU loads and networking workloads on it. We also analyse the resource sharing patterns and their effects amongst VMs of different sizes on the same Host. Finally, having observed a dependency between network performance of a VM and the Host CPU load, we suggest that in a KVM based cloud installation, workload profiling and optimum processor pinning mechanism can be e ectively utilised to regulate network performance of the VMs. The ndings from this research work are applicable to optimising latency oriented VM provisioning in the cloud data centres, which would benefit most latency sensitive mobile cloud applications.
Mobil cloud computing, har anammat ideerna som beräknings avlastning, att en låg latens, höghastighetsnät för att tillfredsställa tjänsternas kvalitet och användbarhet garantier för mobila applikationer. Nätverks prestanda moln baserade på Xen och VMware virtualiseringslösningar har studerats av forskare, även om de har mestadels fokuserat på nätverksgenomströmning och bandbredd statistik. Arbetet är inriktat på mätning och analys av nätverksprestanda i virtuella maskiner i en liten, KVM baserade datacenter, betonar betydelsen av virtualiserings omkostnader i värd-VM latens och så småningom till den totala fördröjningen upplevs av fjärrklienter. Wealso presentera några användbara verktyg som Driftanalyser, VirtoCalc och Trotter som vi utvecklat för att utföra specifika mätningar och analyser. Vårt arbete visar att en ökning av en VM processor arbetsbelastning har direkta konsekvenser för nätverket Round restider. Vi visar också att Virtualiserings omkostnader (VO) har stor betydelse för början till slut latens och kan bidra med upp till 70 % av rundtrippstid mellan värd och VM. Dessutom är vi noga studera Latency grund Virtualiserings Omkostnader som en nätverksprestanda och undersöka effekterna av CPU-belastning och nätverks arbetsbelastning på den. Vi analyserar också de resursdelningsmönster och deras effekter bland virtuella maskiner i olika storlekar på samma värd. Slutligen, efter att ha observerat ett beroende mellan nätverksprestanda i ett VM och värd CPU belastning, föreslar vi att i en KVM baserad moln installation, arbetsbelastning profilering och optimal processor pinning mekanism kan anvandas effektivt för att reglera VM nätverksprestanda. Resultaten från denna forskning gäller att optimera latens orienterade VM provisione i molnet datacenter, som skulle dra störst latency känsliga mobila molnapplikationer.
APA, Harvard, Vancouver, ISO, and other styles
14

Ibn, Khedher Hatem. "Optimization and virtualization techniques adapted to networking." Thesis, Evry, Institut national des télécommunications, 2018. http://www.theses.fr/2018TELE0007/document.

Full text
Abstract:
Dans cette thèse, on présente nos travaux sur la virtualisation dans le contexte de la réplication des serveurs de contenu vidéo. Les travaux couvrent la conception d’une architecture de virtualisation pour laquelle on présente aussi plusieurs algorithmes qui peuvent réduire les couts globaux à long terme et améliorer la performance du système. Le travail est divisé en plusieurs parties : les solutions optimales, les solutions heuristiques pour des raisons de passage à l’échelle, l’orchestration des services, l’optimisation multi-objective, la planification de services dans des réseaux actifs et complexes et l'intégration d'algorithmes dans une plate-forme réelle
In this thesis, we designed and implemented a tool which performs optimizations that reduce the number of migrations necessary for a delivery task. We present our work on virtualization in the context of replication of video content servers. The work covers the design of a virtualization architecture for which there are also several algorithms that can reduce overall costs and improve system performance. The thesis is divided into several parts: optimal solutions, greedy (heuristic) solutions for reasons of scalability, orchestration of services, multi-objective optimization, service planning in complex active networks, and integration of algorithms in real platform. This thesis is supported by models, implementations and simulations which provide results that showcase our work, quantify the importance of evaluating optimization techniques and analyze the trade-off between reducing operator cost and enhancing end user satisfaction index
APA, Harvard, Vancouver, ISO, and other styles
15

Marini, Riccardo. "Software Defined Networking Architectures for LoRaWAN." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2019.

Find full text
Abstract:
This thesis proposes new solutions for LoRaWAN networks taking advantages of Software Defined Networking architectures. In particular, an analysis of the current implementation of the Adaptive Data Rate mechanism developed by LoRaWAN standard, as well as a proposal of a new algorithm, will be provided. This will be addressed by considering both a cloud-based and a fog-based architecture in order to observe differences between the two approaches in a number of different scenarios. The proposed algorithms and the two architectures are compared via numerical results achieved through simulations and experimental tests.
APA, Harvard, Vancouver, ISO, and other styles
16

Zhao, Jian, and 趙建. "Performance modeling and optimization solutions for networking systems." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2014. http://hdl.handle.net/10722/196434.

Full text
Abstract:
This thesis targets at modeling and resolving practical problems using mathematical tools in two representative networking systems nowadays, i.e., peer-to-peer (P2P) video streaming system and cloud computing system. In the first part, we study how to mitigate the following tussle between content service providers and ISPs in P2P video streaming systems: network-agnostic P2P protocol designs bring lots of inter-ISP traffic and increase traffic relay cost of ISPs; in turn, ISPs start to throttle P2P packets, which significantly deteriorates P2P streaming performance. First, we investigate the problem in a mesh-based P2P live streaming system. We use end-to-end streaming delays as performance, and quantify the amount of inter-ISP traffic with the number of copies of the live streams imported into each ISP. Considering multiple ISPs at different bandwidth levels, we model the generic relationship between the volume of inter-ISP traffic and streaming performance, which provides useful insights on the design of effective locality-aware peer selection protocols and server deployment strategies across multiple ISPs. Next, we study a similar problem in a hybrid P2P-cloud CDN system for VoD streaming. We characterize the relationship between the costly bandwidth consumption from the cloud CDN and the inter-ISP traffic. We apply a loss network model to derive the bandwidth consumption under any given chunk distribution pattern among peer caches and any streaming request dispatching strategy among ISPs, and derive the optimal peer caching and request dispatching strategies which minimize the bandwidth demand from the cloud CDN. Based on the fundamental insights from our analytical results, we design a locality-aware, hybrid P2P-cloud CDN streaming protocol. In the second part, we study the profit maximization and cost minimization problems in Infrastructure-as- a- Service (IaaS) cloud systems. The first problem is how a geo-distributed cloud system should price its datacenter resources at different locations, such that its overall profit is maximized over long-term operation. We design an efficient online algorithm for dynamic pricing of VM resources across datacenters, together with job scheduling and server provisioning in each datacenter, to maximize the cloud's profit over the long run. Theoretical analysis shows that our algorithm can schedule jobs within their respective deadlines, while achieving a time-averaged overall profit closely approaching the offline maximum, which is computed by assuming perfect information on future job arrivals is freely available. The second problem is how federated clouds should trade their computing resources among each other to reduce the cost, by exploiting diversities of different clouds' workloads and operational costs. We formulate a global cost minimization problem among multiple clouds under the cooperative scenario where each individual cloud's workload and cost information is publicly available. Taking into considerations jobs with disparate length, a non-preemptive approximation algorithm for leftover job migration and new job scheduling is designed. Given to the selfishness of individual clouds, we further design a randomized double auction mechanism to elicit clouds' truthful bidding for buying or selling virtual machines. The auction mechanism is proven to be truthful, and to guarantee the same approximation ratio to what the cooperative approximation algorithm achieves.
published_or_final_version
Computer Science
Doctoral
Doctor of Philosophy
APA, Harvard, Vancouver, ISO, and other styles
17

Jamaliannasrabadi, Saba. "High Performance Computing as a Service in the Cloud Using Software-Defined Networking." Bowling Green State University / OhioLINK, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1433963448.

Full text
APA, Harvard, Vancouver, ISO, and other styles
18

Pham, Van Quan. "Cloud-native optical network automation platforms." Electronic Thesis or Diss., Institut polytechnique de Paris, 2022. http://www.theses.fr/2022IPPAS005.

Full text
Abstract:
La gestion et le contrôle des communications optiques se transforment pour intégrer de nouvelles fonctionnalités telles que la gestion de réseau basée sur l'intention, l'automatisation du contrôle en boucle fermée et l'orchestration multipartite. Ces fonctionnalités sont motivées par les nouvelles exigences de connectivité entre les centres de données (datacenters) pour permettre le déploiement de futures générations de services telles que les applications au-delà de la 5G (Beyond 5G or B5G) et 6G offertes à la périphérie des réseaux optiques. La prochaine génération d'architectures de gestion et de contrôle des réseaux optiques comportera des principes de mise en réseau définie par logiciel (SDN) en relation avec la désagrégation des futurs systèmes optiques. Les contrôleurs et gestionnaires de réseaux optiques actuels sont intrinsèquement propriétaires et sont donc limités en termes d'ouverture, d'évolutivité et de flexibilité. Cette thèse de doctorat étudie et propose des architectures logicielles disruptives avec : (i) leurs fonctions de contrôle pour les systèmes optiques et (ii) leurs fonctions de gestion pour les services de connexion optique des réseaux optiques ouverts désagrégés. Après avoir presenté les architectures SDN dans le contexte et les contraintes des réseaux de commutation et de transmission optiques, la thèse explique les défis techniques des réseaux optiques actuels évoluant vers le contrôle des réseaux optiques partiellement désagrégés comme première étape de transition; et vers le contrôle des réseaux optiques entièrement désagrégés comme étape ultime. La nouvelles plates-formes logicielles d'automatisation des réseaux optiques avec leurs fonctions de contrôle basées sur des micro-services sont décrites de manière pragmatique en tenant compte des logiciels libres ainsi que de plusieurs forums de standardisation définissant leurs langages, leurs modèles de données et leurs protocoles pour les équipements, la topologie du réseau et les services de communication. Ensuite, la thèse décrit comment les fonctions de contrôle sont conçues comme des fonctions de réseau virtualisées permettant une intégration et un développement continus des plateformes de réseaux optiques natives en nuage (Cloud). Les fonctions de calcul de chemins de canal optique automatisées conçues comme des services sont d'abord abordées. Ces services de calcul de chemin de canal optique sont décrits en expliquant comment les contraintes de routage définies par l'évolution des fonctionnalités des systèmes optiques peuvent être intégrées dans les moteurs de calcul de chemin (PCE). Plusieurs algorithmes PCE pour le routage des canaux optiques et l'allocation sur la grille spectral sont présentés et leurs performances sont comparées en termes d'allocation bonne ou éventuellement optimale sur le spectre optique. A la suite des concepts de fonctions de calcul de chemin de canaux optiques automatisées en conçues comme des services , la thèse propose des fonctions de défragmentation automatisée des canaux optiques conçues comme des services pour réarranger les placements des canaux optiques afin d’obtenir une meilleure et éventuellement optimale utilisation de la grille de spectre optique pour gagner des ressources. A partir des évaluations de ces différentes fonctions de contrôle optique basées sur des conteneurs, plusieurs scénarios d'automatisation du contrôle des canaux optiques sont décrits pour prouver leurs concepts en utilisant un banc réseau dans un laboratoire, et pour démontrer les applications potentielles des VNF optiques. Enfin, la thèse conclut sur la synthèse de ces travaux de recherche et les défis futurs pour rendre le contrôle et la gestion des réseaux optiques plus unifiés et rationalisés afin de permettre aux communications optiques d'être conçues et d'être un atout conçues comme des services de connectivité pour les services de future génération
Optical communication management and control are transforming to integrate new capabilities such as intent-based network management, closed-loop control automation, and multi-stakeholder orchestration. These capabilities are driven by the new connectivity requirements between data centers to enable future generations of services: Beyond 5G (B5G) and 6G applications offered at the edges of optical networks. The next generation of optical network management and control architectures will entail Software-Defined Networking (SDN) principles for the disaggregation of future optical systems. The current optical network controllers and managers are intrinsically proprietary and, consequently, restricted in openness, scalability, and flexibility. This Ph.D. thesis investigates and proposes breakthrough software architectures with: (i) their control functions for the optical systems and (ii) their management functions for optical connection services of Open Disaggregated Optical Networks. After explaining SDN architectures in the context and the constraints of optical switching and transmission networks, the thesis explains the challenges of current optical networks to transition towards the control of Partially Disaggregated Optical Networks as a first step and the control of Fully Disaggregated Optical Networks as the ultimate step. Novel software-defined optical network automation platforms with control functions based on micro-services are described pragmatically, considering open-source software frameworks and several open forums providing their languages and their data models. Their protocols are being developed for devices, network topology, and communication services. Next, the thesis described how control functions are designed as cloud-native network functions (CNF), enabling continuous integration and continuous development of cloud-native optical networking platforms. Automated optical channel path computation functions as services a re first addressed. These optical channel path computation services are described by explaining how the routing constraints defined by the evolutions of optical system capabilities can be integrated into the path computation engines (PCE). Several PCE algorithms for optical channel routing and spectrum allocation are presented, and their performances are compared in terms of reasonable or possibly optimal spectrum allocation.Subsequent to the concepts of automated optical channel path computation functions as services, the thesis proposes automated optical channel defragmentation functions as services to re-arrange the placements of optical channels for better and possibly optimal use of the spectrum grid to gain resources.From the evaluations of these different container-based optical control functions, several optical channel control automation scenarios are described to prove their concepts using a network bench in a lab and demonstrate the potential applications of optical CNFs.Finally, the thesis concludes with the synthesis of these research works and the future challenges to make the control and the management of optical networks more unified and streamlined to enable optical communications to be designed and an asset as connectivity services for future generation services
APA, Harvard, Vancouver, ISO, and other styles
19

Pagola, Moledo Santiago. "Vendor-Independent Software-Defined Networking : Beyond The Hype." Thesis, Linköpings universitet, Databas och informationsteknik, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-157456.

Full text
Abstract:
Software-Defined Networking (SDN) is an emerging trend in networking that offers a number of advantages such as smoother network management over traditional networks. By decoupling the control and data planes from network elements, a huge amount of new opportunities arise, especially in network virtualization. In cloud datacenters, where virtualization plays a fundamental role, SDN presents itself as the perfect candidate to ease infrastructure management and to ensure correct operation. Even if the original SDN ideology advocates openness of source and interfaces, multiple networking vendors offer their own proprietary solutions. In this work, an open-source SDN solution, named Tungsten Fabric, will be deployed in a virtualized datacenter and a number of SDN-related use-cases will be examined. The main goal of this work is to determine whether Tungsten Fabric can deliver the same set of use-cases as a proprietary solution from Juniper, named Contrail Cloud. Finally, this work will give some guidelines on whether open-source SDN is the right candidate for Ericsson.
APA, Harvard, Vancouver, ISO, and other styles
20

Zhang, Jie Zhang. "Designing and Building Efficient HPC Cloud with Modern Networking Technologies on Heterogeneous HPC Clusters." The Ohio State University, 2018. http://rave.ohiolink.edu/etdc/view?acc_num=osu1532737201524604.

Full text
APA, Harvard, Vancouver, ISO, and other styles
21

Zaripov, Behruz. "Analysis of Fog Networking Procedures in Heterogeneous Wireless Networks." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2017.

Find full text
Abstract:
The purpose of this study is to provide a general framework of the latest trends of mobile network architectures. The two main architectures treated in this work are Cloud-RAN and Fog-RAN. We give descriptions of both architectures, then show advantages and disadvantages of them. We mainly focus on the performance of Fog-RAN model taking into account only computation and communication latencies. In order to analyse our Fog-RAN architecture, we measure the impact of traffic produced in our network, by using 3 different policies: Random Policy, Maximum Available Capacity Policy and Nearest Node Policy. Furthermore, we measure the impact of delay by fixing the amount of traffic generated by the network. Numerical results of our considered scenarios show that the maximum available capacity policy outperforms two other polices, when the traffic produced in the network is very high. When the traffic is very low, the best policy is the nearest node one. On the other hand, by fixing the amount of traffic we show that when the delay threshold is from 1-3 ms the Maximum capacity policy performs better than two other policies. When the delay threshold is greater than 5 ms the Nearest Node policy shows better results.
APA, Harvard, Vancouver, ISO, and other styles
22

Ekanayake, Mudiyanselage Wijaya Dheeshakthi. "An SDN-based Framework for QoSaware Mobile Cloud Computing." Thesis, Université d'Ottawa / University of Ottawa, 2016. http://hdl.handle.net/10393/35117.

Full text
Abstract:
In mobile cloud computing (MCC), rich mobile application data is processed at the cloud infrastructure by reliving resource limited mobile devices from computationally complex tasks. However, due to the ubiquitous and mobility nature, providing time critical rich applications over remote cloud infrastructure is a challenging task for mobile application service providers. Therefore, according to the literature, close proximity placement of cloud services has been identified as a way to achieve lower end-to-end access delay and thereby provide a higher quality of experience (QoE) for rich mobile application users. However, providing a higher Quality of Service (QoS) with mobility is still a challenge within close proximity clouds. Access delay to a closely placed cloud tends to be increased over time when users move away from the cloud. However, reactive resource relocation mechanism proposed in literature does not provide a comprehensive mechanism to guarantee the QoS and as well as to minimize service provisioning cost for mobile cloud service providers. As a result, using the benefits of SDN and the data plane programmability with logically centralized controllers, a resource allocation framework was proposed for IaaS mobile clouds with regional datacenters. The user mobility problem was analyzed within SDN-enabled wireless networks and addressed the possible service level agreement violations that could occur with inter-regional mobility. The proposed framework is composed of an optimization algorithm to provide seamless cloud service during user mobility. Further a service provisioning cost minimization criteria was considered during an event of resource allocation and inter-regional user mobility.
APA, Harvard, Vancouver, ISO, and other styles
23

Subrani, Nikolas. "Alternative per la configurazione dell'infrastruttura di rete in piattaforme Openstack." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2016. http://amslaurea.unibo.it/11185/.

Full text
Abstract:
In questo scritto si analizzeranno alcune alternative nella configurazione di rete della piattaforma di cloud computing open source OpenStack. Verrà mostrata un’ installazione in ambiente di laboratorio di un cluster completo basato sulla release Liberty di Openstack, per poi modificarne la componente dedicata al Networking in modo da sfruttare diversi plugin e diversi protocolli. Si osserverà il traffico generato all’interno e verso l’esterno del sistema Openstack in modo da avere un quadro generale del comportamento dell’infrastruttura.
APA, Harvard, Vancouver, ISO, and other styles
24

Cavalcanti, Gustavo Andriolli de Siqueira. "Alocacão ótima de recursos para infraestruturas virtuais confiáveis." Universidade do Estado de Santa Catarina, 2014. http://tede.udesc.br/handle/handle/2043.

Full text
Abstract:
Made available in DSpace on 2016-12-12T20:22:52Z (GMT). No. of bitstreams: 1 Gustavo Andriolli S Cavalcanti.pdf: 1327760 bytes, checksum: bc2279bc162dd158d8e1d52a2a7df585 (MD5) Previous issue date: 2014-07-11
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior
Virtual infrastructures (VIs) sets of virtual machines interconnected by links and routers also virtual consolidated the dynamic provisioning of computing and communication resources and their services providers (InPs) face a challenge in choose the better approach to allocating and reserve these resources. Resource allocation (such as CPU, disk, memory, bandwidth) is a complex problem that needs to satisfy different goals: users expect to run their applications on survivable VIs, while InPs aim to maximize profits, minimize costs and reduce substrate fragmentation. However, there is a dichotomy between maximizing VI survivability, by sparsely allocating resources to decrease the impact of substrate failures, and minimizing substrate fragmentation, by co-locating VIs. In this context, we propose a mixed integer programming model to allocate resources considering the joint coordination of survivability and fragmentation. Experimental results show that it is possible to enhance VI survivability without significantly impacting substrate fragmentation.
Com a consolidação do provisionamento dinâmico de Infraestruturas Virtuais (IVs) - conjuntos de máquinas virtuais interconectadas por enlaces e roteadores também virtuais -, provedores de serviço (InPs) enfrentam um desafio na escolha da melhor abordagem para alocação e reserva de recursos computacionais e de comunicação. ´E fato que a alocação de recursos (como CPU, disco, memória, largura de banda) é um problema complexo que precisa satisfazer diferentes objetivos: usuários esperam executar suas aplicações em IVs eficientes e confiáveis, enquanto InPs objetivam maximizar lucros, minimizar custos e reduzir a fragmentação do substrato físico. Sobretudo, há uma dicotomia entre maximizar a confiabilidade de IVs, alocando recursos esparsos para diminuir o impacto de falhas no substrato, e minimizara fragmentação do substrato, co-alocando IVs. Nesse contexto, ´e proposto um modelo de programação inteira mista para alocar recursos considerando a coordenação conjunta de confiabilidade e fragmentação. Resultados experimentais mostram que é possível aprimorar a confiabilidade de IVs sem impactar significativamente na fragmentação do substrato.
APA, Harvard, Vancouver, ISO, and other styles
25

Wickboldt, Juliano Araújo. "Flexible and integrated resource management for IaaS cloud environments based on programmability." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2015. http://hdl.handle.net/10183/131894.

Full text
Abstract:
Nuvens de infraestrutura como serviço (IaaS) estão se tornando um ambiente habitual para execução de aplicações modernas da Internet. Muitas plataformas de gerenciamento de nuvem estão disponíveis para aquele que deseja construir uma nuvem de IaaS privada ou pública (e.g., OpenStack, Eucalyptus, OpenNebula). Um aspecto comum do projeto de plataformas atuais diz respeito ao seu modelo de controle caixa-preta. Em geral, as plataformas de gerenciamento de nuvem são distribuídas com um conjunto de estratégias de alocação de recursos embutida em seu núcleo. Dessa forma, os administradores de nuvem têm poucas oportunidades de influenciar a maneira como os recursos são realmente gerenciados (e.g., posicionamento de máquinas virtuais ou seleção caminho de enlaces virtuais). Os administradores poderiam se beneficiar de personalizações em estratégias de gerenciamento de recursos, por exemplo, para atingir os objetivos específicos de cada ambiente ou a fim de permitir a alocação de recursos orientada à aplicação. Além disso, as preocupações acerca do gerenciamento de recursos em nuvens se dividem geralmente em computação, armazenamento e redes. Idealmente, essas três preocupações deveriam ser abordadas no mesmo nível de importância por implementações de plataformas. No entanto, ao contrário do gerenciamento de computação e armazenamento, que têm sido amplamente estudados, o gerenciamento de redes em ambientes de nuvem ainda é bastante incipiente. A falta de flexibilidade e suporte desequilibrado para o gerenciamento de recursos dificulta a adoção de nuvens como um ambiente de execução viável para muitas aplicações modernas da Internet com requisitos rigorosos de elasticidade e qualidade do serviço. Nesta tese, um novo conceito de plataforma de gerenciamento de nuvem é introduzido onde o gerenciamento de recursos flexível é obtido pela adição de programabilidade no núcleo da plataforma. Além disso, uma API simplificada e orientada a objetos é introduzida a fim de permitir que os administradores escrevam e executem programas de gerenciamento de recursos para lidar com todos os tipos de recursos a partir de um único ponto. Uma plataforma é apresentada como uma prova de conceito, incluindo um conjunto de adaptadores para lidar com tecnologias de virtualização e de redes modernas, como redes definidas por software com OpenFlow, Open vSwitches e Libvirt. Dois estudos de caso foram realizados a fim de avaliar a utilização de programas de gerenciamento de recursos para implantação e otimização de aplicações através de uma rede emulada usando contêineres de virtualização Linux e Open vSwitches operando sob o protocolo OpenFlow. Os resultados mostram a viabilidade da abordagem proposta e como os programas de implantação e otimização são capazes de alcançar diferentes objetivos definidos pelo administrador.
Infrastructure as a Service (IaaS) clouds are becoming an increasingly common way to deploy modern Internet applications. Many cloud management platforms are available for users that want to build a private or public IaaS cloud (e.g., OpenStack, Eucalyptus, OpenNebula). A common design aspect of current platforms is their black-box-like controlling nature. In general, cloud management platforms ship with one or a set of resource allocation strategies hard-coded into their core. Thus, cloud administrators have few opportunities to influence how resources are actually managed (e.g., virtual machine placement or virtual link path selection). Administrators could benefit from customizations in resource management strategies, for example, to achieve environment specific objectives or to enable application-oriented resource allocation. Furthermore, resource management concerns in clouds are generally divided into computing, storage, and networking. Ideally, these three concerns should be addressed at the same level of importance by platform implementations. However, as opposed to computing and storage management, which have been extensively investigated, network management in cloud environments is rather incipient. The lack of flexibility and unbalanced support for resource management hinders the adoption of clouds as a viable execution environment for many modern Internet applications with strict requirements for elasticity or Quality of Service. In this thesis, a new concept of cloud management platform is introduced where resource management is made flexible by the addition of programmability to the core of the platform. Moreover, a simplified object-oriented API is introduced to enable administrators to write and run resource management programs to handle all kinds of resources from a single point. An implementation is presented as a proof of concept, including a set of drivers to deal with modern virtualization and networking technologies, such as software-defined networking with OpenFlow, Open vSwitches, and Libvirt. Two case studies are conducted to evaluate the use of resource management programs for the deployment and optimization of applications over an emulated network using Linux virtualization containers and Open vSwitches running the OpenFlow protocol. Results show the feasibility of the proposed approach and how deployment and optimization programs are able to achieve different objectives defined by the administrator.
APA, Harvard, Vancouver, ISO, and other styles
26

Ibn, Khedher Hatem. "Optimization and virtualization techniques adapted to networking." Electronic Thesis or Diss., Evry, Institut national des télécommunications, 2018. http://www.theses.fr/2018TELE0007.

Full text
Abstract:
Dans cette thèse, on présente nos travaux sur la virtualisation dans le contexte de la réplication des serveurs de contenu vidéo. Les travaux couvrent la conception d’une architecture de virtualisation pour laquelle on présente aussi plusieurs algorithmes qui peuvent réduire les couts globaux à long terme et améliorer la performance du système. Le travail est divisé en plusieurs parties : les solutions optimales, les solutions heuristiques pour des raisons de passage à l’échelle, l’orchestration des services, l’optimisation multi-objective, la planification de services dans des réseaux actifs et complexes et l'intégration d'algorithmes dans une plate-forme réelle
In this thesis, we designed and implemented a tool which performs optimizations that reduce the number of migrations necessary for a delivery task. We present our work on virtualization in the context of replication of video content servers. The work covers the design of a virtualization architecture for which there are also several algorithms that can reduce overall costs and improve system performance. The thesis is divided into several parts: optimal solutions, greedy (heuristic) solutions for reasons of scalability, orchestration of services, multi-objective optimization, service planning in complex active networks, and integration of algorithms in real platform. This thesis is supported by models, implementations and simulations which provide results that showcase our work, quantify the importance of evaluating optimization techniques and analyze the trade-off between reducing operator cost and enhancing end user satisfaction index
APA, Harvard, Vancouver, ISO, and other styles
27

Soares, João Monteiro. "Integration of the cloud computing paradigm with the opeerator network’s infrastructure." Doctoral thesis, Universidade de Aveiro, 2015. http://hdl.handle.net/10773/14854.

Full text
Abstract:
Doutoramento em Engenharia Informática
The proliferation of Internet access allows that users have the possibility to use services available directly through the Internet, which translates in a change of the paradigm of using applications and in the way of communicating, popularizing in this way the so-called cloud computing paradigm. Cloud computing brings with it requirements at two different levels: at the cloud level, usually relying in centralized data centers, where information technology and network resources must be able to guarantee the demand of such services; and at the access level, i.e., depending on the service being consumed, different quality of service is required in the access network, which is a Network Operator (NO) domain. In summary, there is an obvious network dependency. However, the network has been playing a relatively minor role, mostly as a provider of (best-effort) connectivity within the cloud and in the access network. The work developed in this Thesis enables for the effective integration of cloud and NO domains, allowing the required network support for cloud. We propose a framework and a set of associated mechanisms for the integrated management and control of cloud computing and NO domains to provide endto- end services. Moreover, we elaborate a thorough study on the embedding of virtual resources in this integrated environment. The study focuses on maximizing the host of virtual resources on the physical infrastructure through optimal embedding strategies (considering the initial allocation of resources as well as adaptations through time), while at the same time minimizing the costs associated to energy consumption, in single and multiple domains. Furthermore, we explore how the NO can take advantage of the integrated environment to host traditional network functions. In this sense, we study how virtual network Service Functions (SFs) should be modelled and managed in a cloud environment and enhance the framework accordingly. A thorough evaluation of the proposed solutions was performed in the scope of this Thesis, assessing their benefits. We implemented proof of concepts to prove the added value, feasibility and easy deployment characteristics of the proposed framework. Furthermore, the embedding strategies evaluation has been performed through simulation and Integer Linear Programming (ILP) solving tools, and it showed that it is possible to reduce the physical infrastructure energy consumption without jeopardizing the virtual resources acceptance. This fact can be further increased by allowing virtual resource adaptation through time. However, one should have in mind the costs associated to adaptation processes. The costs can be minimized, but the virtual resource acceptance can be also reduced. This tradeoff has also been subject of the work in this Thesis.
A proliferação do acesso à Internet permite aos utilizadores usar serviços disponibilizados diretamente através da Internet, o que se traduz numa mudança de paradigma na forma de usar aplicações e na forma de comunicar, popularizando desta forma o conceito denominado de cloud computing. Cloud computing traz consigo requisitos a dois níveis: ao nível da própria cloud, geralmente dependente de centros de dados centralizados, onde as tecnologias de informação e recursos de rede têm que ser capazes de garantir as exigências destes serviços; e ao nível do acesso, ou seja, dependendo do serviço que esteja a ser consumido, são necessários diferentes níveis de qualidade de serviço na rede de acesso, um domínio do operador de rede. Em síntese, existe uma clara dependência da cloud na rede. No entanto, o papel que a rede tem vindo a desempenhar neste âmbito é reduzido, sendo principalmente um fornecedor de conectividade (best-effort) tanto no dominio da cloud como no da rede de acesso. O trabalho desenvolvido nesta Tese permite uma integração efetiva dos domínios de cloud e operador de rede, dando assim à cloud o efetivo suporte da rede. Para tal, apresentamos uma plataforma e um conjunto de mecanismos associados para gestão e controlo integrado de domínios cloud computing e operador de rede por forma a fornecer serviços fim-a-fim. Além disso, elaboramos um estudo aprofundado sobre o mapeamento de recursos virtuais neste ambiente integrado. O estudo centra-se na maximização da incorporação de recursos virtuais na infraestrutura física por meio de estratégias de mapeamento ótimas (considerando a alocação inicial de recursos, bem como adaptações ao longo do tempo), enquanto que se minimizam os custos associados ao consumo de energia. Este estudo é feito para cenários de apenas um domínio e para cenários com múltiplos domínios. Além disso, exploramos como o operador de rede pode aproveitar o referido ambiente integrado para suportar funções de rede tradicionais. Neste sentido, estudamos como as funções de rede virtualizadas devem ser modeladas e geridas num ambiente cloud e estendemos a plataforma de acordo com este conceito. No âmbito desta Tese foi feita uma avaliação extensa das soluções propostas, avaliando os seus benefícios. Implementámos provas de conceito por forma a demonstrar as mais-valias, viabilidade e fácil implantação das soluções propostas. Além disso, a avaliação das estratégias de mapeamento foi realizada através de ferramentas de simulação e de programação linear inteira, mostrando que é possível reduzir o consumo de energia da infraestrutura física, sem comprometer a aceitação de recursos virtuais. Este aspeto pode ser melhorado através da adaptação de recursos virtuais ao longo do tempo. No entanto, deve-se ter em mente os custos associados aos processos de adaptação. Os custos podem ser minimizados, mas isso implica uma redução na aceitação de recursos virtuais. Esta compensação foi também um tema abordado nesta Tese.
APA, Harvard, Vancouver, ISO, and other styles
28

Henriksson, Johannes, and Alexander Magnusson. "Impact of using cloud-based SDNcontrollers on the networkperformance." Thesis, Mälardalens högskola, Akademin för innovation, design och teknik, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:mdh:diva-44152.

Full text
Abstract:
Software-Defined Networking (SDN) is a network architecture that differs from traditionalnetwork planes. SDN has tree layers: infrastructure, controller, and application. Thegoal of SDN is to simplify management of larger networks by centralizing control into thecontroller layer instead of having it in the infrastructure. Given the known advantages ofSDN networks, and the flexibility of cloud computing. We are interested if this combinationof SDN and cloud services affects network performance, and what affect the cloud providersphysical location have on the network performance. These points are important whenSDN becomes more popular in enterprise networks. This seems like a logical next step inSDN, centralizing branch networks into one cloud-based SDN controller. These questionswere created with a literature studies and answered with an experimentation method. Theexperiments consist of two network topologies both locally hosted SDN (baseline) and cloudhosted SDN. The topology used Zodiac FX switches and Linux hosts. The following metricswas measured: throughput, latency, jitter, packet loss, and time to add new hosts. Theconclusion is that SDN as a cloud service is possible and does not significantly affect networkperformance. One limitation with this thesis was the hardware, resulting in big fluctuationin throughput and packet loss.
APA, Harvard, Vancouver, ISO, and other styles
29

Li, Ming. "User-Centric Security and Privacy Mechanisms in Untrusted Networking and Computing Environments." Digital WPI, 2011. https://digitalcommons.wpi.edu/etd-dissertations/323.

Full text
Abstract:
"Our modern society is increasingly relying on the collection, processing, and sharing of digital information. There are two fundamental trends: (1) Enabled by the rapid developments in sensor, wireless, and networking technologies, communication and networking are becoming more and more pervasive and ad hoc. (2) Driven by the explosive growth of hardware and software capabilities, computation power is becoming a public utility and information is often stored in centralized servers which facilitate ubiquitous access and sharing. Many emerging platforms and systems hinge on both dimensions, such as E-healthcare and Smart Grid. However, the majority information handled by these critical systems is usually sensitive and of high value, while various security breaches could compromise the social welfare of these systems. Thus there is an urgent need to develop security and privacy mechanisms to protect the authenticity, integrity and confidentiality of the collected data, and to control the disclosure of private information. In achieving that, two unique challenges arise: (1) There lacks centralized trusted parties in pervasive networking; (2) The remote data servers tend not to be trusted by system users in handling their data. They make existing security solutions developed for traditional networked information systems unsuitable. To this end, in this dissertation we propose a series of user-centric security and privacy mechanisms that resolve these challenging issues in untrusted network and computing environments, spanning wireless body area networks (WBAN), mobile social networks (MSN), and cloud computing. The main contributions of this dissertation are fourfold. First, we propose a secure ad hoc trust initialization protocol for WBAN, without relying on any pre-established security context among nodes, while defending against a powerful wireless attacker that may or may not compromise sensor nodes. The protocol is highly usable for a human user. Second, we present novel schemes for sharing sensitive information among distributed mobile hosts in MSN which preserves user privacy, where the users neither need to fully trust each other nor rely on any central trusted party. Third, to realize owner-controlled sharing of sensitive data stored on untrusted servers, we put forward a data access control framework using Multi-Authority Attribute-Based Encryption (ABE), that supports scalable fine-grained access and on-demand user revocation, and is free of key-escrow. Finally, we propose mechanisms for authorized keyword search over encrypted data on untrusted servers, with efficient multi-dimensional range, subset and equality query capabilities, and with enhanced search privacy. The common characteristic of our contributions is they minimize the extent of trust that users must place in the corresponding network or computing environments, in a way that is user-centric, i.e., favoring individual owners/users."
APA, Harvard, Vancouver, ISO, and other styles
30

Bianchi, Stefano. "Design and Implementation of a Cloud Infrastructure for Distributed Scientific Calculation." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2016.

Find full text
Abstract:
Cloud computing enables independent end users and applications to share data and pooled resources, possibly located in geographically distributed Data Centers, in a fully transparent way. This need is particularly felt by scientific applications to exploit distributed resources in efficient and scalable way for the processing of big amount of data. This paper proposes an open so- lution to deploy a Platform as a service (PaaS) over a set of multi- site data centers by applying open source virtualization tools to facilitate operation among virtual machines while optimizing the usage of distributed resources. An experimental testbed is set up in Openstack environment to obtain evaluations with different types of TCP sample connections to demonstrate the functionality of the proposed solution and to obtain throughput measurements in relation to relevant design parameters.
APA, Harvard, Vancouver, ISO, and other styles
31

Hamze, Mohamad. "Autonomie, sécurité et QoS de bout en bout dans un environnement de Cloud Computing." Thesis, Dijon, 2015. http://www.theses.fr/2015DIJOS033/document.

Full text
Abstract:
De nos jours, le Cloud Networking est considéré comme étant l'un des domaines de recherche innovants au sein de la communauté de recherche du Cloud Computing. Les principaux défis dans un environnement de Cloud Networking concernent non seulement la garantie de qualité de service (QoS) et de sécurité mais aussi sa gestion en conformité avec un accord de niveau de service (SLA) correspondant. Dans cette thèse, nous proposons un Framework pour l'allocation des ressources conformément à un SLA établi de bout en bout entre un utilisateur de services Cloud (CSU) et plusieurs fournisseurs de services Cloud (CSP) dans un environnement de Cloud Networking (architectures d’inter-Cloud Broker et Fédération). Nos travaux se concentrent sur les services Cloud de types NaaS et IaaS. Ainsi, nous proposons l'auto-établissement de plusieurs types de SLA ainsi que la gestion autonome des ressources de Cloud correspondantes en conformité avec ces SLA en utilisant des gestionnaires autonomes spécifiques de Cloud. De plus, nous étendons les architectures et les SLA proposés pour offrir un niveau de service intégrant une garantie de sécurité. Ainsi, nous permettons aux gestionnaires autonomes de Cloud d'élargir leurs objectifs de gestion autonome aux fonctions de sécurité (auto-protection) tout en étudiant l'impact de la sécurité proposée sur la garantie de QoS. Enfin, nous validons notre architecture avec différents scénarios de simulation. Nous considérons dans le cadre de ces simulations des applications de vidéoconférence et de calcul intensif afin de leur fournir une garantie de QoS et de sécurité dans un environnement de gestion autonome des ressources du Cloud. Les résultats obtenus montrent que nos contributions permettent de bonnes performances pour ce type d’applications. En particulier, nous observons que l'architecture de type Broker est la plus économique, tout en assurant les exigences de QoS et de sécurité. De plus, nous observons que la gestion autonome des ressources du Cloud permet la réduction des violations, des pénalités et limite l'impact de la sécurité sur la garantie de la QoS
Today, Cloud Networking is one of the recent research areas within the Cloud Computing research communities. The main challenges of Cloud Networking concern Quality of Service (QoS) and security guarantee as well as its management in conformance with a corresponding Service Level Agreement (SLA). In this thesis, we propose a framework for resource allocation according to an end-to-end SLA established between a Cloud Service User (CSU) and several Cloud Service Providers (CSPs) within a Cloud Networking environment (Inter-Cloud Broker and Federation architectures). We focus on NaaS and IaaS Cloud services. Then, we propose the self-establishing of several kinds of SLAs and the self-management of the corresponding Cloud resources in conformance with these SLAs using specific autonomic cloud managers. In addition, we extend the proposed architectures and the corresponding SLAs in order to deliver a service level taking into account security guarantee. Moreover, we allow autonomic cloud managers to expand the self-management objectives to security functions (self-protection) while studying the impact of the proposed security on QoS guarantee. Finally, our proposed architecture is validated by different simulation scenarios. We consider, within these simulations, videoconferencing and intensive computing applications in order to provide them with QoS and security guarantee in a Cloud self-management environment. The obtained results show that our contributions enable good performances for these applications. In particular, we observe that the Broker architecture is the most economical while ensuring QoS and security requirements. In addition, we observe that Cloud self-management enables violations and penalties’ reduction as well as limiting security impact on QoS guarantee
APA, Harvard, Vancouver, ISO, and other styles
32

Chivukula, Krishna Varaynya. "Monitoring and Analysis of CPU load relationships between Host and Guests in a Cloud Networking Infrastructure : An Empirical Study." Thesis, Blekinge Tekniska Högskola, Institutionen för kommunikationssystem, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-10810.

Full text
APA, Harvard, Vancouver, ISO, and other styles
33

Mechtri, Marouen. "Virtual networked infrastructure provisioning in distributed cloud environments." Thesis, Evry, Institut national des télécommunications, 2014. http://www.theses.fr/2014TELE0028/document.

Full text
Abstract:
L'informatique en nuage (Cloud Computing) a émergé comme un nouveau paradigme pour offrir des ressources informatiques à la demande et pour externaliser des infrastructures logicielles et matérielles. Le Cloud Computing est rapidement et fondamentalement en train de révolutionner la façon dont les services informatiques sont mis à disposition et gérés. Ces services peuvent être demandés à partir d’un ou plusieurs fournisseurs de Cloud d’où le besoin de la mise en réseau entre les composants des services informatiques distribués dans des emplacements géographiquement répartis. Les utilisateurs du Cloud veulent aussi déployer et instancier facilement leurs ressources entre les différentes plateformes hétérogènes de Cloud Computing. Les fournisseurs de Cloud assurent la mise à disposition des ressources de calcul sous forme des machines virtuelles à leurs utilisateurs. Par contre, ces clients veulent aussi la mise en réseau entre leurs ressources virtuelles. En plus, ils veulent non seulement contrôler et gérer leurs applications, mais aussi contrôler la connectivité réseau et déployer des fonctions et des services de réseaux complexes dans leurs infrastructures virtuelles dédiées. Les besoins des utilisateurs avaient évolué au-delà d'avoir une simple machine virtuelle à l'acquisition de ressources et de services virtuels complexes, flexibles, élastiques et intelligents. L'objectif de cette thèse est de permettre le placement et l’instanciation des ressources complexes dans des infrastructures de Cloud distribués tout en permettant aux utilisateurs le contrôle et la gestion de leurs ressources. En plus, notre objectif est d'assurer la convergence entre les services de cloud et de réseau. Pour atteindre cela, nous proposons des algorithmes de mapping d’infrastructures virtuelles dans les centres de données et dans le réseau tout en respectant les exigences des utilisateurs. Avec l'apparition du Cloud Computing, les réseaux traditionnels sont étendus et renforcés avec des réseaux logiciels reposant sur la virtualisation des ressources et des fonctions réseaux. En plus, le nouveau paradigme d'architecture réseau (Software Defined Networks) est particulièrement pertinent car il vise à offrir la programmation du réseau et à découpler, dans un équipement réseau, la partie plan de données de la partie plan de contrôle. Dans ce contexte, la première partie propose des algorithmes optimaux (exacts) et heuristiques de placement pour trouver le meilleur mapping entre les demandes des utilisateurs et les infrastructures sous-jacentes, tout en respectant les exigences exprimées dans les demandes. Cela inclut des contraintes de localisation permettant de placer une partie des ressources virtuelles dans le même nœud physique. Ces contraintes assurent aussi le placement des ressources dans des nœuds distincts. Les algorithmes proposés assurent le placement simultané des nœuds et des liens virtuels sur l’infrastructure physique. Nous avons proposé aussi un algorithme heuristique afin d’accélérer le temps de résolution et de réduire la complexité du problème. L'approche proposée se base sur la technique de décomposition des graphes et la technique de couplage des graphes bipartis. Dans la troisième partie, nous proposons un cadriciel open source (framework) permettant d’assurer la mise en réseau dynamique entre des ressources Cloud distribués et l’instanciation des fonctions réseau dans l’infrastructure virtuelle de l’utilisateur. Ce cadriciel permettra de déployer et d’activer les composants réseaux afin de mettre en place les demandes des utilisateurs. Cette solution se base sur un gestionnaire des ressources réseaux "Cloud Network Gateway Manager" et des passerelles logicielles permettant d’établir la connectivité dynamique et à la demande entre des ressources cloud et réseau. Le CNG-Manager offre le contrôle de la partie réseau et prend en charge le déploiement des fonctions réseau nécessaires dans l'infrastructure virtuelle des utilisateurs
Cloud computing emerged as a new paradigm for on-demand provisioning of IT resources and for infrastructure externalization and is rapidly and fundamentally revolutionizing the way IT is delivered and managed. The resulting incremental Cloud adoption is fostering to some extent cloud providers cooperation and increasing the needs of tenants and the complexity of their demands. Tenants need to network their distributed and geographically spread cloud resources and services. They also want to easily accomplish their deployments and instantiations across heterogeneous cloud platforms. Traditional cloud providers focus on compute resources provisioning and offer mostly virtual machines to tenants and cloud services consumers who actually expect full-fledged (complete) networking of their virtual and dedicated resources. They not only want to control and manage their applications but also control connectivity to easily deploy complex network functions and services in their dedicated virtual infrastructures. The needs of users are thus growing beyond the simple provisioning of virtual machines to the acquisition of complex, flexible, elastic and intelligent virtual resources and services. The goal of this thesis is to enable the provisioning and instantiation of this type of more complex resources while empowering tenants with control and management capabilities and to enable the convergence of cloud and network services. To reach these goals, the thesis proposes mapping algorithms for optimized in-data center and in-network resources hosting according to the tenants' virtual infrastructures requests. In parallel to the apparition of cloud services, traditional networks are being extended and enhanced with software networks relying on the virtualization of network resources and functions especially through network resources and functions virtualization. Software Defined Networks are especially relevant as they decouple network control and data forwarding and provide the needed network programmability and system and network management capabilities. In such a context, the first part proposes optimal (exact) and heuristic placement algorithms to find the best mapping between the tenants' requests and the hosting infrastructures while respecting the objectives expressed in the demands. This includes localization constraints to place some of the virtual resources and services in the same host and to distribute other resources in distinct hosts. The proposed algorithms achieve simultaneous node (host) and link (connection) mappings. A heuristic algorithm is proposed to address the poor scalability and high complexity of the exact solution(s). The heuristic scales much better and is several orders of magnitude more efficient in terms of convergence time towards near optimal and optimal solutions. This is achieved by reducing complexity of the mapping process using topological patterns to map virtual graph requests to physical graphs representing respectively the tenants' requests and the providers' physical infrastructures. The proposed approach relies on graph decomposition into topology patterns and bipartite graphs matching techniques. The third part propose an open source Cloud Networking framework to achieve cloud and network resources provisioning and instantiation in order to respectively host and activate the tenants' virtual resources and services. This framework enables and facilitates dynamic networking of distributed cloud services and applications. This solution relies on a Cloud Network Gateway Manager and gateways to establish dynamic connectivity between cloud and network resources. The CNG-Manager provides the application networking control and supports the deployment of the needed underlying network functions in the tenant desired infrastructure (or slice since the physical infrastructure is shared by multiple tenants with each tenant receiving a dedicated and isolated portion/share of the physical resources)
APA, Harvard, Vancouver, ISO, and other styles
34

Foresta, Francesco. "Composizione dinamica di funzioni di rete virtuali in ambienti cloud." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2015. http://amslaurea.unibo.it/8381/.

Full text
Abstract:
Questo documento si interroga sulle nuove possibilità offerte agli operatori del mondo delle Reti di Telecomunicazioni dai paradigmi di Network Functions Virtualization, Cloud Computing e Software Defined Networking: questi sono nuovi approcci che permettono la creazione di reti dinamiche e altamente programmabili, senza disdegnare troppo il lato prestazionale. L'intento finale è valutare se con un approccio di questo genere si possano implementare dinamicamente delle concatenazioni di servizi di rete e se le prestazioni finali rispecchiano ciò che viene teorizzato dai suddetti paradigmi. Tutto ciò viene valutato per cercare una soluzione efficace al problema dell'ossificazione di Internet: infatti le applicazioni di rete, dette middle-boxes, comportano costi elevati, situazioni di dipendenza dal vendor e staticità delle reti stesse, portando all'impossibilità per i providers di sviluppare nuovi servizi. Il caso di studio si basa proprio su una rete che implementa questi nuovi paradigmi: si farà infatti riferimento a due diverse topologie, una relativa al Livello L2 del modello OSI (cioè lo strato di collegamento) e una al Livello L3 (strato di rete). Le misure effettuate infine mostrano come le potenzialità teorizzate siano decisamente interessanti e innovative, aprendo un ventaglio di infinite possibilità per il futuro sviluppo di questo settore.
APA, Harvard, Vancouver, ISO, and other styles
35

Mechtri, Marouen. "Virtual networked infrastructure provisioning in distributed cloud environments." Electronic Thesis or Diss., Evry, Institut national des télécommunications, 2014. http://www.theses.fr/2014TELE0028.

Full text
Abstract:
L'informatique en nuage (Cloud Computing) a émergé comme un nouveau paradigme pour offrir des ressources informatiques à la demande et pour externaliser des infrastructures logicielles et matérielles. Le Cloud Computing est rapidement et fondamentalement en train de révolutionner la façon dont les services informatiques sont mis à disposition et gérés. Ces services peuvent être demandés à partir d’un ou plusieurs fournisseurs de Cloud d’où le besoin de la mise en réseau entre les composants des services informatiques distribués dans des emplacements géographiquement répartis. Les utilisateurs du Cloud veulent aussi déployer et instancier facilement leurs ressources entre les différentes plateformes hétérogènes de Cloud Computing. Les fournisseurs de Cloud assurent la mise à disposition des ressources de calcul sous forme des machines virtuelles à leurs utilisateurs. Par contre, ces clients veulent aussi la mise en réseau entre leurs ressources virtuelles. En plus, ils veulent non seulement contrôler et gérer leurs applications, mais aussi contrôler la connectivité réseau et déployer des fonctions et des services de réseaux complexes dans leurs infrastructures virtuelles dédiées. Les besoins des utilisateurs avaient évolué au-delà d'avoir une simple machine virtuelle à l'acquisition de ressources et de services virtuels complexes, flexibles, élastiques et intelligents. L'objectif de cette thèse est de permettre le placement et l’instanciation des ressources complexes dans des infrastructures de Cloud distribués tout en permettant aux utilisateurs le contrôle et la gestion de leurs ressources. En plus, notre objectif est d'assurer la convergence entre les services de cloud et de réseau. Pour atteindre cela, nous proposons des algorithmes de mapping d’infrastructures virtuelles dans les centres de données et dans le réseau tout en respectant les exigences des utilisateurs. Avec l'apparition du Cloud Computing, les réseaux traditionnels sont étendus et renforcés avec des réseaux logiciels reposant sur la virtualisation des ressources et des fonctions réseaux. En plus, le nouveau paradigme d'architecture réseau (Software Defined Networks) est particulièrement pertinent car il vise à offrir la programmation du réseau et à découpler, dans un équipement réseau, la partie plan de données de la partie plan de contrôle. Dans ce contexte, la première partie propose des algorithmes optimaux (exacts) et heuristiques de placement pour trouver le meilleur mapping entre les demandes des utilisateurs et les infrastructures sous-jacentes, tout en respectant les exigences exprimées dans les demandes. Cela inclut des contraintes de localisation permettant de placer une partie des ressources virtuelles dans le même nœud physique. Ces contraintes assurent aussi le placement des ressources dans des nœuds distincts. Les algorithmes proposés assurent le placement simultané des nœuds et des liens virtuels sur l’infrastructure physique. Nous avons proposé aussi un algorithme heuristique afin d’accélérer le temps de résolution et de réduire la complexité du problème. L'approche proposée se base sur la technique de décomposition des graphes et la technique de couplage des graphes bipartis. Dans la troisième partie, nous proposons un cadriciel open source (framework) permettant d’assurer la mise en réseau dynamique entre des ressources Cloud distribués et l’instanciation des fonctions réseau dans l’infrastructure virtuelle de l’utilisateur. Ce cadriciel permettra de déployer et d’activer les composants réseaux afin de mettre en place les demandes des utilisateurs. Cette solution se base sur un gestionnaire des ressources réseaux "Cloud Network Gateway Manager" et des passerelles logicielles permettant d’établir la connectivité dynamique et à la demande entre des ressources cloud et réseau. Le CNG-Manager offre le contrôle de la partie réseau et prend en charge le déploiement des fonctions réseau nécessaires dans l'infrastructure virtuelle des utilisateurs
Cloud computing emerged as a new paradigm for on-demand provisioning of IT resources and for infrastructure externalization and is rapidly and fundamentally revolutionizing the way IT is delivered and managed. The resulting incremental Cloud adoption is fostering to some extent cloud providers cooperation and increasing the needs of tenants and the complexity of their demands. Tenants need to network their distributed and geographically spread cloud resources and services. They also want to easily accomplish their deployments and instantiations across heterogeneous cloud platforms. Traditional cloud providers focus on compute resources provisioning and offer mostly virtual machines to tenants and cloud services consumers who actually expect full-fledged (complete) networking of their virtual and dedicated resources. They not only want to control and manage their applications but also control connectivity to easily deploy complex network functions and services in their dedicated virtual infrastructures. The needs of users are thus growing beyond the simple provisioning of virtual machines to the acquisition of complex, flexible, elastic and intelligent virtual resources and services. The goal of this thesis is to enable the provisioning and instantiation of this type of more complex resources while empowering tenants with control and management capabilities and to enable the convergence of cloud and network services. To reach these goals, the thesis proposes mapping algorithms for optimized in-data center and in-network resources hosting according to the tenants' virtual infrastructures requests. In parallel to the apparition of cloud services, traditional networks are being extended and enhanced with software networks relying on the virtualization of network resources and functions especially through network resources and functions virtualization. Software Defined Networks are especially relevant as they decouple network control and data forwarding and provide the needed network programmability and system and network management capabilities. In such a context, the first part proposes optimal (exact) and heuristic placement algorithms to find the best mapping between the tenants' requests and the hosting infrastructures while respecting the objectives expressed in the demands. This includes localization constraints to place some of the virtual resources and services in the same host and to distribute other resources in distinct hosts. The proposed algorithms achieve simultaneous node (host) and link (connection) mappings. A heuristic algorithm is proposed to address the poor scalability and high complexity of the exact solution(s). The heuristic scales much better and is several orders of magnitude more efficient in terms of convergence time towards near optimal and optimal solutions. This is achieved by reducing complexity of the mapping process using topological patterns to map virtual graph requests to physical graphs representing respectively the tenants' requests and the providers' physical infrastructures. The proposed approach relies on graph decomposition into topology patterns and bipartite graphs matching techniques. The third part propose an open source Cloud Networking framework to achieve cloud and network resources provisioning and instantiation in order to respectively host and activate the tenants' virtual resources and services. This framework enables and facilitates dynamic networking of distributed cloud services and applications. This solution relies on a Cloud Network Gateway Manager and gateways to establish dynamic connectivity between cloud and network resources. The CNG-Manager provides the application networking control and supports the deployment of the needed underlying network functions in the tenant desired infrastructure (or slice since the physical infrastructure is shared by multiple tenants with each tenant receiving a dedicated and isolated portion/share of the physical resources)
APA, Harvard, Vancouver, ISO, and other styles
36

Harish, Kumar Ivaturi Venkata. "Performance Evaluation of OpenStack with HTTP/3." Thesis, Blekinge Tekniska Högskola, Institutionen för datavetenskap, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-21242.

Full text
Abstract:
In today’s technology industry, cloud refers to servers which can be accessed viainternet, and the software and database applications run on the servers [22]. Whereas cloud computing is a concept of delivering the IT resources via internet which is accessed by the users and companies. In these scenarios OpenStack is the widely used cloud software which controls large pool of IT resources like compute, storage and networking which are managed and provisioned through APIs [11]. In this technology the underlying parts are the transport protocols and web servers used for authenticating and provisioning mechanisms. When stating about its underlying transport protocols, TCP is the default (standard) protocol used behind the functioning of cloud, and HTTP/1.1 version towards actions between Web servers (apache2 and nginx) and browsers [10]. The scope of the thesis is to observe the complexity of replacing TCP with QUIC (Quick UDP Internet Connection) in OpenStack and to observe the performance difference in OpenStack using HTTP/1.1 and HTTP/3. To observe this complexity, Performance Evaluation has been considered the best way in which the performance can be observed from the terminals. The thesis deals with performance of OpenStack with transport protocols from a Web server supporting HTTP/3 feature. We prove that its possible to provide the Keystone API via both HTTP/1.1 and HTTP/3. From our results we see that for simple API access HTTP/3 is faster than the HTTP/1.1 and also when the network is subjected to packet loss. The resultstates that there is a path obtained for OpenStack’s Keystone Service to interact with HTTP/3 and the average request-response time (total time) of HTTP/3 is less than that of HTTP/1.1 for accessing Keystone and Token generation even at defined packet loss rates.
APA, Harvard, Vancouver, ISO, and other styles
37

Tran, Khanh Toan. "Déploiement efficace de services complexes dans l'infrastructure de cloud." Phd thesis, Université d'Evry-Val d'Essonne, 2013. http://tel.archives-ouvertes.fr/tel-00875818.

Full text
Abstract:
Le but du travail réalisé dans cette thèse est de fournir aux fournisseurs de services une solution qui est capable de non seulement déployer les services dans le cloud de façon économique, automatique, mais aussi à grande échelle. La première contribution traite le problème de la construction d'un nouveau service demandé par le client à partir de services disponibles dans le cloud de manière à satisfaire les exigences en termes de qualité de service ainsi qu'en termes de coût. Nous présentons ce problème par un modèle analytique et proposons un algorithme heuristique dont la performance est améliorée de 20-30% par rapport aux autres approches. La seconde contribution est une solution pour déployer les services dans le cloud en considérant les demandes des utilisateurs finaux. Pour assurer qualité de services dans une grande échelle, le service demandé est dupliqué et distribué dans le réseau; chacun de ses réplicas servira les utilisateurs à proximité. Le plan d'approvisionnement selon lequel le service est dupliqué dépend de sa demande, ce qui ne cesse pas de changer en quantité ainsi qu'en distribution, ce qui rend le problème plus compliqué. Nous proposons une solution qui est capable de s'adapter aux changements dans le réseau, y compris ceux des demandes de ses utilisateurs. Enfin, nous proposons un système basé sur OpenStack qui permet de déployer les services complexes dans un cloud qui couvre différente locations (cloud multi-site). A partir d'une demande du client, le système automatiquement calcule le plan d'approvisionnement optimal et le déploie en respectant les contraintes du client.
APA, Harvard, Vancouver, ISO, and other styles
38

Foresta, Francesco. "Integration of SDN frameworks and Cloud Computing platforms: an Open Source approach." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2017. http://amslaurea.unibo.it/14271/.

Full text
Abstract:
As a result of the explosion in the number of services offered over the Internet, network traffic has experienced a remarkable increment and is supposed to increase even more in the few next years. Therefore, Telco operators are investigating new solutions aiming at managing this traffic efficiently and transparently to guarantee the users the needed Quality of Service. The most viable solution is to have a paradigm shift in the networking field: the old and legacy routing will be indeed replaced by something more dynamic, through the use of Software Defined Networking. In addition to it, Network Functions Virtualization will play a key role making possible to virtualize the intermediate nodes implementing network functions, also called middle-boxes, on general purpose hardware. The most suitable environment to understand their potentiality is the Cloud, where resources, as computational power, storage, development platforms, etc. are outsourced and provided to the user as a service on a pay-per-use model. All of this is done in a complete dynamic way, as a result of the presence of the implementation of the above cited paradigms. However, whenever it comes to strict requirements, Telecommunication Networks are still underperforming: one of the cause is the weak integration among these paradigms to reactively intervene to the users' need. It is therefore remarkably important to properly evaluate solutions where SDN and NFV are cooperating actively inside the Cloud, leading to more adaptive systems. In this document, after the description of the state of the art in networking, the deployment of an OpenStack Cloud platform on an outperforming cluster will be shown. In addition, its networking capabilities will be improved via a careful cloud firewalling configuration; moreover, this cluster will be integrated with Open Source SDN frameworks to enhance its services. Finally, some measurements showing how much this approach could be interesting will be provided.
APA, Harvard, Vancouver, ISO, and other styles
39

Al-Kaseem, Bilal R. "Optimised cloud-based 6LoWPAN network using SDN/NFV concepts for energy-aware IoT applications." Thesis, Brunel University, 2017. http://bura.brunel.ac.uk/handle/2438/15642.

Full text
Abstract:
The Internet of Things (IoT) concept has been realised with the advent of Machineto-Machine (M2M) communication through which the vision of future Internet has been revolutionised. IPv6 over Low power Wireless Personal Area Networks (6LoWPAN) provides feasible IPv6 connectivity to previously isolated environments, e.g. wireless M2M sensors and actuator networks. This thesis's contributions include a novel mathematical model, energy-efficient algorithms, and a centralised software controller for dynamic consolidation of programmability features in cloud-based M2M networks. A new generalised joint mathematical model has been proposed for performance analysis of the 6LoWPAN MAC and PHY layers. The proposed model differs from existing analytical models as it precisely adopts the 6LoWPAN specifications introduced by the Internet Engineering Task Force (IETF) working group. The proposed approach is based on Markov chain modelling and validated through Monte-Carlo simulation. In addition, an intelligent mechanism has been proposed for optimal 6LoWPAN MAC layer parameters set selection. The proposed mechanism depends on Artificial Neural Network (ANN), Genetic Algorithm (GA), and Particles Swarm Optimisation (PSO). Simulation results show that utilising the optimal MAC parameters improve the 6LoWPAN network throughput by 52-63% and reduce end-to-end delay by 54-65%. This thesis focuses on energy-efficient data extraction and dissemination in a wireless M2M sensor network based on 6LoWPAN. A new scalable and self-organised clustering technique with a smart sleep scheduler has been proposed for prolonging M2M network's lifetime and enhancing network connectivity. These solutions succeed in overcoming performance degradation and unbalanced energy consumption problems in homogeneous and heterogeneous sensor networks. Simulation results show that by adopting the proposed schemes in multiple mobile sink sensory field will improve the total aggregated packets by 38-167% and extend network lifetime by 30-78%. Proof-of-concept real-time hardware testbed experiments are used to verify the effectiveness of Software-Defined Networking (SDN), Network Function Virtualisation (NFV) and cloud computing on a 6LoWPAN network. The implemented testbed is based on open standards development boards (i.e. Arduino), with one sink, which is the M2M 6LoWPAN gateway, where the network coordinator and the customised SDN controller operated. Experimental results indicate that the proposed approach reduces network discovery time by 60% and extends the node lifetime by 65% in comparison with the traditional 6LoWPAN network. Finally, the thesis is concluded with an overall picture of the research conducted and some suggestions for future work.
APA, Harvard, Vancouver, ISO, and other styles
40

Deval, Niharika. "Empirical Evaluation of Cloud IAAS Platforms using System-level Benchmarks." UNF Digital Commons, 2017. https://digitalcommons.unf.edu/etd/765.

Full text
Abstract:
Cloud Computing is an emerging paradigm in the field of computing where scalable IT enabled capabilities are delivered ‘as-a-service’ using Internet technology. The Cloud industry adopted three basic types of computing service models based on software level abstraction: Infrastructure-as-a-Service (IaaS), Platform-as-a-Service (PaaS), and Software-as-a-Service (SaaS). Infrastructure-as-a-Service allows customers to outsource fundamental computing resources such as servers, networking, storage, as well as services where the provider owns and manages the entire infrastructure. This allows customers to only pay for the resources they consume. In a fast-growing IaaS market with multiple cloud platforms offering IaaS services, the user's decision on the selection of the best IaaS platform is quite challenging. Therefore, it is very important for organizations to evaluate and compare the performance of different IaaS cloud platforms in order to minimize cost and maximize performance. Using a vendor-neutral approach, this research focused on four of the top IaaS cloud platforms- Amazon EC2, Microsoft Azure, Google Compute Engine, and Rackspace cloud services. This research compared the performance of IaaS cloud platforms using system-level parameters including server, file I/O, and network. System-level benchmarking provides an objective comparison of the IaaS cloud platforms from performance perspective. Unixbench, Dbench, and Iperf are the system-level benchmarks chosen to test the performance of the server, file I/O, and network respectively. In order to capture the performance variability, the benchmark tests were performed at different time periods on weekdays and weekends. Each IaaS platform's performance was also tested using various parameters. The benchmark tests conducted on different virtual machine (VM) configurations should help cloud users select the best IaaS platform for their needs. Also, based on their applications' requirements, cloud users should get a clearer picture of which VM configuration they should choose. In addition to the performance evaluation, the price-per-performance value of all the IaaS cloud platforms was also examined.
APA, Harvard, Vancouver, ISO, and other styles
41

Olivi, Matteo. "Design of a Kubernetes-based Software-Defined Network Control Plane." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2020.

Find full text
Abstract:
Negli ultimi anni, Kubernetes è emerso come l’orchestratore di applicazioni a containers dominante. Il suo design è basato su un’API che permette di descrivere in modo dichiarativo lo stato desiderato delle applicazioni e su un piano di controllo che lavora per far convergere lo stato effettivo delle applicazioni verso lo stato desiderato, ottenendo fault-­tolerance, self-­healing ed elevata scalabilità. Questo design pattern si è dimostrato estremamente efficace per la gestione dei container, ma è abbastanza generale da poter essere usato per orchestrare con successo qualsiasi tipo di risorsa virtuale che viene tradizionalmente offerta mediante il paradigma del cloud IaaS. Abbiamo testato questa idea estendendo Kubernetes per fargli gestire, oltre alle usuali applicazioni a containers, delle reti virtuali. Così facendo abbiamo di fatto realizzato il prototipo di un piano di controllo di una Software­Defined Network. Nel fare ciò sono emersi sia punti di forza che debolezze del design pattern di Kubernetes e delle librerie open source che lo supportano. Per verificare che il sistema ottenuto abbia una scalabilità adeguata a quella necessaria nei moderni cloud data centers, abbiamo condotto uno studio di performance.
APA, Harvard, Vancouver, ISO, and other styles
42

Grandi, Stefano. "Sviluppo di Servizi Android per applicazioni Fog Computing." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2017.

Find full text
Abstract:
Il Cloud Computing è stata la risposta alla crescente domanda di capacità di calcolo, di memorizzazione di dati e di flessibilità nel loro approvvigionamento. Ma se da un lato il Cloud ha permesso di sopperire alle limitazioni hardware a cui i dispositivi utente sono sottoposti, questo nuovo paradigma soffre inevitabilmente di quelle che sono le latenze dovute allo scambio di informazioni attraverso Internet. Inoltre la rete Internet, già messa a dura prova dal continuo e crescente traffico dati, non ha beneficiato dell'utilizzo del paradigma Cloud e teme quello che si appresta a diventare di vastissimo ambito di utilizzo, il paradigma dell'Internet of Things (IoT), per il quale si intende una famiglia di tecnologie il cui scopo è rendere qualunque tipo di oggetto, un dispositivo collegato ad Internet. È proprio in questo scenario che nasce il paradigma del Fog Computing, cioè l'idea di costruire la stessa architettura del Cloud, fisicamente più vicina ai client che, a causa delle loro scarse capacità di elaborazione, memoria o autonomia energetica, necessitano di demandare al Cloud l'esecuzione di operazioni costose. Il Fog si pone quindi l'obiettivo di ridurre al minimo la latenza nelle comunicazioni tra i client e i nodi Fog (dispositivi fisici che si occupano dell'esecuzione delle richieste di servizi), permettendo ai client, di non sottostare più agli svantaggi dovuti alla comunicazione Internet che si hanno con il Cloud, e scaricando quindi la rete Internet di traffico in eccesso. Il progetto di tesi ha quindi visto lo sviluppo di una applicazione Android, chiamata Fog Network Application, in grado di creare e gestire una Fog Network, cioè una rete virtuale, interna alla rete locale LAN, alla quale tutti i dispositivi in possesso dell'applicazione siano in grado di connettersi, quindi fornire dei servizi generici e/o tramite la definizione di una interfaccia, e infine richiedere l'utilizzo di tali servizi.
APA, Harvard, Vancouver, ISO, and other styles
43

Amarasinghe, Heli. "Network Resource Management in Infrastructure-as-a-Service Clouds." Thesis, Université d'Ottawa / University of Ottawa, 2019. http://hdl.handle.net/10393/39141.

Full text
Abstract:
Cloud Infrastructure-as-a-Service (IaaS) is a form of utility computing which has emerged with the recent innovations in the service computing and data communication technologies. Regardless of the fact that IaaS is attractive for application service providers, satisfying user requests while ensuring cloud operational objectives is a complicated task that raises several resource management challenges. Among these challenges, limited controllability over network services delivered to cloud consumers is prominent in single datacenter cloud environments. In addition, the lack of seamless service migration and optimization, poor infrastructure utilization, and unavailability of efficient fault tolerant techniques are noteworthy challenges in geographically distributed datacenter clouds. Initially in this thesis, a datacenter resource management framework is presented to address the challenge of limited controllability over cloud network traffic. The proposed framework integrates network virtualization functionalities offered by software defined networking (SDN) into cloud ecosystem. To provide rich traffic control features to IaaS consumers, control plane virtualization capabilities offered by SDN have been employed. Secondly, a quality of service (QoS) aware seamless service migration and optimization framework has been proposed in the context of geo-distributed datacenters. Focus has been given to a mobile end-user scenario where frequent cloud service migrations are required to mitigate QoS violations. Finally, an SDN-based dynamic fault restoration scheme and a shared backup-based fault protection scheme have been proposed. The fault restoration has been achieved by introducing QoS-aware reactive and shared risk link group-aware proactive path computation algorithms. Shared backup protection has been achieved by optimizing virtual and backup link embedding through a novel integer linear programming approach. The proposed solutions significantly improve bandwidth utilization in inter-datacenter networks while recovering from substrate link failures.
APA, Harvard, Vancouver, ISO, and other styles
44

Mani, Sindhu. "Empirical Performance Analysis of High Performance Computing Benchmarks Across Variations in Cloud Computing." UNF Digital Commons, 2012. http://digitalcommons.unf.edu/etd/418.

Full text
Abstract:
High Performance Computing (HPC) applications are data-intensive scientific software requiring significant CPU and data storage capabilities. Researchers have examined the performance of Amazon Elastic Compute Cloud (EC2) environment across several HPC benchmarks; however, an extensive HPC benchmark study and a comparison between Amazon EC2 and Windows Azure (Microsoft’s cloud computing platform), with metrics such as memory bandwidth, Input/Output (I/O) performance, and communication computational performance, are largely absent. The purpose of this study is to perform an exhaustive HPC benchmark comparison on EC2 and Windows Azure platforms. We implement existing benchmarks to evaluate and analyze performance of two public clouds spanning both IaaS and PaaS types. We use Amazon EC2 and Windows Azure as platforms for hosting HPC benchmarks with variations such as instance types, number of nodes, hardware and software. This is accomplished by running benchmarks including STREAM, IOR and NPB benchmarks on these platforms on varied number of nodes for small and medium instance types. These benchmarks measure the memory bandwidth, I/O performance, communication and computational performance. Benchmarking cloud platforms provides useful objective measures of their worthiness for HPC applications in addition to assessing their consistency and predictability in supporting them.
APA, Harvard, Vancouver, ISO, and other styles
45

Butterfield, Ellis H. "Fog Computing with Go: A Comparative Study." Scholarship @ Claremont, 2016. http://scholarship.claremont.edu/cmc_theses/1348.

Full text
Abstract:
The Internet of Things is a recent computing paradigm, de- fined by networks of highly connected things – sensors, actuators and smart objects – communicating across networks of homes, buildings, vehicles, and even people. The Internet of Things brings with it a host of new problems, from managing security on constrained devices to processing never before seen amounts of data. While cloud computing might be able to keep up with current data processing and computational demands, it is unclear whether it can be extended to the requirements brought forth by Internet of Things. Fog computing provides an architectural solution to address some of these problems by providing a layer of intermediary nodes within what is called an edge network, separating the local object networks and the Cloud. These edge nodes provide interoperability, real-time interaction, routing, and, if necessary, computational delegation to the Cloud. This paper attempts to evaluate Go, a distributed systems language developed by Google, in the context of requirements set forth by Fog computing. Similar methodologies of previous literature are simulated and benchmarked against in order to assess the viability of Go in the edge nodes of Fog computing architecture.
APA, Harvard, Vancouver, ISO, and other styles
46

Jesus, Wanderson Paim de. "Network programming as a service : an innovation friendly business model." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2014. http://hdl.handle.net/10183/88885.

Full text
Abstract:
As redes de computadores têm evoluído para acomodar uma grande variedade de serviços, tais como streaming de vídeos de alta qualidade e entrega de conteúdo sensível a atrasos. Estes serviços têm aumentado a demanda por recursos não originalmente considerados na Internet. Com a promessa de atender novas demandas de rede rapidamente, pesquisadores propuseram Redes Programáveis, nas quais o comportamento dos dispositivos de rede pode ser alterado utilizando aplicativos. Entretanto, tal comportamento pode não ser um consenso entre usuários da rede. O surgimento de Redes Virtualizadas superou tal questão, ao permitir a coexistência de múltiplas redes virtuais sobre a mesma infraestrutura física. A fim de se obter redes virtuais isoladas com comportamento programável, foram propostas as Redes Virtuais Programáveis (RVP). Diante dessa nova realidade, os administradores de rede não estão mais olhando unicamente para dispositivos de rede. Eles estão olhando para um sistema composto de dispositivos e aplicativos de rede que definem o comportamento individual de cada rede virtual. Isso requer não apenas novas ferramentas e abordagens de gerenciamento, além disso, exige a revisão de conceitos tradicionais sobre redes. Implementações de RVP são encontradas principalmente em testbeds e ambientes de Computação em Nuvem. Testbeds são muito propícios à inovação, mas possuem fortes limitações no que diz respeito a migração de soluções experimentais para produção. Por outro lado, computação em nuvem é um ótimo ambiente de produção, mas possui restrições de flexibilidade e inovação, uma vez que as soluções de rede adotadas geralmente são proprietárias. Portanto, nesta dissertação introduz-se um novo modelo de negócio que permite a criação de soluções inovadoras em ambientes de produção, a Programabilidade de Redes como um Serviço (NPaaS). Diferente do modelo de negócio de redes tradicionais, onde os usuários finais são apenas consumidores dos serviços de rede já disponíveis, em NPaaS os usuários finais também são capazes de desenvolver e implantar novas soluções de rede. Para apoiar NPaaS, propõe-se uma plataforma de gerenciamento de rede virtual programável, chamada ProViNet. Essa plataforma fornece a arquitetura de software e estratégias necessárias para permitir a implantação e gestão NPaaS. Uma avaliação qualitativa do modelo de negócio NPaaS foi realizada, o resultado foi contrastado com alguns dos modelos de negócio praticados atualmente. Assim, enfatizando a singularidade do NPaaS. Enquanto isso, uma avaliação experimental foi realizada para demonstrar a viabilidade da plataforma ProViNet. Os resultados mostraram que NPaaS representa uma alternativa promissora para ambientes de rede virtual com acesso público, como as nuvens públicas. Além disso, uma avaliação quantitativa do protótipo da plataforma demonstrou a viabilidade técnica e provou que aplicativos de rede desenvolvidos usando BPMN são capazes de executar com desempenho aceitáveis.
Computer networks have evolved to accommodate a wide variety of services, such as streaming of high quality videos and delay-sensitive content delivery. These services have increased the demand for features not originally considered in the Internet. Aiming to address novel network demands quickly, some researchers proposed Programmable Networks, in which network devices behavior can be changed using applications. Notwithstanding, such behavior might not be a consensus between computer network stakeholders. The emergence of Virtualized Networks overcame this issue by allowing the coexistence of multiple virtual networks on the same physical infrastructure. Finally, the convergence of programmability and virtualization techniques are explored within a third concept, the Programmable Virtual Networks (PVN). Faced with this new reality, network administrators are no longer just looking at network devices. They are looking at a system made of virtual devices and applications that define each virtual network behavior. This requires not just new tools and management approaches, over and above that, requires new thinking. PVN deployments are found mostly in shared experimental facilities (also known as testbeds) and Cloud Computing environments. Testbeds are very innovation friendly, but with strong limitations in regards to taking experimental solutions to production. On the other hand, Cloud computing is a great production environment, but presents flexibility and innovation restrictions once network solutions adopted are usually proprietary. Therefore, in this dissertation it is introduced Network Programming as a Service (NPaaS), a new business model that aims to facilitate the conduct of innovative solutions for production environments. Different from traditional network business models, where end-users are just consumers of network services already available, in NPaaS, end-users are also able to develop and deploy novel network solutions. To support NPaaS, Programmable Virtual Network management platform is proposed. Such platform, named ProViNet, provides all architectural and technical features necessary to enable NPaaS deployment and management. A qualitative evaluation of the NPaaS business model was performed, and the result was contrasted with some of the current models, thus, emphasizing the singularity of NPaaS. In the meanwhile, an experimental evaluation was conducted to demonstrate the feasibility of ProViNet platform. Results have shown that NPaaS represent a promising alternative for virtual network environments with public access such as public clouds. Moreover, a quantitative evaluation of the platform prototype demonstrated the technical feasibility and proved that network applications developed using BPMN are able to run with acceptable performance rates.
APA, Harvard, Vancouver, ISO, and other styles
47

Muthiah, Karthika Ms. "Performance Evaluation of Hadoop based Big Data Applications with HiBench Benchmarking tool on IaaS Cloud Platforms." UNF Digital Commons, 2017. https://digitalcommons.unf.edu/etd/771.

Full text
Abstract:
Cloud computing is a computing paradigm where large numbers of devices are connected through networks that provide a dynamically scalable infrastructure for applications, data and storage. Currently, many businesses, from small scale to big companies and industries, are changing their operations to utilize cloud services because cloud platforms could increase company’s growth through process efficiency and reduction in information technology spending [Coles16]. Companies are relying on cloud platforms like Amazon Web Services, Google Compute Engine, and Microsoft Azure, etc., for their business development. Due to the emergence of new technologies, devices, and communications, the amount of data produced is growing rapidly every day. Big data is a collection of large dataset, typically hundreds of gigabytes, terabytes or petabytes. Big data storage and the analytics of this huge volume of data are a great challenge for companies and new businesses to handle, which is a primary focus of this paper. This research was conducted on Amazon’s Elastic Compute Cloud (EC2) and Microsoft Azure platforms using the HiBench Hadoop Big Data Benchmark suite [HiBench16]. Processing huge volumes of data is a tedious task that is normally handled through traditional database servers. In contrast, Hadoop is a powerful framework is used to handle applications with big data requirements efficiently by using the MapReduce algorithm to run them on systems with many commodity hardware nodes. Hadoop’s distributed file system facilitates rapid storage and data transfer rates of big data among the nodes and remains operational even when a node failure has occurred in a cluster. HiBench is a big data benchmarking tool that is used for evaluating the performance of big data applications whose data are handled and controlled by the Hadoop framework cluster. Hadoop cluster environment was enabled and evaluated on two cloud platforms. A quantitative comparison was performed on Amazon EC2 and Microsoft Azure along with a study of their pricing models. Measures are suggested for future studies and research.
APA, Harvard, Vancouver, ISO, and other styles
48

Manco, Filipe José Jesus. "Network infrastructure control for virtual campus." Master's thesis, Universidade de Aveiro, 2013. http://hdl.handle.net/10773/12725.

Full text
Abstract:
Mestrado em Engenharia Electrónica e Telecomunicações
An evolution of the current university’s networking infrastructure and service models has been shown to be necessary to enable the provisioning of innovative services that are able to respond to today’s needs. On this work a network orchestration tool is proposed that, integrated with the OpenStack cloud framework, is able to virtualize any network deployment in a non-disruptive manner, providing a clean upgrade path from the traditional networking to the world of virtualization. The framework is able to extend virtual networks created on the datacenter by OpenStack or other cloud frameworks, to the outside campus. It does so by directly reconfiguring the network devices according to the needs, independently of the device vendor, the type of device or its specificities, and independently of the specific physical network topology. This service is provided to the end user using a cloud like service model, much more flexible than the current one, properly integrated with the cloud framework services. The project was developed with the Aveiro University use cases in mind, but the final result can be applied in many academic or business environments. The framework is presented both from a conceptual perspective, by describing the abstractions and mechanisms created, and from an implementation perspective, provinding the reader the necessary understanding about the framework operation and the integration with OpenStack. The integration of this work with a broader vision for the future of the university’s services is left as future work.
A evolução da actual infraestrutura de rede e modelos de serviço da universidade tem-se mostrado necessária para permitir o fornecimento de serviços inovadores capazes de responder às necessidades do mundo actual. Neste trabalho é proposta uma ferramenta de orquestração de rede que, integrada com a plataforma de cloud OpenStack, é capaz de virtualizar qualquer infraestrutura de rede de forma não disruptiva, proporcionando um modelo de actualização simples dos tradicionais serviços para o novo mundo da virtualização. A framework é capaz de estender as redes virtuais criadas no datacenter pelo OpenStack ou qualquer outra plataforma de cloud, para o campus. Fá-lo reconfigurando directamente os dispositivos de rede de acordo com as necessidades, independentemente do fabricante, do tipo de dispositivo ou das suas especificidades, e independentemente da topologia da rede física. O serviço é fornecido ao utilizador usando um modelo de cloud, muito mais flexivel que o modelo actual, devidamente integrado com os serviços da plataforma de cloud. O projecto foi desenvolvido com os casos de uso da Universidade de Aveiro em mente, mas o resultado final pode ser aplicado em muitos outros ambientes académicos ou empresariais. A framework é apresentada tanto do ponto de vista conceptual, descrevendo as abstrações e mecanismos criados, como do ponto de vista de implementação, dando ao leitor o entendimento necessário acerca da operação da ferramenta e da sua integração com o OpenStack. A integração deste trabalho com uma visão mais abrangente para o futuro dos serviços da universidade é deixada como trabalho futuro.
APA, Harvard, Vancouver, ISO, and other styles
49

Zaman, Faisal Ameen. "VN Embedding in SDN-based Metro Optical Network for Multimedia Services." Thesis, Université d'Ottawa / University of Ottawa, 2017. http://hdl.handle.net/10393/35933.

Full text
Abstract:
Currently a growing number of users depend on the Edge Cloud Computing Paradigm in a Metro Optical Network (MON). This has led to increased competition among the Cloud Service Providers (CPs) to supply incentives for the user through guaranteed Quality of Service (QoS). If the CP fails to guarantee the QoS for the accepted request, then the user will move to another CP. Making an informed decision dynamically in such a sensitive situation demands that the CP knows the user's application requirements. The Software Defined Networking (SDN) paradigm enabled the CP to achieve such desired requirement. Therefore, a framework called Virtual Network Embedding on SDN-based Metro Optical Network (VNE-MON) is proposed in this Thesis. The use of SDN paradigm in the framework guarantees profit to the CP as well as QoS to the user.\par The design concept of the SDN control plane, raises concerns regarding its scalability, reliability and performance compared to a traditionally distributed network. To justify concerns regarding the SDN, the performance of VNE-MON and its possible dependancy on the controller location is investigated. Several strategies are proposed and formulated using Integer Linear Programming to determine the controller location in a MON. Performance results from the assessment of the VNE-MON illustrates that it is more stable compare to GMPLS-based network. It is evident that the controller location's attributes have a significant effect on the efficacy of the accepted VN request.
APA, Harvard, Vancouver, ISO, and other styles
50

FRANÇA, Adriano Henrique de Melo. "Infraestrutura para operações de Offloading computacional em ambiente integrado Cloudlet-SDN com suporte a mobilidade." Universidade federal de Pernambuco, 2016. https://repositorio.ufpe.br/handle/123456789/18624.

Full text
Abstract:
Submitted by Fabio Sobreira Campos da Costa (fabio.sobreira@ufpe.br) on 2017-04-25T12:03:54Z No. of bitstreams: 2 license_rdf: 1232 bytes, checksum: 66e71c371cc565284e70f40736c94386 (MD5) Dissertacao_AdrianoHenrique.pdf: 1956295 bytes, checksum: 38ce5d73db0d44416c8653e58120f11c (MD5)
Made available in DSpace on 2017-04-25T12:03:55Z (GMT). No. of bitstreams: 2 license_rdf: 1232 bytes, checksum: 66e71c371cc565284e70f40736c94386 (MD5) Dissertacao_AdrianoHenrique.pdf: 1956295 bytes, checksum: 38ce5d73db0d44416c8653e58120f11c (MD5) Previous issue date: 2016-08-29
Apesar da grande evolução tecnológica nos hardwares dos dispositivos móveis e nas redes sem fio, ainda existem grandes limitações nesses dispositivos em termos de processamento, capacidade de armazenamento e autonomia de energia, quando comparados aos desktops e servidores. O paradigma de Computação em Nuvem Móvel (MCC – Mobile Cloud Computing) permite estender os recursos computacionais dos dispositivos móveis através da utilização das técnicas de offloading computacional possibilitando um melhor desempenho as aplicações e uma redução no consumo das baterias dos dispositivos. Entretanto, a técnica de offloading nem sempre traz benefícios para o dispositivo móvel em situações de constante mobilidade do usuário, já que cada mudança de rede requer que o processo de offloading seja refeito. Esta dissertação propõe uma solução para otimizar o consumo de energia e o tempo de resposta durante as operações de offloading computacional quando o dispositivo muda de ponto de acesso. A proposta considera um esquema de gerenciamento de mobilidade baseado em Software Defined Networking (SDN) e técnica de caching remoto, que permite ao usuário receber o resultado do offloading no próximo acesso à rede, mesmo que esse fique desconectado por um longo período. A solução foi implementada em um testbed WiFi, com acesso ao ambiente MCC utilizando cloudlet baseada na plataforma OpenStack e integrada ao controlador SDN OpenDaylight. O consumo de energia obtido pela proposta que utiliza SDN/OpenFlow para o gerenciamento de mobilidade chegou a ser 11,33 vezes menor e a velocidade de processamento foi 3,23 vezes maior que do ambiente tradicional. O sistema de caching remoto, apesar de se mostrar útil em relação à rápida entrega dos resultados já processados, elevou consideravelmente o consumo de energia da bateria. A técnica de caching remoto é indicada para os casos nos quais a aplicação envia à cloudlet um grande volume de dados para ser processado e o nível da bateria do dispositivo encontra-se em estado não crítico ou quando o usuário enfrenta um longo período sem comunicação com a cloudlet.
Although the great technological evolution in the mobile devices hardware and wireless networks, remains significant limitations of these devices regarding processing, storage, and energy, when compared to desktops and servers. The paradigm of Mobile Cloud Computing (MCC) allows to extend the computational resources of the mobile devices through the use of computational offloading techniques, achieving a better performance on the part of the applications and a reduction in the battery consumption of the devices. The offloading technique does not always bring benefits to a mobile device in situations of high mobility since each network change requires the execution of the offloading process. This dissertation proposes a solution to optimize energy consumption and response times during the computational offloading operations when the device change of access points (AP). To this end, the proposal considers for such, a mobility management scheme based on SDN (Software Defined Networking) and a remote caching technique, that allows the user to receive the result from offloading in the next AP, even if he stays disconnected for an extended period. The solution was implemented in one Wi-Fi testbed, with access to the MCC environment using cloudlet based on the OpenStack platform and integrated with the OpenDaylight SDN controller. The achieved reduction of energy consumption for the mobility management proposal arrived to be 11.33 times lower, and the processing speed was 3.23 times bigger that of the traditional environment. The remote caching system, although useful in fast delivering the already processed results, considerably raised the battery energy consumption. Thus, the applicability of remote caching limits it to the cases where the application sends to the cloudlet an enormous volume of data to be processed and the battery level of the device is not critical or when the user faces an extended period without communication with the cloudlet.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography