design and use of virtualization technology in cloud computing pdf

Design and use of virtualization technology in cloud computing pdf

File Name: design and use of virtualization technology in cloud computing .zip
Size: 2441Kb
Published: 10.05.2021


Network Virtualization Technology for Cloud Services

Search form

Cloud computing pdf

You need Adobe Reader 7.

Explanation: Cloud computing is a computing technique in which applications are accessed by common internet protocols and networking standards. Explanation: Whenever any intranet becomes large enough in size that a diagram is not able to differentiate the individual physical system, so at that stage intranet also becomes known as a cloud. Hadoop can be considered as a cloud.


You need Adobe Reader 7. If Adobe Reader is not installed on your computer, click the button below and go to the download site. The network virtualization technology needed to effectively and efficiently construct and operate a cloud is already in place. In this article, we introduce the trends in the latest network virtualization technology for the cloud environment and its fields of application.

In recent years, the development of server virtualization technology has led to changes in the cloud services environment, such as more efficient usage of physical servers high aggregation , sharing of hardware resources such as server and network multitenacy , and the need for practical migration.

For networks within datacenters, attention has been drawn to problems such as the explosive increase in the number of medial access control MAC addresses and virtual local area networks VLANs , the construction of layer 2 L2 networks across offices, and network migration. Technology for solving these problems has taken two major directions: architectures that extend existing technology and network equipment commoditization through virtualization.

Although this approach is expected to be widely used in large, next-generation datacenter networks, there are problems such as strong dependence on the switch vendor, insufficient interworking with network equipment other than switches e. The latter approach, on the other hand, establishes a logical network independently of the physical network by logically integrating functions such as those of network equipment other than switches into a standard switch called an OpenFlow switch.

As a result, a carrier can expect lower construction costs for datacenter networks using commodity network equipment. OpenFlow was initiated by the Open Network Foundation ONF , a promotional organization that has been active in moving the technology toward a practical stage. ONF has focused on the flexibility of network programmability reconfiguration of the network in connection with applications , which is considered to be a powerful network virtualization technology for the cloud environment.

In this article, we explain the need for network virtualization as well as the technology itself and its application areas. As described above, multitenancy requires the ability to provide an isolated network for each cloud services user, together with network flexibility and agile construction and configuration changes for high aggregation and migration needs.

There are, however, difficult problems in satisfying those requirements with existing VLANs and products. In datacenters, the tagged VLAN is generally used to isolate each user's network, but there are two problems with this approach. One is the limited capacity of the tagged VLAN. A VLAN-ID must be unique over the entire datacenter network, so it is impossible to accommodate more than users at the same time. The other problem is the use of proprietary specifications by the vendors of the products used in datacenters.

For the products currently used in datacenters, most vendors use proprietary specifications in both network design and setup. For setup methods, individual vendors have proprietary control protocol specifications etc.

It is thus necessary to assign the VLAN-ID to match the vendor specifications, and the control protocol used must also match the vendor specifications. Furthermore, as the numbers of servers and switches in datacenters increase to more than hundreds of units, the use of products from multiple vendors to avoid vendor lock-in and reduce costs increases the difficulty of datacenter network design and management. For cloud services, virtual machine migration among multiple datacenters in times measured in hours is necessary in order to take advantage of night-time electricity rates and to respond to disasters.

Specifically, services can be migrated without interruption on a virtual machine that uses TCP transmission control protocol , UDP user datagram protocol , or other protocols, and the network configuration can be changed in cooperation with the migration within a very short time. As a step toward solving the above two problems, a new standard network virtualization technology that does not use VLANs and that eliminates vendor dependence is needed for cloud services.

OpenFlow originated as technology for an academic network at Stanford University in , but it is now being studied by the OpenFlow Switch Consortium as a network control technology. The basic concept of OpenFlow is that a controller performs central control by distributing programs to switches that conform to the OpenFlow specifications. Each switch then operates according to the program Fig.

Unlike the conventional Internet, in which the various types of network equipment exchange path information and select paths autonomously, the controller performs all control centrally, and each switch operates according to the instructions given to it. This scheme is referred to as controller-data separation. Switch control in OpenFlow. OpenFlow control is specified as combinations of rules and actions.

Rules identify the packets to be processed. It is thus possible to specify the evaluation conditions for L1—L4 header contents for packets whose TCP port number is 80, for example.

Actions specify operations to be performed on the packets that match the rules. Specifically, it is possible to rewrite the header so that the packets are transferred to different ports or to specify that they are to be discarded, etc.

For example, an action can specify that packets that arrive at a particular port number are to be discarded. Another way to regard this is that the controller can change a switch into a router, firewall, or load balancer as needed by sending a simple program to it. The flexibility of being able to control anything by programs is one reason that OpenFlow has been attracting attention. NTT is also a participant. An interesting feature of ONF is that the board members are representatives of companies that operate large-scale datacenters, which is to say they are network user companies rather than network vendors.

With the appearance of ONF, OpenFlow has taken on a commercial aspect in addition to its previous image of an academic system and it is now attracting much attention. Considering routing, there are two schemes for using OpenFlow: hop-by-hop and overlay Fig. Hop-by-hop and overlay schemes. In the hop-by-hop routing scheme, the controller knows all of the switches and designs paths service-by-service.

Each switch operates according to instructions so as to repeatedly forward packets in a relay scheme that delivers the packets to their final destination. Although this scheme makes free use of the advantages of OpenFlow, each switch must hold all of the path information, so scalability may be a problem.

It is suitable for the construction of small-scale networks, but application to large-scale networks requires measures against path congestion etc.

In the overlay scheme, the controller does not control all of the paths, but uses the tunneling technique described later to control the communicating end points, a practice that is referred to as edge networking. With this scheme, the controller and the various switches need to know only the source and destination of the communication; the path is handled by the conventional routing mechanism.

This approach enables the amount of routing data to be managed to be kept down to a realistic level, even for a large-scale network. Early introduction to actual services is expected to be more feasible for the overlay scheme than for the hop-by-hop scheme.

Next, we explain the implementation of an overlay virtual network with L2-over-L3 tunneling, which is one kind of overlay scheme. The overlay virtual network is implemented by encapsulating users' L2 frames inside L3 packets to achieve L2-over-L3 tunneling Fig. The three main points are explained below.

In this technique, an OpenFlow switch within a hypervisor is equipped with an L2-over-L3 tunnel endpoint function and a tunnel is established between two hypervisors. The connection between the on-premises environment locally operated and a hypervisor is also established by setting up an OpenFlow switch-based virtual gateway that has a tunnel endpoint function.

User isolation is implemented with a function that assigns a user ID to each user and encapsulates the data within the tunnel endpoint function of the OpenFlow switch. The switching is then performed by the OpenFlow switch's switching function using both the virtual and physical interfaces and the user ID. The use of OpenFlow for the switch control protocol makes it possible to develop a hardware controller that can control the products of multiple vendors. That allows a reduction in equipment costs for servers and switches through multivendor sourcing.

The development of this hardware also enables reductions in maintenance and operation costs. Overlay virtual network. Next, we describe a few fields of application for overlay virtual networks as virtual network technology. After the Great East Japan Earthquake on March 11, , disaster recovery and business continuity plans that involve the backup of data used in offices have gained attention. Disaster recovery countermeasures require remote copying of data among distant offices and migration between virtual machines at different locations.

The conventional movement of virtual machines involves the connection of special-purpose machines between offices and requires full network setup to be performed at both locations.

That took months to accomplish in the past. By contrast, virtual network functions enable end users themselves to perform a live migration of a virtual machine to a remote location in minutes, which enables smooth disaster recovery. Evaluation tests have confirmed that smooth disaster recovery measures can be implemented in this manner.

After the Great East Japan Earthquake in March , which disrupted the electricity supply, attention turned to ways of reducing the power consumption of cloud services. The number of physical servers, and thus power consumption, can be reduced by concentrating the processing achieved by server virtualization using cloud services. Nevertheless, the situation surrounding virtual machine use is changing over time, so having virtual servers running on a single physical server is not necessarily the optimum arrangement of physical servers and virtual servers.

Partial movement of a virtual machine among physical servers according to the virtual machine operation state enables concentrated processing that is always optimal to be achieved and power can be conserved by powering down empty physical servers. Furthermore, using the virtual network for migration between remote locations allows flexible operation, such as partial movement of virtual machines to areas that have a large surplus of power.

Desktop as a service DaaS puts the user desktop environment in the cloud so that inexpensive personal computers or smart phones can be used for the user environment while maintaining the same high degree of operability provided by a local desktop environment. Furthermore, the provision of new services by using DaaS in combination with virtual networks is being studied.

For example, it would be possible to place the desktop environments of corporate employees of affiliated companies in the cloud and also build logical networks between arbitrary employee desktops on demand. By setting up shared servers and chat servers, etc.

Virtual desktop service. In the past, it has been necessary to set up virtual desktops and a VPN for each project, and the end users had to access the virtual desktop of each particular project. With the combination of DaaS and virtual networks, on the other hand, the virtual desktops can be collected together for each user, and the end users only need to switch among the virtual desktops of the projects in which they are participating on demand.

We have introduced the network virtualization technology needed for the cloud environment. In future work, we plan to investigate interworking between network virtualization technology within datacenters and VPN services in broadband networks, as well as gateway technology for maintaining quality and service level agreement guarantees.

Introduction In recent years, the development of server virtualization technology has led to changes in the cloud services environment, such as more efficient usage of physical servers high aggregation , sharing of hardware resources such as server and network multitenacy , and the need for practical migration.

Need for network virtualization As described above, multitenancy requires the ability to provide an isolated network for each cloud services user, together with network flexibility and agile construction and configuration changes for high aggregation and migration needs. He received the B. Recently, he has been interested in enterprise cloud network systems. He is a member of the Information Processing Society of Japan. Since , he has been interested in enterprise cloud network system and studying the architecture and construction of virtual networks in a cloud environment.

She received the B. Since , she has been interested in enterprise cloud network systems and is studying the architecture and construction of virtual networks in a cloud environment.

Network Virtualization Technology for Cloud Services

To browse Academia. Skip to main content. By using our site, you agree to our collection of information through the use of cookies. To learn more, view our Privacy Policy. Log In Sign Up. Download Free PDF.

This certificate program is designed for individuals who have some experience or training in systems and network operations. It is designed to provide expertise in virtualization and cloud computing technologies to prepare you for employment in the fast growing fields of virtualization and cloud computing. This program presents the knowledge and skills needed to develop and implement virtualization and cloud computing systems and infrastructure for business and industry. Lec: 3 Lab: 0 Credit:3 This course is a study of directory services covering the planning, design, installation, configuration and administration of a network directory structure. This course covers the objectives of the associated Microsoft certification exam. Learn More. Lec: 3 Lab: 0 Credit:3 This course provides students with the fundamental knowledge and concepts of the Linux operating system, including command line functions, file systems, user and group administration, process management, text editors, and network applications.

Search this site. Advances in Financial Planning and Forecasting: v. American Architecture PDF. Anarchy PDF. Anatomia i fizjologia czlowieka PDF.

Search form

In computing, virtualization or virtualisation sometimes abbreviated v12n , a numeronym is the act of creating a virtual rather than actual version of something, including virtual computer hardware platforms, storage devices , and computer network resources. Virtualization began in the s, as a method of logically dividing the system resources provided by mainframe computers between different applications. Since then, the meaning of the term has broadened. Hardware virtualization or platform virtualization refers to the creation of a virtual machine that acts like a real computer with an operating system.

Research on the Virtualization Technology in Cloud Computing Environment

Это был Джабба. ГЛАВА 107 Сьюзан понятия не имела, сколько прошло времени. Жжение в горле заставило ее собраться с мыслями.

Cloud computing pdf

В его голосе слышалось беспокойство.  - Быть может, Хейл был прав, говоря, что система резервного питания подает недостаточное количество фреона. - А как же автоматическое отключение.

Через несколько месяцев оба начали подозревать, что обрели нечто такое, что может продлиться всю жизнь. Они были вместе уже два года, когда Дэвид вдруг сделал ей предложение. Это случилось во время поездки на уик-энд в Смоки-Маунтинс. Они лежали на широкой кровати под балдахином в Стоун-Мэнор. О кольце он позаботиться не успел, слова пришли сами. Именно это и нравилось ей в нем - спонтанность решений.

Search form

 - Чтоб мы не надоедали. - Значит, я не могу сойти. Парень захохотал. - Доедешь до конечной остановки, приятель. Через пять минут автобус, подпрыгивая, несся по темной сельской дороге. Беккер повернулся к панку.

 Я действую по инструкции, сэр.  - Пилот повернулся и скрылся в кабине. Дверца за ним захлопнулась. Беккер спустился вниз, постоял, глядя на самолет, потом опустил глаза на пачку денег в руке. Постояв еще некоторое время в нерешительности, он сунул конверт во внутренний карман пиджака и зашагал по летному полю.

Multiple Choice Question of Cloud Computing

Он слишком долго обделял. И ради. Чтобы увидеть, как какой-то молодой профессор украл его мечту.


  • Teri A. 13.05.2021 at 04:22

    Cloud computing is rapidly expanding in its applications and capabilities through various parts of society. Utilizing different types of virtualization technologies can​.

  • Ovzumuper 13.05.2021 at 06:37

    Energy efficient buildings with solar and geothermal resources pdf essays in english on current topics in india pdf

  • Lotye R. 14.05.2021 at 19:44

    Skip to main content.

  • Hollie C. 19.05.2021 at 07:08

    Cloud Computing is defined as storing and accessing of data and computing services over the internet.


Leave a reply