Virtualizing the Enterprise: An overview

Virtualizing the Enterprise: An overview

Summary: Server, storage, network, desktop and application virtualization allows the creation of flexible, scalable, manageable and secure IT infrastructure that utilises the underlying hardware to maximum efficiency. Some of these technologies are in the early stages of development while others are mature and widely deployed.

SHARE:

Virtualization has its roots in the mainframe era of the 1960s, and is particularly associated with IBM, whose CP-67/CMS was the first commercial mainframe operating system to support a virtual machine architecture in 1968. (CP stands for Control Program, while CMS stands for Console Monitor System; CP created the virtual machines, which ran the user-facing CMS.)

PC-era virtualization kicked off in 1987 with Insignia Solutions' SoftPC, which allowed DOS programs to run on Unix workstations. A Mac version that also supported Windows applications appeared in 1989, followed by SoftWindows bundles containing SoftPC and a copy of Windows. Another notable PC virtualisation pioneer was Connectix, whose Virtual PC and Virtual Server products were acquired by Microsoft in 2003 and re-released in 2006 and 2004 respectively. VMware, the current market leader in virtualization, released its first product, VMware Workstation, in 1999.

VMware Workstation is still widely used today, but the company's most significant releases were arguably its 2001 server virtualisation products — GSX Server and ESX Server. GSX Server is an example of a Type 2 (or hosted) hypervisor that, like the Workstation product, runs on top of a conventional operating system. ESX Server is a more efficient Type 1 (or 'bare metal') hypervisor that runs directly on the underlying hardware. From these beginnings, VMware has evolved a market-leading ecosystem of virtualization products based around its vSphere platform.

Virtualization has become the key technology underpinning 'cloud-era' IT infrastructure because the ability to deploy virtual instances of servers, desktop PCs, storage devices and network resources allows existing hardware to be utilised efficiently, and makes for much better manageability, flexibility and scalability.

Before we look at detail at the main areas where virtualization plays a key role, it's worth noting a key finding from the recent Cisco 2013 Global IT Impact Survey, which canvassed 1,321 IT professionals in 13 countries. When asked which were the most difficult IT projects to deploy over the past year, the top three cited directly involve virtualisation: moving applications to the cloud; centralizing/virtualizing the data center; and VDI (Virtual Desktop Infrastructure):

cisco-gitis-difficult
Source: 2013 Cisco IT Global IT Impact Survey

Virtualisation may be the foundation of modern IT infrastructure, but today's IT professionals evidently find it challenging to get to grips with on the ground.

Even though virtualization is a specialist IT discipline, it's a little surprising to find that, in a July 2013 study sponsored by a group of Cisco partners, 40 percent of knowledge workers and 54 percent of non-IT workers had not even heard of it. Moreover, 80 percent of decision-makers (senior management, VPs and SVPs) were unaware of how virtualization will benefit their businesses, and only 34 percent of workers believed they were using virtualization in their workplaces.

Server virtualization

Server virtualization makes perfect sense both for in-house data centres and for service providers in the public cloud. In both cases, utilising servers to maximum efficiency and having the ability to scale virtual server provision rapidly as workload demand ebbs and flows is vital to keeping costs down and (if you're a service provider) revenues up.

A recent survey by market research company Vanson Bourne on behalf of networking vendor Brocade shows the level of penetration of server virtualization in company data centres. This study canvassed 1,750 IT decision-makers and office workers in companies ranging in size from 500 to over 3,000 employees in Europe (UK, France, Germany) and the USA:

vb-brocade-sv-1
Source: Infrastructure and connectivity issues in organisations today (Brocade/Vanson Bourne, June 2013)

Brocade's survey found that, on average, companies currently virtualize 46 percent of their servers, with more than half virtualizing at least 40 percent of their servers. In two years' time, the respondents expect 59 percent of their servers to be virtualized, on average, with three quarters of companies virtualizing at least 40 percent of their servers. Although it's well established, server virtualization is clearly still on the rise.

Another recent survey, conducted by storage virtualization specialist DataCore, canvassed 477 IT professionals worldwide, from a broad range of company sizes and industry sectors, and found VMware (vSphere) to be the clear leader in server virtualization software, followed by Microsoft (HyperV) and Citrix (XenServer):

datacore-sv-hyperv
Source: The State of Virtualization and the Impact of Storage (DataCore, March 2013)

Gartner's latest Magic Quadrant for x86 Server Virtualization Infrastructure (June 2013) backs up this finding, placing VMware and Microsoft in the 'leaders' quadrant, Citrix and Oracle as sole occupants of 'visionaries' and 'challengers' respectively, with Parallels and Red Hat in the 'niche players' corner:

gmq-x86-sv
The June 2013 Gartner Magic Quadrant for x86 Server Virtualization Infrastructure.

Parallels' core market is service providers aimed at SMEs, while Red Hat Enterprise Virtualization is most often used in organisations with an existing Linux-based IT infrastructure.

Storage virtualization

Storage has always been important, but in the era of Big Data and real-time business analytics, it's becoming a core component of enterprise IT, and one in which virtualization has an increasingly important part to play. The key advantage of storage virtualization, as in other spheres, is to free IT managers from orchestrating multiple units of proprietary hardware by inserting a layer of software — a storage hypervisor — that creates easily manageable virtual 'pools' of efficiently utilised network storage.

These days, companies use a wide range of storage hardware, including high-end, mid-range and low-end hard disk arrays, solid-state and hybrid drives and flash memory on PCIe cards. The ability to manage these heterogeneous resources via software that can easily create and migrate virtual disks, and automatically match server workloads to the most appropriate type of storage, is highly desirable. Tiered virtual storage solutions also make vital business tasks like backup and disaster recovery a lot more manageable.

DataCore's recent State of Virtualization survey found that, of the 477 respondents, 66 percent (315) did not use external storage hypervisor software (such as DataCore's SANsymphony-V) to manage their storage. However, the 34 percent that did offered multiple reasons why they felt it was a good solution:

datacore-storage-hyper-2
Source: The State of Virtualization and the Impact of Storage (DataCore, March 2013)

Reflecting the changing storage market, Gartner has recently (March 2013) combined a number of segmented Magic Quadrants into one, which covers General-Purpose Disk Arrays. The 'leaders' quadrant contains a sextet of familiar companies: EMC, NetApp, Hitachi, IBM, Dell and HP:

gmq-disk-arrays
The March 2013 Gartner Magic Quadrant for General-Purpose Disk Arrays.

All of these disk array vendors offer storage virtualization solutions — EMC's VPLEX and NetApp's Data ONTAP are examples. Other storage companies worth following are flash specialists like Fusion-io, LSI and Violin Memory.

Network virtualization

The latest IT infrastructure component to come under the virtualization spotlight is the network. Network virtualization, or Software-Defined Networking (SDN) as it's generally known, decouples the control layer from the physical hardware that routes the network packets. SDN makes it easier for IT managers to configure heterogeneous network components and allows network traffic to be shaped to ensure that, where possible, demanding workloads get the required bandwidth, low latency and quality of service.

SDN will often work in conjunction with another developing technology, network fabrics, which are designed to allow any node on the network to connect to any other over multiple physical switches. Some vendors see SDN and network fabrics as tightly integrated, while others — including the Open Networking Foundation (ONF), curator of the OpenFlow SDN protocol — view them as independent.

According to Brocade's June 2013 survey, both network fabrics and SDN are technologies for the future, with only around a fifth of companies currently using them — although 60 percent or more are investigating or planning to investigate network fabrics and SDN:

brocade-fabric-sdn
Source: Infrastructure and connectivity issues in organisations today (Brocade/Vanson Bourne, June 2013)

A broadly similar picture emerges from Cisco's Global IT Impact Survey, just under half of whose respondents are currently evaluating SDN. Nearly all of these evaluators are planning to implement the technology within the next two years:

cisco-gitis-sdn
Source: 2013 Cisco IT Global IT Impact Survey

Both Cisco and Brocade asked their survey respondents about the drivers for investing in software-defined networking, revealing a broad range of perceived benefits:

cisco-sdn-drivers
Source: 2013 Cisco IT Global IT Impact Survey
brocade-sdn-drivers
Source: Infrastructure and connectivity issues in organisations today (Brocade/Vanson Bourne, June 2013)

Software-defined networking is likely to make an impact first in the data centre, but is also applicable to campus networks and enterprise WANs. This market is too embryonic for specific vendor comparisons to have been made by market research firms like Gartner (which is still in the process of defining its taxonomy). However, most of the leading data center network infrastructure vendors are now offering SDN solutions. Recent Plexxi-sponsored research forecasts that the SDN market will be worth $35 billion by 2018, representing 40 percent of all network spending. The same April 2013 study estimates that there are 225 SDN companies, from a starting point of zero in 2009. Network virtualization is clearly in the establishment phase, but is already experiencing rapid growth.

Desktop virtualization

Desktop virtualization involves running virtual machines, each with their own desktop OS instance, on a server and accessing them from remote client devices (PCs, notebooks, thin clients, tablets or smartphones) over local- or wide-area networks. There are several delivery models, ranging from cloud-hosted desktop-as-a-service, to managed hosting, to on-premises deployment.

Advantages of desktop virtualization compared to traditional desktop computing include data security (data is stored in the data centre rather than on the local device), manageability, wide device support (you can run a virtual desktop on a BYOD tablet or a smartphone if you want to), easier backup and disaster recovery, and lower operational costs (client device troubleshooting, for example).

Disadvantages include the need for an always-on network connection with high enough bandwidth and low enough latency to deliver an acceptable desktop experience for users (particularly if they need to use graphically demanding applications), licensing issues (particularly if you're running Windows) and — if you're considering an in-house deployment — the capital cost of server, storage and networking infrastructure.

Before desktop virtualization came to prominence there was, and remains, the Terminal Services model, in which multiple remote desktop sessions are hosted by a single server operating system. With Windows Server 2012, Microsoft supports three flavours of desktop virtualization: Personal Desktops (each user gets their own 'stateful' virtual desktop); Pooled Desktops (multiple users share a 'stateless' virtual desktop) and Remote Desktop Sessions (formerly known as Terminal Services). Advantages of the terminal services model include rapid application deployment, security and low support costs.

Desktop virtualization is another technology that most businesses are just getting to grips with (remember, it came above BYOD and deploying SaaS application in the list of 'difficult' IT projects in Cisco's Global IT Impact Survey quoted earlier). In DataCore's March 2013 State of Virtualization survey, 55 percent of organisations had not undertaken a desktop virtualization project and only 11 percent had virtualized more than a quarter of their desktops:

dvi-datacore-2
Source: The State of Virtualization and the Impact of Storage (DataCore, March 2013)

As well as VMware (View), Citrix (XenDesktop) and Microsoft (Windows Server 2012), leading desktop virtualization vendors include Desktone, Unidesk, Virtual Bridges, Quest Software (now owned by Dell), MokaFive, NComputing, Oracle and Red Hat:

idc-mktscape-dvi
IDC MarketScape: Worldwide Client Virtualization 2012 Vendor Analysis

Application virtualization

Application virtualization decouples the installation of an application from the endpoint that's running it. This can be done via the desktop or server virtualization solutions described above, or by using application streaming technology: in the latter case, packaged applications are stored a central server, downloaded over the network and cached on the client. Generally speaking, only a small percentage of an entire application needs to be delivered to the client before it can launch, with additional functionality being delivered on demand.

Whatever the delivery model, application virtualization makes for much easier app deployment and management compared to installing and maintaining local installations on multiple endpoints. This is especially true when it comes to tricky areas like licensing and, on client endpoints, application conflicts.

In DataCore's State of Virtualization survey, nearly one in five (19 percent) of respondents reported that all of their mission-critical applications would be virtualized by the end of this year, with over half (58 percent) saying that 80 percent or more would be virtualized. Heading the hit-list is a trio of Microsoft enterprise stalwarts: SQL Server, Exchange and SharePoint:

app-virt-datacore
Source: The State of Virtualization and the Impact of Storage (DataCore, March 2013)

Leading application virtualization vendors include Cameyo, Citrix (XenApp), Evalaze, Numecent, Novell (ZenWorks), Microsoft (App-V), Roozz, Spoon (formerly Xencode), Symantec (Endpoint Virtualization) and VMware (ThinApp).

Conclusion

Today's server, storage, network, desktop and application virtualization technologies provide IT managers with the means to create flexible, scalable, manageable and secure infrastructure that utilises the underlying hardware to maximum efficiency. Some of these technologies — notably software-defined networking — are in the early stages of development while others, such as server virtualization, are mature and widely deployed.

IT professionals may still be getting to grips with virtualization in many businesses, and their users and senior management may be a little hazy on the technology and its benefits, but it's here to stay and is undoubtedly the future of business IT.

Topics: Virtualizing the Enterprise, Networking, Servers, Virtualization

About

Charles has been in tech publishing since the late 1980s, starting with Reed's Practical Computing, then moving to Ziff-Davis to help launch the UK version of PC Magazine in 1992. ZDNet came looking for a Reviews Editor in 2000, and he's been here ever since.

Kick off your day with ZDNet's daily email newsletter. It's the freshest tech news and opinion, served hot. Get it.

Talkback

2 comments
Log in or register to join the discussion
  • Virtualization will drive changes in SDN and Big Data too

    This is good overview of all the virtualization initiatives. Thanks for providing the article.

    I was moderating a panel this week with VMWare, Citrix, Savvis, and a networking incubator called The Fabric where the topic was basically SDN and Big Data. The general thesis was that an automated SDN framework will rely heavily on data analytics, and the sheer volume of data in a large distributed environment classifies as a Big Data problem.

    What was interesting, though, as it relates to this article is the impact of virtualization on the problem. The edge of the network is no longer defined by the physical boxes but by the virtual devices. This means that the network is much larger than before, and when you include other infrastructure and application data, the problem is exponentially larger than it would have been even a few years ago.

    It could be that virtualization is going to drive massive new technology requirements in the analytics space, and the panel thought this was likely the biggest hurdle to moving to a more orchestrated, No-Ops (completely automated) environment.

    For reference, the event I mentioned is described in a blog I wrote about it in the runup to the panel: http://www.plexxi.com/2013/07/big-data-and-sdn-isnt-it-inevitable/

    -Mike Bushong (@mbushong)
    Plexxi
    Mike Bushong
  • Shocking to be honest

    We went 100% all virtual for Windows Servers 3-4 years ago on VMware. 80% of our users have Wyse Thin clients that either go to virtualized Windows Terminal servers or virtualized windows desktops, VDI.

    It is old news these days. I can imagine working in IT and not knowing about it.
    paebin2s