Skip to main content

Server Virtualization drives cross domain management

A new set of challenges
Without the tools to optimize and manage end-to-end virtualization, IT is unable to do system-wide capacity planning, lacks visibility into server and storage resource allocation and utilization, has difficulty troubleshooting and managing change, and can no longer carry out historical analysis. The impact can be severe.

Challenge #1: System-wide capacity planning becomes impossible
The data center exists to maximize application performance, and IT keeps service level agreements (SLAs) to track its success. Meeting service level objectives requires IT to optimize resource capacity and planning, but the growth of complexity in mixed environments has made this increasingly difficult to do. In an age where virtual machines can be created easily, manual capacity planning across the infrastructure becomes impossible. The inability to do accurate capacity planning forces IT to over-provision resources. This is particularly true with storage, but extends to other domains as IT is forced to provision more performance and bandwidth while trying to stay ahead of application demand. Without the means to correlate critical applications with the underlying infrastructure, IT is taking shots in the dark.

Challenge #2: Utilization and relationships become murky
Even a purely physical network is a challenging entity, given its hundreds or thousands of computing objects that make up a complex web of relationships. Add in virtualization as an additional layer of abstraction, and complexity grows by leaps and bounds.

Maintaining visibility into server and storage allocation and utilization in this environment is a very challenging assignment. Manual correlation and root-cause analysis can become impossible, even though the need for correlation and visibility are growing across the infrastructure. Individual devices come with their own diagnostics, but device-level information cannot solve the problem of correlating computing behavior from a variety of sources.

Challenge #3: Device failure leads to SLA failure
In a virtualized infrastructure, failed devices kick off multiple device failures downstream, resulting in a diagnostic and predictive nightmare. For example, in a compact physical network a failed host bus adapter (HBA) sends out an alert, making it simple to find and replace. But in a complex network with virtualized layers, a failed HBA is not the only element alerting IT—so are the dozens (or more) of devices that the HBA failure is affecting downstream.

With all the alerts coming in at once, manually tracking down the problem ranges from time-consuming to impossible. Worse, the alerts do not only reach a single server or storage administrator. Because devices and applications throughout the infrastructure are affected by the failure, specialized IT administrators and database administrators are receiving alerts as well. And because the affected applications are failing or slowing down, SLAs go unmet and end-user calls start pouring in.

Challenge #4: Historical analysis withers in the face of virtual mobility
In a virtualized environment, virtual server resource allocation can change at the drop of a hat. However, if historical trending is tied to the physical machine via the WWN (and most are) then the trending is cut loose from the VM and storage usage analysis becomes impossible. This is a significant loss, leaving IT without a way to track SLA success rates across virtualized infrastructure or to report on resource usage patterns. The lack of historical analysis makes it increasingly difficult to plan for growth and to protect SLAs.



Link

Comments

Popular posts from this blog

Security: VMware Workstation 6 vulnerability

vulnerable software: VMware Workstation 6.0 for Windows, possible some other VMware products as well type of vulnerability: DoS, potential privilege escalation I found a vulnerability in VMware Workstation 6.0 which allows an unprivileged user in the host OS to crash the system and potentially run arbitrary code with kernel privileges. The issue is in the vmstor-60 driver, which is supposed to mount VMware images within the host OS. When sending the IOCTL code FsSetVoleInformation with subcode FsSetFileInformation with a large buffer and underreporting its size to at max 1024 bytes, it will underrun and potentially execute arbitrary code. Security focus

Splunk that!

Saw this advert on Slashdot and went on to look for it and found the tour pretty neat to look at. Check out the demo too! So why would I need it? WHY NOT? I'd say. As an organization grows , new services, new data comes by, new logs start accumulating on the servers and it becomes increasingly difficult to look at all those logs, leave alone that you'd have time to read them and who cares about analysis as the time to look for those log files already makes your day, isn't it? Well a solution like this is a cool option to have your sysadmins/operators look at ONE PLACE and thus you don't have your administrators lurking around in your physical servers and *accidentally* messing up things there. Go ahead and give it a shot by downloading it and testing it. I'll give it a shot myself! Ok so I went ahead and installed it. Do this... [root@tarrydev Software]# ./splunk-Server-1.0.1-linux-installer.bin to install and this (if you screw up) [root@tarrydev Software]# /op

Virtualization is hot and sexy!

If this does not convince you to virtualize, believe me, nothing will :-) As you will hear these gorgeous women mention VMware, Akkori, Pano Logic, Microsoft and VKernel. They forgot to mention rackspace ;-) virtualization girl video I'm convinced, aren't you? Check out their site as well!