As organizations move more of their IT infrastructure and applications into private clouds, they are entering into uncharted territory. The roadmap isn’t always clear, and it’s easy to make mistakes – mistakes that in hindsight look obvious, but which can be hard to predict if you lack experience with cloud infrastructures.
However, this isn’t completely uncharted territory. Just as most experts will tell you to avoid the first version of any new gadget (let someone else work the kinks out), organizations that have moved slowly to the cloud can now benefit from the experiences of early adopters.
Here are five private cloud pitfalls you should avoid:
1. Believing that consolidating servers through virtualization will eliminate over-provisioning.
Apartments.com started to virtualize its development and testing servers back in 2007 using VMware. However, Apartments.com soon realized that they need to push even further in order to stay competitive in the increasingly cutthroat online housing market. Matt Stratton, Director of Technology Operations at Apartments.com, noted that even with a consolidated infrastructure, they didn’t have the tools necessary to create an automated, self-service environment.
The IT staff also asked for tools that would help them better monitor server performance and detect potential server health issues before they impacted performance. The status quo was that when a server reached approximately 70 percent capacity, IT ordered more servers. Of course, in a highly cyclical market like apartment rentals, this led to plenty of servers sitting idle after the peak apartment-hunting season was over.
To address this problem, Apartments.com switched from VMware to Microsoft’s Hyper-V technology, which is included in the Windows Server 2008 R2 operating system. They migrated several hundred VMware virtual machines (VMs) to Hyper-V in late 2011 and early 2012. Now, Apartments.com is upgrading to Microsoft System Center 2012, which is designed specifically for managing private cloud environments.
System Center 2012 includes Server App-V, a tool that allows IT teams to create virtual application packages that can be copied to any computer that has a Server App-V Agent on it, without requiring a local installation. This reduces the number of images IT has to manage, speeds software deployment and improves availability.
The goal, which Apartments.com expects to achieve soon, is to give developers the ability to provision their own VMs using a template and a self-service portal. By automating much of the process, IT can eliminate most of the manual work involved in provisioning and de-provisioning virtual resources, which means that if servers are sitting idle, they can be de-provisioned and freed up for something else, rather than just waiting for the next year’s peak season.
2. Believing that de-provisioning VMs is easy to do.
Just because you can de-provision a VM and free it up for something else in theory, doesn’t mean you’ll be able to do so in practice.
“How can you put a bullet in a virtual machine without knowing who owns it?” asked Paul Martin, Systems Engineering Lead, EMEA, for Embotics, a private cloud management provider. “Most IT administrators don’t know who owns it and will end up erring on the side of caution, which causes other problems.”
Without the proper tools in place to identify “zombies and VMs that have had no log-in, or have not been powered on for a certain number of days” and to identify the owners of those orphaned VMs, IT pros won’t risk getting rid of them. It’s too big of a political risk. Who knows what toes you might be stepping on?
“Having an owner assigned from day one makes this whole process easier. It’s important to ensure that all new VMs have ownership assigned and that your private cloud management solution is able to apply these on deployment and also retrospectively,” Martin said.
3. Forgetting to update chargeback tools along with your infrastructure.
Aston University, based in Birmingham, UK, began virtualizing its infrastructure back in 2004, when it initially consolidated a set of ten finance applications. This initial small environment eventually grew into the start of a full-fledged private cloud initiative.
Aston U. moved their first customers onto their private cloud in 2009, billing them through a SharePoint manual chargeback system. Service chargebacks allowed IT to fund the gradual expansion of the cloud infrastructure. A “cloud service first” corporate directive was established in 2010, with all new services going automatically to the cloud environment unless there was a strong reason why they shouldn’t.
All of the university’s schools of study and support departments now utilize the service. As their private cloud environment grew, however, the IT server team started to experience problems with capacity and with their manual chargeback system.
To address these problems, they investigated solutions from VMware and Veeam, before settling on the V-Commander private cloud management suite from Embotics. V-Commander’s chargeback portal helped them transform the chargeback process from a manual, error-prone process into an automated, simplified one.
Capacity management and resource optimization features also identified existing VMs that were either sprawled or over- or under-resourced, and an assisted placement feature now automatically guides new services to hosts that have available capacity.
Aston uses built-in decommissioning workflows to automate service renewals. Now, as each VM reaches its renewal date, an e-mail that includes the cost of the service (automatically calculated by V-Commander) is sent to the customer, who has until the actual renewal date to approve the billing. If the billing is approved, the expiration date is reset. If not, the service is automatically shut down, and that resource is freed up for someone else.