The emergence of public cloud providers has dramatically altered the playing field for hardware vendors of every type. Amazon Web Services (AWS) and its competitors opened Pandora’s Box by introducing the world to a completely programmatic, scalable and pay-as-you-go way to procure and utilize network, compute and storage resources on a global scale. They have disrupted many layers of the technology industry; from applications being written to the way companies interact with the infrastructure being used to support those applications.
Nowhere is this disruption easier to see than the virtualization ecosystem. For most of the last decade, hypervisor companies like VMware, Citrix, Microsoft and others worked hand-in-hand with hardware manufacturers like Cisco, NetApp, EMC and Dell to define the infrastructure foundation as well as the virtualized abstraction layer that sat underneath the entirety of the client/server era. These companies provided a direct link between the enterprise applications, the hypervisor and the hardware. They owned the traditional datacentre construct.
It’s that construct, since rebranded as “private cloud,” that is directly under attack from public cloud providers and this will be the battlefield for the heart and soul of enterprise IT for the next decade.
The response to the public cloud threat has been varied. Interestingly, erstwhile competitors Microsoft and VMware reacted similarly. This is because they were both uniquely positioned to create a software-defined solution to the problem.
For both companies, the response started with existing enterprise workloads. One of the largest challenges of the AWS public cloud is the fact that getting workloads, and especially data, into and out of an enterprise environment can be technically challenging, making migration potentially expensive and risky. As companies with extensive and hard-won experience running mission-critical enterprise workloads, Microsoft and VMware came to much the same conclusion: build a public cloud using their existing platform and allow customers and developers to leverage all of the investment they’ve made in their own data centres as they selectively move workloads outside their own data centres. Thus, Microsoft Azure and VMware vCHS were born. Both are clouds that customers can move workloads to without having to rewrite or re-architect them.
Unfortunately, the traditional data centre infrastructure is now the weak link in this new software-defined world. In each of the public clouds referenced, the focus has been on the abstraction layer and how it interacts with end users. What’s missing is how the abstraction layer and the applications sitting on it interact with the infrastructure directly. And although there have been attempts at this (VAAI is a good example of VMware), it’s a rudimentary exchange at best.
So in summary, we have an outside force, AWS and public cloud, being the primary catalyst driving change into the enterprise, yet very little of that change is happening below the cloud management or hypervisor layer. Why? Why is it important that the infrastructure layer become more of an asset to the stack? What would that look like?
The question of why is simple: it’s really hard to take legacy hardware architecture and retrofit it into something agile and programmatic. In some cases it’s just a new concept that requires a hardware refresh, but in many cases, especially around storage, it requires a complete reimagining of the platform. It’s no coincidence that most of the innovation in this agile infrastructure space is being done by startups who have no legacy customers, technical debt or margins to deal with.
Why is it important? Because while the best hardware is boring hardware, it’s still a critical part to providing a flexible, reliable and high-performance foundation to handle applications that matter to enterprises. There are times where the best way to handle the demands of an application, or more importantly multiple applications at once, is in hardware. This is true at the network layer, where the manipulation of packets benefits from proximity to processing resources, the compute layer, where apps can benefit from having specialized GPU resources to handle unique requirements, and most especially at the storage layer.
Storage services can have the most dramatic impact on workload performance, yet are often implemented in such a way that they have no direct relationship with those workloads. Services like compression, deduplication and quality-of-service (QoS) are usually “on or off” features when it comes to storage arrays. Best case, a storage administrator will create a volume or LUN, choose the features that need to be enabled, and then a virtualization admin will map that volume to a data store. Perhaps the virtualization team will create manual storage profiles that define the features offered by that data store, but placing and migrating VMs remains a manual process, and they will not have the ability to map application policy equally across the hypervisor and hardware layers.
Enterprises have come to expect fundamental features from the public cloud space: simple architecture, linear scaling, API availability and granular application of services. These features allow infrastructure to respond to the increased requirements of a workload natively, without the overhead of a bolt-on orchestration engine. They enable the Next Generation Data Centre, one in which the hardware, the hypervisor and the application itself all play an integrated, coordinated role in providing the performance and availability demanded by the enterprise.
No matter where your workloads run, the rise of public cloud has ushered in an era of computing defined by a seamless, programmatic experience. The battle for the enterprise soul is beginning, and inside or outside the firewall, the clouds that can best adapt to the demands of the workloads they are supporting will be best positioned for success.