Your submission was sent successfully! Close

Thank you for signing up for our newsletter!
In these regular emails you will find the latest updates from Canonical and upcoming events where you can meet our team.Close

Thank you for contacting our team. We will be in touch shortly.Close

  1. Blog
  2. Article

nickbarcet
on 12 December 2012


Nigel Beighton, International VP of Technology, Rackspace

Cloud computing is changing everything we know. Immense mathematical problems like the sequencing of the human genome and the discovery of the Higgs Boson simply wouldn’t have been solved without access to massive computing power, precisely when it was needed. Closer to home, the mobile services we use every day depend just as much on the cloud. The introduction of cloud computing is as much a revolution as the introduction of personal computing was, nearly 30 years ago.

The catalyst for this revolution was a combination of business need and technological opportunity. There was a need to change to the economics of business IT. Businesses wanted to cut their capital expenditure on IT; they wanted the ability to buy it as a pay-as-you-go service that would scale on demand. The opportunity to answer that need came from a technology we’ve been using for over 10 years: virtualisation. Cloud took virtualisation and added a whole new world of on-demand scalability and delivery over the Internet. The public cloud was born.

Naturally, CIOs soon wanted ways to mirror the flexibility of the public cloud within their organisations. Hence the growth in private clouds – this represented the same infrastructure as the public cloud, but was used for the flexible delivery of IT resources within the walls of the enterprise.

In my view, every CIO needs a private cloud in his back pocket. When I was a CIO myself, I hated unplanned projects that suddenly demanded new chunks of infrastructure, because it would often be hard to react fast enough. With a private cloud, however, applications can literally be moved around your infrastructure to make space for new services. With some prudent management and the correct implementation this can be like a ‘Get Out of Jail Free’ card in Monopoly; it enables you to provide the flexibility your business requires.

The next big (hybrid)

The next big thing in cloud architecture, however, is the hybrid model.

Today, many companies are looking at hybrid models because of security concerns. They want to keep sensitive data behind their firewalls and use the public cloud where they can. This makes sense, but these security concerns will be short-lived. They are based on first-generation cloud issues that will be addressed in the next three years – even if it takes the regulators time to catch up.

In my opinion, the bigger hybrid story is what has popularly become known as cloud-bursting. Sometimes summarised as, ‘own the base, rent the spike’, it involves running your applications on your own private cloud infrastructure, but bursting some workloads to a public cloud at times of high demand.

Cloud-bursting makes economic sense, because it does lower your costs. But it needs applications that have been written from an architectural perspective to work that way. In the future, it will be normal for applications to intelligently auto-scale, by bursting from private clouds to public clouds. But it will take time for today’s applications to be re-engineered accordingly by the software programmers who oversee them. Enterprise software that has been developed in-house just doesn’t get re-written every year. In fact, it typically has a lifespan of five to eight years.

For the packaged software industry, this spells an even more profound change. How will auto-scaling cloud apps be licensed? Will there be new, separate software packages written especially for cloud? We are on the cusp of radical change in the structure of the software industry – just one of the effects of the cloud revolution.

Nevertheless, every CIO I talk to seems to want a hybrid model in the near future. That’s why I think that, while large-scale cloud bursting might still seem a long way off for some businesses, three years from now, there will be very few enterprises who don’t use it.

So this is why OpenStack is so important. None of this can happen without portability of applications and workloads. That portability will require open standards and APIs for vendors and enterprises to develop to. Today, OpenStack is the single best route to a cloud based on open standards – and it is the platform that looks most likely to continue to conform with those standards as they evolve.

What the cloud – and specifically OpenStack – promises is a consistent way to deploy around the world. Apps and databases will change over the next five years. The way we have handled data over the last 10 years will change. Not just the growth of NoSQL. Even today’s clustering is not the ideal solution for cloud.

It’s easy to pick up cloud and quickly understand how it works. Log on to a public cloud and for a few pounds, you can play around and get experience without buying servers and licences. OpenStack, API, tools – it’s all there. That ability to get experience at low cost, with little delay and little red tape in setting up environments etc. Take a small group of developers, give them a public account and let them go.

With many of the world’s largest technology companies committed to the OpenStack project, its code is already mature enough for production deployment. While there is still plenty of room for growth in the OpenStack ecosystem – in terms of monitoring, management and tracking, for example – many of today’s most forward-looking companies are deploying it now.

A prescription for OpenStack

My recommendation to any enterprise CTO is simply to start using OpenStack. There are now several OpenStack-based public clouds available, from companies like Rackspace. So go and play with it, examine it; see whether your ecosystem is ready for it. A lot depends on your application strategy, of course. But the benefits are there for those companies who need it at scale and even if you don’t, you soon will. With the cloud revolution gathering pace, OpenStack is just too interesting to ignore.

Want to receive this content direct in your inbox? Register here.

Related posts


Tytus Kurek
26 April 2023

ChatGPT uncertain about the future of cloud computing

Cloud and server Article

ChatGPT has been the talk of the town for more than four months now. As the first ever artificial intelligence (AI) -powered chatbot, it has quickly gained immense popularity, helping students, engineers and even executives generate content, write and debug code and run market analyses. But could ChatGPT be used for anything other than na ...


Tytus Kurek
17 March 2023

What is cloud repatriation?

Cloud and server Article

Cloud repatriation is undoubtedly one of the hottest trends in the cloud infrastructure space as of 2023. It enables organisations to regain control of their cloud spend, workloads and data. According to a report on cloud repatriation by 451 Research group, 48% of IT decision makers confirmed that they had moved their applications away fr ...


Tytus Kurek
16 January 2023

Cloud optimisation: cut your 2023 cloud spending by two-thirds

Cloud and server Article

Cloud optimisation enables organisations to significantly lower their cloud spending while ensuring the desired performance and necessary compliance. It’s a process that every business should adopt when choosing cloud infrastructure as a foundation for their applications. It’s been proven that applying cloud optimisation best practices le ...