Tag Archive | hp

Data – The new Rock’n’Roll

“ Data is the new oil”

“The most valuable currency in the world is not money, it’s information”

– A couple of great quotes written by people much more eloquent than me. However I do have one of my own ;

Data is the new rock’n’roll

Just as rock’n’roll transformed music scene the use, and future potential use, of information is dramatically changing the landscape of a data centre. Historically the storage array was effectively the drummer of the band, required but sitting fairly quietly in the background, and whilst a vital component it was not necessarily the first thing people thought of when putting the band together. Even now, if you look at a picture of any band, the drummer is the one hanging about aimlessly in the background, try naming the drummer in any large and well-known bands; it’s much harder than you think. And so it was with storage and data; the storage array would sit somewhere towards the back of the datacentre whilst the shiny servers were the visible component, and the items that got the most attention.

As we hit 2013 that all changes; the storage array is the Kylie of the datacentre, it’s the sexiest piece of equipment in there. And so it should be given that upwards of 40% of a customer’s IT budget is spent simply on provisioning the capacity to house data.

At Computacenter, we’ve made a large investment in our Solution Centre. Whats sits in the front row now? Of course it’s the data arrays; with the latest technology from EMC, HP, HDS, IBM and NetApp all showcased. Why is it front row? Obviously as it’s the most important component of any solution nowadays. And of course, it looks sexy, or is that just me?

The storage array is now front and centre, it’s the first component to be designed when re-architecting an environment. Why? Simply because a customer’s data is their most valuable asset, it’s transforming the way people do business; it’s changing the way we interact with systems and even each other, your data is now the lead singer in the band.

Data is the one thing that is getting attention within the business; it’s the one thing you have making the front pages of “Heat” magazine – Where’s it going? What’s it doing? Is it putting on weight? Is it on a diet? What clothes is it in? Should it be in rehab? But as the manager of the data (or the band) there is one simple question that you want answered; how do I make money out of it?

And that, dear reader, is the $64,000 question. The good news is that is becoming ever more possible to use your data as a revenue generation tool, we are only starting to see business value being generated from data, as 2013 progresses we will see some niche players mature (and possibly be acquired), we’ll see an increased push from the mainstream vendors and we’ll start to see ways of manipulating and using data that we just couldn’t contemplate when the storage was simply providing the rhythm section.

Even converged systems, the boy bands of the decade, which perform in harmony always have one better singer than the rest, well he’s the data

So: Compute, Networking, and Software, the gauntlet is down; Data is the new rock God, it’s the Mick Jagger to your Charlie Watts, you want the crown back? Come and get it, but for now it’s all mine.

All the data architects out there can join me as I sing (with apologies to Liam & Noel) “…Tonight, I’m a rock’n’roll star!”

How can a good enough network really be good enough?

A quick look at the current popular enterprise networking infrastructure platforms and they all seem to suffer from a similar predicament – almost without exception the functionality is good, reliability levels are high and performance (in relevant terms) delivers against expectations.

The reasons for this rather stable state include a networking journey to date that embraced the pain of interoperability and standardisation many years ago, the common use of high performance off the shelf network processing asics (with a few notable vendor exceptions) and until recently no real need to change the status quo.

After numerous years of highly effective network solution design by the extensively trained and highly talented network engineers, that embraced inherent technology limitations and extracted maximum performance we now have our “good enough” networks. I reiterate that there are many great network engineers that underpin the largest enterprises in the world, make complex networking “just work” and deliver business outcome after outcome – helping in many cases to hide that fact that below the surface all is not as well as it may seem.  

But surely, if you were given a blank sheet of paper and networking / security designs were architected with a clean view of the vendor landscape plus tomorrows business outcomes as well as today’s, would you still design yesterdays way? If the business outcomes of today and definitely tomorrow differ from the network usage approach of yesteryear surely good enough can’t still be “good enough”.

A five year old network designed and configured for large volumes of direct connected network servers with one Gigabit interfaces surely won’t be good enough for a densely consolidated converged infrastructure requiring multiple ten Gigabit network interfaces. Equally a multi layer network topology originally configured for hundreds and potentially thousands of physical servers, with multiple physical network interfaces has very different operational and performance characteristics to a distributed switch, hypervisor virtualised network layer.

The stage is set for good enough (or worse) networks to be evolved in line with tomorrow’s application and business requirements. Software defined networks (SDN) underpinned by the open standards aligned with OpenFlow and Openstack protocols and frameworks may in time enable the granular levels of flexibility and capability required to personalise today’s “good enough” general purpose networked infrastructure footprint into outcome specific networked topologies. This blog was set to discuss the well crafted Cisco ONE strategy that leverages the value delivered by OpenFlow and Openstack and clearly positions a customer journey that leverages existing technologies interfaced with the emerging software network footprints and equally the highly innovative HP VAN software aligned network play that leverages IMC and IRF tightly woven into those same open network software foundations, to deliver tangible application aligned networking.

But both of those great stories may now be somewhat pale when compared to VMware shock acquisition of Nicira. Put simply the worlds dominant x86 hypervisor vendor now includes a highly regarded SDN networking core that can be leveraged in numerous and as yet unannounced ways that could potentially paint a new picture for enterprise networking. (save this for another blog).

So “Good enough networks” in the not too distant future may become a thing of the past. Will they ever be “perfect networks”, unlikely due to the ever changing nature of business and increasing levels of complexity, but could they become much closer aligned with the levels of flexibility and adaptability and cost effectiveness currently sought by enteprise network customers. “Quite possibly”…….

And then they will be more than “Good enough”.

Until next time

Colin W

Twitter: @Colinwccuk