If you were to draw a line graph of world records for the fastest marathon, 100m sprint or triathlon times over the past 100 years, they would be top-left to bottom right. Why? There are multiple reasons. Firstly, athletes train smarter than before: they embark on strict fitness programmes and diets which, when combined, optimise their performance.
Secondly, there are more runners participating than ever before, making the competition fierce and leaving no space for complacency. Thirdly, today’s athlete has access to state-of-the-art technology which monitors performance, both retrospectively and in real-time.
Running a datacentre is much the same: businesses are smarter and more efficient than before. Gone are the days when throwing extra resources at the datacentre would suffice. Instead, today’s IT managers need to carefully consider where they allocate their resources if they are to optimise performance. The necessity of looking at the whole IT infrastructure is due to a combination of new pressures brought about by growing volumes of unstructured data, and ever-evolving, complex IT environments that are continuously impacted by virtualisation, cloud computing, BYOD, rich media and video streaming to name just a few. The traditional ways of coping with an increased need for performance are no longer acceptable, and costly over-provisioning is no longer an option to address performance spikes or degradation. Managers want to stop wasting their budget and other resources and demand to know what’s happening across the entire data centre.
A professional runner can only succeed on the world stage by relying on technology and tools that monitor their performance. The same is true for an IT infrastructure. The tools that give visibility into what’s under the IT systems hood are already available from most vendors although the insight they provide is limited. The problem is that these vendor-specific tools are biased, as they are restricted to that particular vendor’s products. This limits the level of monitoring and troubleshooting that can be done across the entire environment, and does not enable an accurate, unbiased, system-wide view across the infrastructure as a whole.
Most organisations try to manage their infrastructures using the tools provided by their virtual server, fabric and array provider. The problem with such tools is that they do not provide information in real-time because if they did it would result in latency as they polled the devices too often. This means that the reports generated are based on an average over time from information gathered usually at one to five-minute intervals. The other issue with device dependency is that the vendors operating the virtual system can take weeks to report on service levels as getting logs from each device can take days.
The Virtual Instruments VirtualWisdom IPM platform can be easily installed on the back of Traffic Access Points or TAPs; it is compatible with all vendor hardware and has won several awards for its non-disruptive, real-time monitoring and analysis capabilities.
Most data centres have built-in redundancy to cope with spikes in demand for capacity: by using VirtualWisdom, IT managers can see precisely where latency and performance degradation occurs, and they can proactively identify and address any traffic glitches before they become major issues. With VirtualWisdom this granular and comprehensive insight into the IT infrastructure is achieved without impacting application performance or end user experience and it doesn’t add any ‘load’ on to the system as polling does. Once implemented VirtualWisdom reads the Fibre Channel protocol in real time, end-to-end regardless of vendor equipment.
By having in-depth levels of insight datacentre managers can truly plan, and manage their IT environments, and ensure that they remain perpetually agile. In a similar manner, with VirtualWisdom, IT infrastructure managers can see how applications are performing, spot any existing and upcoming bottlenecks, review the level of utilisation of each component and understand whether these can be optimised to reduce cost. Once armed with this information, it is possible to eliminate downtime and latency, create application-aligned SLAs, and set in motion new initiatives such as tiering to ensure that each component (server, switch, fabric, storage) is perpetually optimised, and the IT infrastructure is therefore running as cost effectively as possible.
A marathon runner 100 years ago would have had heavier trainers than his counterpart today; he or she may have focused on one area of performance while ignoring another. Without monitoring tools they would almost certainly over provision in one area while neglecting another.
Likewise, if you don’t TAP and monitor your infrastructure, you are like the athletes of old; over provisioning, spending budget unnecessarily and allocating resources to the most important applications whether they need them or not. You are also susceptible to unplanned latency and outages. No runner today would train without their monitoring in place, they just wouldn’t be competitive enough, and likewise today’s leading businesses have had their data centres tapped because this gives them a huge competitive advantage. Why would anyone want to build a system they have no visibility into, and that can’t guarantee application performance?
Today’s professional athlete will have a team of dedicated specialists working to improve their performance. The dietician, the coach, the physiotherapist will all work in harmony, responding to analytics to ensure optimal performance is achieved. Similarly to a team of sports specialists, different people within IT need different views of what is going on in an infrastructure. But what is happening currently in many organisations is that, instead of the various teams having different views of the same data set, they have different perspectives of their own data. This is a historical issue that creates challenges between what certain people “see” in the storage network, leading to unnecessary finger pointing. This is limiting, as it does not provide the entire team with visibility into the bigger picture. The database administrator, for example, might experience performance problems, but the tools used to assess database performance are quite different and often unfamiliar to the team that gets data from the storage.
Our latest platform, VirtualWisdom4, is akin to having a team of specialists and it is the only product to do so. VirtualWisdom delivers the industry’s first entity-centric approach enabling application-aware infrastructure performance management. Using this approach, organisations can logically group system-wide resources from physical devices to application workloads, providing everyone from IT operations to line of business leaders with the insight they need to understand how resources and applications are performing. IT teams can now customise entities to show all of the resources supporting a specific application, business unit or tier of service, allowing individuals and teams to immediately see the information relevant to them.
Preparing for a triathlon and managing a complex IT infrastructure are two very different occupations. But there are some useful comparisons and lessons for IT managers to be drawn from professional runners. Most crucially, the industry can learn that continually adding additional storage resources will not fix any underlying issues in the long term. In both disciplines, IT management and running, at the heart of success is the ability to find a balance.
For runners, this means finding the right levels of fitness training, physiotherapy and carbohydrates in-take. Equally, what underpins an efficient IT datacentre is the ability to allocate resources efficiently, which requires one to take sight of the entire infrastructure’s performance and to fully understand how its resources and applications are performing.