META Report: Planned or Unplanned, It's All Downtime
The art of high-availability (HA) planning -- that is, keeping your IT system up and running as much as 99.999 percent of the time -- is to balance potential business cost with the definite cost of insuring against system downtime.
Through 2003, database management system vendors will continue to improve and promote HA options -- including online utilities, improved replication, and scale-out clusters -- to meet expected user demand. Through 2005, organizations will struggle to implement and enforce processes, such as change management, designed to mitigate the need for planned downtime and the impact that planned changes have on unplanned downtime. This will be due to an overreliance on redundant infrastructures and change management tools without devoting proper attention to internal people, processes, and responsibilities.
Solving unplanned downtime is about eliminating single points of failure. This goal is a priority among infrastructure planners, and certainly redundant application infrastructures are a prerequisite of any HA strategy. However, a recent outage at one of the major stock exchanges shows that simply having two of everything cannot eliminate all downtime (such as that caused by a change made to application code during a planned maintenance window). Failover may occur, but if the failover node is also running corrupted software, it too will fail.
While this appears to the outside world as a technology glitch, the truth is that software (unlike hardware) never wears out, so when there is a problem, it is likely a human or process issue, not a technology issue. Organizations must recognize that people, process, and infrastructure are all interdependent facets of an HA solution. In fact, the people and process issues comprise at least 80 percent of the solution, with infrastructure (i.e., redundancy) assuming the remainder.
Most IT organizations need look no further than their own mainframe environment for inspiration. Although it is universally accepted as more reliable and available than distributed platforms, the mainframe is not inherently more reliable. It simply has more defined processes, procedures, roles, and responsibilities (e.g., scheduling groups, change management groups, or security groups) than the distributed platform. Centralized infrastructure support models are a first step that every IT organization must consider when developing an HA strategy.
|Other Recent META Reports|
|How To Ensure IT Projects Boost Profits Taking Ordering to the Web The Hazards Of Subscription Licensing Room At the Data Center?|
Planned downtime is reserved for performing necessary jobs, such as batch jobs, system performance tuning, or application or system code fixes and upgrades; the common denominator is typically a change of some kind. Change, while unavoidable, is by far the number-one factor in the cause of downtime. How, then, do best-practice IT organizations address the issue of HA and planned change?
1. Invest in Change Infrastructure.
Organizations must invest in infrastructure to support change. On the pure hardware and software side of infrastructure, this means a well-planned and well-managed quality assurance (QA) test environment.
On the people and process side, a change management group should be established, with responsibility to coordinate, serialize, document, communicate, and approve change promotion (e.g., test to QA, QA to production) for all groups effecting change to an application. This group should require a detailed process flow for applying the change and detailed plans to fall back or recover (including expected time frames). Included should be a listing of users, applications, and jobs potentially affected by the proposed change.
Plans should be communicated to all interested parties within a reasonable time before any scheduled change. This enables colleagues to evaluate any related or conflicting changes they may be contemplating, or to request a delay if the impact of the change is ill-timed based on application usage. It is also important that communication take place early and often in the event of failure after the change is applied. This has proven critical in reducing the actual user impact of a failure, as well as end-user frustration.
3. Reduce Change Mass.
The first rule in problem determination is to understand what has changed in the system. Organizations should introduce change in small, related batches if possible. Introducing too much change at one time will inevitably lengthen the application's mean time to recovery by lengthening problem determination.
4. Work With Vendors.
Managing all the potential vendor software patches (code that fixes reported bugs) and all the different ways vendors package, deliver, and upgrade patches can be intimidating. Organizations should contact vendors or review their Web sites for available fixes on a monthly basis.
- Apply patches to the QA environment first.
- Apply vendor-recommended patches as soon as possible.
- Keep a list of patches required for the specific environment and apply as needed.
- Avoid patch versions that are constantly changing. View patches whose versions have not changed within two months as safe to apply.
- Research user feedback on new service packs (e.g., bundles of patches) via Internet bulletin boards to judge how others have fared. Avoid being the first to apply one if possible.
5. Protect Personnel.
Organizations must keep the people happy that make the change process work. Personnel should not be overmanaged and should be given the authority and responsibility to do their jobs. Organizations should motivate and reward them for meeting service-level agreements by paying bonuses and giving additional pay for on-call duties. Organizations should keep personnel educated and current by sending them to a related conference or course once a year.
Business Impact: Application availability greatly affects an organization's bottom line, brand equity, and valuation. Budget dollars should be spent on the people that plan and execute change processes in the high-availability infrastructure.
Bottom Line: Planned downtime is about managing change and accounts for as much as 90 percent of all application downtime. To manage change, organizations must recognize this and focus high-availability planning resources on addressing the issue of people and process.
Charlie Garry is a consultant for META Group, an IT consulting firm based in Stamford, Conn.