As far back as the mid-1990s, companies began bucking the IT decentralization trend by moving Unix and Windows servers into their data centers. But many data center managers are still struggling to coordinate systems management processes between their mainframes and the newer servers.
The big roadblock is getting business managers and various IT groups to agree that common mainframelike-systems operation procedures are needed in areas such as software change management and the scheduling of batch processing jobs, said a half-dozen attendees at last week's spring conference of the American Federation of Computer Operations Management (AFCOM) here.
"We still have to get our hands around open systems," said Pete Lillo, manager of data center operations at NCCI Holdings Inc., a Boca Raton, Fla.-based company that collects and processes data for firms in the workers' compensation industry.
NCCI's data center currently houses a mainframe, 20 Unix servers and about 130 Windows-based systems. Early last year, Lillo said, a new CIO persuaded managers at the company to adopt a unified change management process.
But there's more to be done, Lillo said. NCCI puts its application source code data in four separate repositories that were set up for the different types of systems. A proposal to develop a single repository or provide a common view of the existing data has received preliminary approval but is still in the evaluation stage, Lillo said.
Lillo is also seeking funding for enterprise job scheduling software. He said there's no common scheduling tool now, which means that data center workers at NCCI "really don't manage [the process] the way I'd like it to be managed."
Other IT managers at the semiannual conference run by AFCOM, an Orange, Calif.-based association for data center professionals, said they face similar challenges.
Coordinating operation support procedures on Unix and Windows servers "is the biggest problem we have," said David Sandberg, chief of the business management division at the Defense Information Systems Agency's data center in Ogden, Utah. The facility controls 140 servers and a pair of mainframes. Sandberg said he's looking to create a team that would take charge of tracking the configurations of all of the systems and potentially set up a common change management process. But nothing is definite yet, he added.
Gary Yeck, an analyst at Meta Group Inc. in Stamford, Conn., said much of the resistance to adopting more rigid systems operation procedures boils down to a distrust of mainframe methods.
"Walk into any data center and talk to the distributed [systems] people, and they think mainframe is a dirty word," Yeck said. "They don't want to carry forward any of the disciplines we've learned from the mainframe world."
But some IT managers are finding ways to get around that resistance. For example, Royal Bank of Canada next weekend is due to finish converting the last of 43 server-based applications as part of an 18-month project that was designed to let its data center personnel use mainframe tools to do multiplatform job scheduling and workload balancing.
Business units didn't have to sign up, said Patrick Morassutti, manager of enterprise scheduling and batch automation at the Toronto-based bank. But if they didn't, they would have to monitor overnight batch jobs themselves. "They get scared when you say that," he said. "No one's opted out yet."