Thursday, June 17, 2010

CollabNet delivers Subversion Edge to ease deployment, administration

When you think of the cloud, Agile application lifecycle management may not be the first thing that comes to mind. CollabNet is hoping to change that with a beta distribution today that combines Subversion, Apache and ViewVC with a web-based management interface.

Dubbed CollabNet Subversion Edge, the graphical user interface (GUI) works to simplify installation, administration, use, and governance of the software stack. It does this with a built-in update mechanism that alerts users when new software components are ready to install. Users can then install those updates directly from the browser.

The idea is to make it faster and easier to deploy Subversion servers by reducing administration time. And the promised end result is cost savings, says CollabNet in Brisbane, Calif. The graphical user interface on the open-source software is central to the promise. The interface aims to eliminate admin errors, make management of employee and external partner development teams more secure, and provide operational analytics.

In other CollabNet news this week, HP Software and Solutions announced a partnership with CollabNet to fill out HP's ALM capabilities and enhance testing and quality assurance capabilities.

CollabNet CEO Bill Portelli says the new product fits in with the company’s mission to advance the way in which Agile teams develop software. He went on to say that the company listened to the market and worked closely with global talent on the open source project.

“What’s clear is that Subversion installation, administration, security, and governance have historically been command-line-driven, requiring special expertise and up-to-the-minute knowledge,” Portelli says, adding that CollabNet Subversion Edge is a response to that challenge.

A closer look at Subversion Edge

In effort to make it easier to deploy Subversion servers, CollabNet Subversion Edge includes a certified bundle of all the software components of a fully functioning, Apache-based Subversion stack. This bundle effectively does away the need to search the web for software versions that work together. Then there’s the administration console that installs, manages, and updates the entire stack.

Beyond the deployment, the web-based user interface also works to streamline common administrative tasks, like creating and managing Subversion repositories, starting and stopping the Subversion server, and monitoring server health. The software automatically rotates and archives the logs, and an integrated auto-update feature lets users know when compatible updates are available for any of the distribution components.

On the security and governance front, CollabNet Subversion Edge offers web-based, role-based access control to manage all of the software’s administration features. It also lets admins manage internal and external user access to the Subversion repositories to bolster governance.

Free downloads of the beta version are available immediately on the CollabNet site. General availability is expected later this summer.
BriefingsDirect contributor Jennifer LeClaire provided editorial assistance and research on this post. She can be reached at http://www.linkedin.com/in/jleclaire and http://www.jenniferleclaire.com.
You may also be interested in:

Tuesday, June 15, 2010

HP's Robin Purohit unpacks Business Service Management 9 as way to address complexity in hybrid data centers

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Welcome to a special BriefingsDirect podcast, an interview with Robin Purohit, Vice President and General Manager of the Software Products Business Unit for HP Software & Solutions, conducted by Dana Gardner, Principal Analyst at Interarbor Solutions.

The one-on-one discussion comes to you from the HP Software Universe 2010 Conference in Washington D.C. We're here the week of June 14, 2010, to explore some major enterprise software and solutions trends and innovations making news across HP’s ecosystem of customers, partners, and developers.

Here are some excerpts:
Purohit: Our customers are dealing with some of the most significant combination of changes in IT technologies and paradigms that I had ever seen. There's a whole new way of developing applications like Agile development, the real acceleration of virtualization from your desktop and test environments to production workloads, and all the evaluations of where the cloud and software as a service (SaaS) fits and how that supports the enterprise applications.

All these things and more are colliding at once. What customers are saying is, "How do we take advantage of these new technology shifts and new ways of dealing with technology to get dramatic impact in cost, but not increase the risk? How can I do things faster and cheaper, but do things right?"

What they're looking for is a way to somehow simplify and automate the use of all these technologies and their processes using management software, so they get the most they can of these new paradigm shifts, while they keep up what the business wants them to do.

Compared to a year ago, the active interest of our clients in cloud computing has just exploded. Last year was a curiosity for many senior IT executives. It was something on the horizon, but this year there's really an active evaluation.

Most customers are looking initially at something a little safer, meaning a private cloud approach, where there is either new stack of infrastructure and applications are run for them by somebody else on their site or at some off-site operation. That seems to be the predominant paradigm.

Piece of a puzzle

The challenge is that that set of cloud services, that private service, is really just a piece of a puzzle to run their business operation. It's usually slice of infrastructure or certain class of application that’s part of the larger critical business service for their company.

What they have to do is to figure out how to take advantage of that, target the right workload where it's okay to take that risk, select the right partner, but then make sure that all of the instrumentation of making sure they are getting what they wanted out of it is actually integrated with the rest of their operation. Otherwise, it's just another thing to manage.

The challenge for IT is how they enable that level of innovation at the right pace, but make sure that it's all very well governed -- simple things like getting what we pay for in terms of performance, capability, and the capacity.

That whole notion of cloud governance is one of the most critical things to get right near term, so that the IT guys can keep up with the business guys.



We're sourcing some sort of elastic-like services. And, by the way, is that environment secure, so we're not putting the business at risk? Then, if I want to change to a different cloud provider or another private provider, how do I do that in a fairly nimble way, without having to re-architect everything that I have done.

That whole notion of cloud governance is one of the most critical things to get right near-term, so that the IT guys can keep up with the business guys, but all the risk is there, it's still going be on the line.

There are two really big cost drains in IT. One is that when they roll out your applications, the majority of the time, things don't work well when they initially roll something out. If you think of Agile and the pace at which now new application innovations are bring rolled out, it really means that you have to get things right the first time. The first thing is to tackle that problem, as you go to these hybrid models.

Chaotic environment

The second thing is that most companies still are trying to get a handle on a right way of simplifying and automating their operation in a very chaotic environment. A typical data center is dealing with 900 changes a month. They might get a million incidents over a couple months and each one of those incidents could cost up to $80 plus labor. So, you can just imagine how chaotic and expensive it is just to run day-by-day.

It's really critical for both, the hand-off of the applications to operation, as well as this running the daily operations. This is incredibly automated and simplified and all focused on the impact of all these things that the business wanted in the first place.

If we do that right and then make extensions into all of these same core processes to accommodate a SaaS model, a private cloud model, or even ultimately a public cloud model, without having to change all of that, you are going to be able to bridge from today to the future. You'll be getting all that benefit and actually keep reducing your cost, because you want to keep doing this innovation in a sustainable way.

Arrival of HP BSM 9.0


This has been a great release, and we're incredibly proud of it. BSM 9 is our solution for end-to-end monitoring of services in the data center. It's been a great business for us, and we have a break-through release that we revealed to our customers this week.

It's anchored on what we call the runtime service model. A service model is basically a real-time map of everything from the business transactions of the businesses running to all of the software that makes up that composite applications for the service, and all of the infrastructure -- whether it be physical or virtual, on-premise or off-premise -- that supports all of that application.

All of that together -- knowing how it's connected, what the health of it is, what's changing in it so, you can actually make sure it's all running exactly the way the business expects -- is really critical.

If you can imagine what we've talked about with virtualization and the rate of change there, people optimizing virtual workloads, new application coming on as being fired in the data center with Agile and maybe some outsourced environments and private/public clouds, that service model better be real time and up to date all the time.

That’s the real break through. Before we had a service model that was really linked to the configuration that we thought was running. Now that they have everything up to date in real time with all of this increased velocity, it's really critical.

So, we've rolled that out, and it's now the backbone for all of our end-to-end monitoring. The other thing I'd stress is that once you have that, especially, in this very fast-paced environment, you can really increase the levels of automation.

What you've seen before

When you detect an event, making sure you know exactly what was going on at the time of the event, you can help people diagnose it and probably help solve it, because most of the times these things are based on what you've seen before.

We've taken all of our world-class automation technology, wrapped right into this end-to-end monitoring solution to automate everything possible. We think this can drive dramatic reduction in the cost of operations.

The last thing, I’d emphasize is that there are a lot of people involved in solving these problems, and running these operations. What’s important is that all of them have a very personalized UI that looks and feels like a modern application, but os all based on one version of the truth of what’s going on. We made major improvements, just overhauling the way all of this is presented in a very rich Web 2.0 way, but also in a way that’s targeted to the needs of every single user in operations.

What we're trying to do in HP is not just worry about the data center. We're trying to help customers really adapt and morph these applications into the new world. Most customers are shifting their IT focus on innovating around the application. This means that more of their people who they have internally are creating new IT in the form of a new application for a sales person or new customer-facing portal. That’s going to drive more revenue.

What we're really trying to do is to help them bridge that world, which is very innovation centric, into this new hybrid world, which has to be very operationally tight. A couple of things that we have also announced this week have gotten great feedback. One is a new capability called Application Deployment Manager, which is basically an extension to our industry relating automation capabilities.

That’s going to allow us to bridge all that upstream innovation, make sure it’s designed and tested correctly, then hand it off in an automated way into production, and run on these new optimized hybrid environment.



It really allows development, QA, and operations to coordinate hand-offs of applications in a very well prescribed way, so that they can make sure that what they designed gets handed off and rolled out into the production environment in a very crisp automated way and a way that represents the best practices and everything that’s been learned in a QA cycle. That was a big step forward.

We've also worked up-stream. We've extended our Quality Management Solution to tackle the requirements problem, linking business developers and QA together, and opened up that environment, so that it's much easier to integrate with source code management tools and development tools from folks like CollabNet.

CollabNet is one of the industry's leading development tools providers. As announced, we've integrated with that new open interface. We also support any software out there in that environment. That’s going to allow us to bridge all that upstream innovation, make sure it’s designed and tested correctly, then hand it off in an automated way into production, and run on these new optimized hybrid environment. So, we really are talking about the whole problem, which is really the thing that our customers are most excited about.

For me, this release is an extremely proud moment. This has been the vision that we've had for some time, particularly around BSM 9, being able to bring all these points of monitoring information together into a simple, powerful way to solve these big business problems. What’s changed though is that the necessity to do that now in this new, rapidly changing environment. So, all this new technology becomes even more important to our customers.

For us, particularly, BSM 9 is vision being turned into reality at just the right time for our customers. That’s really the most exciting thing for me about what we did here this week.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in:

HP's Anton Knolmar on seeking innovations to enterprise IT challenges at Software Universe conference

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Welcome to a special BriefingsDirect podcast series, coming to you from the HP Software Universe 2010 Conference in Washington D.C. We're here the week of June 14, 2010, to explore some major enterprise software and solutions trends and innovations making news across HP’s ecosystem of customers, partners, and developers.

We're now joined by Anton Knolmar, Vice President of Marketing for HP Software & Solutions. The interview is conducted by Dana Gardner, Principal Analyst at Interarbor Solutions.

Here are some excerpts:
Knolmar: There are a lot of things going on at the moment around virtualization, cloud, outsourcing, on-premise, off-premise. I think that over the next five or six years, there will be an even greater disruptions in how organizations adopt and use technology.

If you look around, there are a couple of facts which are really critical. There will be a 100 percent increase in the number of virtual machines from 2009 to 2012 and a 43 percent growth in virtualized applications. Mobile devices will grow even more.

A lot of things are happening at the moment around those different areas, around mobile devices, cloud, and virtualization, as well as around the information explosion for the foreseeable future.

Ahead of the game

D
efinitely, from an HP and from an HP Software & Solutions perspective, we want to be ahead of this game and provide the appropriate level for our customers, so that they can be future ready with whatever we provide them.

They can deploy there services. They can deploy them better. They can deploy it all more simply. They can integrate more simply. And, they can use this stuff to give them, our customers, a competitive advantage against their competitors.

Technologies like virtualization and cloud represent the biggest disruption in the technology environment since client-server. But, unlike client-server, the entire enterprise is not just going on one service delivery method or another. We believe that enterprises will have hybrid technologies, as well as a hybrid application environment.

This hybrid environment will be created from enterprise sourcing services from a variety of service delivery models. It will require a set of tools that can manage the service irrespective of where it comes from, either from in-house, physical, virtual, outsource, or via the cloud.

These new delivery models will be directly related to how they can manage and how they can automate them, irrespective of where they are sourced or where they are running.



The ability to benefit from these advances is where our customers are struggling. These new delivery models will be directly related to how they can manage and how they can automate them, irrespective of where they are sourced or where they are running.

That’s one key piece of our announcement. What we can get across from this event in Washington is to explain our customers how we can help them to speed up time-to-innovation by reducing the risk. On the other hand, they can get ready by building a management environment that is ready for the next big thing.

Also, we can explain to our customers how we can simplify, integrate, and automate to gain, as I mentioned before, a competitive advantage from the new technologies. What’s clear to everyone is that one size fits no one. So, enterprises will need to have multiple sourcing options for their applications.

HP Software Universe has quite a history. It’s not the first time that we're running this. A big thing for this year, especially for Americans, is that we're moving out of Las Vegas, where we were for the last couple of years, to Washington.

First, we wanted to get a different staging and we wanted to attract our public-sector customers. That’s an important thing for us, because we do a lot of business here. That’s one of the reasons we've moved to Washington.

What we're also doing for the first time is web streaming our content to different parts of the world, so that we really can reach out much more broadly. We're also building up a kind of HP Software & Solutions community. We have other ways of doing this and are using the social media capabilities as well. [Search for conference goings-on at Twitter on #HPSWU.]

We're connecting our customer community in a better way to bring those pieces together, even the different persona levels. Basically, we're drilling down from a CIO level, via the VP or IT manager, down to the other areas, where we have more of a practitioner level.

For this show specifically, people can even follow us on Twitter and on Facebook. It’s really a big thing for us to be investing in these kinds of new areas and reaching out as broadly as we can do here to the different target audiences and using all these new capabilities which are out there.

More than 200 track sessions


We're really excited, as we have a fully packed agenda from Tuesday until Friday with mainstage sessions and 200 track sessions. We have a dedicated Executive Track and we have a combined solution showcase, where we have our HP product experts, our service professionals, and our key partners together.

We have awards of excellence and partner summits. So, we're really trying to get the entire ecosystem that has already deployed our solutions, from a customer and a partner perspective here, but also prospects in Washington. I'm happy to tell you that all the hotel rooms are booked. So, we're fully packed in the middle of Washington to try to bring across the best of what we have from our Software and Solutions portfolio.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in:

New HP products take aim at managing complexity in 'hybrid data center' era

WASHINGTON - The HP Software Universe 2010 conference is in full swing here this week, highlighted by a slew of HP products and services focused on giving IT leaders the means to view and control their rapidly escalating complexity.

Among the most significant announcements Tuesday is the next iteration of the HP Business Service Management (BSM) software suite. HP BSM 9.0 works on automating comprehensive management across applications lifecycles, with new means to bring a common approach to hybrid-sourced app delivery models. Whether apps are supported from virtualized infrastructures, on-premises stacks, private clouds, public clouds, software as a service (SaaS) sources, or outsourced IT -- they need to managed with commonality. [Disclosure: HP is a sponsor of BriefingsDirect podcasts.]

"Our customers are dealing with some of the most significant combination of changes in IT technologies and paradigm that I have ever seen," said Robin Purohit, Vice President and General Manager of HP Software Products. "So whether it's a whole new way of developing applications like Agile, the real acceleration of virtualization, test environments moving into production workloads, and all of the evaluations of where the cloud and SaaS fits -- and then how they support all the enterprise applications."

[See an interview with Purohit on BSM 9.0, or listen to it as a podcast.]

HP’s core message around the BSM 9.0 is clear: To help companies automate apps and services management amid complexity so they can reinvest in innovation. The economics of innovation -- of being able to do more in terms of results without the luxury of spending a lot more -- has to be factored into the management matrix.

“We did a study last October that showed our clients believe innovation is going to help them even through uncertain economic times—and they see technology as central to their ability to succeed in a changing environment,” said Paul Muller, vice president of Strategic Marketing, Software Products, HP Software & Solutions. “There is some skepticism within the business community that IT is ready to make those changes.”

HP’s view of the hybrid world

HP wants to help IT combat that skepticism by equipping them with HP BSM 9.0. The solutions not only address hybrid delivery models, but also what Muller calls the “consumerization of IT,” referring to people who use non-company-owned devices on a company network. As Muller sees it, employees expect to have the same dependable experience while working from home as they do at work.

“We believe organizations will struggle to deliver the quality outcomes expected of them unless they are able to deal with the increase rate of change that occurs when you deploy virtualization or cloud technologies,” Muller said. “It’s a change that allows for innovation, but it’s also a change that creates opportunity for something to go wrong.”

Indeed, Bill Veghte, HP Executive Vice President for HP Software and Solutions, said that three major trends -- all game changers on their own -- are converging around IT: virtualization, cloud, and mobile.

"We need to continuously simplify" management to head off rapid complexity acceleration aroud this confluence of trends. He pointed to the need for gaining a comprehensive view into hybrid IT operations, automation for management and remediation, and "simply expressing" the views on what is going on it IT, regardless of the location or types of services.

"Users want a unified view in a visually compelling way, and they want to be able to take action on it," said Veghte.

At the center of Wednesday’s announcement is what HP proposes as the solution to this challenge: HP BSM 9.0. The software offers several features that should cause companies that work with hybrid IT environments to take a closer look. One of those features is automated operations that work to reduce troubleshooting costs and hasten repair time. BSM 9.0 also offers cloud-ready and virtualized operations that aim to reduce security risks with strategic management services.

"The active interest of our clients in cloud computing has just exploded," said Purohit. "I think last year was a curiosity for many senior IT executives something on the horizon, but this year it's really an active evaluation. I think most customers are looking initially at something a little safer, meaning a private cloud approach, where there is either new stack of infrastructure and applications are run for them by somebody else on their site, or at some off-site operation. So that seems to be the predominant new paradigm."

Purohit described BSM 9 as "a breakthrough" for coming to grips with the "hybrid data center."

It’s the great trap because if you don’t know what infrastructure your application is depending on from one minute to the next, you can’t troubleshoot it when something goes wrong.


Indeed, integration is a running theme with HP BSM 9.0. The solution offers a single, integrated view for IT to manage enterprise services in hybrid environments, while new collaborative operations promise to boost efficiency with an integrated view for service operations management. Every IT operations user receives contextual and role-based information through mobile devices and other access points for faster resolution.

"BSM 9 is our solution for end-to-end monitoring of services in the data center. It's been a great business for us, and we now have a break-through release that we reveled to our customers today, that’s anchored around what we call the Runtime Service Model," said Purohit.

"So a service model is basically a real-time map of everything from the business transactions of the businesses running, to all of the software that makes up that composite applications for the service, and all of the infrastructure whether it be physical or virtual or on-premise or off-premise that supports all of that application," said Purohit.

"So all of that together -- knowing how it's connected, what the health of it is, what's changing in it so, you can actually make sure it's all running exactly the way the business expects -- is really critical," he said.

The run-time service model works to save time by improving organizational service impact analysis and troubleshooting processing times, said HP. The HP Business Availability Center (BAC) 9.0 offers an integrated user experience as well as applications monitoring and diagnostics with HP’s twist on the run-time service model.

“The rate of change in the way infrastructure elements relate to each other -- or even where they are from one minute to the next -- means we’ve moved from an environment where you could scan your infrastructure weekly and still be quite accurate to workloads shifting minute by minute,” Muller said. “It’s the great trap because if you don’t know what infrastructure your application is depending on from one minute to the next, you can’t troubleshoot it when something goes wrong.”

Other elements of BPM 9.0 include the BAC Anywhere service that lets organizations monitor external web apps from anywhere, even outside the firewall, from a single integrated console. HP Operations Manager i 9.0 promises to improve IT service performance by way of “smart plug-ins” that automatically discover application changes and updates them in the run-time service model. Finally, HP Network Management Center 9.0 gives aims to give companies better network visibility by connecting virtual services, physical networks and public cloud services together.

HP’s expanded universe

In other HP news, the company announced software that aims to accelerate application testing while reducing the risks associated with new delivery models. Dubbed HP Test Data Management, HP also promises the new solution lowers costs associated with application testing and ensures sensitive data doesn’t violate compliance regulations.

The improvement helps simplify and accelerate testing data preparation, an important factor in making tests and quality more integral to applications development and deployment, again, across a variety of infrastructure models.

Along with HP Test Data Management, HP launched new integrations between HP Quality Center and the Collabnet TeamForge with the goal of improving communication and collaboration among business analysts, project managers, developers and quality assurance teams.

The integration with CollabNet, built largely on Apache Subversion, will help further bind the "devops" process, said Purohit. "The result is better apps," he said.

And as part of its work to help clients maximize software investments, HP also rolled HP Solution Management Services. This offering is a converged portfolio of software support and consulting services that offers a single point of accountability to manage enterprise software investments.
BriefingsDirect contributor Jennifer LeClaire provided editorial assistance and research on this post. She can be reached at http://www.linkedin.com/in/jleclaire and http://www.jenniferleclaire.com.
You may also be interested in:

HP Data Protector, a case study on scale and completeness for total enterprise data backup and recovery

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript. Sponsor: HP.

Welcome to a special BriefingsDirect podcast series coming to you from the HP Software Universe 2010 Conference in Washington, DC. We're here the week of June 14, 2010 to explore some major enterprise software and solutions trends and innovations making news across HP's ecosystem of customers, partners, and developers.

Our topic for this live conversation focuses on the challenges and progress in conducting massive and comprehensive backups of enterprise live data, applications, and systems. We'll take a look at how HP Data Protector is managing and safeguarding petabytes of storage per week across HP's next-generation data centers.

The case-study on HP's ongoing experiences sheds light on how enterprises can consolidate their storage and backup efforts to improve response and recovery times, while also reducing total costs.

To learn more about high-performance enterprise scale storage and reliable backup, please welcome Lowell Dale, a technical architect in HP's IT organization. The interview is moderated by Dana Gardner, Principal Analyst at Interarbor Solutions.

Here are some excerpts:
Dale: One of the things that everyone is dealing with these days is the growth of data. Although we have a lot of technologies out there that are evolving -- virtualization and the globalization-effect -- what we're dealing with on the backup and recovery side is an aggregate amount of data that's just growing year after year.

Some of the things that we're running into are the effects of consolidation. For example, we end up trying to backup databases that are getting larger and larger. Some of the applications and servers that consolidate will end up being more of a challenge for some of the services such as backup and recovery. It's pretty common across the industry.

In our environment at HP, we're running about 93,000-95,000 backups per week with an aggregate data volume of about 4 petabytes of backup data and 53,000 run-time hours. That's about 17,000 servers worth of backup across 14 petabytes of storage.

It's pretty much every application that HP's business is run upon. It doesn’t matter if it's enterprise warehousing or data warehousing or if it's internal things like payroll or web-facing front-ends like hp.com. It's the whole slew of applications that we have to manage.

The storage technologies are managed across two different teams. We have a storage-focused team that manages the storage technologies. They're currently using HP Surestore XP Disk Array and EVA as well. We have our Fibre Channel networks in front of those. In the team that I work on, we're responsible for the backup and recovery of the data on that storage infrastructure.

We're using the Virtual Library Systems that HP manufactures as well as the Enterprise System Libraries (ESL). Those are two predominant storage technologies for getting data to the data protection pool.

One of the first things we had to do was simplify, so that we could scale to the size and scope that we have to manage. You have to find and simplify configuration and architecture as much as possible, so that you can continue to grow out scale.

We had to take a step-wise approach on how we adopted virtual tape library and what we used it for. Virtual tape libraries were one of the things that we had to figure out. What was the use-case scenario for virtual tape? It's not easy to switch from old technology to something new and go 100 percent at it.

We first started with a minimal amount of use-cases and little by little, we started learning what that was really good for. We’ve evolved the use case even more, so that in our next generation design that will move forward. That’s just one example.

We're still using physical tape for certain scenarios where we need the data mobility to move applications or enable the migration of applications and/or data between disparate geographies.



HP Data Protector 6.11 is the current release that we are running and deploying in our next generation. Some of the features with that release that are very helpful to us have to do with checkpoint recoveries.

For example, if the backup or resource should fail, we have the ability with automation to go out and have it pick up where it left off. This has helped us in multi-fold ways. If you have a bunch of data that you need to get backed up, you don’t want to start over, because it’s going to impact the next minute or the next hour of demand.

Not only that, but it’s also helped us be able to keep our backup success rates up and our tickets down. Instead of bringing a ticket to light for somebody to go look at it, it will attempt a few times for a checkpoint recovery. After so many attempts, then we’ll bring light to the issue so that someone would have to look at.

Data Protector also has a very powerful feature called object copy. That allowed us to maintain our retention of data across two different products or technologies. So, object copy was another one that was very powerful.

There are also a couple of things around the ability to do the integration backups. In the past, we were using some technology that was very expensive in terms of using of disk space on our XPs, and using split-mirror backups. Now, we're using the online integrations for Oracle or SQL and we're also getting ready to add SharePoint and Microsoft Exchange.

Now, we're able to do online backups of these databases. Some of them are upwards of 23 terabytes. We're able to do that without any additional disk space and we're able to back that up without taking down the environment or having any downtime. That’s another thing that’s been very helpful with Data Protector.

Scheduling overhead

With VMs increasing and the use case for virtualization increasing, one of the challenges is trying to work with scheduling overhead tasks. It could be anywhere from a backup to indexing to virus scanning and whatnot, and trying to find out what the limitations and the bottlenecks are across the entire ecosystem to find out when to run certain overhead and not impact production.

That’s one of the things that’s evolving. We are not there yet, but obviously we have to figure out how to get the data to the data protection pool. With virtualization, it just makes it a little bit more interesting.

Nowadays, we can bring up a fairly large-scale environment, like an entire data center, within a matter of months -- if not weeks. This is how long it would take us. The process from there moves toward how we facilitate setting up backup policies and schedules, and even that’s evolving.

Right now, we're looking at ideas and ways to automate that, so that' when a server plugs in, basically it’ll configure itself. We're not there yet, but we are looking at that. Some of the things that we’ve improved upon are how we build out quickly and then turn around and set up the configurations, as that business demand is then turned around and converted into backup demand, storage demand, and network demand. We’ve improved quite a bit on that front.

Being able to bring that backup success rate up is key. Some of the things that we’ve done with architecture and the product -- just the different ways for doing process -- has helped with that backup success rate.

The other thing that it's helped us do is that we’ve got a team now, which we didn’t have before, that’s just focused on analytics, looking at events before they become incidents.

With some of the evolving technologies and some of the things around cloud computing, at the end of the day, we'll still need to mitigate downtime, data loss, logical corruption, or anything that would jeopardize that business asset.

With cloud computing, if we're using the current technology today with peak base backup, we have to get the data copied over to a data protection pool. There still would be the same approach of trying to get that data. If there is anything to keep up with these emerging technologies, for example, maybe we approach data protection a little bit differently and spread the load out, so that it’s somewhat transparent.

Some of the things we need to see and we may start seeing in the industry are load management and how loads from different types of technologies talk to each other. I mentioned virtualization earlier. Some of the tools with content-awareness and indexing has overhead associated with it.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript. Sponsor: HP.

You may also be interested in: