Tuesday, December 14, 2010

Case study: Automated client management from HP helps Vodafone standardize in 30 countries

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Learn more. Sponsor: HP.

Barcelona -- Welcome to a special BriefingsDirect podcast series coming to you from the HP Software Universe 2010 Conference in Barcelona.

We were here earlier this month to explore some major enterprise software and solutions, trends and innovations making news across HP’s ecosystem of customers, partners, and developers. [See more on HP's new ALM 11 offerings.]

This customer case-study from the conference focuses on Vodafone and how they worked toward improved PC client management and automation of client management. I interviewed two executives from their IT organization, Michael Janssen, the Manager of Deployment Automation with Vodafone , and Michael Schroeder, also Manager of Deployment Automation, both based at Vodafone group in Düsseldorf, Germany.

Here are some excerpts:
Janssen: Vodafone had independent countries operating their [IT] environments by themselves. So, we had 30 countries worldwide with all the solutions in place. That meant 30 times software deployment, 30 times application packaging, 30 times Active Directory, and so on.

Vodafone decided in 2006 to go for a global IT project and centralization in terms of PC client automation. It came down to us to evaluate the current solutions in place in all these countries and then come up with a solution which would be the best solution for the new global environment. That was our main problem.

Standardization and reducing cost

If you're starting a centralization process, then it’s all about standardization and reducing cost. That meant reducing cost by reducing effort of the solutions and make as much as possible automated and self-service. That was the main reason we started this exercise.

Schroeder: The most important thing was that administration should be very easy. It shouldn’t be too complex in the end and it should fit every need in every country. At that time, we had a whole zoo of hardware and software products. We had about 8,000 different software applications in place at that time. We tried to reduce that as much as we could.

The most important thing was that administration should be very easy. It shouldn’t be too complex in the end and it should fit every need in every country.



The overall number of clients in Vodafone is 65,000, and at the moment, we've finished the transition for 52,000 clients. Nearly 80 percent is done after four years. Of course, there is a long wait with the smaller countries, and we need to migrate 15 other countries that are still in the loop.

In the past, in each of these 30 countries, we had one to four people working within the client automation environments. Today, we have five people left doing that globally. You can imagine 30 times a minimum of two persons. That was 60 people working for client deployment, and that's now reduced to five for the global solution.

Always pros and cons

There are always pros and cons with standardization and with centralization. The consensus takes a little bit longer, because there are no strict processes to bring new applications. But, the main advantage is that much of the applications are already there for any country. We test it once and can deploy to many, instead of doing this 30 times, like we did that in the past, and we avoid any double spend of money.

Then, of course, with the global environment, the main advantage is that now we are all connected, which was not possible in the past, because all the networks were independent and all the applications were independent. There was no unified messaging or anything like that. This is the major benefit of the global environment.

Security is one big thing we're now dealing with. For example, if we are talking about client automation, we're talking about patch management as well. We're able to bring out patches -- for example, security patches from Microsoft -- within two days, if it’s a real hot-fix, or even within 24 hours, if it’s a major issue.

Countries that used HP Client Automation had much higher success rate, 90 percent or higher, in deploying application and patches, than the others.



Janssen: First, there was the evolution phase, where we studied all the countries. What were the products that they used in the past? Then we decided what was the best way forward. For us, that was a major split between countries that already used the HP Client Automation solution and the other countries that used other deployment suites.

That was also one of the major criteria for the final decision. Countries that used HP Client Automation had much higher success rate, 90 percent or higher, in deploying application and patches, than the others, where they were on average at 70 percent. So, this was the first big decision point.

The second was countries using HP Client Automation had less operational staff than the others. It was mainly one to two full-time employees fewer than in countries that operated with other tools.

Policy-based technology

Schroeder: If we're talking about the Client Automation Suite from HP, we're talking about policy-based or a desired state technology. That is one of the criteria. Everything is done every day. For example, if you're trying to deploy applications to clients, this is done every day. It's controlled every day, managed every day, and without any admin or user interaction. That’s a great point for us.

Janssen: What I can recommend is that there are two main issues that you need to overcome. First, you only can deploy what you receive from the business. We already were experienced in the Vodafone-Germany organization, where we did the same exercise five years ago. You need to have a strict software standardization process in place. There is one main rule for that.

Also, in the global environment, that means that if there is a business application, then the business needs to have an application owner for that. Otherwise, the application does not exist in the whole company.

We gave that function or that responsibility back to the business, and now they're all responsible and they finally approve before application goes live.



The application owner is responsible for the whole application lifecycle, including describing the application installation documents, doing the final testing and approval after packaging, his responsibility is to look after security issues of the application, look after upgrades or version or release changes, and so on.

It's not not the packaging team, the client team, or the central IT team that is responsible for all the applications and their functionality. We gave that function or that responsibility back to the business, and now they're all responsible and they finally approve before application goes live.

Schroeder: We have in place self-service, which is a web application. You can go to a store and choose different applications to install on your machine, depending on your needs. You can choose an application, just click a box, and the application request goes to your line manager who has to approve the license costs, if there are any. Then, the policy will go back to your machine and the installation of this specific application goes straight to your machine. The user experience with it is very good.

Janssen: The self-service web shop is not only for software. We use that also for other user needs, like access rights, permissions on some projects, mobile device management and so on. This is a global web shop solution, but very effective. It avoids any help desk calls for new applications, paperwork to approve licenses, and so on. It’s very efficient and, of course, one of our main parts of this new global solution.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Learn more. Sponsor: HP.

You may also be interested in:

Thursday, December 9, 2010

WAN governance and network unification make or break successful cloud and hybrid computing implementations

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Learn more. Sponsor: Ipanema Technologies.

Get a free white paper on WAN Governance for Cloud Computing.

Get the free Cloud Networking Report.

The latest BriefingsDirect podcast discussion focuses on the rapidly escalating complexity and consequent need for network management innovation in the age of hybrid computing.

And that's because the emphasis nowadays is on "networks," not "network." Long gone are the days when a common and controlled local area network (LAN) served as the workhorse for critical applications and data delivery.

With the increased interest in cloud, software as a service (SaaS), and mobile computing, applications are jockeying across multiple networks, both in terms of how services are assembled, as well in how users in different environments access and respond to these critical applications.

Indeed, cloud computing forces a collapse in the gaps between the former silos of private, public, and personal networking domains. Since the network management and governance tasks have changed and continue to evolve rapidly, so too must the ways in which solutions and technologies address the tangled networks environment we all now live and work in.

Automated network unification and pervasive wide area networking (WAN) governance are proving essential to ensure quality, scale, and manage security across all forms of today's applications use. Join us as we explore the new and future path to WAN governance and to better understand how Ipanema Technologies is working to help its customers make clear headway, so that the next few years bring about a hybrid cloud computing opportunity and not a hastening downward complexity spiral.

We're here now to discuss the new reality of networks and applications delivery performance with Peter Schmidt, Chief Technology Officer, North America, for Ipanema Technologies, and David White, Vice President of Global Business Development at Ipanema. The panel is moderated by BriefingsDirect's Dana Gardner, Principal Analyst at Interarbor Solutions.

Here are some excerpts:
Schmidt: As soon as you start using multiple networks, you're in the cloud, because now you're making use of resources that are outside the control of your own IT organization and your service provider. Whether people think about it or not, just by adding a second network, they're taking their first steps into the cloud.

Anybody who carries a smartphone is experiencing the personal, private, public boundary of operations themselves. But what seems natural to somebody carrying an iPhone or Blackberry is a tremendous challenge to the traditional models of IT.

Even as little as three years ago, the focus was on how to get the most performance for your applications out of your single MPLS network. I am talking enterprises where all of their applications are hosted on their property. They’ve got a single MPLS network from one service provider and they're still struggling to deliver reliable application performance across the infrastructure.

Now, we throw in multiple places to host applications. You have SaaS, Salesforce, and Google Docs. You have platform as a service (PaaS) and infrastructure as a service (IaaS). People’s critical applications can be hosted in numerous locations, many of which are beyond their control. Then, as I mentioned, these are being accessed via multiple networks, and you have the legacy MPLS plus the Internet.

There are increasing numbers or diversity of models of those networks, whether the Internet connection gets to a service provider POP and then via MPLS to their own data center, or what is the impact of content delivery networks? So we've got a situation where enterprises who are struggling to master the complexity with one data center and one network are now using multiple data centers and multiple networks. Something is going to have to give.

White: This is also all focused once again on the branch office. We’ve had server consolidation where we try to remove any type of issues for the branch and remove intelligence from the branch. As cloud computing has come in, we are now putting more stress on the branch.

We're not necessarily putting intelligence out there, but we're having 2, 3, 4, 5, or more networks, all coming into the branch at the same time, and that traffic has to be managed. It’s something a lot of people haven’t thought about.

When you look at the announcements that have been coming out and the hype on cloud in the industry, it's all focused on the data center. That’s because most of the vendors say, "That’s where the big bucks are being made. We are going to make money out of the data center."

Ipanema, on the other hand, is focused on application acceleration, and in order to do that, you have to take care of what goes on in the branch and manage it.

At a high level, the first thing you have do is provide some type of WAN governance, simply meaning that we are going to make sure that you have taken care of the management of your business. Because that’s what WAN governance means -- providing the type of control over your business to allow it to continue to be productive, as you're making changes to your WAN.

Simply put, you first of all have to find out what's going on in the network.



Simply put, you first of all have to find out what's going on in the network. You have to understand what's happening on those 4, 5, or 6 different flows that are all going in from different sources to your branch. You have to be able to control those flows and manage them, so that you don't have your edge device or edge router getting congested.

You have to be able to guarantee performance and, very importantly, you also have to then unify, balance, and optimize the performance over those multiple network points that are coming into your branch.

If you're doing it the right way, at least what we would say is the right way, it needs to be dynamic, automatic and, in Ipanema terminology, autonomic, meaning that not only does it happen automatically, but the network learns and manages itself. It doesn’t require extra human intervention.

Schmidt: The way the enterprise is going to get its arms around this increasingly complex environment is not through throwing people at it. Throwing people at network management has never worked and, now that the environment is more complex, it's going to work even less.

The whole point of cloud is that you're going to use virtualization and automation to bring up instances of servers quickly and automatically, and that's where this order of magnitude improvement potential comes from. But, if you don't want the multiple networks to be the bottleneck, then you have to apply automation in that domain as well. That's what we've done. We've made it possible for the network to run itself to meet the businesses’ objectives.

The effect that has in a branch office with multiple network connections is really to hide all the complexity that that multiplicity brings, because the system is managing them all in a unified way. That's what we're getting at when we're talking about network unification. The details that bedeviled traditional management just kind of disappear.

WAN governance is what the CIO wants to buy. CIOs don’t want to buy a WAN, and they certainly don't want to buy WAN optimization controllers. What they want to buy is reliable application performance across their infrastructure with the best possible performance and lowest possible cost. My high-level definition of WAN governance is that it's the technology and techniques that allow the CIO to buy that.

iPhone app

W
e're about to release our first iPhone app to provide an interface into our central management system, and it's terrific. It's exactly the kind of thing the CIO would want to have in their hand. That just shows the value of pushing IT to be democratized and put into the hands of all of the people tied to the enterprise.

The Ipanema system is designed to provide the full control by giving the enterprise IT organization not just visibility in reporting on every user's access to their IT infrastructure, but also to automatically control all of that traffic in accordance with various policies.

Get a free white paper on WAN Governance for Cloud Computing.

Get the free Cloud Networking Report.

We don't see any other way around it. You're not going to do this manually. You've got to build smarter systems. We happen to think that we are a huge piece of that puzzle in terms of how we control things at the network level.

White: And we look at this WAN governance as really a piece of ISO standard for IT governance, which is an official ISO standard. There is a section in there on WAN governance. In a way, it talks about what you have to do to manage your wide area.

Ipanema strongly believes the WAN governance is really a standard that should be put on the books, but isn't yet. If you're really going to have governance over your IT, since the network is a strategic asset to promote enterprise customers, you need to have governance over the wide area as well.

Schmidt: Ipanema has pioneered a unique approach that stems from the idea that all that matters is that end users are able to get good performance from their applications, because that’s when they are most productive. When application performance slows down, end users start surfing the web. So, ensuring the performance of the application is critical. That’s what the enterprise needs to reorient itself toward.

The fundamental input into our system is a list of applications and their performance. The system itself is intelligent enough to monitor and dynamically control all of the traffic to achieve those objectives on behalf of the business. So, it’s imposing the business’s will on the network.

White: It starts with our three founders who got together and took a look at what the needs were from an application perspective. Their goal was to find a way to ensure that, as users, we all had the performance we needed and that enterprises could deliver performance from an application perspective to their users.

That’s what they started out with. Then they took a look at how you would deliver that service and recognized the best way to provide for the delivery of the right type of consistent application performance is to do it over the wide area and to look what happens over the WAN.

They were very visionary in recognizing that application performance over the wide area is going to be the single most critical piece of the puzzle, when it comes to taking a look at how we as users of enterprise deliver service and do it in conjunction with major service providers and network providers, because they are the ones that deliver the wide area connections.

When they started out, they were told that they were wrong and weren't looking at it the right way. When you see what’s happened to the network and how it’s evolved, particularly now that we are moving into the cloud generation, they were focused exactly in the right area. Although we have a lot of new features, the basic architecture has been there for years and it’s been proven in major service provider networks and is installed on a global basis.

The basic architecture has been there for years and it’s been proven in major service provider networks and is installed on a global basis.



Schmidt: There are a couple of things that are the secret sauce, but the easiest one to explain probably is the fact that our appliances actually cooperate with each other, and this is unique. Our appliances know about not just the traffic that’s impinging on their network interfaces, but they actually know about the flows that are active everywhere on the network.

It’s actually not that that simple. They really only need to know about the flows that might conflict with the flows that they are managing. But conceptually, every device on the network knows about all the other flows it needs to know about. They are constantly communicating with each other -- what flows are active and what performance those flows are getting from the infrastructure, which includes the whole WAN, but also the data center and the service. So what does that enable?

Global perspective

Sharing this information means that all of the decisions made by an individual device are made from a global perspective. They're no longer making a local optimization decision. They each run the same algorithm and can come to the same result. And that result is a globally optimum traffic mix on the network.

When I say globally optimum, that’s a valid technical term as opposed to a marketing term, because the information has been collected globally from the entire system. In terms of optimum, what I mean is the best possible performance from the most applications using the given network infrastructure and its status at that point in time.

White: The point I'd like to make is that it's absolutely impossible to measure it in a cloud environment as an enterprise network manager, because you only see a piece of the network. Unless you’ve done something different, which is what we provide, than the way you are going to look at your network, if you are looking at it the way you’ve done for the last 10 or 20 years, there is no way that you can see everything.

The closing point here is that the first step is visibility into the network, and the next step is providing the control. You need to do that in the cloud environment, and that's what Ipanema does.

Get a free white paper on WAN Governance for Cloud Computing.

Get the free Cloud Networking Report.

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Learn more. Sponsor: Ipanema Technologies.

You may also be interested in:

Wednesday, December 8, 2010

The Open Group bridges SOA definitions gap with release of new technical standard SOA Ontology

In an effort to bridge the communication gap between business and IT over service oriented architecture (SOA), The Open Group today released a new ontology that helps better define the concepts, terms, and semantics involved in SOA.

The SOA Ontology Technical Standard bridges different architecture, engineering, business and marketing domains, providing common terminology and concept mapping that business and technical people can use to discuss problems and opportunities.

It will also serve as a foundation for further work in domain specific areas by supplying a consistent framework that can be reused as SOA projects evolve. [Disclosure: The Open Group is a sponsor of BriefingsDirect podcasts.]

“The release of the SOA Ontology will significantly benefit the industry considering the increased use of SOA within organizations, especially due to the rise of cloud adoption,” said Dr. Chris Harding, the Open Group’s SOA Work Group forum director. “It’s critical for business and technical executives across disciplines and organizations to have a lingua franca for SOA to ensure the success of their deployments. As with all of the concepts and models developed through the SOA work group, we anticipate the ontology to be a living document that will be updated as the industry evolves and SOA concepts are further refined.”

Foundational model

In fact, the SOA Ontology Technical Standard has already been used as the foundational business model for the SOA Repository Artifact Model and Protocol (S-RAMP), a proposed standard for repositories and their use in SOA environments to publish, search for, and retrieve a wide variety of technical documents.

SOA discussion between IT professionals and business people, and among IT professionals themselves, often bog down because of a lack of clarity in the terms and concepts used. The ontology is designed to focus the discussions by providing a common framework where everyone is clear on the terms of the discussion.

The release of the SOA Ontology will significantly benefit the industry considering the increased use of SOA within organizations, especially due to the rise of cloud adoption.



The confusion started early, according to Claus Jensen, IBM’s Senior Technical Staff Member & Chief Architect for SOA-BPM-EA Technical Strategy. "In most cases, the business side is confused, because nobody gave them a definition," he said. "Over time, people brought in slightly different definitions and conflicting messages. Taking out the uncertainty is never a bad thing."

The SOA Ontology is designed for use by:
  • Business people, to give them a deeper understanding of SOA concepts and how they are used in the enterprise and its environment
  • Architects, as metadata for architectural artifacts
  • Architecture methodologists, as a component of SOA meta-models
  • System and software designers, for guidance in terminology and structure
The SOA Ontology is also intended for use in conjunction with other industry standards and can also be used by computing systems to create modeling tools, automate standard terms and relationships, clarify working assumptions, and enable interoperability.

The SOA Ontology Technical Standard is available for free and may be downloaded from the Open Group website.

You may also be interested in:

Jonathan Priestley recaps the news and events at HP's Software Universe 2010 in Barcelona

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Barcelona -- Welcome to a special BriefingsDirect podcast from the HP Software Universe 2010 Conference in Barcelona, an interview with Jonathan Priestley, Director of Marketing for HP Software and Solutions, Europe, Middle-East, and Africa (EMEA).

We were at the conference last week to explore some major enterprise software and solutions, trends and innovations, making news across HP’s ecosystem of customers, partners, and developers. [Disclosure: HP is a sponsor of BriefingsDirect podcasts.]

Dana Gardner, Principal Analyst at Interarbor Solutions, moderated the recap discussion coming on the heels of the roll-out of HP’s application lifecycle management (ALM) news, the release of ALM 11. [See more on HP's new ALM 11 offerings.]

Here are some excerpts:
Priestley: HP ALM 11 is the culmination of two years worth of engineering time. This is a massive innovation for the marketplace and for us.

For the first time, you have something that goes not just across the traditional software lifecycle, as we know it -- develop, test, deploy, and manage -- but across the entire lifecycle of an application.

This includes one of those things that we don’t like to talk about in IT, which is retirement, when we actually get rid of something, because we are not real good at that in this industry. Of course, it is a key part as we go forward. One of the fundamental reasons organizations can’t reduce their data or application load is because they never get to the point of being able to retire anything.

One of the key benefits of ALM is that we’ve unified a complete view, a single view, of that whole process.



One of the key benefits of ALM is that we’ve unified a complete view, a single view, of that whole process. Now, the people who are responsible for each of those siloed areas -- the business analysts, developers, testers, and operational people -- who have to support and maintain the applications can actually see, from one place, how each of those areas hand off to each other to ensure that they can manage it effectively.

And, of course, the final piece is how I make sure I’ve got security engineered into the application before I deploy it. That whole piece fits together under ALM.

So we have a really good story that ties together the history of our organically built technology, as well as the acquisitions that we’ve made. We’ve constructed this around those areas that we see our customers working with in their IT organizations today, as they deliver the solutions that drive their businesses. We see that breaking out into these areas.

Supporting services

Build -- how you actually put the applications that support the services out there. Operate -- how you supply the infrastructure that supports those applications. Secure -- how you make sure that you’ve got the right kind of security in place. Store -- how you make sure that this enormous information explosion we're all dealing with today is capable and managed. Finally, analyze -- how you take the business data that you're collecting in real-time and get decision-making data and information out of it.

Let me tell you about main stage events, because we had some really good presenters on main stage this year. We began with Anton Knolmar.

He started out with a traditional flamenco dance, and you really had to be there to see that, in all his glory, as he came out with a rose between his teeth. It was a super start to the event and very appropriate here in Barcelona.

We moved on to Robin Purohit, the Vice President and General Manager of the R&D organization within software, to take us through the new strategy and vision for HP software. This was a really big one for us, because it’s the first time that the whole strategy and vision has had a public airing. It was exciting to see customers’ reaction to that.

We followed Robin with Jonathan Rende, Vice President and General Manager of our Application Solutions. This was really the news point for our main stage, the launch of Application Lifecycle Management 11 (ALM 11). This is probably the biggest news point we’ve had in the last six months or so. And we saved it especially for the show.

We were furthermore celebrating our 10-year anniversary of Software Universe by showing just how the HP software solutions can help with ALM, of course -- that was our focus for the show -- but also how to use the solutions in a cloud environment, which is also a big message we're trying to get across today.

We were celebrating our ten-year anniversary by showing just how the HP software solutions can help with ALM, but also how to use the solutions in a cloud environment.



Ulrich Pfeiffer, our CTO in EMEA, who sets up the whole main stage event, pulled together pre-sales people from across the region. You could hear all the different accents and nationalities working together, and it was the perfect example of what our customer organizations face in working with multi nationalities across the geography.

If we go back to the strategy and vision, we talked about that in terms of the areas in which we see the IT people within an organization working. But, if you take that up a level and think about the challenges that these large enterprising business and government are facing, we see that breaking down into five core areas, something we call Converged Infrastructure. This is all of the elements of the infrastructure working together.

Working together

Enterprise security -- not just thinking about the individual components, but looking at it from across the entire delivery mechanism.

Application transformation -- something that all our customers are telling us they are facing today, and obviously, that’s one of the key areas where ALM fits in.

Information optimization -- another key, when we talk about the information explosion and the challenges with that. This ensures that you can manage not just data, but the information you want to get out of that hybrid delivery.

It addresses probably the hottest topic in the industry, if I want to put my enterprise resources on premise, running locally, versus putting it somewhere in a cloud, which I am running privately, or putting it in a public cloud. Our expectation is that all large enterprises will be facing those kinds of sourcing decisions and we call that "hybrid delivery."

We also put a lot of effort into a new executive program, and I am calling it a program, versus a track, this year because instead of thinking as a single standalone event, we're thinking of it as a program that runs across the entire year. The launch point has been here at Software Universe.

We started out by taking an agenda from what our CIOs have told us are the key things that they are looking to try and get some help with.

The first is innovation, how they actually put innovation into their own organizations and how can they remove the barriers to that innovation. This is a key one, because it’s very difficult for them to figure out how to put innovation into their own organizations, but particularly how to remove the barriers to innovation. We’ve done a lot of work around showing them how to go through a discovery exercise to look for those opportunities.

The second thing is that the CIOs themselves are always measuring themselves against their peer groups, and that’s something else we’ve helped them with today. We're bringing in some expertise from outside to look at what are the skill-sets that make the perfect CIO, because we see that role changing.

We talked a bit about hybrid delivery and sourcing options. The actual skill-set roles of a CIO could very well change, and that’s one of the things we’ve tried to explore with them today.

CIO annual report

A
nd as we go across the rest of the year, we'll be looking at something like an annual report. Every company or business today delivers an annual report on how they perform. CIOs are looking for the same kind of the thing themselves. How do they produce an annual report that shows how IT is performing and what kind of service it’s delivering back to the business? That’s something we will be taking that journey with them through the whole year.

So we're really excited about that program, and we’ve had some great feedback from the CIOs, that this is really hitting the hot spots for them, particularly on innovation and how they plan themselves going forward.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in:

Energy-efficient Ethernet switch modules from HP help cut power use and TCO

HP today announced the availability of new energy-efficient Ethernet switches, designed to help companies deal with the data explosion, while, at the same time, cutting power needs and reducing the total cost of ownership (TCO) by some 51 percent.

The company says that it's E-Series zl modules are the first devices that conform to the IEEE Energy Efficient Ethernet (EEE) standard 802.3az. The switches reduce power consumption by automatically entering sleep mode when no traffic is being transmitted. [Disclosure: HP is a sponsor of BriefingsDirect podcasts.]

Many companies today are caught between competing pressures brought by data and power. Faced with moving ever increasing data faster and faster, they find the power to do that, as well as deal with the associated cooling, expensive and increasingly unavailable.

Sleep mode

The modules, which fit into existing switches, address this by going into sleep mode, as will EEE-connected devices in the absence of traffic. Because most network traffic comes in bursts, these devices are good candidates for energy efficiency.

The modules also address another concern, which is latency in the network. According to Jay Mellman, Director of Network Product and Solutions Marketing, the modules function with an instant-on capability and will wake up immediately upon detecting traffic, eliminating latency.

In a study conducted by the Tolly Group, power and cooling savings from the energy-efficient modules ranged from 17 to 26 percent, when compared to traditional switches. The lower operating costs, combined with lower acquisition costs resulted in a decrease of up to 51 percent in TCO.

HP assisted in developing the EEE standard and the company plans to incorporate it across multiple devices. including servers, laptops, and wireless devices.

You may also be interested in:

Tuesday, December 7, 2010

Salesforce revs up the cloud data tier era with Database.com

San Francisco -- Salesforce.com at its Dreamforce conference here today debuted a database in the cloud service, Database.com, that combines attractive heterogeneous features for a virtual data tier for developers of all commercial, technical and open source persuasions.

Salesforce.com is upping the ante on cloud storage services by going far beyond the plain vanilla elastic, pay-as-you-go variety of database and storage services that have come to the market in the past few years, with Amazon Web Services as a leading offering. [Disclosure: Salesforce.com helped defray the majority of travel costs for me to attend Dreamforce this week.]

Database.com offers as a service a cross-language, cross-platform, elastic pricing data tier that should be smelling sweet to developers and -- potentially -- enterprise architects. Salesforce, taking a page from traditional database suppliers like Oracle, IBM, Microsoft and SAP/Sybase, recognizes that owning the data tier -- regardless of where it is -- means owning a long-term keystone to IT.

If the new data tier in the cloud service is popular, it could disrupt not only the traditional relational database market, but also the development/PaaS market, the Infrastructure as a Service market, and the middleware/integration markets.

Even more fascinating is the prospect of Database.com becoming a new data services resource darling of open source developers, just as they are losing patience and interest in MySQL, now under the stewardship of Oracle since it bought Sun Microsystems. This is a core constituency that is in flux and is being courted assiduously by Amazon, VMWare, Google and others.

IBM, Microsoft and Oracle will need to respond to Database.com -- as will Google, Amazon and VMware -- first for open source, mobile and start-up developers and later -- to an yet uncertain degree -- enterprise developers, systems integrators and more conservative ISVs.

So, be sure, the race is on generally to try and provide the best cloud data and increasingly integrated PaaS services at the best cost that proves its mettle in terms of performance, security, reliability and ease of use. If recent cloud interest and adoption are any indication, this could be killer cloud capability that becomes a killer IT capability.

Database.com already raises the stakes for cloud storage providers and wannabes: the value-add to the plain vanilla storage service has to now entice developers with meeting or exceeding Salesforce.com. By catering to a wide swath of tools, frameworks, IT platforms and mobile device platforms, Salesforce.com is heading off the traditional vendors at trying to (not too fast) usher their installed bases to that own vendors brand of hybrid and cloud offerings. Think a lock-in on the ground segue to a lock-in in the sky slick trick.

Both Amazon and Salesforce.com have proven that developers are not timid about changing how they attain value and resources. This may well prove true of how to access and procure data tier services, too, which makes the vendors slick cloud segue trick all the more tricky. Instead of going to the DBA for data services, all stripes of developers could just as easily (maybe more easily) fire up a value-added Database.com instance and support their apps fast and furious.

The stakes are high on attracting the developers, of course, because the more data that Salesforce attracts with Database.com, the more integration and analytics they can offer -- which then attracts even more data and applications -- and developer allegiance -- and so on and so on. It's a value-add assemblage activity that Salesforce has already shown aptness with Force.com.

What remains to be seen is if this all vaults Salesforce.com beyond it roots as a CRM business applications SaaS provider and emerging PaaS ecosystem supporter for good. If owning the cloud data tier proves as instrumental to business success (as evidenced by Oracle's consistence in generating envious profit margins) as the on-premises, distributed computing DB business -- well, Salesorce.com is looking at a massive business opportunity. And, like the Internet in general, it can easily become an early winner takes all affair.

It's now a race for scale and value, that cloud-based data accumulation can become super sticky, with the lock-in coming from inescapably attractive benefits, not technical or license lock-in. Can you say insanely good data services? Sure you can. But more easily said than done. Safesforce has to execute well and long on its audacious new offering.

But what if? Like a rolling mountainside snowball gathering mass, velocity and power, Database.com could quickly become more than formidable because of the new nature of data in the cloud. Because the more data from more applications amid and among more symbiotic and collaborative ecosystems, the more insights, analytics and instant marketing prowess the hive managers (and hopefully users) gain.

In the end, Database.com could become a pervasive business intelligence services engine, something that's far more valuable than a cheap and easy data tier in the sky for hire.

Friday, December 3, 2010

Case study: AIG insurance group leverages ALM to attain IT performance architecture advantage

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Barcelona -- Welcome to a special BriefingsDirect podcast series coming to you from the HP Software Universe 2010 Conference in Barcelona.

We're here in the week of November 29, 2010 to explore some major enterprise software and solutions, trends and innovations making news across HP’s ecosystem of customers, partners, and developers. [See more on HP's new ALM 11 offerings.]

This customer case-study from the conference focuses on AIG-Chartis insurance and how their business has benefited from ongoing application transformation and modernization projects.

To learn more about AIG-Chartis insurance’s innovative use of IT consolidation and application lifecycle management (ALM) best practices, I interviewed Abe Naguib, Director of Global Performance Architecture and Infrastructure Engineering at AIG-Chartis in Jersey City, NJ.

Here are some excerpts:
Naguib: AIG is a global insurance firm, supporting worldwide international insurance of different varieties.

We're structured with 1,500 companies and roughly about eight lines of businesses that manage those companies. Each group has their own CIO, CTO, COO structure, and I report to the global CTO.

What we look at is supporting their global architecture and performance behavioristics, if you will. One of the key things is how to federate the enterprise in terms of architecture and performance, so that we can standardize the swing over into the Java world, as well as middleware and economy of scale.

When I came on board to standardize architecture, and I saw there was a proliferation of various middleware technologies. As we started going along, we thought about how to standardize that architecture.

As we faced more and more applications coming into the Java middleware world, we found that there’s a lot of footprint waste and there’s a lot of delivery cycles that are also slipped and wasted. So, we saw a need to control it.

After we started the architectural world, we also started the production support world and a facility for testing these environments. We started realizing, again, there were things that impacted business service level agreements (SLAs), economy of scale, even branding. So, we asked, how do we put it together?

One of the key things is, as we started the organizational performance, we were part of QA, but then we realized that we had to change our business strategy, and we thought about how to do that. One key thing is we changed our mindset from a performance testing practice to a performance engineering practice, and we've evolved now to performance architecture.

The engineering practice was focused on testing and analyzing, providing some kind of metrics. But, the performance architectural world now has influence into strategies, design practices, and resolution issues. We're currently a one-man or one-army team, kind of a paratrooper level. We're multi-skilled, from architecture, to performance, to support, and we drive resolution in the organization.

We also saw that resolution had to happen quickly and effectively. Carnegie Mellon did a study about five years ago and it said that post-live application resolution of performance issues was seven times the cost of pre-live [performance application resolution].

In other words, we realized that the faster we resolved issues, the faster to market, the faster we can address things, the less disruption to the delivery practices.

Too many people involved

In normal firefighting mode, architecture is involved, development is involved, and infrastructure is involved. What ends up happening is there are too many people involved. We're all scrambling, pointing fingers, looking at logs. So, we figured that the faster we get to resolution, the better for everyone to continue the train on the track.

... I have experience with Quality Center and the improvements that have gone on over the years. Because of our focus, we built our paradigm out of QA and into the performance world, and we started focusing on improving that process.

The latest TruClient product, which is a LoadRunner product, has been a massive groundbreaking point solution. In the last two years, frankly, with HP and Mercury getting adjusted, there’s been kind of a lag, but I have to give kudos to the team. [See more on HP's new ALM 11 offerings.]

One of the key things is that they have opened up their doors in terms of the delivery, in terms of their roadmap. I've worked extensively for the last roughly year with their product development team, and they have done quite a bit of improvement in their solution.

Good partnership role

They have also improved their service support model; the help desk actually resolves questions a lot faster. And we also have a good partnership role, and we actually work with things that we see, and to the influence of their roadmap as well.

This TruClient product has been phenomenal. One of the key things we're seeing now is BPM solutions are more Ajax-based, and there are so many varieties of Ajax frameworks out there than we know how to deal with. One of the key things with the partnership is that we're able to target what we need, they are able to deliver, and we are able to execute.

LoadRunner and TruClient allow us to get in front of the console, work with the business team, capture their typical use cases in a day-in-the-life scenario, and automate that. That gets buy-in and partnership with the business.

We're also able to execute a test case now and bring that in front of the IT side and show them the actual footprint from a business perspective and the impact and the benefits. What ends up happening is that now we're bringing the two teams together. So, we're bridging the gap basically from execution.

... We also started working with the CIOs to figure out a strategy to develop a service-level target, if you will. As we went along, we began working with the development teams to build a relationship with the architectural teams and the infrastructure teams.

We became more of a team model, building more of a peace-maker model. We regrouped the organization, so that rather than resolve and point fingers at each other, we resolved issues a lot faster.

Now, we're able to address the issue. We call it "isolate, identify, and resolve." At that point, if it’s a database issue, we work directly with the DBA. If it’s an infrastructure or architecture issue, we work directly with that group. We basically cut the cycle down in the last two or three years by about 70 percent.

A lot more CIOs have started bringing in more applications. We see a trend growth internally of roughly about 20-30 percent every year.



Because there is a change in our philosophy, in our strategy to focus more on business value, a lot more CIOs have started bringing in more applications. We see a trend growth internally of roughly about 20-30 percent every year.

I have a staff of nine. So, it’s a very agile, focused team, and they're very delivery-conscious. They're very business value-conscious, and we translate our data, the metrics that we capture, into business KPIs and infrastructure KPIs.

Because of that metric, the CIOs love what we do, because we make them look good with the business, which helps foster the relationship with the business, which helps them justify transformation in the future.

There is a new paradigm now, they call it the "Escalator Message." In 60 seconds or less, we can talk to a CIO, CTO, COO, or CFO about our strategy and how we can help them shift from the firefighting mode to more of an architecture mode.

If that’s the case, the more they can salvage their delivery, the more they can salvage their effective costs, and the more they can now shift to more of an IT-sensitive solutions shop. That helps build a business relationship and helps improve their economy of scale.

I would definitely send the message out to think in business value. Frankly, nobody really cares as much about the footprint cost, until they start realizing the dollars that are spent.

Also, now, business wants to see us more involved from the IT side, in terms of solutions, top-line improvements, and bottom-line improvements. As the performance teams expand and mature and we have the right toolsets, innovative toolsets like TruClient, we're able to now shift the cost of waste into a cost of improvements, and that’s been a huge factor in the last couple of years.

Last, I would say that in 8,000+ engagements -- we're actually closing in on now 10,000 events this year -- we've seen roughly $127 million in infrastructure savings that we have recouped. Again, that helps to benefit the firm. Instead of waste, now we're able to leverage that into more improvement side.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in:

Case study: Enel Green Power uses PPM to gain visibility, orchestrate myriad energy activities across 16 countries

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Barcelona -- Welcome to a special BriefingsDirect podcast series coming to you from the HP Software Universe 2010 Conference in Barcelona.

We're here in the week of November 29, 2010 to explore some major enterprise software and solutions, trends and innovations making news across HP’s ecosystem of customers, partners, and developers. [See more on HP's new ALM 11 offerings.]

This customer case-study from the conference focuses on Enel Green Power and how their Italian utility business has benefited from improved management of core business processes and gained visibility into new energy projects, also adhering to compliance through better planning and the ability to scope out new projects comprehensively. [Disclosure: HP is a sponsor of BriefingsDirect podcasts.]

To learn about Enel Green Power’s innovative use of project and portfolio management (PPM), I interviewed Massimo Ferriani, CIO of Enel Green Power in Rome.

Here are some excerpts:
Ferriani: Enel Green Power is one of the leaders in the renewables market ... We're in all the most mature technologies such as hydro, geothermal, wind, and solar.

If you think about a matrix to cross technologies and countries, we have a lot of trouble, because we operate four technologies in 16 countries.

It's difficult because we have more than 300 plants all around the world. So, it's an asset portfolio that we have to operate, and we have to reduce the risks.

When we decided to deploy IT platforms, we didn’t think that it was a good idea to deploy conventional-generation IT platforms, but to build up new platforms more fitted to renewables' needs.

We thought about the main objective in deploying these platforms and said, "Okay, maybe we have to deploy platforms that permit us to minimize the portfolio risk, in order to know exactly what production should be." For us, knowing the production is a condition.

We have to know production, and we have to know exactly the production that we're promising to sell to the market.

The business strategy is to manage centrally and operate locally. IT had to follow the strategy. Our main IT platforms are developed with the objective to be global. Global doesn’t mean managing everything centralized, but to manage the IT platform as centralized, because it's better for synergies and in terms of costs. But, because we have to fit local needs, we we have to localize these platforms in 16 countries.

For PPM, as well, we decided to have a global, centralized, unique platform, in order to gather and collect all the data that we get from the field. This is one of the problems that we frequently have because, in effect, the operation is located everywhere. And, it’s not easy to collect information from each field operation.

It’s important to have global IT platforms, because one of our main objectives is that all our people have to work in the same way.



We have lot of plants in the middle of nowhere -- in the middle of the Nevada desert and in the middle of the Mato Grosso in Brazil. We have to gather information from these plants. So, it’s important to have global IT platforms, because one of our main objectives is that all our people have to work in the same way.

It’s also important to set the main goal of the PPM solution. Now, the PPM solution lets Enel Green Power manage its own worldwide portfolio initiatives, both business development side and the plant construction Phase 2, because we have to remember that the business development hands over the construction of the project.

We have to do it through building a unique centralized integrated platform, valuable to all the countries, designed to certify the market value of the pipeline and the potential future production related to that pipeline. For us, it's absolutely important to forecast better, to make budgets, and so on. It had to be designed to support people, our colleagues, in activities like planning, project development, reporting, document management, and so on.

Setting the main goals

So when we decided to deploy this platform, we had a lot of work for a couple of reasons.

First of all, because we wanted to develop an integrated in-house platform in order to map the ... core processes of the project, and at the same time to implement algorithms to develop a portfolio evaluation.

The second was to investigate adopting a standard solution available on the market that allowed us, with little customization, to fit the need of the business. It's important to underline that, when we started this project, it was the end of May, 2010. We already knew we were going to have an IPO. We didn’t know the time exactly, but we had to be ready for the end of October, the estimated date of the IPO.

We adopted the HP solution, because the HP people convinced us that with a minimal set of customization we would be ready for the end of October -- and we did it.

We chose HP because of the ... strong automation in the collection of the data. As I said before, also important for us were simplicity and flexibility. Also, with reference to our geographical distribution everywhere, the adoption of a solution supported with global support was another constraint and was absolutely important.

We needed a standard technology accessible from a lot of countries and with integration with other applications that we have, for example Microsoft Project. We also required scalability and platform growth -- and HP has a strength on this point -- because we are adopting a web service architecture. And, we wanted the viability of a unique homogenous view of mandating KPIs.

For us, the flexibility was one of the three main strengths on this platform and the reasons we chose HP.


We're only in the first phase in order to support the IPO and to support the certification of the market value of the pipeline. But, the main benefits of this platform for the business are acquisition and centralization of the data.

For us, the flexibility was maybe one of the three main strengths on this platform and the reasons we chose HP. But, the best one, as I said before, was the minimum customization we needed in order to fit the first phase. It’s not easy to have only three months time to set 64 workflows, because the local business development wants to fit their workflow on these needs.

It’s important for the automation to monitor all the steps of the workflow, of the individual steps of the process, to manage the workflow authorization of the individual steps, and monitor progress of the individual steps. All these data have to support us in order to plan the strategy. So, there are plenty of benefits and maybe more benefits in the future with the evolution of this platform.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in: