Wednesday, December 8, 2010

The Open Group bridges SOA definitions gap with release of new technical standard SOA Ontology

In an effort to bridge the communication gap between business and IT over service oriented architecture (SOA), The Open Group today released a new ontology that helps better define the concepts, terms, and semantics involved in SOA.

The SOA Ontology Technical Standard bridges different architecture, engineering, business and marketing domains, providing common terminology and concept mapping that business and technical people can use to discuss problems and opportunities.

It will also serve as a foundation for further work in domain specific areas by supplying a consistent framework that can be reused as SOA projects evolve. [Disclosure: The Open Group is a sponsor of BriefingsDirect podcasts.]

“The release of the SOA Ontology will significantly benefit the industry considering the increased use of SOA within organizations, especially due to the rise of cloud adoption,” said Dr. Chris Harding, the Open Group’s SOA Work Group forum director. “It’s critical for business and technical executives across disciplines and organizations to have a lingua franca for SOA to ensure the success of their deployments. As with all of the concepts and models developed through the SOA work group, we anticipate the ontology to be a living document that will be updated as the industry evolves and SOA concepts are further refined.”

Foundational model

In fact, the SOA Ontology Technical Standard has already been used as the foundational business model for the SOA Repository Artifact Model and Protocol (S-RAMP), a proposed standard for repositories and their use in SOA environments to publish, search for, and retrieve a wide variety of technical documents.

SOA discussion between IT professionals and business people, and among IT professionals themselves, often bog down because of a lack of clarity in the terms and concepts used. The ontology is designed to focus the discussions by providing a common framework where everyone is clear on the terms of the discussion.

The release of the SOA Ontology will significantly benefit the industry considering the increased use of SOA within organizations, especially due to the rise of cloud adoption.



The confusion started early, according to Claus Jensen, IBM’s Senior Technical Staff Member & Chief Architect for SOA-BPM-EA Technical Strategy. "In most cases, the business side is confused, because nobody gave them a definition," he said. "Over time, people brought in slightly different definitions and conflicting messages. Taking out the uncertainty is never a bad thing."

The SOA Ontology is designed for use by:
  • Business people, to give them a deeper understanding of SOA concepts and how they are used in the enterprise and its environment
  • Architects, as metadata for architectural artifacts
  • Architecture methodologists, as a component of SOA meta-models
  • System and software designers, for guidance in terminology and structure
The SOA Ontology is also intended for use in conjunction with other industry standards and can also be used by computing systems to create modeling tools, automate standard terms and relationships, clarify working assumptions, and enable interoperability.

The SOA Ontology Technical Standard is available for free and may be downloaded from the Open Group website.

You may also be interested in:

Jonathan Priestley recaps the news and events at HP's Software Universe 2010 in Barcelona

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Barcelona -- Welcome to a special BriefingsDirect podcast from the HP Software Universe 2010 Conference in Barcelona, an interview with Jonathan Priestley, Director of Marketing for HP Software and Solutions, Europe, Middle-East, and Africa (EMEA).

We were at the conference last week to explore some major enterprise software and solutions, trends and innovations, making news across HP’s ecosystem of customers, partners, and developers. [Disclosure: HP is a sponsor of BriefingsDirect podcasts.]

Dana Gardner, Principal Analyst at Interarbor Solutions, moderated the recap discussion coming on the heels of the roll-out of HP’s application lifecycle management (ALM) news, the release of ALM 11. [See more on HP's new ALM 11 offerings.]

Here are some excerpts:
Priestley: HP ALM 11 is the culmination of two years worth of engineering time. This is a massive innovation for the marketplace and for us.

For the first time, you have something that goes not just across the traditional software lifecycle, as we know it -- develop, test, deploy, and manage -- but across the entire lifecycle of an application.

This includes one of those things that we don’t like to talk about in IT, which is retirement, when we actually get rid of something, because we are not real good at that in this industry. Of course, it is a key part as we go forward. One of the fundamental reasons organizations can’t reduce their data or application load is because they never get to the point of being able to retire anything.

One of the key benefits of ALM is that we’ve unified a complete view, a single view, of that whole process.



One of the key benefits of ALM is that we’ve unified a complete view, a single view, of that whole process. Now, the people who are responsible for each of those siloed areas -- the business analysts, developers, testers, and operational people -- who have to support and maintain the applications can actually see, from one place, how each of those areas hand off to each other to ensure that they can manage it effectively.

And, of course, the final piece is how I make sure I’ve got security engineered into the application before I deploy it. That whole piece fits together under ALM.

So we have a really good story that ties together the history of our organically built technology, as well as the acquisitions that we’ve made. We’ve constructed this around those areas that we see our customers working with in their IT organizations today, as they deliver the solutions that drive their businesses. We see that breaking out into these areas.

Supporting services

Build -- how you actually put the applications that support the services out there. Operate -- how you supply the infrastructure that supports those applications. Secure -- how you make sure that you’ve got the right kind of security in place. Store -- how you make sure that this enormous information explosion we're all dealing with today is capable and managed. Finally, analyze -- how you take the business data that you're collecting in real-time and get decision-making data and information out of it.

Let me tell you about main stage events, because we had some really good presenters on main stage this year. We began with Anton Knolmar.

He started out with a traditional flamenco dance, and you really had to be there to see that, in all his glory, as he came out with a rose between his teeth. It was a super start to the event and very appropriate here in Barcelona.

We moved on to Robin Purohit, the Vice President and General Manager of the R&D organization within software, to take us through the new strategy and vision for HP software. This was a really big one for us, because it’s the first time that the whole strategy and vision has had a public airing. It was exciting to see customers’ reaction to that.

We followed Robin with Jonathan Rende, Vice President and General Manager of our Application Solutions. This was really the news point for our main stage, the launch of Application Lifecycle Management 11 (ALM 11). This is probably the biggest news point we’ve had in the last six months or so. And we saved it especially for the show.

We were furthermore celebrating our 10-year anniversary of Software Universe by showing just how the HP software solutions can help with ALM, of course -- that was our focus for the show -- but also how to use the solutions in a cloud environment, which is also a big message we're trying to get across today.

We were celebrating our ten-year anniversary by showing just how the HP software solutions can help with ALM, but also how to use the solutions in a cloud environment.



Ulrich Pfeiffer, our CTO in EMEA, who sets up the whole main stage event, pulled together pre-sales people from across the region. You could hear all the different accents and nationalities working together, and it was the perfect example of what our customer organizations face in working with multi nationalities across the geography.

If we go back to the strategy and vision, we talked about that in terms of the areas in which we see the IT people within an organization working. But, if you take that up a level and think about the challenges that these large enterprising business and government are facing, we see that breaking down into five core areas, something we call Converged Infrastructure. This is all of the elements of the infrastructure working together.

Working together

Enterprise security -- not just thinking about the individual components, but looking at it from across the entire delivery mechanism.

Application transformation -- something that all our customers are telling us they are facing today, and obviously, that’s one of the key areas where ALM fits in.

Information optimization -- another key, when we talk about the information explosion and the challenges with that. This ensures that you can manage not just data, but the information you want to get out of that hybrid delivery.

It addresses probably the hottest topic in the industry, if I want to put my enterprise resources on premise, running locally, versus putting it somewhere in a cloud, which I am running privately, or putting it in a public cloud. Our expectation is that all large enterprises will be facing those kinds of sourcing decisions and we call that "hybrid delivery."

We also put a lot of effort into a new executive program, and I am calling it a program, versus a track, this year because instead of thinking as a single standalone event, we're thinking of it as a program that runs across the entire year. The launch point has been here at Software Universe.

We started out by taking an agenda from what our CIOs have told us are the key things that they are looking to try and get some help with.

The first is innovation, how they actually put innovation into their own organizations and how can they remove the barriers to that innovation. This is a key one, because it’s very difficult for them to figure out how to put innovation into their own organizations, but particularly how to remove the barriers to innovation. We’ve done a lot of work around showing them how to go through a discovery exercise to look for those opportunities.

The second thing is that the CIOs themselves are always measuring themselves against their peer groups, and that’s something else we’ve helped them with today. We're bringing in some expertise from outside to look at what are the skill-sets that make the perfect CIO, because we see that role changing.

We talked a bit about hybrid delivery and sourcing options. The actual skill-set roles of a CIO could very well change, and that’s one of the things we’ve tried to explore with them today.

CIO annual report

A
nd as we go across the rest of the year, we'll be looking at something like an annual report. Every company or business today delivers an annual report on how they perform. CIOs are looking for the same kind of the thing themselves. How do they produce an annual report that shows how IT is performing and what kind of service it’s delivering back to the business? That’s something we will be taking that journey with them through the whole year.

So we're really excited about that program, and we’ve had some great feedback from the CIOs, that this is really hitting the hot spots for them, particularly on innovation and how they plan themselves going forward.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in:

Energy-efficient Ethernet switch modules from HP help cut power use and TCO

HP today announced the availability of new energy-efficient Ethernet switches, designed to help companies deal with the data explosion, while, at the same time, cutting power needs and reducing the total cost of ownership (TCO) by some 51 percent.

The company says that it's E-Series zl modules are the first devices that conform to the IEEE Energy Efficient Ethernet (EEE) standard 802.3az. The switches reduce power consumption by automatically entering sleep mode when no traffic is being transmitted. [Disclosure: HP is a sponsor of BriefingsDirect podcasts.]

Many companies today are caught between competing pressures brought by data and power. Faced with moving ever increasing data faster and faster, they find the power to do that, as well as deal with the associated cooling, expensive and increasingly unavailable.

Sleep mode

The modules, which fit into existing switches, address this by going into sleep mode, as will EEE-connected devices in the absence of traffic. Because most network traffic comes in bursts, these devices are good candidates for energy efficiency.

The modules also address another concern, which is latency in the network. According to Jay Mellman, Director of Network Product and Solutions Marketing, the modules function with an instant-on capability and will wake up immediately upon detecting traffic, eliminating latency.

In a study conducted by the Tolly Group, power and cooling savings from the energy-efficient modules ranged from 17 to 26 percent, when compared to traditional switches. The lower operating costs, combined with lower acquisition costs resulted in a decrease of up to 51 percent in TCO.

HP assisted in developing the EEE standard and the company plans to incorporate it across multiple devices. including servers, laptops, and wireless devices.

You may also be interested in:

Tuesday, December 7, 2010

Salesforce revs up the cloud data tier era with Database.com

San Francisco -- Salesforce.com at its Dreamforce conference here today debuted a database in the cloud service, Database.com, that combines attractive heterogeneous features for a virtual data tier for developers of all commercial, technical and open source persuasions.

Salesforce.com is upping the ante on cloud storage services by going far beyond the plain vanilla elastic, pay-as-you-go variety of database and storage services that have come to the market in the past few years, with Amazon Web Services as a leading offering. [Disclosure: Salesforce.com helped defray the majority of travel costs for me to attend Dreamforce this week.]

Database.com offers as a service a cross-language, cross-platform, elastic pricing data tier that should be smelling sweet to developers and -- potentially -- enterprise architects. Salesforce, taking a page from traditional database suppliers like Oracle, IBM, Microsoft and SAP/Sybase, recognizes that owning the data tier -- regardless of where it is -- means owning a long-term keystone to IT.

If the new data tier in the cloud service is popular, it could disrupt not only the traditional relational database market, but also the development/PaaS market, the Infrastructure as a Service market, and the middleware/integration markets.

Even more fascinating is the prospect of Database.com becoming a new data services resource darling of open source developers, just as they are losing patience and interest in MySQL, now under the stewardship of Oracle since it bought Sun Microsystems. This is a core constituency that is in flux and is being courted assiduously by Amazon, VMWare, Google and others.

IBM, Microsoft and Oracle will need to respond to Database.com -- as will Google, Amazon and VMware -- first for open source, mobile and start-up developers and later -- to an yet uncertain degree -- enterprise developers, systems integrators and more conservative ISVs.

So, be sure, the race is on generally to try and provide the best cloud data and increasingly integrated PaaS services at the best cost that proves its mettle in terms of performance, security, reliability and ease of use. If recent cloud interest and adoption are any indication, this could be killer cloud capability that becomes a killer IT capability.

Database.com already raises the stakes for cloud storage providers and wannabes: the value-add to the plain vanilla storage service has to now entice developers with meeting or exceeding Salesforce.com. By catering to a wide swath of tools, frameworks, IT platforms and mobile device platforms, Salesforce.com is heading off the traditional vendors at trying to (not too fast) usher their installed bases to that own vendors brand of hybrid and cloud offerings. Think a lock-in on the ground segue to a lock-in in the sky slick trick.

Both Amazon and Salesforce.com have proven that developers are not timid about changing how they attain value and resources. This may well prove true of how to access and procure data tier services, too, which makes the vendors slick cloud segue trick all the more tricky. Instead of going to the DBA for data services, all stripes of developers could just as easily (maybe more easily) fire up a value-added Database.com instance and support their apps fast and furious.

The stakes are high on attracting the developers, of course, because the more data that Salesforce attracts with Database.com, the more integration and analytics they can offer -- which then attracts even more data and applications -- and developer allegiance -- and so on and so on. It's a value-add assemblage activity that Salesforce has already shown aptness with Force.com.

What remains to be seen is if this all vaults Salesforce.com beyond it roots as a CRM business applications SaaS provider and emerging PaaS ecosystem supporter for good. If owning the cloud data tier proves as instrumental to business success (as evidenced by Oracle's consistence in generating envious profit margins) as the on-premises, distributed computing DB business -- well, Salesorce.com is looking at a massive business opportunity. And, like the Internet in general, it can easily become an early winner takes all affair.

It's now a race for scale and value, that cloud-based data accumulation can become super sticky, with the lock-in coming from inescapably attractive benefits, not technical or license lock-in. Can you say insanely good data services? Sure you can. But more easily said than done. Safesforce has to execute well and long on its audacious new offering.

But what if? Like a rolling mountainside snowball gathering mass, velocity and power, Database.com could quickly become more than formidable because of the new nature of data in the cloud. Because the more data from more applications amid and among more symbiotic and collaborative ecosystems, the more insights, analytics and instant marketing prowess the hive managers (and hopefully users) gain.

In the end, Database.com could become a pervasive business intelligence services engine, something that's far more valuable than a cheap and easy data tier in the sky for hire.

Friday, December 3, 2010

Case study: AIG insurance group leverages ALM to attain IT performance architecture advantage

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Barcelona -- Welcome to a special BriefingsDirect podcast series coming to you from the HP Software Universe 2010 Conference in Barcelona.

We're here in the week of November 29, 2010 to explore some major enterprise software and solutions, trends and innovations making news across HP’s ecosystem of customers, partners, and developers. [See more on HP's new ALM 11 offerings.]

This customer case-study from the conference focuses on AIG-Chartis insurance and how their business has benefited from ongoing application transformation and modernization projects.

To learn more about AIG-Chartis insurance’s innovative use of IT consolidation and application lifecycle management (ALM) best practices, I interviewed Abe Naguib, Director of Global Performance Architecture and Infrastructure Engineering at AIG-Chartis in Jersey City, NJ.

Here are some excerpts:
Naguib: AIG is a global insurance firm, supporting worldwide international insurance of different varieties.

We're structured with 1,500 companies and roughly about eight lines of businesses that manage those companies. Each group has their own CIO, CTO, COO structure, and I report to the global CTO.

What we look at is supporting their global architecture and performance behavioristics, if you will. One of the key things is how to federate the enterprise in terms of architecture and performance, so that we can standardize the swing over into the Java world, as well as middleware and economy of scale.

When I came on board to standardize architecture, and I saw there was a proliferation of various middleware technologies. As we started going along, we thought about how to standardize that architecture.

As we faced more and more applications coming into the Java middleware world, we found that there’s a lot of footprint waste and there’s a lot of delivery cycles that are also slipped and wasted. So, we saw a need to control it.

After we started the architectural world, we also started the production support world and a facility for testing these environments. We started realizing, again, there were things that impacted business service level agreements (SLAs), economy of scale, even branding. So, we asked, how do we put it together?

One of the key things is, as we started the organizational performance, we were part of QA, but then we realized that we had to change our business strategy, and we thought about how to do that. One key thing is we changed our mindset from a performance testing practice to a performance engineering practice, and we've evolved now to performance architecture.

The engineering practice was focused on testing and analyzing, providing some kind of metrics. But, the performance architectural world now has influence into strategies, design practices, and resolution issues. We're currently a one-man or one-army team, kind of a paratrooper level. We're multi-skilled, from architecture, to performance, to support, and we drive resolution in the organization.

We also saw that resolution had to happen quickly and effectively. Carnegie Mellon did a study about five years ago and it said that post-live application resolution of performance issues was seven times the cost of pre-live [performance application resolution].

In other words, we realized that the faster we resolved issues, the faster to market, the faster we can address things, the less disruption to the delivery practices.

Too many people involved

In normal firefighting mode, architecture is involved, development is involved, and infrastructure is involved. What ends up happening is there are too many people involved. We're all scrambling, pointing fingers, looking at logs. So, we figured that the faster we get to resolution, the better for everyone to continue the train on the track.

... I have experience with Quality Center and the improvements that have gone on over the years. Because of our focus, we built our paradigm out of QA and into the performance world, and we started focusing on improving that process.

The latest TruClient product, which is a LoadRunner product, has been a massive groundbreaking point solution. In the last two years, frankly, with HP and Mercury getting adjusted, there’s been kind of a lag, but I have to give kudos to the team. [See more on HP's new ALM 11 offerings.]

One of the key things is that they have opened up their doors in terms of the delivery, in terms of their roadmap. I've worked extensively for the last roughly year with their product development team, and they have done quite a bit of improvement in their solution.

Good partnership role

They have also improved their service support model; the help desk actually resolves questions a lot faster. And we also have a good partnership role, and we actually work with things that we see, and to the influence of their roadmap as well.

This TruClient product has been phenomenal. One of the key things we're seeing now is BPM solutions are more Ajax-based, and there are so many varieties of Ajax frameworks out there than we know how to deal with. One of the key things with the partnership is that we're able to target what we need, they are able to deliver, and we are able to execute.

LoadRunner and TruClient allow us to get in front of the console, work with the business team, capture their typical use cases in a day-in-the-life scenario, and automate that. That gets buy-in and partnership with the business.

We're also able to execute a test case now and bring that in front of the IT side and show them the actual footprint from a business perspective and the impact and the benefits. What ends up happening is that now we're bringing the two teams together. So, we're bridging the gap basically from execution.

... We also started working with the CIOs to figure out a strategy to develop a service-level target, if you will. As we went along, we began working with the development teams to build a relationship with the architectural teams and the infrastructure teams.

We became more of a team model, building more of a peace-maker model. We regrouped the organization, so that rather than resolve and point fingers at each other, we resolved issues a lot faster.

Now, we're able to address the issue. We call it "isolate, identify, and resolve." At that point, if it’s a database issue, we work directly with the DBA. If it’s an infrastructure or architecture issue, we work directly with that group. We basically cut the cycle down in the last two or three years by about 70 percent.

A lot more CIOs have started bringing in more applications. We see a trend growth internally of roughly about 20-30 percent every year.



Because there is a change in our philosophy, in our strategy to focus more on business value, a lot more CIOs have started bringing in more applications. We see a trend growth internally of roughly about 20-30 percent every year.

I have a staff of nine. So, it’s a very agile, focused team, and they're very delivery-conscious. They're very business value-conscious, and we translate our data, the metrics that we capture, into business KPIs and infrastructure KPIs.

Because of that metric, the CIOs love what we do, because we make them look good with the business, which helps foster the relationship with the business, which helps them justify transformation in the future.

There is a new paradigm now, they call it the "Escalator Message." In 60 seconds or less, we can talk to a CIO, CTO, COO, or CFO about our strategy and how we can help them shift from the firefighting mode to more of an architecture mode.

If that’s the case, the more they can salvage their delivery, the more they can salvage their effective costs, and the more they can now shift to more of an IT-sensitive solutions shop. That helps build a business relationship and helps improve their economy of scale.

I would definitely send the message out to think in business value. Frankly, nobody really cares as much about the footprint cost, until they start realizing the dollars that are spent.

Also, now, business wants to see us more involved from the IT side, in terms of solutions, top-line improvements, and bottom-line improvements. As the performance teams expand and mature and we have the right toolsets, innovative toolsets like TruClient, we're able to now shift the cost of waste into a cost of improvements, and that’s been a huge factor in the last couple of years.

Last, I would say that in 8,000+ engagements -- we're actually closing in on now 10,000 events this year -- we've seen roughly $127 million in infrastructure savings that we have recouped. Again, that helps to benefit the firm. Instead of waste, now we're able to leverage that into more improvement side.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in:

Case study: Enel Green Power uses PPM to gain visibility, orchestrate myriad energy activities across 16 countries

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Barcelona -- Welcome to a special BriefingsDirect podcast series coming to you from the HP Software Universe 2010 Conference in Barcelona.

We're here in the week of November 29, 2010 to explore some major enterprise software and solutions, trends and innovations making news across HP’s ecosystem of customers, partners, and developers. [See more on HP's new ALM 11 offerings.]

This customer case-study from the conference focuses on Enel Green Power and how their Italian utility business has benefited from improved management of core business processes and gained visibility into new energy projects, also adhering to compliance through better planning and the ability to scope out new projects comprehensively. [Disclosure: HP is a sponsor of BriefingsDirect podcasts.]

To learn about Enel Green Power’s innovative use of project and portfolio management (PPM), I interviewed Massimo Ferriani, CIO of Enel Green Power in Rome.

Here are some excerpts:
Ferriani: Enel Green Power is one of the leaders in the renewables market ... We're in all the most mature technologies such as hydro, geothermal, wind, and solar.

If you think about a matrix to cross technologies and countries, we have a lot of trouble, because we operate four technologies in 16 countries.

It's difficult because we have more than 300 plants all around the world. So, it's an asset portfolio that we have to operate, and we have to reduce the risks.

When we decided to deploy IT platforms, we didn’t think that it was a good idea to deploy conventional-generation IT platforms, but to build up new platforms more fitted to renewables' needs.

We thought about the main objective in deploying these platforms and said, "Okay, maybe we have to deploy platforms that permit us to minimize the portfolio risk, in order to know exactly what production should be." For us, knowing the production is a condition.

We have to know production, and we have to know exactly the production that we're promising to sell to the market.

The business strategy is to manage centrally and operate locally. IT had to follow the strategy. Our main IT platforms are developed with the objective to be global. Global doesn’t mean managing everything centralized, but to manage the IT platform as centralized, because it's better for synergies and in terms of costs. But, because we have to fit local needs, we we have to localize these platforms in 16 countries.

For PPM, as well, we decided to have a global, centralized, unique platform, in order to gather and collect all the data that we get from the field. This is one of the problems that we frequently have because, in effect, the operation is located everywhere. And, it’s not easy to collect information from each field operation.

It’s important to have global IT platforms, because one of our main objectives is that all our people have to work in the same way.



We have lot of plants in the middle of nowhere -- in the middle of the Nevada desert and in the middle of the Mato Grosso in Brazil. We have to gather information from these plants. So, it’s important to have global IT platforms, because one of our main objectives is that all our people have to work in the same way.

It’s also important to set the main goal of the PPM solution. Now, the PPM solution lets Enel Green Power manage its own worldwide portfolio initiatives, both business development side and the plant construction Phase 2, because we have to remember that the business development hands over the construction of the project.

We have to do it through building a unique centralized integrated platform, valuable to all the countries, designed to certify the market value of the pipeline and the potential future production related to that pipeline. For us, it's absolutely important to forecast better, to make budgets, and so on. It had to be designed to support people, our colleagues, in activities like planning, project development, reporting, document management, and so on.

Setting the main goals

So when we decided to deploy this platform, we had a lot of work for a couple of reasons.

First of all, because we wanted to develop an integrated in-house platform in order to map the ... core processes of the project, and at the same time to implement algorithms to develop a portfolio evaluation.

The second was to investigate adopting a standard solution available on the market that allowed us, with little customization, to fit the need of the business. It's important to underline that, when we started this project, it was the end of May, 2010. We already knew we were going to have an IPO. We didn’t know the time exactly, but we had to be ready for the end of October, the estimated date of the IPO.

We adopted the HP solution, because the HP people convinced us that with a minimal set of customization we would be ready for the end of October -- and we did it.

We chose HP because of the ... strong automation in the collection of the data. As I said before, also important for us were simplicity and flexibility. Also, with reference to our geographical distribution everywhere, the adoption of a solution supported with global support was another constraint and was absolutely important.

We needed a standard technology accessible from a lot of countries and with integration with other applications that we have, for example Microsoft Project. We also required scalability and platform growth -- and HP has a strength on this point -- because we are adopting a web service architecture. And, we wanted the viability of a unique homogenous view of mandating KPIs.

For us, the flexibility was one of the three main strengths on this platform and the reasons we chose HP.


We're only in the first phase in order to support the IPO and to support the certification of the market value of the pipeline. But, the main benefits of this platform for the business are acquisition and centralization of the data.

For us, the flexibility was maybe one of the three main strengths on this platform and the reasons we chose HP. But, the best one, as I said before, was the minimum customization we needed in order to fit the first phase. It’s not easy to have only three months time to set 64 workflows, because the local business development wants to fit their workflow on these needs.

It’s important for the automation to monitor all the steps of the workflow, of the individual steps of the process, to manage the workflow authorization of the individual steps, and monitor progress of the individual steps. All these data have to support us in order to plan the strategy. So, there are plenty of benefits and maybe more benefits in the future with the evolution of this platform.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in:

Wednesday, December 1, 2010

HP Software GM Jonathan Rende on how ALM enables IT to modernize businesses faster

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Barcelona -- Welcome to a special BriefingsDirect podcast from the HP Software Universe 2010 Conference in Barcelona, an interview with Jonathan Rende, Vice President and General Manager for Applications Business at HP Software.

We're here the week of November 29, 2010 to explore some major enterprise software and solutions, trends and innovations, making news across HP’s ecosystem of customers, partners, and developers. [Disclosure: HP is a sponsor of BriefingsDirect podcasts.]

Dana Gardner, Principal Analyst at Interarbor Solutions, moderated the discussion just after the roll-out of HP’s big application lifecycle management (ALM) news, the release of ALM 11. [See more on HP's new ALM 11 offerings.]

Here are some excerpts:
Rende: Over the last 25 years that I've been in the business, I've seen two or three such waves [of applications refresh] happen. Every seven to 10 years, the right combination of process and technology changes come along, and it becomes economically the right thing to do for an IT organization to take a fresh look at their application portfolio.

What’s different now than in the previous couple of cycles is that there is no lack of business applications out there. With those kind of impacts and requirements and responsibilities on the business, the agility and innovation of an application, is now synonymous with the agility and innovation of the applications themselves in the business.

It’s not really the case that the people building, provisioning, testing, and defining the applications are lacking or don’t know what they're doing. It’s mostly that the practices and processes they're engaged in are antiquated.

What I mean by that is that today, acquiring or delivering applications in a much more agile manner requires a ton more collaboration and transparency between the teams. Most processes and systems supporting those processes just aren’t set up to do that. We're asking people to do things that they don’t have the tools or wherewithal to complete.

Lifecycle roles

N
ot only are we bringing together -- through collaboration, transparency, linking, and traceability -- the core app lifecycle roles of business analysts, quality performance, security professionals, and developers, but we're extending that upstream to program management office and project managers. We're extending it upstream to architects. Those are very important constituents upstream who are establishing the standards and the stacks and the technologies that will be used across the organization.

Likewise, downstream, we're extending this to the areas of service management and service mangers who sit on help desks who need to connect. Their lifeblood is the connection with defects. Similarly, people in operations who monitor applications today need to be linked into all the information coming upstream along with those dealing with change and new releases happening all the time.

[ALM advances] extend upstream much further to a whole group of people -- and also downstream to a whole group of audiences.

Number one, they need to be able to share important information. There’s so much change that happens from the time an application project or program begins to the time that it gets delivered. There are a lot of changing requirements, changing learnings from a development perspective, problems that are found that need to be corrected.

All of that needs to be very flexible and iterative. You need those teams to be able to work together in very short cycles, so that they can effectively deliver, not only on time, but many times even more quickly than they did in the past. That’s what’s needed in an organization.

There isn’t a single IT organization in the world that doesn’t have a mixed environment, from a technology perspective.



On top of that, there isn’t a single IT organization in the world that doesn’t have a mixed environment, from a technology perspective. Most organizations don’t choose just Visual Studio to write their applications in -- or just Java. Many have a combination of either of those, or both of those, along with packaged applications off-the-shelf.

So, one of the big requirements is heterogeneity for those applications, and the management of those applications from a lifecycle approach should be accommodating of any environment. That’s a big part of what we do.

You have to be able to maintain and manage all of the information in one place, so that it can be linked, and so you can draw the right, important information in understanding how one activity affects another.

But that process, that information that you link, has to be independent of specific technology stacks. We believe that, over the past few years, not only have we created that in our quality solutions, in our performance solutions, but now we have added to that with our ALM 11 release -- the same concepts but in a much broader sense.

Integrating to other environments

B
y bringing together those core roles that I mentioned before, we've been able to do that from a requirements perspective, independent of [deployment] stack -- and from a development environment. We integrate to other environments, whether it’s a Microsoft platform, a Java platform, or from CollabNet. The use-cases that we've supported work in all of those environments very tightly -- between requirements and tests -- and pull that information all together in one place.

A business analyst or a subject matter expert who is generating requirements, captures all that information from what he hears of what’s needed, the business processes that need to built, the application, and the way it should work. He captures all of that information, and it needs to reside in one single place. However, if I'm a developer, I need to work off of a list of a set of tasks that build to those requirements.

It’s important that I have a link to that. It’s important that my priorities that I put in place then map to the business needs of those requirements. At the same time, if I'm in quality-, performance-, and security-assurance, I also need to understand the priority of those.

So, while those requirements will fit in one place, they'll change and they'll evolve. I need to be able to understand how that impacts my test plans that I am building.

With ALM 11, we're already seeing returns where organizations are able to cut the delivery time, the time from the inception of the project to the actual release of that project, by 50 percent.



If you look at some of the statistics that are thrown around from third parties that do this research on an annual basis: In almost two-thirds of projects today, application projects still fail. Then, you look at what benefits can be put in place, if you put together the right kind of an approach, system, and automation that supports that approach.

Cutting cost of delivery

We're seeing organizations similarly cut the cost of releasing an application, that whole delivery process -- cut the cost of delivery in half. And, that’s not to mention side benefits that really have a far more reaching impact later on, identifying and eliminating on creation up to 80 percent of the defects that would typically be found in production.

With ALM 11, we're already seeing returns where organizations are able to cut the delivery time, the time from the inception of the project to the actual release of that project, by 50 percent.

As a lot of folks who are close to this will know, finding a defect in production can be up to 500 times more expensive to fix than if you address it when it’s created during the development and the test process. Some really huge benefits and metrics are already coming from our customers who are using ALM 11.

Again, if you go back to the very beginning topic that we discussed, there isn’t a business, there isn’t a business activity, there isn’t a single action within corporate America that doesn’t rely on applications. Those applications -- the performance, the security, and the reliability of those systems -- are synonymous with that of the business itself.

If that’s the case, allowing organizations to deploy business critical processes in half the time, at half the cost, at a much higher level of quality, with a much reduced risk only reflects well on the business, and it’s a necessity, if you are going to be a leader in any industry.

There are so many different options of how people can deploy or choose to operate and run an application -- and those options are also available in the creation of those applications themselves. ALM 11 runs through on-premise deployment, or also through our software as a service (SaaS), so will allow flexibility.

Deep software DNA

S
oftware and our software business are increasingly important. If you look at the leadership within the company today, our new CEO has a very deep software DNA. Bill Veghte, who came in from Microsoft, has 20 plus years. The rest of the leadership team here also has 20 plus years in enterprise software.

Aside from the business metrics that are so beneficial in software versus other businesses, there is just a real focus on making enterprise software one of the premier businesses within all of HP. You're starting to see that with investments and acquisitions, but also the investment in, more importantly, organic development and what’s coming out.

So, it’s clearly top of list and top of mind when it comes to HP. Our new CEO, Leo Apotheker, has been very clear on that since he came in.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in:

Tuesday, November 30, 2010

HP's new ALM 11 helps guide IT through shifting landscape of modern application development and service requirements

Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

Welcome to a special BriefingsDirect podcast series, coming to you from the HP Software Universe 2010 Conference in Barcelona the week of November 29, 2010. We're here to explore some major enterprise software and solutions, trends and innovations, making news across HP’s ecosystem of customers, partners, and developers. [See more on HP's new ALM 11 offerings.]

To learn more about HP’s application life-cycle management (ALM) news -- and its customer impact from the conference -- please welcome Mark Sarbiewski, Vice President of Product Marketing for HP applications. The discussion is moderated by BriefingsDirect's Dana Gardner, Principal Analyst at Interarbor Solutions.

Here are some excerpts:
Sarbiewski: The legacy approach is not going to be the right path for delivering modern applications. We’ve been hard at work for a couple of years now, recasting and re-inventing our portfolio to match the modern approach to software, going through them one-by-one.

You’ve got changes in how you are organized. You’ve got changes in the approach that people are taking. And, you’ve got brand-new technology in the mix and new ways of actually constructing applications. All of these hold great promise, but great challenges too. That's clashing with the legacy approach that people in the past took in building software.

We talk to our customers about this all of the time. It boils down to the same old changes that we see sort of every 10 years. A new technology comes into play with all its great opportunity and problems, and we revisit how we do this. In the last several years, it’s been about how do I get a global team going, focused on potentially a brand-new process and approach.

What are the new technologies that everybody is employing? We’ve got rich Internet technologies, Web 2.0 approaches and our technology is there. For composite applications, we’ve built a variety of capabilities that help people understand how to make the performance right with those technologies, keep the security and the quality high, while keeping the speed up.

So everything from how do we do performance testing in that environment to testing things that don’t have interfaces. And how do we understand the impact of change on the systems like that? We’ve built capabilities that help people move to Agile as a process approach, things like fundamentally changing how they can do exploratory testing, and how they can bring in automation much sooner in the process of performance, quality, and security.

Lastly, we’ve been very focused on creating a single, unified system that scales to tens of thousands of users. And, it’s a web-based system, so that wherever the team members are located, even if they don’t work for you, they can become a harmonious part of the overall team, 24-hour cycles around the globe. It speeds everything up, but it also keeps everyone on the same page. It’s that kind of anytime, anywhere access that’s just required in this modern approach to software.

How is software really supported?

When I talk to customers, I ask them, how they're supporting software. If we talk about software delivery, it's fundamentally a team sport. There isn't a single stakeholder that does it all. They all have to play and do their part.

When they tell me they’ve got requirements management in Microsoft Word, Excel, or maybe even a requirements tool, and they have a bug database for this, test management for that, and this tool here, on the surface it looks like they fitted everybody with a tool and it must be good. Right?

The problem is that the work is not isolated. You might be helping each individual stakeholder out a little bit, but you're not helping the team.



The problem is that the work is not isolated. You might be helping each individual stakeholder out a little bit, but you're not helping the team. The team’s work relates to each other. When requirements get created or changed, it's the ripple effect. What tests have to be modified or newly created? What code then has to be modified? When that code gets checked in, what tests has to be run? It’s the ripple effect of the work we talk about it as workflow automation. It's also the insight to know exactly where you are.

When the real question of how far am I on this project or what quality level am I at -- am I ready to release -- needs to be answered in the context of everyone’s work, I have to understand how many requirements are tested? Is my highest priority stuff working against what code?

So, you see the team aspects of it. There is so much latency in a traditional approach. Even if each player has their own tool, it's how we get that latency out and the finger-pointing and the mis-communication that also results. We take all that out of that process and, lo and behold, we see our customers cutting their delivery times in half, dropping their defect rates by 80 percent or more, and actually doing this more cheaply with fewer people.

In requirements management, one of the big new things that we’ve done is allow the import of business process models (BPMs) into the system. Now, we’ve got the whole business process flow that’s pulled right into the system. It can be pulled right from the systems like Eris or anything that’s putting in the standard business process modeling language (BPML) right into the system.

Business processes-focused

Now, everyone who accesses ALM 11 can see the actual business process. We can start articulating that this is the highest priority flow. This step of the business process, maybe it's check credit or something like that, is an external thing but it's super-important. So, we’ve got to make sure we really test the heck out of that thing. [See more on HP's new ALM 11 offerings.]

Everyone is aligned around what we’re doing, and all the requirements can be articulated in that same priority. The beautiful thing now about having all this in one place is that work connects to everything else. It connects to the test I set up, the test I run, the defects I find, and I can link it even back to the code, because we work with the major development tools like Visual Studio, Eclipse, and CollabNet.

It's hugely important that we connect into the world of developers. They're already comfortable with their tools. We just want to integrate with that work, and that’s really what we’ve done. They become part of the workflow process. They become part of the traceability we have.

What we hear from our customers is that the coolest new technology they want to work with is also the most problematic from a performance standpoint.

The bottom line is that the coolest new Web 2.0 front ends can now be very easily performance tested.



Modern requirements

We went back to the drawing board and reinvented how well we can understand these great new Web 2.0 technologies, in particular Ajax, which is really pervasive out there. We now can script from within the browser itself.

The big breakthrough there is if the browser can understand it, we can understand it. Before, we were sort of on the outside looking in, trying to figure out what a slider bar really did, and when a slider bar was moved what did that mean.

Now, we can generate a very readable script. I challenge anybody. Even a businessperson can understand, when they're clicking through an application, what gets created for the performance testing script.

We parameterize it. We can script logic there. We can suggest alternate steps. The bottom line is that the coolest new Web 2.0 front ends can now be very easily performance tested. So we don't end up in that situation where it's great, you did a beautiful rich job, and it's such a compelling interface, but only works when 10 people are hitting the application. We've got to fix that problem.

It speeds everything up, because it's so readable and quick. And it just works seamlessly. We've tested against the top 40 websites, and they are out there out using all this great new technology and it's working flawlessly.

Lots of pieces

If you think about a composite application, it's really made up of lots of pieces. There are application services or components. The idea is that if I’ve got something that works really well and I can reuse it as part of and combine it with maybe a few other things or in a couple of new pieces and I get new capability, I've saved money. I’ve moved faster and I'm delivering innovation to the business in a much better, quicker way and it should be rock-solid, because I can trust these components.

The challenge is, I'm not making up software made of lots of bits and pieces. I need to test every individual aspect of it. I need to test how they communicate together and I need to do end-to-end testing.

If I try to create composite apps and reuse all this technology, but it takes me ten times longer to test, I haven’t achieved my ultimate goal which was cheaper, faster and still high quality. So Unified Functional Testing is addressing that very challenge.

We've got Service Test which actually is incredible visual canvas for how I can test things that don't have an interface. One of the big challenges with something that doesn't have an interface is that I can't test it manually, because there are no buttons to push. It's all kind of under the covers. But, we have a wonderful, easy, brand-new reinvented tool here called Service Test that takes care of all that. [See more on HP's new ALM 11 offerings.]

That’s connected and integrated with our functional testing product that allows you to test everything end-to-end in the GUI level. The beautiful thing about our approach is you get to do that end-to-end, GUI level type of testing and the non-GUI stuff all from one solution and you report out all the testing that you get done.

Bring in a lot of automation to speed it up, keep the quality high and the time down low and you get to see it all kind of come together in one place.



So again, bring in a lot of automation to speed it up, keep the quality high and the time down low and you get to see it all kind of come together in one place.

Sprinter is not even a reinvention. It's brand-new thinking about how we can do manual testing in an Agile world. Think of that Instant-On world. It's such a big change when people move to an Agile delivery approach. Everyone on the team now plays kind of a derivative role of what they used to do. Developers take a part of testing, and quality folks have to jump in super-early. It's just a huge change.

What Sprinter brings is a toolset for that tester, for that person who is jumping in, getting right after the code to give immediate feedback, and it's a toolset that allows that tester to automatically figure out what test apps are supposed to go through to drop in data instead of typing it in. I don't have to type it anymore. I can just use an Excel spreadsheet and I can start ripping through screens and tests really fast, because I'm not testing whether it can take the input. I'm testing whether it processes it right. [See more on HP's new ALM 11 offerings.]

Cool tools

A
nd when I come across an error, there's a tool that allows me to capture those screens, annotate them, and send that back to the developer. What’s our goal when we find a defect? The goal is to explain exactly what was done to create the defect and exactly where it is. There are a whole bunch of cool tools around that.

The last point I’d make about this is called Mirror Testing. It’s super-important. It’s imperative that things like websites actually work across the variety of browsers and operating environments and operating systems, but testing all those combinations is very painful.

Mirror Testing allows the system to work in the background, while someone is testing, say on XP and Internet Explorer, five other systems, different combinations will be driven on the exact same test. I'm sitting in front of it, doing my testing, and in the background, Safari is being tested or Firefox. [See more on HP's new ALM 11 offerings.]

If there is an error on that system, I see it, I mark it, and I send it right away, essentially turning one tester into six. It's really great breakthrough thinking on the part of R&D here and a huge productivity bump.

What we hear from our customers is that they really do want their lives to be simplified, and the conclusion that they have come to in many cases is Post-It Notes, emails, and Word docs. It seems simpler at first and then it quickly falls apart at scale. Conversely, if you have tools that you can only work with in one particular environment, and most enterprises have a lot of those, you end up with a complex mess.

Companies have said, "I have a set of development tools. I probably have some SAP, maybe some Oracle. I’ve got built-in .NET, with Microsoft. I do some Eclipse stuff and I do Java. I’ve got those but if you can work with those and if you can help me get a common approach to requirements, to managing tests, functional performance, security, manage my overall project, and integrate with those tools, you’ve made my life easier."

When we talk about being environment agnostic, that’s what we mean. Our goal is to support better than anyone else in the market the variety of environments that enterprises have. The developers are happy where they are. We want them as part of the process, but we don’t want to yank them out of their environment to participate. So our goal again is to support those environments and connect into that world without disrupting the developer.

And, the other piece that you mentioned is just as important. Most customers aren’t taking one uniform approach to software. They know they’ve got different types of projects. I’ve got some big infrastructure software projects that I am not going to do all the time and I am not going to release every 30 days and a waterfall approach or a sequential approach is perfect for that.

Rock solid

I want to make sure it’s rock solid, that I can afford to take that type of an approach, and it's the right approach. For a whole host of other projects, I want to be much more agile. I want to do 60-day releases or 90-day releases or even more, and it makes sense for those projects. What I don’t want, they tell us, I don’t want every team inventing their own approach for Waterfall, Agile, or custom approaches. I want to be able to help the teams follow a best-practice approach.

As far as the workflow, they can customize it. They can have an Agile best practice, a Waterfall best practice, and even another one if they want. The system helps the team do the right thing and get a common language, common approach, all that stuff. That’s the process kind of agnostic belief we have.

The great news is that today you can download all the solutions that we’ve talked about for trials. We have some online demos that you can check out as well. There are a lot of white papers and other things. You can literally pull the software 30 minutes from now and see what I'm talking about.

On the licensing side, we believe that the simplest approach is a concurrent license, which we have on most of the products that we’ve got here. For all the modules that we’ve been talking about, if you have a concurrent license to the system, you can get any of the modules. And, it’s a nice floating license. You don’t have to count up everybody in your shop and figure out exactly who is going to be using what module.

The concurrent license model is very flexible, nice approach. It’s one we’ve had in the past. We're carrying it forward and we’ll look to continue to simplify and make it easier for customers to understand all the great capabilities and how to simply license so that they can get their teams to their modules for the capability they need.
Listen to the podcast. Find it on iTunes/iPod and Podcast.com. Read a full transcript or download a copy. Sponsor: HP.

You may also be interested in: