We performed a comparison between OpenText ALM / Quality Center and OpenText LoadRunner Enterprise based on real PeerSpot user reviews.
Find out what your peers are saying about Microsoft, Atlassian, Nutanix and others in Application Lifecycle Management (ALM) Suites."The setup is pretty straightforward."
"The initial setup is straightforward. It's not too hard to deploy."
"What's most valuable in Micro Focus ALM Quality Center is that it's useful for these activities: test designing, test planning, and test execution."
"As a system administrator, HPE ALM can be flexibly configured so that it can accommodate a variety of defined project lifecycles and test methodologies."
"The most valuable feature of Micro Focus ALM Quality Center is its support for many automation technologies."
"You can plan ahead with all the requirements and the test lab set it up as a library, then go do multiple testing times, recording the default that's in the system."
"Micro Focus ALM Quality Center is quite stable."
"The most valuable Quality Center feature, I find, is the solution's integration with some of our automation tools. For us, the ability to capture and record and the ease of use from a user perspective, are all key."
"IP Spoofing can be done using Performance Center."
"Probably its prime advantage, it provides a centralized location for testing."
"Now that LoadRunner integrates with Dynatrace and other monitoring tools, it simplifies the process of integration into a company, taking merely five minutes to set up. This ease of integration allows for quick comparison of monitoring and performance results, a feature I highly appreciate."
"It allows you to work out how well you are doing project-wise because you see the number of scripts done, the number of tests run, and whether you have mapped all your requirements to it."
"We implemented through the vendor, who used highly-skilled professionals."
"The product is very user-friendly."
"The most beneficial features of the solution are flexibility and versatility in their performance."
"We can measure metrics like hits per second and detect deviations or issues through graphs. We can filter out response times based on timings and identify spikes in the database or AWS reports."
"The support is not good and the documentation is not consistent."
"One drawback is that ALM only launches with the IE browser. It is not supporting the latest in Chrome... It should be launched for all of the latest browsers."
"It can be quite clunky, and it can easily be configured badly, which I've seen in a couple of places. If it is configured badly, it can be very hard to use. It is not so easy to integrate with other products. I've not used Micro Focus in a proper CI/CD pipeline, and I haven't managed to get that working because that has not been my focus. So, I find it hard. I've often lost the information because it had committed badly. It doesn't commit very well sometimes, but that might have to do with the sites that I was working at and the way they had configured it."
"HPE ALM’s out-of-the-box reporting can be perceived as rigid and limited, to an extent."
"If they could improve their BPT business components that would be good"
"There are always new features and more support for new and legacy technology architectures with each release. But the bad news is a growing list of long-standing issues with the product rarely gets addressed."
"Client-side ActiveX with patch upgrades"
"Sometimes I do run my queries from the admin login. However, if I want to reassess all my test cases, then I am still doing this in a manual manner. I write SQL queries, then fire them off. Therefore, a library of those SQL queries would help. If we could have a typical SQL query to change the parameters within test cases, then this is one aspect I can still think that could be included in ALM. Though they would need to be analyzed and used in a very knowledgeable way."
"Canned reports are always a challenge and a question with customers because customers want to see sexy reports."
"I think better or more integration with some of the monitoring tools that we're considering."
"A room for improvement in Micro Focus LoadRunner Enterprise is that it should take multiple exhibitions for a particular scenario and have automatic trending for that. This will be a very useful feature that lets users look into how many exhibitions happened for the scenario and their performance, and you should be able to see the data within the Performance Center dashboard. For example, there's one scenario I'm focusing on multiple times in a month, and if I check five times, there's no way for me to see the trend and find out how it went with those five exhibitions. It would be great if the Performance Center has a view of all five exhibitions, particularly transaction by transaction, and how they happened. If Micro Focus LoadRunner Enterprise shows you the time trends, information about one exhibition to another, and how each performed, it'll be an immense feature, and that should be visible to every user. Reporting should be simpler in Micro Focus LoadRunner Enterprise. If I did a scenario with one exhibition now, and I did that scenario again, then I should be able to schedule that scenario for the exhibition, and if that scenario is executed multiple times, there should be the option to turn it into a single view that shows you all the transactions, how the performance was, what the trend graph is for a particular time, etc."
"The product's scalability must be improved."
"New features have been added in latest version and need to be improved with the DevOps integration."
"I believe the data that demonstrates the automated correlations should be corrected."
"I know there are integrations with continuous testing. It's got tie-ins to some of the newer tools to allow continuous testing. I'd love to see us not have to customize it, but for it to be out of the box."
"The installation has not been straightforward, and we have had so many problems. We have had to re-install, try to install on a different machine, etc. We have not been able to launch the LRE server itself yet."
More OpenText ALM / Quality Center Pricing and Cost Advice →
More OpenText LoadRunner Enterprise Pricing and Cost Advice →
OpenText ALM / Quality Center is ranked 6th in Application Lifecycle Management (ALM) Suites with 197 reviews while OpenText LoadRunner Enterprise is ranked 5th in Performance Testing Tools with 81 reviews. OpenText ALM / Quality Center is rated 8.0, while OpenText LoadRunner Enterprise is rated 8.4. The top reviewer of OpenText ALM / Quality Center writes "Offers features for higher-end traceability and integration with different tools but lacks in scalability ". On the other hand, the top reviewer of OpenText LoadRunner Enterprise writes "Saves time and effort, and makes it easy to set up scenarios and execute tests". OpenText ALM / Quality Center is most compared with Microsoft Azure DevOps, OpenText ALM Octane, Jira, Tricentis qTest and TestRail, whereas OpenText LoadRunner Enterprise is most compared with OpenText LoadRunner Professional, OpenText LoadRunner Cloud, OpenText Silk Performer, Tricentis NeoLoad and Akamai CloudTest.
We monitor all Application Lifecycle Management (ALM) Suites reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.