We performed a comparison between OpenText LoadRunner Enterprise and Perfecto based on real PeerSpot user reviews.
Find out in this report how the two Performance Testing Tools solutions compare in terms of features, pricing, service and support, easy of deployment, and ROI."The product is very user-friendly."
"It is pretty easy to do test execution and results analysis. When it comes to scenario settings, LoadRunner Enterprise has an extra edge over other testing tools in the industry. The scenario setup is easy, and in terms of execution, we have a clear idea of what is happening"
"It is mostly user-friendly and usable."
"It's a very powerful tool."
"The fact that you can have tens of thousands of virtual users and just expand an army of load generators to hammer on whatever application you're testing."
"For me, LoadRunner stands out, especially with its reporting capabilities, the graphs that can be generated, and the unique feature of measuring our application's response alongside our infrastructure metrics, such as CPU, memory, or disk usage, all presented in graph form. This is something other applications struggle to match."
"With LoadRunner Enterprise, doing various types of performance testing, load testing, and automation testing has been very helpful for some of the teams."
"What I like most in Micro Focus LoadRunner Enterprise is the comparison between two different exhibitions which gives value to my company. I also like that the solution is user-friendly, especially in terms of making specific changes. For example, in the past, you can't see the changes when you upload scripts into the Performance Center, but now, it has that visibility, so whenever you want, you can change the script in the Performance Center. I also like that Micro Focus LoadRunner Enterprise is the only tool you can utilize for all your needs, even for different protocols and scripting. The solution also has the latest features, for example, networkability, where it can, within the UI, follow the waterfall model. You can use the insights in the Performance Center of Micro Focus LoadRunner Enterprise to address or test URLs that usually take up much time."
"We are able to offer a quality product that has been tested fully, which improves our customer satisfaction. That is a good thing. It has also reduced our IT infrastructure cost. We don't have to spend a lot on setting up infrastructure, which becomes redundant or obsolete very soon. It helps in offsetting that cost."
"It saves on the cost and effort of having to maintain our own virtual testing environment. Even our onshore team is not in the city that we work in, so that helps a lot. Even if we didn't invest a lot in getting multiple devices, having to share those devices would become a hassle."
"The automated test reporting functionality is the most valuable feature. We use the CI Dashboard. It's very important as it is the main reporting tool for our automated tests."
"One of the good things about Perfecto is the scalability that it provides."
"The automation piece is the most valuable feature. Every time we had a new version of either OS or an application, we found that being able to automate the testing across different devices is very valuable."
"I also like the reporting functions. We are constantly downloading these reports and sharing them with our final customers. They help us understand what kind of bugs are happening through the applications. The recording feature is handy because it lets us see a video of the process we run through the pipeline and discover the point at which the automation is breaking."
"The reporting feature is really tough to find in some of the other products that are competitors. Having your CITB type dashboard, where we can see the test results and see recordings of each test that passed or failed, is probably one of the distinguishing aspects of Perfecto."
"Perfecto has affected our software quality in a good way. It has allowed us to execute on-demand and on-choice. We also track the number of issues that we find in the product. Every single day, we tag the issues that we found. For example, if something was found by automation, that means it was found by a Perfecto execution. Over time, we realized the real value in tracking those numbers. We can see now that we have clearly been finding issues earlier. It has allowed us to catch our defects earlier, thus improving the quality of our applications."
"The cost of the solution is high and can be improved."
"I think better or more integration with some of the monitoring tools that we're considering."
"The reporting has room for improvement."
"New features have been added in latest version and need to be improved with the DevOps integration."
"More real-time monitoring should be available for the system under test."
"Sometimes, the code is not generated when we record the scripts in the backend."
"Third-party product integrations could be a little more slickly handled."
"Micro Focus LoadRunner Enterprise needs to add more features for Citrix performance-based applications testing. This was one of the challenges we observed. Additionally, we experienced some APIs challenges."
"The monitoring features, in particular network traffic monitoring, could be improved."
"Previously, we used the cradle. Every time the mobile was blocking it, we would have to ask Perfecto to provide another one. That took a lot of time away from us."
"I'm hoping they can support on-premises instances. We have been working on a JIRA integration with Perfecto—and I'm extremely impressed that they have that—but at this time they're not supporting onsite JIRA instances, which is what we have."
"It would be nice if there were some kind of AI to compile a list of available devices. Currently, we have to look at the web interface to see the available devices, but the pipelines can't do it on their own there. We always need to do this manually, so it would be better if this feature were automated."
"I would like to see the inclusion of machine learning features. If we can have that, it will be a better tool."
"Its performance should be improved. Anything to speed up the user interface would be a great help. We've had a lot of pain with their migration from a product UI that was based on Adobe Flash to the new product that is based on HTML5. Migrations like that seemed to be very painful and not a real smooth process. We're still sort of recovering from that migration from old technology to new, and we haven't got all the functionality ported over that we used to have on the old UI."
"There could be some improvements done on the interface. At times, there has been a bit of a struggle when finding things on the interface. A UI revamp would be a better option in future. That UI hasn't changed much in a long time, so I think they could just make it a bit better so that people could find stuff easily and intuitively."
"There was a discussion about having the capability to export the test results to a certain tool that we use in our project. If that were added it would be great not having to manually take screenshots, put them in a document, and share them on the different test management tools."
More OpenText LoadRunner Enterprise Pricing and Cost Advice →
OpenText LoadRunner Enterprise is ranked 5th in Performance Testing Tools with 81 reviews while Perfecto is ranked 8th in Performance Testing Tools with 23 reviews. OpenText LoadRunner Enterprise is rated 8.4, while Perfecto is rated 8.4. The top reviewer of OpenText LoadRunner Enterprise writes "Saves time and effort, and makes it easy to set up scenarios and execute tests". On the other hand, the top reviewer of Perfecto writes "Its reporting allows us to have a clear view regarding what tests have been executed". OpenText LoadRunner Enterprise is most compared with OpenText LoadRunner Professional, OpenText LoadRunner Cloud, OpenText Silk Performer, Tricentis NeoLoad and Apache JMeter, whereas Perfecto is most compared with BrowserStack, Sauce Labs, Appium, AWS Device Farm and Katalon Studio. See our OpenText LoadRunner Enterprise vs. Perfecto report.
See our list of best Performance Testing Tools vendors.
We monitor all Performance Testing Tools reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.