We performed a comparison between Apica and OpenText LoadRunner Enterprise based on real PeerSpot user reviews.
Find out what your peers are saying about Datadog, Dynatrace, New Relic and others in Application Performance Monitoring (APM) and Observability."You can tell from the operational space of people who are using and consuming this data that they are more integrated. It is not dependent on one team anymore. It saves a lot of time by capturing and pinpointing the exact problem that is happening quickly. We have moved from getting escalations manually to getting escalations synthetically."
"I like the transcript download feature. And with UI scripting, it's helpful that Apica handles a lot of the backend work automatically. I don't have to tag everything manually, though I can tag elements later if needed. It's really good at recording the steps."
"Our application SREs do script checks in such a way that closely mimic our customers' actions using the platform. Because there are so many different ways and options to be able to configure checks to closely mirror your applications' capabilities, it provides a lot of optionality for teams to create the right type of check that can notify when there are any issues. At the end of the day, we want our monitoring tools to be able to catch any outage before our customers do. This is where Apica Synthetic does a great job."
"It helps with releases because we monitor them in staging. We can tell if something is critically wrong before it gets into production, e.g., if it was load related or function related and also what was different in the dev stage. It then alerts us straightaway inside of our production monitors once it has been released. Therefore, it has improved how we run our systems since we monitor multiple environments."
"It is easy to set up and configure."
"With the ZebraTester, the ability to have and store dynamic variables, when setting up the monitors, means you can extract that value and use it in a subsequent service call. This is something that has made our lives easier... This is one of the features that I like the most because it helps us in configuring these services, in a certain flow, without the need to re-record the whole thing."
"As always, within the IT industry, everybody's always looking to upgrade and update everything else like that. Apica has been one of those things but it's really hard to replace because it offers us the unique capability to see what the customer is seeing. A lot of other ones can do Selenium script and things like that, but there's a lot in Apica that we use right now. We utilize a lot of the scenario options in Apica right now, and there's a lot of other ones that do parts of it, but it doesn't do everything that Apica does."
"There are several features that are really good. The first one is the flexibility and the advanced configuration that Apica offers when it comes to configuring synthetic checks. It provides the ability to customize how the check should be performed and it is very flexible in the number of synthetic locations that it can use. It allows us to run scripts from different locations all over the world, and they have a really good number of these locations."
"For me, LoadRunner stands out, especially with its reporting capabilities, the graphs that can be generated, and the unique feature of measuring our application's response alongside our infrastructure metrics, such as CPU, memory, or disk usage, all presented in graph form. This is something other applications struggle to match."
"The most beneficial features of the solution are flexibility and versatility in their performance."
"We implemented through the vendor, who used highly-skilled professionals."
"It is also good for reporting purposes, which would be most familiar for QC and UFT users."
"I think the number one feature everybody likes is the capability to easily generate virtual users as well as the reporting."
"The product is very user-friendly."
"LoadRunner Enterprise's most valuable features are load simulation and creating correlation for parameters."
"It is mostly user-friendly and usable."
"The having to install an application on your desktop to utilize something like ZebraTester is a little cumbersome. It would be nice to see that become a web-based application. Having the documentation a little more accessible, and easier to digest by people who are just learning how to use the framework, especially when it comes to more complex or more edge-based cases would be really helpful to have."
"Learning the tool has always been a little difficult from a scripting perspective because the framework is proprietary and unique. Once we became used to what it does and how to perform it, then it became easier for my team and me. I would like to see some of the testing steps be part of a more well-known language, like Java or Python. That would be a big improvement."
"When it comes to the way the internal agent is installed, because you can install an application on a server, I would love to see the application Docker-ized. If you could install internal agents using Docker or using containers, it would be easier for us to manage them and spin up internal agents."
"We could use more detailed information in the request and response sections."
"Alerting needs improvement. It's a little noisy. It needs some better options. Currently, they have an issue, when you set up a synthetic monitor, you can set up where it's monitoring from, a data center that Apica owns."
"The customer service and support were a little slow to respond. The browser sometimes checks alerts on unknown issues like latency from Apica's side."
"The initial screen on their dashboard could have a bit more data, but this is a small thing. It could have more data, so we do not need to drill down to a screen behind that initial information. I would like them to get a little better on the user interfaces that we need to go into."
"The accuracy of alerts can be improved a little bit. Right now, it's pretty good in terms of alerting pretty quickly about failures or changes in response times. However, what we have seen happen is the number of alerts that we are getting is very frequent, and we would like to tone down the number of alerts. That's the only trouble we have. Apica could tone down those settings because there is no option for us to tone it down to a level that would reduce the alerts to a minimum. As a platform, it does send us good alerts, but it could be improved a bit."
"OpenText LoadRunner Enterprise doesn't support some mainframe protocols. We had to build scripts to access the interface."
"Sometimes, the code is not generated when we record the scripts in the backend."
"The solution is a very expensive tool when compared with other tools."
"They had wanted to change the GUI to improve the look and feel. However, since that time, we see a lot of hanging issues."
"In Micro Focus LoadRunner Enterprise, I need to spend a lot of time training people, while on other low-code or no-code platforms, I need not invest that much time."
"Currently, when we try open LRE we encounter cookie banner issues. However, I'm not sure if it is within the enterprise solution or with the vendors."
"It would be good if we could look forward at the future technology needs we have. I would like to see Micro Focus provide more customer awareness around how LoadRunner can fulfill requirements with Big Data use cases, for example, where you do performance testing at the scale of data lakes... when it comes to technologies our company has yet to adopt, I would like to see an indication from Micro Focus of how one does performance testing and what kinds of challenges can we foresee. Those kinds of studies would really help us."
"A room for improvement in Micro Focus LoadRunner Enterprise is that it should take multiple exhibitions for a particular scenario and have automatic trending for that. This will be a very useful feature that lets users look into how many exhibitions happened for the scenario and their performance, and you should be able to see the data within the Performance Center dashboard. For example, there's one scenario I'm focusing on multiple times in a month, and if I check five times, there's no way for me to see the trend and find out how it went with those five exhibitions. It would be great if the Performance Center has a view of all five exhibitions, particularly transaction by transaction, and how they happened. If Micro Focus LoadRunner Enterprise shows you the time trends, information about one exhibition to another, and how each performed, it'll be an immense feature, and that should be visible to every user. Reporting should be simpler in Micro Focus LoadRunner Enterprise. If I did a scenario with one exhibition now, and I did that scenario again, then I should be able to schedule that scenario for the exhibition, and if that scenario is executed multiple times, there should be the option to turn it into a single view that shows you all the transactions, how the performance was, what the trend graph is for a particular time, etc."
More OpenText LoadRunner Enterprise Pricing and Cost Advice →
Earn 20 points
Apica is ranked 55th in Application Performance Monitoring (APM) and Observability with 4 reviews while OpenText LoadRunner Enterprise is ranked 5th in Performance Testing Tools with 81 reviews. Apica is rated 8.4, while OpenText LoadRunner Enterprise is rated 8.4. The top reviewer of Apica writes "Offers transcript download feature and easy to set up and configure tests but not very user friendly". On the other hand, the top reviewer of OpenText LoadRunner Enterprise writes "Saves time and effort, and makes it easy to set up scenarios and execute tests". Apica is most compared with Dynatrace, Datadog, AppDynamics, Apache JMeter and OpenText LoadRunner Cloud, whereas OpenText LoadRunner Enterprise is most compared with OpenText LoadRunner Cloud, OpenText LoadRunner Professional, OpenText Silk Performer, Tricentis NeoLoad and Apache JMeter.
We monitor all Application Performance Monitoring (APM) and Observability reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.