We performed a comparison between BlazeMeter and OpenText LoadRunner Enterprise based on real PeerSpot user reviews.
Find out in this report how the two Performance Testing Tools solutions compare in terms of features, pricing, service and support, easy of deployment, and ROI."It has helped us simulate heavy load situations so we can fix performance issues ahead of time."
"It's a great platform because it's a SaaS solution, but it also builds the on-premises hosting solutions, so we have implemented a hybrid approach. BlazeMeter sets us up for our traditional hosting platforms and application stack as well as the modern cloud-based or SaaS-based application technologies."
"With the help of the Mock Services, we are overcoming everything. Wherever we are facing issues, whether they will be long term or temporary, by implementing the Mock Services we can bypass the faulty components that are not needed for our particular testing."
"The baseline comparison in BlazeMeter is very easy, especially considering the different tests that users can easily compare."
"It supports any number of features and has a lot of tutorials."
"Using cloud-based load generators is highly valuable to us, as we can test from outside our network and increase load generation without having to upscale our hardware as much. The cloud load generator is there when we need it and is the feature we leverage the most."
"It is a stable solution. When we compare BlazeMeter with other tools in the market, I can say that the solution's overall performance has also been very good in our company."
"I really like the recording because when I use the JMeter the scripting a lot of recording it takes me a lot of time to get used to. The BlazeMeter the recording is quick."
"The most valuable feature is the Vuser protocols."
"It is mostly user-friendly and usable."
"The fact that you can have tens of thousands of virtual users and just expand an army of load generators to hammer on whatever application you're testing."
"LoadRunner Enterprise's best feature is the detailed reporting structure."
"The initial setup was straightforward. I was able to download everything myself without any IT support."
"The solution is a very user-friendly tool, especially when you compare it to a competitor like BlazeMeter."
"We have a centralized delivery team and we are able to meet enterprise requirements, which include different types of protocols that are involved, including scripting. The technology supports that and enables us to have a wider range of testing. Enterprise-level testing is something that we are satisfied with."
"The most valuable part of the product is the way you can scale the basic testing easily."
"The product currently doesn't allow users to run parallel thread groups, making it an area that should be considered for improvement."
"Having more options for customization would be helpful."
"The tool fails to offer better parameterization to allow it to run the same script across different environments, making it a feature that needs a little improvement."
"Lacks an option to include additional users during a test run."
"The scanning capability needs improvement."
"A possible improvement could be the integration with APM tools."
"My only complaint is about the technical support, where sometimes I found that they would not read into and understand the details of my question before answering it."
"BlazeMeter needs more granular access control. Currently, BlazeMeter controls everything at a workspace level, so a user can view or modify anything inside that workspace depending on their role. It would be nice if there was a more granular control where you could say, "This person can only do A, B, and C," or, "This user only has access to functional testing. This user only has access to mock services." That feature set doesn't currently exist."
"The reporting has room for improvement."
"OpenText LoadRunner Enterprise needs to improve reporting."
"When we have a new application, recording the application is a pretty tough task. We have tried multiple things. We do scripting or try to record with different settings and on different machines. We try to record multiple times, but we do not know why it is recording and why it is not recording. We do the same thing on different machines. It sometimes records, and at other times, it does not. That is one of the major concerns."
"A room for improvement in Micro Focus LoadRunner Enterprise is that it should take multiple exhibitions for a particular scenario and have automatic trending for that. This will be a very useful feature that lets users look into how many exhibitions happened for the scenario and their performance, and you should be able to see the data within the Performance Center dashboard. For example, there's one scenario I'm focusing on multiple times in a month, and if I check five times, there's no way for me to see the trend and find out how it went with those five exhibitions. It would be great if the Performance Center has a view of all five exhibitions, particularly transaction by transaction, and how they happened. If Micro Focus LoadRunner Enterprise shows you the time trends, information about one exhibition to another, and how each performed, it'll be an immense feature, and that should be visible to every user. Reporting should be simpler in Micro Focus LoadRunner Enterprise. If I did a scenario with one exhibition now, and I did that scenario again, then I should be able to schedule that scenario for the exhibition, and if that scenario is executed multiple times, there should be the option to turn it into a single view that shows you all the transactions, how the performance was, what the trend graph is for a particular time, etc."
"It would be good if we could look forward at the future technology needs we have. I would like to see Micro Focus provide more customer awareness around how LoadRunner can fulfill requirements with Big Data use cases, for example, where you do performance testing at the scale of data lakes... when it comes to technologies our company has yet to adopt, I would like to see an indication from Micro Focus of how one does performance testing and what kinds of challenges can we foresee. Those kinds of studies would really help us."
"For such an experienced team as mine, who have been with the product for over ten years, sometimes working with technical support is not that easy."
"It would be beneficial if LoadRunner could optimize resource usage, especially for protocols that require significant resources, like TrueClient, which interacts directly with the UI. If they could improve resource usage, like ingest or for the load generator, using less CPU or RAM memory, that would be great. That's where I have problems."
"Micro Focus LoadRunner Enterprise needs to add more features for Citrix performance-based applications testing. This was one of the challenges we observed. Additionally, we experienced some APIs challenges."
More OpenText LoadRunner Enterprise Pricing and Cost Advice →
BlazeMeter is ranked 4th in Performance Testing Tools with 41 reviews while OpenText LoadRunner Enterprise is ranked 5th in Performance Testing Tools with 81 reviews. BlazeMeter is rated 8.2, while OpenText LoadRunner Enterprise is rated 8.4. The top reviewer of BlazeMeter writes "Reduced our test operating costs, provides quick feedback, and helps us understand how to build better test cases". On the other hand, the top reviewer of OpenText LoadRunner Enterprise writes "Saves time and effort, and makes it easy to set up scenarios and execute tests". BlazeMeter is most compared with Apache JMeter, Tricentis NeoLoad, OpenText LoadRunner Cloud, OpenText LoadRunner Professional and BrowserStack, whereas OpenText LoadRunner Enterprise is most compared with OpenText LoadRunner Cloud, OpenText LoadRunner Professional, OpenText Silk Performer, Tricentis NeoLoad and Apache JMeter. See our BlazeMeter vs. OpenText LoadRunner Enterprise report.
See our list of best Performance Testing Tools vendors and best Load Testing Tools vendors.
We monitor all Performance Testing Tools reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.