We performed a comparison between BlazeMeter and OpenText LoadRunner Cloud based on real PeerSpot user reviews.
Find out in this report how the two Performance Testing Tools solutions compare in terms of features, pricing, service and support, easy of deployment, and ROI."The most valuable feature of the solution is its ability to run high loads and generate reports."
"I really like the recording because when I use the JMeter the scripting a lot of recording it takes me a lot of time to get used to. The BlazeMeter the recording is quick."
"BlazeMeter can be used for both API and performance testing, it is a multi-facility tool."
"It's a great platform because it's a SaaS solution, but it also builds the on-premises hosting solutions, so we have implemented a hybrid approach. BlazeMeter sets us up for our traditional hosting platforms and application stack as well as the modern cloud-based or SaaS-based application technologies."
"The orchestration feature is the most valuable. It's like the tourist backend component of BlazeMeter. It allows me to essentially give BlazeMeter multiple JMeter scripts and a YAML file, and it will orchestrate and execute that load test and all those scripts as I define them."
"For me, the best part is that we can graphically see the test result at runtime. It helps us understand the behavior of the application during all stages of the test."
"BlazeMeter's most valuable feature is its cloud-based platform for performance testing."
"Its most valuable features are its strong community support, user-friendly interface, and flexible capacity options."
"The beauty of LoadRunner Cloud is that we can use the load generator that is hosted by us on-premises, and we also have the option to use their hosted load generator. If it is a public-hosted application, we can also use their public-hosted load generator, but in our case, all our applications are hosted in our data center, so we are using the on-premise load generator. We have the option to deploy those load generators as we want."
"The usability and ability to integrate with other solutions is quite good. When I use it in on Azure, then Red Hat is the most likely solution I use. When I use AWS, then I tend to use Lambda functions. In either case, it works well and you can use it either way."
"The TCO has been optimized along with the total ROI."
"The product supports a wide variety of technology compared to any other tool."
"The most valuable feature is that you can create an infrastructure on-demand and do performance testing with it."
"One of LoadRunner's standout features is its extensive support for various TechStacks and protocols."
"The most valuable feature is having load generators in countries where we don’t have access to them."
"The initial setup was straightforward."
"I believe that data management and test server virtualization are things that Perforce is working on, or should be working on."
"One problem, while we are executing a test, is that it will take some time to download data. Let's say I'm performance testing with a high-end load configuration. It takes a minimum of three minutes or so to start the test itself. That's the bad part of the performance testing... every time I rerun the same test, it is downloaded again... That means I have to wait for three to four minutes again."
"Version controlling of the test cases and the information, the ability to compare the current version and the previous version within Runscope would be really nice. The history shows who made the changes, but it doesn't compare the changes."
"The Timeline Report panel has no customization options. One feature that I missed was not having a time filter, which I had in ELK. For example, there are only filter requests for a time of less than 5 seconds."
"Scalability is an area of concern in BlazeMeter, where improvements are required."
"I don't think I can generate a JMX file unless I run JMeter, which is one of my concerns when it comes to BlazeMeter."
"From a performance perspective, BlazeMeter needs to be improved...BlazeMeter has not found the extensions for WebSockets or Java Applet."
"The reporting capabilities could be improved."
"One area of improvement in the software's support is the replaying of captured data within the development environment. It would be beneficial if the replay feature could accurately mimic what the actual application is doing for better analysis and testing."
"Its scripting features need improvement."
"There is a steep learning curve for the product, too."
"Reporting and analysis need improvement. Compared to the old school LoadRunner Windows application, the reporting and analysis are mediocre in LoadRunner Cloud."
"The product must provide agents to monitor servers."
"An area for improvement is analytics on why response times are slow from certain countries."
"Their documentation is not technical enough for us. We would like to have much deeper technical documentation so that we can self-serve without constantly having to go back to them and ask."
"Improvements to the reporting would be good."
BlazeMeter is ranked 4th in Performance Testing Tools with 41 reviews while OpenText LoadRunner Cloud is ranked 6th in Performance Testing Tools with 39 reviews. BlazeMeter is rated 8.2, while OpenText LoadRunner Cloud is rated 8.2. The top reviewer of BlazeMeter writes "Reduced our test operating costs, provides quick feedback, and helps us understand how to build better test cases". On the other hand, the top reviewer of OpenText LoadRunner Cloud writes "Enterprise modeling, server maintenance, and competitive pricing". BlazeMeter is most compared with Apache JMeter, Tricentis NeoLoad, OpenText LoadRunner Professional, Perfecto and BrowserStack, whereas OpenText LoadRunner Cloud is most compared with OpenText LoadRunner Enterprise, Tricentis NeoLoad, OpenText LoadRunner Professional, Apache JMeter and OpenText UFT One. See our BlazeMeter vs. OpenText LoadRunner Cloud report.
See our list of best Performance Testing Tools vendors and best Load Testing Tools vendors.
We monitor all Performance Testing Tools reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.