We performed a comparison between OpenText LoadRunner Cloud and OpenText UFT Digital Lab based on real PeerSpot user reviews.
Find out in this report how the two Performance Testing Tools solutions compare in terms of features, pricing, service and support, easy of deployment, and ROI."The fact that the solution supports multiple protocols such as open source, VuGen, TruWeb, TruClient, and SAP is very important because these protocols help us to concentrate on what is really needed to produce performance tests. If something is not supported, you have to use other tools or find other ways of assimilating loads."
"It's a fast product, so you don't have much trouble in terms of maintenance overhead. You don't want to just look into configuring load generators, look for upgrades, and end up having that take up a lot of your time. With this solution, you just log in and you start using it. This means that there is a huge benefit in terms of the overhead of maintaining the infrastructure and the maintenance effort."
"Both the professional and cloud versions of Micro Focus LoadRunner use the same scripting or programming to execute performance modeling operations. This feature allows users to use various programming languages such as Java, C, or C++, which can run within either of the two environments. This flexibility in the programming language is a strong point of the software."
"The most valuable feature is the ability to configure browser settings for different operating systems and on different versions without the need to install every single version on each machine and to manage them."
"The record and playback feature is the most valuable feature. It's all driven by the script, so it's a script-based tool where the background tracing starts. Java's background process does a lot of tracing. The process starts in the background. It sees what peaks of volumes that the process can handle. It's easy to use because it's script based, record, and playback. I"
"It's fast, easy to use, has a user-friendly UI, and you can split users."
"The initial setup was straightforward."
"The TruClient feature is the most valuable for us. An application with testing can only be scripted using TruClient, so it's part web-based, but it also has its own protocol combined with HTTP and HTML. So many other tools do not recognize this specific proprietary protocol. Using TruClient, we can still create scripts that cover everything that we need to cover."
"It is a complete solution for mobile application testing."
"The solution is easy to use. There are features to orchestrate mobile testing, including mobile testing automation. You can test different devices at the same time."
"The most valuable feature of this solution is virtualization."
"The product is easy to use."
"There are numerous valuable features such as automation, the ones that facilitate importing and synchronization capabilities between our platform, Jira, and Azure DevOps."
"The fact that it allows users to test on real mobile devices instead of emulators is something that projects have told us is beyond compare."
"For automation testing, the tool provides the record and playback option, which helps with object detection easily."
"One area of improvement in the software's support is the replaying of captured data within the development environment. It would be beneficial if the replay feature could accurately mimic what the actual application is doing for better analysis and testing."
"One area for improvement in LoadRunner Cloud, especially for agile models, is its limited support for functional testing alongside its robust non-functional testing capabilities."
"We did have some challenges with the initial implementation."
"There are three modules in the system that are different products packaged into one, and they can sometimes be difficult to figure out, so they should be better integrated with each other."
"The product must provide agents to monitor servers."
"I'd like to see more ability to dive more deeply into the configuration."
"The product price could be more affordable."
"Sometimes, you are utilizing one of the low generators, then all of a sudden if you discontinue from one project, it actually deletes the entire low generator."
"For the most part, the key challenge is ensuring that customers fully utilize the product as intended and adopt the appropriate frameworks to implement the solutions effectively."
"We need to scale devices easily. Some customers would like to loop in AWS or other cloud providers to check if their devices have the cloud factor. OpenText UFT Digital Lab needs to improve it."
"The documentation and user interface both need improvement."
"I would like to see more integration with automation tools."
"We like to host the tools centrally. We would need them to be multi-tenants, so different projects could log on and have their own set of devices and their own set of apps, and they wouldn't see data from other projects that are using it."
"The product's object detection method needs to be improved since it can help testers do perfect testing."
"They should introduce a pay-per-use subscription model."
OpenText LoadRunner Cloud is ranked 6th in Performance Testing Tools with 39 reviews while OpenText UFT Digital Lab is ranked 6th in Mobile App Testing Tools with 16 reviews. OpenText LoadRunner Cloud is rated 8.2, while OpenText UFT Digital Lab is rated 7.4. The top reviewer of OpenText LoadRunner Cloud writes "Enterprise modeling, server maintenance, and competitive pricing". On the other hand, the top reviewer of OpenText UFT Digital Lab writes "Robust solution for application lifecycle management with numerous valuable features". OpenText LoadRunner Cloud is most compared with OpenText LoadRunner Enterprise, Tricentis NeoLoad, OpenText LoadRunner Professional, BlazeMeter and Apache JMeter, whereas OpenText UFT Digital Lab is most compared with OpenText UFT One, Appium, Perfecto, AWS Device Farm and Sauce Labs. See our OpenText LoadRunner Cloud vs. OpenText UFT Digital Lab report.
We monitor all Performance Testing Tools reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.