We performed a comparison between CrossBrowserTesting and OpenText Silk Test based on real PeerSpot user reviews.
Find out what your peers are saying about Tricentis, OpenText, Perforce and others in Functional Testing Tools."It has increased the speed of our regression testing."
"This solution helps lower the overhead cost associated with buying multiple devices."
"CBT has made it easier to troubleshoot issues across devices when we do not have actual access to those specific devices. I even opt for CBT sometimes when we do have access to the device just because it is easier."
"The screen shot portal is essential for an easy way to run tests across hundreds of browsers and retrieve screenshots which then indicate success or failure."
"The ability to replay sessions is valuable for tracking down issues."
"CrossBrowserTesting allows us to test our site with real-world devices in real-world scenarios and find what we're missing."
"The extensive range of products available to simulate is something I have come to appreciate as it has resulted in an ability to broaden the scope of our tests."
"It was the perfect solution that saved us time and money to perform web viewing tests on real devices, which allowed our team to correct multiple failures in devices."
"Scripting is the most valuable. We are able to record and then go in and modify the script that it creates. It has a lot of generative scripts."
"A good automation tool that supports SAP functional testing."
"The ability to develop scripts in Visual Studio, Visual Studio integration, is the most valuable feature."
"The scalability of the solution is quite good. You can easily expand the product if you need to."
"The feature I like most is the ease of reporting."
"The statistics that are available are very good."
"The major thing it has helped with is to reduce the workload on testing activities."
"Sometimes, some of their instances fail, particularly in older versions of browsers."
"It would be useful if we can run the live-testing test cases on multiple platforms at the same time, instead of waiting for one session to finish."
"We had some issues with the onboarding process and the cloud conductivity could improve."
"Elements of 'real' mobile/tablet testing could be sped up."
"A wider range of physical devices with more browser versions in the Selenium Grid would be great to insure users with out-of-date devices are able to interact with our sites."
"This solution would benefit from faster testing and support for more devices."
"I have experienced some lagging issues, and it does not seem like all of the testing environments are configured the same."
"A problem that we are facing quite often is related to the network connection. Tests can fail if the remote CrossBrowserTesting's VM has connection problems. This happens mostly with browsers of Internet Explorer family which work on Windows OS."
"We moved to Ranorex because the solution did not easily scale, and we could not find good and short term third-party help. We needed to have a bigger pool of third-party contractors that we could draw on for specific implementations. Silk didn't have that, and we found what we needed for Ranorex here in the Houston area. It would be good if there is more community support. I don't know if Silk runs a user conference once a year and how they set up partners. We need to be able to talk to somebody more than just on the phone. It really comes right down to that. The generated automated script was highly dependent upon screen position and other keys that were not as robust as we wanted. We found the automated script generated by Ranorex and the other key information about a specific data point to be more robust. It handled the transition better when we moved from computer to computer and from one size of the application to the other size. When we restarted Silk, we typically had to recalibrate screen elements within the script. Ranorex also has some of these same issues, but when we restart, it typically is faster, which is important."
"Everything is very manual. It's up to us to find out exactly what the issues are."
"The support for automation with iOS applications can be better."
"The solution has a lack of compatibility with newer technologies."
"Could be more user-friendly on the installation and configuration side."
"The pricing is an issue, the program is very expensive. That is something that can improve."
"They should extend some of the functions that are a bit clunky and improve the integration."
CrossBrowserTesting is ranked 28th in Functional Testing Tools while OpenText Silk Test is ranked 26th in Functional Testing Tools. CrossBrowserTesting is rated 9.0, while OpenText Silk Test is rated 7.6. The top reviewer of CrossBrowserTesting writes "Static screenshots are the feature most often used, because they are a simple method of detecting problems". On the other hand, the top reviewer of OpenText Silk Test writes "Stable, with good statistics and detailed reporting available". CrossBrowserTesting is most compared with BrowserStack, Bitbar, Tricentis Tosca, LambdaTest and Sauce Labs, whereas OpenText Silk Test is most compared with Selenium HQ, OpenText UFT One, OpenText UFT Developer, Apache JMeter and froglogic Squish.
See our list of best Functional Testing Tools vendors and best Regression Testing Tools vendors.
We monitor all Functional Testing Tools reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.