"Agility is highly flexible. It can do much more than what our client is doing with it. They use it in a defined way. Some at that company have a much broader knowledge of agile and SAFe, but they're given applications and a mandated way to work. We had to work within their parameters and provide an accurate transition so the data would be mapped and pushed through."
"For visualization capabilities, the automation capabilities make it possible to support the different personas. The features and capabilities are excellent and come with excellent support."
"It can generate reports showing a burndown chart, burnup chart, and the planned vs actual velocity."
"The setup is pretty straightforward."
"Within Quality Center, you have the dashboard where you can monitor your progress over different entities. You can build your own SQL query segments, and all that data is there in the system, then you can make a dashboard report."
"The stability is very good."
"I love linking/associating the requirements to a test case. That's where I get to know my requirement coverage, which helps a lot at a practical level. So, we use the traceability and visibility features a lot. This helps us to understand if there are any requirements not linked to any test case, thus not getting tested at all. That missing link is always very visible, which helps us to create our requirement traceability matrix and maintain it in a dynamic way. Even with changing requirements, we can keep on changing or updating the tool."
"The tools could be useful if we were utilizing them more effectively"
"The initial setup is straightforward. It's not too hard to deploy."
"ALM Quality Center's best features are the test lab, requirement tab, and report dashboard."
"Having used the tool before, I like the use of parameters, being able to do exports and reports of the data for monitoring of executions, and the defect management as well. I feel satisfaction in that area."
"In my work as a contractor, it's always frustrating when a client has multiple software applications that don't talk to each other and they all perform the same function. That presents a huge challenge between their IT groups."
"The user interface can be improved by adding Save, Edit, Add, Cancel, and Return buttons to the popup windows that are displayed when you click on a child item."
"The machine learning features are a new capability but could be improved. This is being worked by Digital.ai currently. Multicolor simulation, specifically, could be improved."
"The uploading of test scripts can get a little cumbersome and that is a very sensitive task. They could improve on that a lot. It's really important that this gets better as I'm loading close to a thousand test scripts per cycle."
"One drawback is that ALM only launches with the IE browser. It is not supporting the latest in Chrome... It should be launched for all of the latest browsers."
"It is pricey."
"When it came to JIRA and Agile adoption, that was not really easy to do with ALM. I tried, but I was not able to do much on that... There is room for improvement in the way it connects to and handles Agile projects."
"Sometimes I do run my queries from the admin login. However, if I want to reassess all my test cases, then I am still doing this in a manual manner. I write SQL queries, then fire them off. Therefore, a library of those SQL queries would help. If we could have a typical SQL query to change the parameters within test cases, then this is one aspect I can still think that could be included in ALM. Though they would need to be analyzed and used in a very knowledgeable way."
"If the solution could create a lighter, more flexible tool with more adaptability to new methodologies such as agile, it would be great."
"The UFT tests don't work very well and it seems to depend on things as simple as the screen resolution on a machine that I've moved to."
"There were multiple modules and stuff to the solution so maybe the requirements can map to test scripts. It can't map to test steps. If you've got a process that's set up and you've got multiple test scripts that are in it, each script has to be linked to the requirement and the whole set can't be. If we're doing process-driven testing, it's more difficult to do it at the script level, which is what we're finding from a traceability perspective."
More Micro Focus ALM Quality Center Pricing and Cost Advice →
Digital.ai Agility is ranked 11th in Application Lifecycle Management (ALM) Suites with 3 reviews while Micro Focus ALM Quality Center is ranked 6th in Application Lifecycle Management (ALM) Suites with 31 reviews. Digital.ai Agility is rated 9.0, while Micro Focus ALM Quality Center is rated 7.0. The top reviewer of Digital.ai Agility writes "It's a flexible solution that scales from traditional agile to SAFe". On the other hand, the top reviewer of Micro Focus ALM Quality Center writes "Makes it easy to go back and execute the same test every time with automation". Digital.ai Agility is most compared with Jira, Rally Software, Jira Align, IBM Engineering Workflow Management and ServiceNow IT Business Management, whereas Micro Focus ALM Quality Center is most compared with Micro Focus ALM Octane, Microsoft Azure DevOps, Jira, Tricentis qTest and Micro Focus UFT One. See our Digital.ai Agility vs. Micro Focus ALM Quality Center report.
See our list of best Application Lifecycle Management (ALM) Suites vendors.
We monitor all Application Lifecycle Management (ALM) Suites reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.