We performed a comparison between OpenText ALM / Quality Center and OpenText Silk Test based on real PeerSpot user reviews.
Find out what your peers are saying about Atlassian, Microsoft, Nutanix and others in Application Lifecycle Management (ALM) Suites."I love to use this solution with single projects. It has helped our productivity. With the metrics that I receive, I can put them onto the management model so I can see them there. It has reduced our time for project management and controls by 20 percent."
"ALM Quality Center is a reliable, consolidated product."
"It allows us to easily make linkage and dependencies, with plenty of integrations."
"Produces good reports and has a great traceability feature."
"It has a good response time."
"The AI and functionality interface are useful."
"The most valuable feature of Micro Focus ALM Quality Center is the alignment of the test to the execution and the linking of the defects to the two. It automatically links any issues you have to the test."
"It provides visibility on release status and readiness."
"The statistics that are available are very good."
"The ability to develop scripts in Visual Studio, Visual Studio integration, is the most valuable feature."
"The feature I like most is the ease of reporting."
"The major thing it has helped with is to reduce the workload on testing activities."
"Scripting is the most valuable. We are able to record and then go in and modify the script that it creates. It has a lot of generative scripts."
"The scalability of the solution is quite good. You can easily expand the product if you need to."
"A good automation tool that supports SAP functional testing."
"Cross project reporting is limited to similar database schemas"
"The support is not good and the documentation is not consistent."
"We would like to have support for agile development."
"Only Internet Explorer is supported. That is a big problem. They don't support Chrome and Firefox and so on."
"Quality Center's ability to connect all the different projects to reflect status and progress is quite complicated. We may develop something because there are so many projects. Right now, I have to do something which Quality Center is really not designed for: over reporting. This is a very big problem right now. We may develop some controls, but it is problem at the moment. I love Quality Center for individual projects to work with it. However, if you have a lot of projects for Quality Manager to do cross reporting on many projects, then it's almost impossible. It takes a lot of time."
"I'm looking at more towards something more from a DevOps perspective. For example, how to pull the DevOps ecosystem into the Micro Focus ALM."
"It can be quite clunky, and it can easily be configured badly, which I've seen in a couple of places. If it is configured badly, it can be very hard to use. It is not so easy to integrate with other products. I've not used Micro Focus in a proper CI/CD pipeline, and I haven't managed to get that working because that has not been my focus. So, I find it hard. I've often lost the information because it had committed badly. It doesn't commit very well sometimes, but that might have to do with the sites that I was working at and the way they had configured it."
"When it came to JIRA and Agile adoption, that was not really easy to do with ALM. I tried, but I was not able to do much on that... There is room for improvement in the way it connects to and handles Agile projects."
"We moved to Ranorex because the solution did not easily scale, and we could not find good and short term third-party help. We needed to have a bigger pool of third-party contractors that we could draw on for specific implementations. Silk didn't have that, and we found what we needed for Ranorex here in the Houston area. It would be good if there is more community support. I don't know if Silk runs a user conference once a year and how they set up partners. We need to be able to talk to somebody more than just on the phone. It really comes right down to that. The generated automated script was highly dependent upon screen position and other keys that were not as robust as we wanted. We found the automated script generated by Ranorex and the other key information about a specific data point to be more robust. It handled the transition better when we moved from computer to computer and from one size of the application to the other size. When we restarted Silk, we typically had to recalibrate screen elements within the script. Ranorex also has some of these same issues, but when we restart, it typically is faster, which is important."
"They should extend some of the functions that are a bit clunky and improve the integration."
"Everything is very manual. It's up to us to find out exactly what the issues are."
"The support for automation with iOS applications can be better."
"The pricing is an issue, the program is very expensive. That is something that can improve."
"The solution has a lack of compatibility with newer technologies."
"Could be more user-friendly on the installation and configuration side."
More OpenText ALM / Quality Center Pricing and Cost Advice →
Earn 20 points
OpenText ALM / Quality Center is ranked 6th in Application Lifecycle Management (ALM) Suites with 197 reviews while OpenText Silk Test is ranked 26th in Functional Testing Tools. OpenText ALM / Quality Center is rated 8.0, while OpenText Silk Test is rated 7.6. The top reviewer of OpenText ALM / Quality Center writes "Offers features for higher-end traceability and integration with different tools but lacks in scalability ". On the other hand, the top reviewer of OpenText Silk Test writes "Stable, with good statistics and detailed reporting available". OpenText ALM / Quality Center is most compared with Microsoft Azure DevOps, OpenText ALM Octane, Jira, Tricentis qTest and Zephyr Enterprise, whereas OpenText Silk Test is most compared with Selenium HQ, OpenText UFT One, OpenText UFT Developer, Apache JMeter and froglogic Squish.
We monitor all Application Lifecycle Management (ALM) Suites reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.