What is our primary use case?
It's performing well. We typically use it for its intended purpose: application monitoring, identifying trends of release-over-release of our solutions, of average response time, impacts, server metrics.
How has it helped my organization?
Specifically, for the subteams that I'm on, it's very much allowed us to get real insights into the impacts of our changes. It's allowed us to see pre- and post-metrics and be able to provide dashboards for our business. It's been extremely valuable.
One aspect of development is the concept of continuous improvements. There are key screens in our applications. We get identification, through Dynatrace on its own, that these are our top ten slow preforming screens. They choose to invest in optimizing those, and then we can show them the outcome through all the dashboards.
What is most valuable?
In the AppMon, offering, currently, would be just the PurePath analysis, being able to deep-dive into call chains.
What needs improvement?
The session replay. That's probably the biggest. I think that's the struggle right now, the ability to reproduce the customer's behavior. "Oh, I had a spinning wheel," or "I observed a error." And you're wondering, "Okay. How? How did you do that?" and they say, "I don't remember." Being able to replay exactly, the exact screen movements and everything, it's very indicative of a good feature.
One of my key focus areas does deal with performance testing, and Andy, here at the Perform 2018 conference, had a good session on performance testing. But it was a lot of utilizing a custom thing he built and "hooks" there, but they should make hooks into some of the more modern performance testing tools a little easier. I think that would go a long way.
What do I think about the stability of the solution?
We actually experimented with Managed initially, very early in its introduction, and it very much came off as a beta product, because some of the core capabilities were just generally throwing errors. But in what I've seen in the demonstrations here at the Perform 2018 conference, it has gotten a lot more polished. The AppMon part itsself is also better polished. So, I think it's the nature of trying to rewrite from the ground up.
What do I think about the scalability of the solution?
I know, at least from an AppMon perspective, there's very much a limitation on memory and things of that nature. I think with the move towards the Managed product that they have some better opportunities there from a scaling perspective.
How are customer service and technical support?
I, myself, have not directly used technical support. We have a team that actually focuses on directly tooling. They're typically a go-between with support.
Which solution did I use previously and why did I switch?
What other advice do I have?
When it comes to the nature of digital complexity, I think AI is significant when it comes to IT's ability to scale in the cloud and manage performance problems, especially in regards to shifting the need for manual observations, and in terms of identifying system degradation, etc. I think it's extremely valuable in terms of being able to anticipate potential issues, as opposed to the typical reactive identification of issues. I'd rather an AI system find it before a customer communicates such.
Regarding one solution that could provide real answers, as opposed to just top-level data, I don't think that's a possibility. Unfortunately, I don't think that there's always a one-solution-fits-all to any problem.
If a friend said he was looking to adopt an APM solution I would tell him, "Use what we have available in the enterprise, which is Dynatrace."
Disclosure: I am a real user, and this review is based on my own experience and opinions.